[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <200803221225.26728.bernd-schubert@gmx.de>
Date: Sat, 22 Mar 2008 12:25:26 +0100
From: Bernd Schubert <bernd-schubert@....de>
To: linux-kernel@...r.kernel.org
Subject: deadline unfairness
Hello,
some it seems the deadline scheduler is rather unfair. Below is an example
of md-raid6 initialization of md3, md4 and md5. All three md-devices
do share the same blockdevices (we have patched md to allow
parallel rebuild of shared block devices, since for us the cpu is the
bottleneck and not the block device).
All rebuilds started basically at the same time, as you can see, md3 is already
done and now md4 rebuilds substantially faster than md5.
md5 : active raid6 sdk3[0] sde3[5] sdi3[4] sdm3[3] sdc3[2] sdg3[1]
6834869248 blocks level 6, 256k chunk, algorithm 2 [6/6] [UUUUUU]
[=============>.......] resync = 65.8% (1124909328/1708717312) finish=272.2min speed=35734K/sec
md4 : active raid6 sdk2[0] sde2[5] sdi2[4] sdm2[3] sdc2[2] sdg2[1]
6834869248 blocks level 6, 256k chunk, algorithm 2 [6/6] [UUUUUU]
[===============>.....] resync = 77.6% (1327362312/1708717312) finish=123.9min speed=51283K/sec
md3 : active raid6 sdk1[0] sde1[5] sdi1[4] sdm1[3] sdc1[2] sdg1[1]
6834869248 blocks level 6, 256k chunk, algorithm 2 [6/6] [UUUUUU]
Reducing write_expire to 2000ms did improve the situation a bit, but noop and
the other schedulers are still by far more fair.
Here with noop:
md5 : active raid6 sdk3[0] sde3[5] sdi3[4] sdm3[3] sdc3[2] sdg3[1]
6834869248 blocks level 6, 256k chunk, algorithm 2 [6/6] [UUUUUU]
[=============>.......] resync = 67.3% (1150741776/1708717312) finish=216.8min speed=42875K/sec
md4 : active raid6 sdk2[0] sde2[5] sdi2[4] sdm2[3] sdc2[2] sdg2[1]
6834869248 blocks level 6, 256k chunk, algorithm 2 [6/6] [UUUUUU]
[===============>.....] resync = 79.3% (1355377160/1708717312) finish=134.8min speed=43659K/sec
md3 : active raid6 sdk1[0] sde1[5] sdi1[4] sdm1[3] sdc1[2] sdg1[1]
6834869248 blocks level 6, 256k chunk, algorithm 2 [6/6] [UUUUUU]
This is basically with a 2.6.22 kernel + lustre + md-backports, but nothing
done to the scheduler.
Cheers,
Bernd
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists