[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20090904231129.GA3689@redhat.com>
Date: Fri, 4 Sep 2009 19:11:29 -0400
From: Vivek Goyal <vgoyal@...hat.com>
To: Ryo Tsuruta <ryov@...inux.co.jp>
Cc: linux-kernel@...r.kernel.org, dm-devel@...hat.com,
Jens Axboe <jens.axboe@...cle.com>,
Alasdair G Kergon <agk@...hat.com>,
Morton Andrew Morton <akpm@...ux-foundation.org>,
Nauman Rafique <nauman@...gle.com>,
Gui Jianfeng <guijianfeng@...fujitsu.com>,
Rik Van Riel <riel@...hat.com>,
Moyer Jeff Moyer <jmoyer@...hat.com>,
Balbir Singh <balbir@...ux.vnet.ibm.com>
Subject: Re: Regarding dm-ioband tests
On Fri, Sep 04, 2009 at 01:02:28PM +0900, Ryo Tsuruta wrote:
> Hi Vivek,
>
> Vivek Goyal <vgoyal@...hat.com> wrote:
> > Hi Ryo,
> >
> > I decided to play a bit more with dm-ioband and started doing some
> > testing. I am doing a simple two dd threads doing reads and don't seem
> > to be gettting the fairness. So thought will ask you what's the issue
> > here. Is there an issue with my testing procedure.
[ CCing relevant folks on thread as what does fairness mean is becoming
interesting]
>
> Thank you for testing dm-ioband. dm-ioband is designed to start
> throttling bandwidth when multiple IO requests are issued to devices
> simultaneously, IOW, to start throttling when IO load exceeds a
> certain level.
>
What is that certain level? Secondly what's the advantage of this?
I can see disadvantages though. So unless a group is really busy "up to
that certain level" it will not get fairness? I breaks the isolation
between groups.
> Here is my test script that runs multiple dd threads on each
> directory. Each directory stores 20 files of 2GB.
>
> #!/bin/sh
> tmout=60
>
> for nr_threads in 1 4 8 12 16 20; do
> sync; echo 3 > /proc/sys/vm/drop_caches
>
> for i in $(seq $nr_threads); do
> dd if=/mnt1/ioband1.${i}.0 of=/dev/null &
> dd if=/mnt2/ioband2.${i}.0 of=/dev/null &
> done
> iostat -k 1 $tmout > ${nr_threads}.log
> killall -ws TERM dd
> done
> exit 0
>
> Here is the result. The average throughputs of each device are
> according to the proportion of the weight settings when the number of
> thread is over four.
>
> Average thoughput in 60 seconds [KB/s]
>
> ioband1 ioband2
> threads weight 200 weight 100 total
> 1 26642 (54.9%) 21925 (45.1%) 48568
> 4 33974 (67.7%) 16181 (32.3%) 50156
> 8 31952 (66.2%) 16297 (33.8%) 48249
> 12 32062 (67.8%) 15236 (32.2%) 47299
> 16 31780 (67.7%) 15165 (32.3%) 46946
> 20 29955 (66.3%) 15239 (33.7%) 45195
>
> Please try to run the above script on your envirionment and I would be
> glad if you let me know the result.
I ran my simple dd test again with two ioband deviecs of weight 200
(ioband1) and 100 (ioband2)respectively. I launched four sequential dd
readers on ioband2 and and one sequential reader in ioband1.
Now if we are providing isolation between groups then ioband1 should get
double the bandwidth of ioband1. But that does not happen. Following is
the output of "dmsetup table" command.
Fri Sep 4 18:02:01 EDT 2009
ioband2: 0 40355280 ioband 1 -1 0 0 0 0 0 0
ioband1: 0 37768752 ioband 1 -1 0 0 0 0 0 0
Fri Sep 4 18:02:08 EDT 2009
ioband2: 0 40355280 ioband 1 -1 2031 32 250280 0 0 0
ioband1: 0 37768752 ioband 1 -1 1484 0 186544 0 0 0
Fri Sep 4 18:02:12 EDT 2009
ioband2: 0 40355280 ioband 1 -1 3541 64 437192 0 0 0
ioband1: 0 37768752 ioband 1 -1 2802 0 352728 0 0 0
Fri Sep 4 18:02:16 EDT 2009
ioband2: 0 40355280 ioband 1 -1 5200 87 644144 0 0 0
ioband1: 0 37768752 ioband 1 -1 4003 0 504296 0 0 0
Fri Sep 4 18:02:20 EDT 2009
ioband2: 0 40355280 ioband 1 -1 7632 111 948232 0 0 0
ioband1: 0 37768752 ioband 1 -1 4494 0 566080 0 0 0
This seems to be breaking the isolation between two groups. Now if there
is one bad group with lots of readers and writers of lower weight, it will
overwhelm a group of higher weight with 1-2 readers running or some random
readers running etc.
If there are lots of readers running in a group and then a small file
reader comes in a different group of higher prio, it will not get any
fairness and latency of file read will be very high. But one would expect
that groups will provide isolation and latency of small file reader will
not increase with number of readers in a low prio group.
I also ran your test of doing heavy IO in two groups. This time I am
running 4 dd threads in both the ioband devices. Following is the snapshot
of "dmsetup table" output.
Fri Sep 4 17:45:27 EDT 2009
ioband2: 0 40355280 ioband 1 -1 0 0 0 0 0 0
ioband1: 0 37768752 ioband 1 -1 0 0 0 0 0 0
Fri Sep 4 17:45:29 EDT 2009
ioband2: 0 40355280 ioband 1 -1 41 0 4184 0 0 0
ioband1: 0 37768752 ioband 1 -1 173 0 20096 0 0 0
Fri Sep 4 17:45:37 EDT 2009
ioband2: 0 40355280 ioband 1 -1 1605 23 197976 0 0 0
ioband1: 0 37768752 ioband 1 -1 4640 1 583168 0 0 0
Fri Sep 4 17:45:45 EDT 2009
ioband2: 0 40355280 ioband 1 -1 3650 47 453488 0 0 0
ioband1: 0 37768752 ioband 1 -1 8572 1 1079144 0 0 0
Fri Sep 4 17:45:51 EDT 2009
ioband2: 0 40355280 ioband 1 -1 5111 68 635696 0 0 0
ioband1: 0 37768752 ioband 1 -1 11587 1 1459544 0 0 0
Fri Sep 4 17:45:53 EDT 2009
ioband2: 0 40355280 ioband 1 -1 5698 73 709272 0 0 0
ioband1: 0 37768752 ioband 1 -1 12503 1 1575112 0 0 0
Fri Sep 4 17:45:57 EDT 2009
ioband2: 0 40355280 ioband 1 -1 6790 87 845808 0 0 0
ioband1: 0 37768752 ioband 1 -1 14395 2 1813680 0 0 0
Note, it took me more than 20 seconds (since I started the threds) to
reach close to desired fairness level. That's too long a duration. Again
random readers or small file readers are compeltely out of picture for
any kind of fairness or are not protected at all with dm-ioband
controller.
I think there are serious issues with the notion of fairness and what kind of
isolation dm-ioband provide between groups and it should be looked into.
Thanks
Vivek
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists