[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20121211150234.GC5580@redhat.com>
Date: Tue, 11 Dec 2012 10:02:34 -0500
From: Vivek Goyal <vgoyal@...hat.com>
To: Tejun Heo <tj@...nel.org>
Cc: Zhao Shuai <zhaoshuai@...ebsd.org>, axboe@...nel.dk,
ctalbott@...gle.com, rni@...gle.com, linux-kernel@...r.kernel.org,
cgroups@...r.kernel.org, containers@...ts.linux-foundation.org
Subject: Re: performance drop after using blkcg
On Tue, Dec 11, 2012 at 06:47:18AM -0800, Tejun Heo wrote:
> Hello,
>
> On Tue, Dec 11, 2012 at 09:43:36AM -0500, Vivek Goyal wrote:
> > I think if one sets slice_idle=0 and group_idle=0 in CFQ, for all practical
> > purposes it should become and IOPS based group scheduling.
>
> No, I don't think it is. You can't achieve isolation without idling
> between group switches. We're measuring slices in terms of iops but
> what cfq actually schedules are still time slices, not IOs.
I think I have not been able to understand your proposal. Can you explain
a bit more.
This is what CFQ does in iops_mode(). It will calculate the number of
requests dispatched from a group and scale that number based on weight
and put the group back on service tree. So if you have not got your
fair share in terms of number of requests dispatched to the device,
you will be put ahead in the queue and given a chance to dispatch
requests first.
Now couple of things.
- There is no idling here. If device is asking for more requests (deep
queue depth) then this group will be removed from service tree and
CFQ will move on to serve other queued group. So if there is a dependent
reader it will lose its share.
If we try to idle here, then we have solved nothing in terms of
performance problems. Device is faster but your workload can't cope
with it so you are artificially slowing down the device.
- But if all the contending workloads/groups are throwing enough IO
traffic on the device and don't get expired, they they should be able
to dispatch number of requests to device in proportion to their weight.
So this is effectively trying to keep track of number of reqeusts
dispatched from the group instead of time slice consumed by group and
then do the scheduling.
cfq_group_served() {
if (iops_mode(cfqd))
charge = cfqq->slice_dispatch;
cfqg->vdisktime += cfq_scale_slice(charge, cfqg);
}
Isn't it effectively IOPS scheduling. One should get IOPS rate in proportion to
their weight (as long as they can throw enough traffic at device to keep
it busy). If not, can you please give more details about your proposal.
Thanks
Vivek
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists