[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20100520185709.GB3078@redhat.com>
Date: Thu, 20 May 2010 14:57:09 -0400
From: Vivek Goyal <vgoyal@...hat.com>
To: Corrado Zoccolo <czoccolo@...il.com>
Cc: linux kernel mailing list <linux-kernel@...r.kernel.org>,
Jens Axboe <jens.axboe@...cle.com>,
Moyer Jeff Moyer <jmoyer@...hat.com>
Subject: Re: [PATCH] cfq-iosched: Revert the logic of deep queues
On Thu, May 20, 2010 at 10:50:24AM -0400, Vivek Goyal wrote:
> On Thu, May 20, 2010 at 04:01:55PM +0200, Corrado Zoccolo wrote:
> > On Thu, May 20, 2010 at 3:18 PM, Vivek Goyal <vgoyal@...hat.com> wrote:
> > > On Thu, May 20, 2010 at 01:51:49AM +0200, Corrado Zoccolo wrote:
> > > Hi Corrado,
> > >
> > > Deep queues can happen often on high end storage. One case I can think of is
> > > multiple kvm virt machines running and doing IO using AIO.
> > >
> > > I am not too keen on introducing a tunable at this point of time. Reason
> > > being that somebody having a SATA disk and driving deep queue depths is
> > > not very practical thing to do. At the same time we have fixed a theoritical
> > > problem in the past. If somebody really runs into the issue of deep queue
> > > starving other random IO, then we can fix it.
> > >
> > > Even if we have to fix it, I think instead of a tunable, a better solution
> > > would be to expire the deep queue after one round of dispatch (after
> > > having dispatched "quantum" number of requests from queue). That way no
> > > single sync-noidle queue will starve other queues and they will get to
> > > dispatch IO very nicely without intorducing any bottlenecks.
> >
> > Can you implement this solution in the patch? It seems that this will
> > solve both the performance issue as well as non-reintroducing the
> > theoretical starvation problem.
> > If we don't mind some more tree operations, the queue could be expired
> > at every dispatch (if there are other queues in the service tree),
> > instead of every quantum dispatches, to cycle through all no-idle
> > queues more quickly and more fairly.
>
> Alright. Following is a copile tested only patch. I have yet to do the
> testing to make sure it works. But I think it should address your concern
> of a deep queue starving other shallow sync-noidle queues.
>
> Does this one look good?
Well, this patch does not work. The reason being that over a period of
time we drive deeper queue depths (due to all the queues doing IO at
same time) and that means somebody doing 1 random read is stuck somewhere
in that 32 requests dispatched to disk and it does not issue next IO
till previous one is completed.
So even if you expire the deep queue early, it will be selected again and
it will do more dispatches (cfq quantum seems to work only for one round
and for next round we don't keep track how many requests have already been
disptached from this queue).
Even your solution of tunable and turning it off by default on NCQ will
not work as most of the SATA disks are now NCQ enabled. Its hard to find
non-NCQ hardware.
CFQ's tunable and everything is so much geared towards SATA disks that it
is hard to use it with servers. Now IO controller feature is primarily
useful for servers and it is dependent on CFQ. This seems to be a bad
situation. Now CFQ has to scale so that one can comfortably use with
high end storage without losing performance. So far folks will easily
switch to dealine but that take IO controller out of the picture.
Autotuning seems to be only reliable way. Default can be set right only
for one type of usage and other type of users will be left with the
exercise of setting the tunables right. Sadly, getting autotuning right
is hard.
Vivek
>
> Index: linux-2.6/block/cfq-iosched.c
> ===================================================================
> --- linux-2.6.orig/block/cfq-iosched.c
> +++ linux-2.6/block/cfq-iosched.c
> @@ -313,7 +313,6 @@ enum cfqq_state_flags {
> CFQ_CFQQ_FLAG_sync, /* synchronous queue */
> CFQ_CFQQ_FLAG_coop, /* cfqq is shared */
> CFQ_CFQQ_FLAG_split_coop, /* shared cfqq will be splitted */
> - CFQ_CFQQ_FLAG_deep, /* sync cfqq experienced large depth */
> CFQ_CFQQ_FLAG_wait_busy, /* Waiting for next request */
> };
>
> @@ -342,7 +341,6 @@ CFQ_CFQQ_FNS(slice_new);
> CFQ_CFQQ_FNS(sync);
> CFQ_CFQQ_FNS(coop);
> CFQ_CFQQ_FNS(split_coop);
> -CFQ_CFQQ_FNS(deep);
> CFQ_CFQQ_FNS(wait_busy);
> #undef CFQ_CFQQ_FNS
>
> @@ -2377,6 +2375,17 @@ static int cfq_dispatch_requests(struct
> cfq_class_idle(cfqq))) {
> cfqq->slice_end = jiffies + 1;
> cfq_slice_expired(cfqd, 0);
> + } else if (cfqq_type(cfqq) == SYNC_NOIDLE_WORKLOAD
> + && cfqq->service_tree->count > 1
> + && cfqq->slice_dispatch >= cfq_prio_to_maxrq(cfqd, cfqq)/2) {
> + /*
> + * Expire a sync-noidle queue immediately if it has already
> + * dispatched many requests. This will make sure one deep
> + * sync-noidle queue will not starve other shallow sync-noidle
> + * queues.
> + */
> + cfqq->slice_end = jiffies + 1;
> + cfq_slice_expired(cfqd, 0);
> }
>
> cfq_log_cfqq(cfqd, cfqq, "dispatched a request");
> @@ -3036,11 +3045,8 @@ cfq_update_idle_window(struct cfq_data *
>
> enable_idle = old_idle = cfq_cfqq_idle_window(cfqq);
>
> - if (cfqq->queued[0] + cfqq->queued[1] >= 4)
> - cfq_mark_cfqq_deep(cfqq);
> -
> if (!atomic_read(&cic->ioc->nr_tasks) || !cfqd->cfq_slice_idle ||
> - (!cfq_cfqq_deep(cfqq) && CFQQ_SEEKY(cfqq)))
> + CFQQ_SEEKY(cfqq))
> enable_idle = 0;
> else if (sample_valid(cic->ttime_samples)) {
> if (cic->ttime_mean > cfqd->cfq_slice_idle)
> @@ -3593,11 +3599,6 @@ static void cfq_idle_slice_timer(unsigne
> */
> if (!RB_EMPTY_ROOT(&cfqq->sort_list))
> goto out_kick;
> -
> - /*
> - * Queue depth flag is reset only when the idle didn't succeed
> - */
> - cfq_clear_cfqq_deep(cfqq);
> }
> expire:
> cfq_slice_expired(cfqd, timed_out);
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists