[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4DF1B035.7080009@tao.ma>
Date: Fri, 10 Jun 2011 13:48:37 +0800
From: Tao Ma <tm@....ma>
To: Vivek Goyal <vgoyal@...hat.com>
CC: linux-kernel@...r.kernel.org, Jens Axboe <axboe@...nel.dk>
Subject: Re: CFQ: async queue blocks the whole system
On 06/10/2011 02:27 AM, Vivek Goyal wrote:
> On Thu, Jun 09, 2011 at 11:44:21PM +0800, Tao Ma wrote:
>
> [..]
>>> CFQ in general tries not to drive too deep a queue depth in an effort
>>> to improve latencies. CFQ is generally recommened for slow SATA drives
>>> and dispatching too many requests from a single queue can only serve to
>>> increase the latency.
>> ok, so do you mean that for a fast drive, cfq isn't recommended and
>> deadline is always prefered? ;) We have a SAS with queue_depth=128, so
>> it should be a fast drive I guess. :)
>
> I think in general that has been true in my experience. SSDs are still
> ok with CFQ because that sets nonrotational flag and cuts down on
> idling. But if it is a rotational media which can handle multiple
> parallel requests at a time you might have better throughput results
> with deadline.
Thank you for the advice.
>
> [..]
>>> Its latency vs throughput tradeoff.
>> ok, so it seems that all these are designed, not a bug. Thanks for the
>> clarification.
>>
>> btw, reverting the patch doesn't work. I can still get the livelock.
>
> Can you give following patch a try and see if it helps. On my system this
> does allow CFQ to dispatch some writes once in a while.
Sorry, this patch doesn't work in my test.
Regards,
Tao
>
> thanks
> Vivek
>
> ---
> block/cfq-iosched.c | 25 ++++++++++++++++++++++++-
> 1 file changed, 24 insertions(+), 1 deletion(-)
>
> Index: linux-2.6/block/cfq-iosched.c
> ===================================================================
> --- linux-2.6.orig/block/cfq-iosched.c 2011-06-09 11:44:40.000000000 -0400
> +++ linux-2.6/block/cfq-iosched.c 2011-06-09 14:04:01.036983301 -0400
> @@ -303,6 +303,8 @@ struct cfq_data {
>
> /* Number of groups which are on blkcg->blkg_list */
> unsigned int nr_blkcg_linked_grps;
> +
> + unsigned long last_async_dispatched;
> };
>
> static struct cfq_group *cfq_get_next_cfqg(struct cfq_data *cfqd);
> @@ -2063,6 +2065,10 @@ static void cfq_dispatch_insert(struct r
>
> cfqd->rq_in_flight[cfq_cfqq_sync(cfqq)]++;
> cfqq->nr_sectors += blk_rq_sectors(rq);
> +
> + if (!cfq_cfqq_sync(cfqq))
> + cfqd->last_async_dispatched = jiffies;
> +
> cfq_blkiocg_update_dispatch_stats(&cfqq->cfqg->blkg, blk_rq_bytes(rq),
> rq_data_dir(rq), rq_is_sync(rq));
> }
> @@ -3315,8 +3321,25 @@ cfq_should_preempt(struct cfq_data *cfqd
> * if the new request is sync, but the currently running queue is
> * not, let the sync request have priority.
> */
> - if (rq_is_sync(rq) && !cfq_cfqq_sync(cfqq))
> + if (rq_is_sync(rq) && !cfq_cfqq_sync(cfqq)) {
> + unsigned long async_delay = 0;
> +
> + async_delay = jiffies - cfqd->last_async_dispatched;
> +
> + /*
> + * CFQ is heavily loaded in favor of sync queues and that
> + * can lead to starvation of async queues. If it has been
> + * too long since last async request was dispatched, don't
> + * preempt async queue
> + *
> + * Once we have per group async queues, this will need
> + * modification.
> + */
> + if (async_delay > 2 * HZ)
> + return false;
> +
> return true;
> + }
>
> if (new_cfqq->cfqg != cfqq->cfqg)
> return false;
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists