[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1403683129-10814-7-git-send-email-mgorman@suse.de>
Date: Wed, 25 Jun 2014 08:58:49 +0100
From: Mel Gorman <mgorman@...e.de>
To: Linux Kernel <linux-kernel@...r.kernel.org>,
Linux-MM <linux-mm@...ck.org>,
Linux-FSDevel <linux-fsdevel@...r.kernel.org>
Cc: Johannes Weiner <hannes@...xchg.org>, Jens Axboe <axboe@...nel.dk>,
Jeff Moyer <jmoyer@...hat.com>,
Dave Chinner <david@...morbit.com>,
Mel Gorman <mgorman@...e.de>
Subject: [PATCH 6/6] cfq: Increase default value of target_latency
The existing CFQ default target_latency results in very poor performance
for larger numbers of threads doing sequential reads. While this can be
easily described as a tuning problem for users, it is one that is tricky
to detect. This patch updates the default to benefit smaller machines.
Dave Chinner points out that it is dangerous to assume that people know
how to tune their IO scheduler. Jeff Moyer asked what workloads even
care about threaded readers but it's reasonable to assume file,
media, database and multi-user servers all experience large sequential
readers from multiple sources at the same time.
It's a bit depressing to note how much slower this relatively simple case
is in comparison to 3.0. The following is from tiobench on a mid-range
desktop using ext3 as the test filesystem although it's known other
filesystems experience similar trouble.
3.16.0-rc2 3.16.0-rc2 3.0.0
lessdirty cfq600 vanilla
Min SeqRead-MB/sec-1 140.79 ( 0.00%) 140.43 ( -0.26%) 134.04 ( -4.79%)
Min SeqRead-MB/sec-2 118.08 ( 0.00%) 118.18 ( 0.08%) 120.76 ( 2.27%)
Min SeqRead-MB/sec-4 108.47 ( 0.00%) 110.84 ( 2.18%) 114.49 ( 5.55%)
Min SeqRead-MB/sec-8 87.20 ( 0.00%) 92.40 ( 5.96%) 98.04 ( 12.43%)
Min SeqRead-MB/sec-16 68.98 ( 0.00%) 76.68 ( 11.16%) 79.49 ( 15.24%)
The full series including this patch brings performance within an acceptable
distance of 3.0.0-vanilla considering that read latencies and fairness are
generally better now at the cost of overall throughput.
Here is the very high-level view of the iostats
3.16.0-rc2 3.16.0-rc2 3.0.0
lessdirty cfq600 vanilla
Mean sda-avgqusz 935.48 957.28 1000.70
Mean sda-avgrqsz 575.27 579.85 600.71
Mean sda-await 4405.00 4471.12 4887.67
Mean sda-r_await 82.43 87.95 108.53
Mean sda-w_await 13272.23 10783.67 11599.83
Mean sda-rrqm 14.12 10.14 19.68
Mean sda-wrqm 1631.24 1744.00 11999.46
Max sda-avgqusz 2179.79 2238.95 2626.78
Max sda-avgrqsz 1021.03 1021.97 1024.00
Max sda-await 15007.79 13600.51 24971.00
Max sda-r_await 897.78 893.09 5308.00
Max sda-w_await 207814.40 179483.79 177698.47
Max sda-rrqm 68.40 45.60 73.30
Max sda-wrqm 19544.00 19619.20 58058.40
await figures are generally ok. Average wait times are still acceptable
and the worst-case read wait times are ok. Queue sizes and request sizes
generally look ok. It's worth noting that the iostats are generally *far*
better than 3.0.
Signed-off-by: Mel Gorman <mgorman@...e.de>
---
block/cfq-iosched.c | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
index cadc378..876ae44 100644
--- a/block/cfq-iosched.c
+++ b/block/cfq-iosched.c
@@ -32,7 +32,7 @@ static int cfq_slice_async = HZ / 25;
static const int cfq_slice_async_rq = 2;
static int cfq_slice_idle = HZ / 125;
static int cfq_group_idle = HZ / 125;
-static const int cfq_target_latency = HZ * 3/10; /* 300 ms */
+static const int cfq_target_latency = HZ * 6/10; /* 600 ms */
static const int cfq_hist_divisor = 4;
/*
--
1.8.4.5
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists