[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <54C08CA6.8050101@fb.com>
Date: Wed, 21 Jan 2015 22:37:42 -0700
From: Jens Axboe <axboe@...com>
To: Huang Ying <ying.huang@...el.com>, Christoph Hellwig <hch@....de>
CC: LKML <linux-kernel@...r.kernel.org>, LKP ML <lkp@...org>
Subject: Re: [LKP] [block] 34b48db66e0: +3291.6% iostat.sde.wrqm/s
On 01/21/2015 06:21 PM, Huang Ying wrote:
> FYI, we noticed the below changes on
>
> commit 34b48db66e08ca1c1bc07cf305d672ac940268dc ("block: remove artifical max_hw_sectors cap")
>
> testbox/testcase/testparams: lkp-ws02/fileio/600s-100%-1HDD-btrfs-64G-1024f-seqrewr-sync
>
> c2661b806092d8ea 34b48db66e08ca1c1bc07cf305
> ---------------- --------------------------
> %stddev %change %stddev
> \ | \
> 47176 ± 2% -67.3% 15406 ± 4% softirqs.BLOCK
> 1110 ± 44% -51.0% 544 ± 35% sched_debug.cpu#8.curr->pid
> 22 ± 33% -48.9% 11 ± 43% sched_debug.cpu#1.cpu_load[0]
> 91 ± 43% +125.0% 204 ± 32% sched_debug.cfs_rq[4]:/.blocked_load_avg
> 17 ± 46% -65.2% 6 ± 31% sched_debug.cfs_rq[1]:/.runnable_load_avg
> 105 ± 43% +102.6% 213 ± 32% sched_debug.cfs_rq[4]:/.tg_load_contrib
> 163 ± 35% +62.6% 265 ± 27% sched_debug.cfs_rq[16]:/.blocked_load_avg
> 183 ± 29% +51.4% 277 ± 26% sched_debug.cfs_rq[16]:/.tg_load_contrib
> 1411 ± 31% -42.5% 812 ± 32% sched_debug.cpu#6.curr->pid
> 57565068 ± 15% +66.8% 96024066 ± 17% cpuidle.C1E-NHM.time
> 94625 ± 9% -32.5% 63893 ± 4% cpuidle.C3-NHM.usage
> 200 ± 14% -22.8% 155 ± 24% sched_debug.cfs_rq[8]:/.tg_load_contrib
> 244 ± 33% -39.0% 149 ± 11% sched_debug.cfs_rq[6]:/.blocked_load_avg
> 265 ± 31% -38.4% 163 ± 9% sched_debug.cfs_rq[6]:/.tg_load_contrib
> 4959 ± 9% -18.2% 4058 ± 1% slabinfo.kmalloc-128.active_objs
> 4987 ± 9% -18.6% 4058 ± 1% slabinfo.kmalloc-128.num_objs
> 19 ± 8% -19.7% 15 ± 14% sched_debug.cpu#0.cpu_load[1]
> 662307 ± 7% -12.6% 579108 ± 3% cpuidle.C6-NHM.usage
> 3028 ± 7% -12.3% 2656 ± 1% slabinfo.ext4_extent_status.num_objs
> 3028 ± 7% -12.3% 2656 ± 1% slabinfo.ext4_extent_status.active_objs
> 4.87 ± 0% +3291.6% 165.07 ± 0% iostat.sde.wrqm/s
> 1006 ± 0% +120.3% 2216 ± 0% iostat.sde.avgrq-sz
So these two above tells us that we are doing way more write merges per
second, and that the average request size has roughly doubled from 1006
to 2216 - both are excellent news.
> 466 ± 0% +115.9% 1007 ± 0% iostat.sde.await
> 466 ± 0% +115.9% 1007 ± 0% iostat.sde.w_await
Service time roughly doubled, must be mostly stream time.
> 301 ± 0% -52.7% 142 ± 0% iostat.sde.w/s
About half the number of writes completed, but from the stats above,
those writes are more than double. 1006 * 301 < 2216 * 142, so again,
this looks good.
> 2230 ± 2% -8.2% 2048 ± 2% vmstat.system.in
And a nice reduction in irq rate, also nice. Way less software irqs from
the first few lines, also a win.
--
Jens Axboe
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists