[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20101117175900.0d7878e5.akpm@linux-foundation.org>
Date: Wed, 17 Nov 2010 17:59:00 -0800
From: Andrew Morton <akpm@...ux-foundation.org>
To: Dave Chinner <david@...morbit.com>
Cc: Wu Fengguang <fengguang.wu@...el.com>, Jan Kara <jack@...e.cz>,
Christoph Hellwig <hch@....de>,
"Theodore Ts'o" <tytso@....edu>,
Chris Mason <chris.mason@...cle.com>,
Peter Zijlstra <a.p.zijlstra@...llo.nl>,
Mel Gorman <mel@....ul.ie>, Rik van Riel <riel@...hat.com>,
KOSAKI Motohiro <kosaki.motohiro@...fujitsu.com>,
linux-mm <linux-mm@...ck.org>,
"linux-fsdevel@...r.kernel.org" <linux-fsdevel@...r.kernel.org>,
LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH 00/13] IO-less dirty throttling
On Thu, 18 Nov 2010 12:40:51 +1100 Dave Chinner <david@...morbit.com> wrote:
>
> There's no point
> waking a dirtier if all they can do is write a single page before
> they are throttled again - IO is most efficient when done in larger
> batches...
That assumes the process was about to do another write. That's
reasonable on average, but a bit sad for interactive/rtprio tasks. At
some stage those scheduler things should be brought into the equation.
>
> ...
>
> Yeah, sorry, should have posted them - I didn't because I snapped
> the numbers before the run had finished. Without series:
>
> 373.19user 14940.49system 41:42.17elapsed 612%CPU (0avgtext+0avgdata 82560maxresident)k
> 0inputs+0outputs (403major+2599763minor)pagefaults 0swaps
>
> With your series:
>
> 359.64user 5559.32system 40:53.23elapsed 241%CPU (0avgtext+0avgdata 82496maxresident)k
> 0inputs+0outputs (312major+2598798minor)pagefaults 0swaps
>
> So the wall time with your series is lower, and system CPU time is
> way down (as I've already noted) for this workload on XFS.
How much of that benefit is an accounting artifact, moving work away
from the calling process's CPU and into kernel threads?
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists