[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1159786807.5574.14.camel@localhost.localdomain>
Date: Mon, 02 Oct 2006 16:30:07 +0530
From: Ashwin Chaugule <ashwin.chaugule@...unite.com>
To: Peter Zijlstra <a.p.zijlstra@...llo.nl>
Cc: Andrew Morton <akpm@...l.org>, linux-kernel@...r.kernel.org,
Rik van Riel <riel@...hat.com>
Subject: Re: [RFC][PATCH 0/2] Swap token re-tuned
On Mon, 2006-10-02 at 09:35 +0200, Peter Zijlstra wrote:
> Being frustrated with these results - I mean the idea made sense, so
> what is going on - I came up with this answer:
>
> Tasks owning the swap token will retain their pages and will hence swap
> less, other (contending) tasks will get less pages and will fault more
> frequent. This prio mechanism will favour exactly those tasks not
> holding the token. Which makes for token bouncing.
>
Right. But, with the token bouncing around, effectively the RSS of the
processes at that time will keep increasing, and they should be able to
spend more time on execution than i/o. And meanwhile the priorities of
the tasks that were contending for the token, but didnt get it, will
increment. So since the fairness is preserved, all the tasks should get
their fair share for execution and it should result in a speedup as
compared to the current upstream implementation. I took a time
instrumentation of the vanilla 2.6.18 kernel build with -j 4 and I've
posted up the results in the previous mail. I'm testing on an ibm t42
1.69Ghz 64M system.
> So while I agree it would be nice to get rid of all magic variables
> (holding time in the current impl) this proposed solution hasn't
> convinced me (for one it introduces another).
>
> (for the interrested, the various attempts I tried are available here:
> http://programming.kicks-ass.net/kernel-patches/swap_token/ )
Cool!
Had you applied these patches when you posted your test results ?
Thanks
Ashwin
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists