[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20070416112440.GA20014@elte.hu>
Date: Mon, 16 Apr 2007 13:24:40 +0200
From: Ingo Molnar <mingo@...e.hu>
To: William Lee Irwin III <wli@...omorphy.com>
Cc: Willy Tarreau <w@....eu>,
"Eric W. Biederman" <ebiederm@...ssion.com>,
Nick Piggin <npiggin@...e.de>, linux-kernel@...r.kernel.org,
Linus Torvalds <torvalds@...ux-foundation.org>,
Andrew Morton <akpm@...ux-foundation.org>,
Con Kolivas <kernel@...ivas.org>,
Mike Galbraith <efault@....de>,
Arjan van de Ven <arjan@...radead.org>,
Thomas Gleixner <tglx@...utronix.de>,
Jiri Slaby <jirislaby@...il.com>,
Alan Cox <alan@...rguk.ukuu.org.uk>
Subject: Re: [Announce] [patch] Modular Scheduler Core and Completely Fair Scheduler [CFS]
* William Lee Irwin III <wli@...omorphy.com> wrote:
> On Sun, Apr 15, 2007 at 09:57:48PM +0200, Ingo Molnar wrote:
> > Oh I was very much testing "CPU bandwidth allocation as influenced by
> > nice numbers" - it's one of the basic things i do when modifying the
> > scheduler. An automated tool, while nice (all automation is nice)
> > wouldnt necessarily show such bugs though, because here too it needed
> > thousands of running tasks to trigger in practice. Any volunteers? ;)
>
> Worse comes to worse I might actually get around to doing it myself.
> Any more detailed descriptions of the test for a rainy day?
the main complication here is that the handling of nice levels is still
typically a 2nd or 3rd degree design factor when writing schedulers. The
reason isnt carelessness, the reason is simply that users typically only
care about a single nice level: the one that all tasks run under by
default.
Also, often there's just one or two good ways to attack the problem
within a given scheduler approach and the quality of nice levels often
suffers under other, more important design factors like performance.
This means that for example for the vanilla scheduler the distribution
of CPU power depends on HZ, on the number of tasks and on the scheduling
pattern. The distribution of CPU power amongst nice levels is basically
a function of _everything_. That makes any automated test pretty
challenging. Both with SD and with CFS there's a good chance to actually
formalize the meaning of nice levels, but i'd not go as far as to
mandate any particular behavior by rigidly saying "pass this automated
tool, else ...", other than "make nice levels resonable". All the other
more formal CPU resource limitation techniques are then a matter of
CKRM-alike patches, which offer much more finegrained mechanisms than
pure nice levels anyway.
so to answer your question: it's pretty much freely defined. Make up
your mind about it and figure out the ways how people use nice levels
and think about which aspects of that experience are worth testing for
intelligently.
Ingo
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists