[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1162219080.2948.21.camel@laptopd505.fenrus.org>
Date: Mon, 30 Oct 2006 15:38:00 +0100
From: Arjan van de Ven <arjan@...radead.org>
To: Kyle Moffett <mrmacman_g4@....com>
Cc: Linus Torvalds <torvalds@...l.org>,
"Adam J. Richter" <adam@...drasil.com>, akpm@...l.org,
bunk@...sta.de, greg@...ah.com, linux-kernel@...r.kernel.org,
linux-pci@...ey.karlin.mff.cuni.cz, matthew@....cx, pavel@....cz,
shemminger@...l.org
Subject: Re: [patch] drivers: wait for threaded probes between initcall
levels
> I admit the complexity is a bit high, but since the maximum nesting
> is bounded by the complexity of the hardware and the number of
> busses, and the maximum memory-allocation is strictly limited in the
> single-threaded case this could allow 64-way systems to probe all
> their hardware an order of magnitude faster than today without
> noticeably impacting an embedded system even in the absolute worst case.
how much of this complexity goes away if you consider the
scanning/probing as a series of "work elements", and you end up with a
queue of work elements that threads can pull work off one at a time (so
that if one element blocks the others just continue to flow). If you
then find, say, a new PCI bus you just put another work element to
process it at the end of the queue, or you process it synchronously. Etc
etc.
All you need to scale then is the number of worker threads on the
system, which should be relatively easy to size....
(check every X miliseconds if there are more than X outstanding work
elements, if there are, spawn one new worker thread if the total number
of worker threads is less than the system wide max. Worker threads die
if they have nothing to do for more than Y miliseconds)
Oh and... we have a concept for this already, or at least mostly, via
the work queue mechanism :)
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists