[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20090530204038.GA20875@alecto.austad.us>
Date: Sat, 30 May 2009 22:40:38 +0200
From: Henrik Austad <henrik@...tad.us>
To: Peter Zijlstra <peterz@...radead.org>
Cc: GeunSik Lim <leemgs1@...il.com>, finarfin@...amos.org,
linux-kernel@...r.kernel.org
Subject: Re: SCHED_EDF infos
On Sat, May 30, 2009 at 09:15:57PM +0200, Peter Zijlstra wrote:
> On Fri, 2009-05-08 at 11:10 +0200, Henrik Austad wrote:
> > > In fact, I also don't have perfect know how to solve PI in Multicore.
> > > [...]
> > > > deadline inversion will be a problem, in fact, whatever you chooose to
> > > > be the 'key' for picking tasks (priority, niceness, deadlines, wind
> > > > direction, <whatever>), you can pretty much take that and add a
> > > > -inversion after it. :)
> >
> > No, PI is going to be deadly no matter what you do.
>
> Right, we would need to extend the Priority Inheritance Protocol to
> include everything the regular scheduling functions operate on.
>
> That is, we can reduce scheduling to a single order operator that orders
> all the available tasks, such that t_n < t_n+1.
>
> For pure EDF that would be a comparison on deadlines (and available
> bandwidth), for FIFO on static priority and for CFS something based on
> the virtual runtimes of the involved tasks. For the combined set of
> these scheduling classes the comparator uses the class hierarchy to
> order between them.
>
> Lets call the full set of data that is used to determine this order a
> task's key.
>
> If we then substitute this key for the static priority of the classic
> PIP and use this generic comparison operator, it can be extended to
> cover arbitrary complex scheduling functions.
I think you can do this in pick_next_task, actually.
I wonder, will it be enough to add a single task_struct *blocker to task_struct?
Then you will end up with a list (or tree of tasks) all ending in a single task
that holds the required resource. Say task A,B,C,D and E holds some critical
resource, then you can end up with (awful ascii-art to starboard!):
A -> B -> E
^
|
C -> D
So, whenever A-D is being scheduled, E will be picked by pick_next_task as it
detects that it is blocked, so it follows the blocker untill it finds an
unblocked task. That task must then be the holder of the resource.
Now, the 'only' thing that needs to be done, is to let the kernel update the
blocker with the correct task, and detect the moment it releases the resource.
Or am I missing some crazy racecondition here?
> This is a bit like the Proxy Execution Protocol, where we leave the
> blocked task in the runqueue, but run another task in its stead. The key
> point is that it donates the full task state as relevant to the
> scheduling function, or even more directly, it uses the scheduler
> itself, to solve the Priority Inversion.
come to think of it, this should be doable in pfair. Basically, the task can be
picked as you suggest, and it would not interfere with *any* other tasks other
than the task it blocks. And, if you do it recursively, for each task it blocks
(that depend on the resource or the release of another blocked task), it will
gradually consume more and more of the timeslices. So you will not get unbounded
execution for non-pfair tasks.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists