lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:	Sat, 30 May 2009 23:15:47 +0200
From:	Peter Zijlstra <peterz@...radead.org>
To:	Henrik Austad <henrik@...tad.us>
Cc:	GeunSik Lim <leemgs1@...il.com>, finarfin@...amos.org,
	linux-kernel@...r.kernel.org
Subject: Re: SCHED_EDF infos

On Sat, 2009-05-30 at 22:40 +0200, Henrik Austad wrote:
> On Sat, May 30, 2009 at 09:15:57PM +0200, Peter Zijlstra wrote:
> > On Fri, 2009-05-08 at 11:10 +0200, Henrik Austad wrote:
> > > > In fact, I also  don't have perfect know how to solve PI in Multicore.
> > > > [...]
> > > > > deadline inversion will be a problem, in fact, whatever you chooose to 
> > > > > be the 'key' for picking tasks (priority, niceness, deadlines, wind 
> > > > > direction, <whatever>), you can pretty much take that and add a 
> > > > > -inversion after it. :)
> > > 
> > > No, PI is going to be deadly no matter what you do.
> > 
> > Right, we would need to extend the Priority Inheritance Protocol to
> > include everything the regular scheduling functions operate on.
> > 
> > That is, we can reduce scheduling to a single order operator that orders
> > all the available tasks, such that t_n < t_n+1.
> > 
> > For pure EDF that would be a comparison on deadlines (and available
> > bandwidth), for FIFO on static priority and for CFS something based on
> > the virtual runtimes of the involved tasks. For the combined set of
> > these scheduling classes the comparator uses the class hierarchy to
> > order between them.
> > 
> > Lets call the full set of data that is used to determine this order a
> > task's key.
> > 
> > If we then substitute this key for the static priority of the classic
> > PIP and use this generic comparison operator, it can be extended to
> > cover arbitrary complex scheduling functions.
> 
> I think you can do this in pick_next_task, actually.
> 
> I wonder, will it be enough to add a single task_struct *blocker to task_struct? 
> 
> Then you will end up with a list (or tree of tasks) all ending in a single task 
> that holds the required resource. Say task A,B,C,D and E holds some critical 
> resource, then you can end up with (awful ascii-art to starboard!):
> 
>    A -> B -> E
>              ^
>              |
>         C -> D
> 
> So, whenever A-D is being scheduled, E will be picked by pick_next_task as it 
> detects that it is blocked, so it follows the blocker untill it finds an 
> unblocked task. That task  must then be the holder of the resource.
> 
> Now, the 'only' thing that needs to be done, is to let the kernel update the 
> blocker with the correct task, and detect the moment it releases the resource.
> 
> Or am I missing some crazy racecondition here?

Except for the bit where you avoid multiple cpus trying to run the same
task :-)

It's tracktable, but it really complicates the matter. But yes, PEP is
very attractive.

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ