[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20071006012437.c0fd3a8a.pj@sgi.com>
Date: Sat, 6 Oct 2007 01:24:37 -0700
From: Paul Jackson <pj@....com>
To: "Paul Menage" <menage@...gle.com>
Cc: akpm@...ux-foundation.org, nickpiggin@...oo.com.au,
a.p.zijlstra@...llo.nl, serue@...ibm.com, clg@...ibm.com,
linux-kernel@...r.kernel.org, ebiederm@...ssion.com,
rientjes@...gle.com, svaidy@...ux.vnet.ibm.com, xemul@...nvz.org,
containers@...ts.osdl.org, balbir@...ux.vnet.ibm.com
Subject: Re: [PATCH] task containersv11 add tasks file interface fix for
cpusets
Paul Menage wrote:
> What was wrong with my suggestion from a couple of emails back? Adding
> the following in cpuset_attach():
>
> struct cgroup_iter it;
> struct task_struct *p;
> cgroup_iter_start(cs->css.cgroup, &it);
> while ((p = cgroup_iter_next(cs->css.cgroup, &it)))
> set_cpus_allowed(p, cs->cpus_allowed);
> cgroup_iter_end(cs->css.cgroup, &it);
This isn't working for me.
The key kernel routine for updating a tasks cpus_allowed
cannot be called while holding a spinlock.
But the above loop holds a spinlock, css_set_lock, between
the cgroup_iter_start and the cgroup_iter_end.
I end up generating complaints of:
BUG: scheduling while atomic
when I invoke the set_cpus_allowed() above.
Should css_set_lock be a mutex? Locking changes like that
can be risky.
Or perhaps there should be another callback, called only by
attach() requests back to the same group. Likely cpusets would
be the only subsystem interested in plugging that callback.
That, or my original patch, which calls the attach routine
even if re-attaching to the current cgroup ...
--
I won't rest till it's the best ...
Programmer, Linux Scalability
Paul Jackson <pj@....com> 1.925.600.0401
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists