[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1192822530.9471.22.camel@lappy>
Date: Fri, 19 Oct 2007 21:35:30 +0200
From: Peter Zijlstra <a.p.zijlstra@...llo.nl>
To: Steven Rostedt <rostedt@...dmis.org>
Cc: LKML <linux-kernel@...r.kernel.org>,
RT <linux-rt-users@...r.kernel.org>,
Linus Torvalds <torvalds@...ux-foundation.org>,
Andrew Morton <akpm@...ux-foundation.org>,
Ingo Molnar <mingo@...e.hu>,
Thomas Gleixner <tglx@...utronix.de>,
Gregory Haskins <ghaskins@...ell.com>,
Paul Jackson <pj@....com>
Subject: Re: [patch 6/8] pull RT tasks
On Fri, 2007-10-19 at 21:24 +0200, Peter Zijlstra wrote:
> On Fri, 2007-10-19 at 14:43 -0400, Steven Rostedt wrote:
> > plain text document attachment (rt-balance-pull-tasks.patch)
>
> > +static int pull_rt_task(struct rq *this_rq)
> > +{
> > + struct task_struct *next;
> > + struct task_struct *p;
> > + struct rq *src_rq;
> > + int this_cpu = this_rq->cpu;
> > + int cpu;
> > + int ret = 0;
> > +
> > + assert_spin_locked(&this_rq->lock);
> > +
> > + if (likely(!atomic_read(&rt_overload)))
> > + return 0;
>
> This seems to be the only usage of rt_overload. I'm not sure its worth
> keeping it around for this.
Ingo just brought up a good point. With large smp (where large is >64)
this will all suck chunks.
rt_overload will bounce around the system, and the rto_cpumask updates
might already hurt.
The idea would be to do this per cpuset, these naturally limit the
migraiton posibilities of tasks and would thus be the natural locality
to break this data structure.
> > + next = pick_next_task_rt(this_rq);
> > +
> > + for_each_cpu_mask(cpu, rto_cpumask) {
> > + if (this_cpu == cpu)
> > + continue;
>
> ...
>
> > + }
> > +
> > + return ret;
> > +}
>
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists