[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1528123050.7898.99.camel@surriel.com>
Date: Mon, 04 Jun 2018 10:37:30 -0400
From: Rik van Riel <riel@...riel.com>
To: Srikar Dronamraju <srikar@...ux.vnet.ibm.com>,
Peter Zijlstra <peterz@...radead.org>
Cc: Ingo Molnar <mingo@...nel.org>,
LKML <linux-kernel@...r.kernel.org>,
Mel Gorman <mgorman@...hsingularity.net>,
Thomas Gleixner <tglx@...utronix.de>
Subject: Re: [PATCH 04/19] sched/numa: Set preferred_node based on best_cpu
On Mon, 2018-06-04 at 05:59 -0700, Srikar Dronamraju wrote:
> * Peter Zijlstra <peterz@...radead.org> [2018-06-04 14:23:36]:
>
> > OK, the above matches the description, but I'm puzzled by the
> > remainder:
> >
> > >
> > > - if (ng->active_nodes > 1 &&
> > > numa_is_active_node(env.dst_nid, ng))
> > > - sched_setnuma(p, env.dst_nid);
> > > + if (nid != p->numa_preferred_nid)
> > > + sched_setnuma(p, nid);
> > > }
> >
> > That seems to entirely loose the active_node thing, or are you
> > saying
> > best_cpu already includes that? (Changelog could use a little help
> > there
> > I suppose)
>
> I think checking for active_nodes before calling sched_setnuma was a
> mistake.
>
> Before this change, we may be retaining numa_preferred_nid to be the
> source node while we select another node with better numa affinity to
> run on.
Sometimes workloads are so large they get spread
around to multiple NUMA nodes.
In that case, you do NOT want all the tasks of
that workload (numa group) to try squeezing onto
the same load, only to have the load balancer
randomly move tasks off of that node again later.
How do you keep that from happening?
--
All Rights Reversed.
Download attachment "signature.asc" of type "application/pgp-signature" (489 bytes)
Powered by blists - more mailing lists