[<prev] [next>] [day] [month] [year] [list]
Message-ID: <tip-zm9pzuxeiebax9vrhx9bbt3i@git.kernel.org>
Date: Sat, 19 May 2012 04:34:29 -0700
From: tip-bot for Peter Zijlstra <a.p.zijlstra@...llo.nl>
To: linux-tip-commits@...r.kernel.org
Cc: linux-kernel@...r.kernel.org, hpa@...or.com, mingo@...nel.org,
torvalds@...ux-foundation.org, a.p.zijlstra@...llo.nl,
pjt@...gle.com, cl@...ux.com, riel@...hat.com,
akpm@...ux-foundation.org, bharata.rao@...il.com,
aarcange@...hat.com, Lee.Schermerhorn@...com,
suresh.b.siddha@...el.com, danms@...ibm.com, tglx@...utronix.de
Subject: [tip:sched/numa] sched/numa: Add {mems,cpus}_allowed tests
Commit-ID: 2cdd8bfaad5ffb0e6c96aa566ea9c1f81a6d6989
Gitweb: http://git.kernel.org/tip/2cdd8bfaad5ffb0e6c96aa566ea9c1f81a6d6989
Author: Peter Zijlstra <a.p.zijlstra@...llo.nl>
AuthorDate: Tue, 8 May 2012 13:12:07 +0200
Committer: Ingo Molnar <mingo@...nel.org>
CommitDate: Sat, 19 May 2012 12:55:26 +0200
sched/numa: Add {mems,cpus}_allowed tests
Make the NUMA balancer respect task_struct::{mems,cpus}_allowed.
Since these are per task we require all tasks in the process/group
to allow the target node otherwise we fail the migration.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@...llo.nl>
Cc: Suresh Siddha <suresh.b.siddha@...el.com>
Cc: Paul Turner <pjt@...gle.com>
Cc: Dan Smith <danms@...ibm.com>
Cc: Bharata B Rao <bharata.rao@...il.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@...com>
Cc: Christoph Lameter <cl@...ux.com>
Cc: Rik van Riel <riel@...hat.com>
Cc: Andrea Arcangeli <aarcange@...hat.com>
Cc: Andrew Morton <akpm@...ux-foundation.org>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>
Link: http://lkml.kernel.org/n/tip-zm9pzuxeiebax9vrhx9bbt3i@git.kernel.org
Signed-off-by: Ingo Molnar <mingo@...nel.org>
---
kernel/sched/numa.c | 51 ++++++++++++++++++++++++++++-----------------------
1 files changed, 28 insertions(+), 23 deletions(-)
diff --git a/kernel/sched/numa.c b/kernel/sched/numa.c
index cfa49b7..77fa7d4 100644
--- a/kernel/sched/numa.c
+++ b/kernel/sched/numa.c
@@ -27,7 +27,7 @@ struct numa_ops {
void (*mem_migrate)(struct numa_entity *ne, int node);
void (*cpu_migrate)(struct numa_entity *ne, int node);
- u64 (*cpu_runtime)(struct numa_entity *ne);
+ bool (*can_migrate)(struct numa_entity *ne, int node);
bool (*tryget)(struct numa_entity *ne);
void (*put)(struct numa_entity *ne);
@@ -210,19 +210,41 @@ static void process_mem_migrate(struct numa_entity *ne, int node)
lazy_migrate_process(ne_mm(ne), node);
}
-static u64 process_cpu_runtime(struct numa_entity *ne)
+static bool __task_can_migrate(struct task_struct *t, u64 *runtime, int node)
+{
+#ifdef CONFIG_CPUSETS
+ if (!node_isset(node, t->mems_allowed))
+ return false;
+#endif
+
+ if (!cpumask_intersects(cpumask_of_node(node), tsk_cpus_allowed(t)))
+ return false;
+
+ *runtime += t->se.sum_exec_runtime; // @#$#@ 32bit
+
+ return true;
+}
+
+static bool process_can_migrate(struct numa_entity *ne, int node)
{
struct task_struct *p, *t;
+ bool allowed = false;
u64 runtime = 0;
rcu_read_lock();
t = p = ne_owner(ne);
if (p) do {
- runtime += t->se.sum_exec_runtime; // @#$#@ 32bit
+ allowed = __task_can_migrate(t, &runtime, node);
+ if (!allowed)
+ break;
} while ((t = next_thread(t)) != p);
rcu_read_unlock();
- return runtime;
+ /*
+ * Don't bother migrating memory if there's less than 1 second
+ * of runtime on the tasks.
+ */
+ return allowed && runtime > NSEC_PER_SEC;
}
static bool process_tryget(struct numa_entity *ne)
@@ -248,7 +270,7 @@ static const struct numa_ops process_numa_ops = {
.mem_migrate = process_mem_migrate,
.cpu_migrate = process_cpu_migrate,
- .cpu_runtime = process_cpu_runtime,
+ .can_migrate = process_can_migrate,
.tryget = process_tryget,
.put = process_put,
@@ -634,23 +656,6 @@ calc_imb:
return node;
}
-static bool can_move_ne(struct numa_entity *ne)
-{
- /*
- * XXX: consider mems_allowed, stinking cpusets has mems_allowed
- * per task and it can actually differ over a whole process, la-la-la.
- */
-
- /*
- * Don't bother migrating memory if there's less than 1 second
- * of runtime on the tasks.
- */
- if (ne->nops->cpu_runtime(ne) < NSEC_PER_SEC)
- return false;
-
- return true;
-}
-
static void move_processes(struct node_queue *busiest_nq,
struct node_queue *this_nq,
struct numa_imbalance *imb)
@@ -689,7 +694,7 @@ static void move_processes(struct node_queue *busiest_nq,
goto next;
}
- if (!can_move_ne(ne))
+ if (!ne->nops->can_migrate(ne, this_nq->node))
goto next;
__dequeue_ne(busiest_nq, ne);
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists