These shouldn't ever get in.. Signed-off-by: Peter Zijlstra --- kernel/sched/numa.c | 41 ++++++++++++++++++++++++++++++++++++----- 1 file changed, 36 insertions(+), 5 deletions(-) --- a/kernel/sched/numa.c +++ b/kernel/sched/numa.c @@ -219,7 +219,9 @@ static u64 process_cpu_runtime(struct nu rcu_read_lock(); t = p = ne_owner(ne); if (p) do { - runtime += t->se.sum_exec_runtime; // @#$#@ 32bit + u64 tmp = t->se.sum_exec_runtime; + trace_printk("pid: %d ran: %llu ns\n", t->pid, tmp); + runtime += tmp; // @#$#@ 32bit } while ((t = next_thread(t)) != p); rcu_read_unlock(); @@ -518,7 +520,8 @@ static void update_node_load(struct node * If there was NUMA_FOREIGN load, that means this node was at its * maximum memory capacity, record that. */ - set_max_mem_load(node_pages_load(nq->node)); + set_max_mem_load(node_pages_load(nq->node) + + node_page_state(nq->node, NR_FREE_PAGES)); } enum numa_balance_type { @@ -556,6 +559,10 @@ static int find_busiest_node(int this_no cpu_load = nq->remote_cpu_load; mem_load = nq->remote_mem_load; + trace_printk("node_load(%d/%d): cpu: %ld, mem: %ld abs_cpu: %ld abs_mem: %ld\n", + node, this_node, cpu_load, mem_load, + nq->cpu_load, node_pages_load(node)); + /* * If this node is overloaded on memory, we don't want more * tasks, bail! @@ -580,6 +587,12 @@ static int find_busiest_node(int this_no } } + trace_printk("cpu_node: %d, cpu_load: %ld, mem_load: %ld, sum_cpu_load: %ld\n", + cpu_node, max_cpu_load, cpu_mem_load, sum_cpu_load); + + trace_printk("mem_node: %d, cpu_load: %ld, mem_load: %ld, sum_mem_load: %ld\n", + mem_node, mem_cpu_load, max_mem_load, sum_mem_load); + /* * Nobody had overload of any kind, cool we're done! */ @@ -626,6 +639,9 @@ static int find_busiest_node(int this_no imb->mem = (long)(node_pages_load(node) - imb->mem_load) / 2; } + trace_printk("busiest_node: %d, cpu_imb: %ld, mem_imb: %ld, type: %d\n", + node, imb->cpu, imb->mem, imb->type); + return node; } @@ -663,6 +679,9 @@ static void move_processes(struct node_q struct numa_entity, numa_entry); + trace_printk("numa_migrate(%d <- %d): candidate: %p\n", + this_nq->node, busiest_nq->node, ne); + ne_cpu = ne->nops->cpu_load(ne); ne_mem = ne->nops->mem_load(ne); @@ -672,20 +691,27 @@ static void move_processes(struct node_q * on the other end. */ if ((imb->type & NUMA_BALANCE_CPU) && - imb->cpu - cpu_moved < ne_cpu / 2) + imb->cpu - cpu_moved < ne_cpu / 2) { + trace_printk("fail cpu: %ld %ld %ld\n", imb->cpu, cpu_moved, ne_cpu); goto next; + } /* * Avoid migrating ne's when we'll know we'll push our * node over the memory limit. */ if (max_mem_load && - imb->mem_load + mem_moved + ne_mem > max_mem_load) + imb->mem_load + mem_moved + ne_mem > max_mem_load) { + trace_printk("fail mem: %ld %ld %ld %ld\n", + imb->mem_load, mem_moved, ne_mem, max_mem_load); goto next; + } } - if (!can_move_ne(ne)) + if (!can_move_ne(ne)) { + trace_printk("%p, can_move_ne() fail\n", ne); goto next; + } __dequeue_ne(busiest_nq, ne); __enqueue_ne(this_nq, ne); @@ -702,6 +728,11 @@ static void move_processes(struct node_q cpu_moved += ne_cpu; mem_moved += ne_mem; + trace_printk("numa_migrate(%d <- %d): cpu_load: %ld mem_load: %ld, " + "cpu_moved: %ld, mem_moved: %ld\n", + this_nq->node, busiest_nq->node, + ne_cpu, ne_mem, cpu_moved, mem_moved); + if (imb->cpu - cpu_moved <= 0 && imb->mem - mem_moved <= 0) break; -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/