[<prev] [next>] [day] [month] [year] [list]
Message-ID: <tip-c186fafe9aba87c1a93df8c7120a6ae01fe435ad@git.kernel.org>
Date: Wed, 23 Feb 2011 12:03:58 GMT
From: tip-bot for Peter Zijlstra <a.p.zijlstra@...llo.nl>
To: linux-tip-commits@...r.kernel.org
Cc: linux-kernel@...r.kernel.org, hpa@...or.com, mingo@...hat.com,
ncrao@...gle.com, a.p.zijlstra@...llo.nl, efault@....de,
suresh.b.siddha@...el.com, tglx@...utronix.de, mingo@...e.hu,
venki@...gle.com
Subject: [tip:sched/core] sched: Clean up remnants of sd_idle
Commit-ID: c186fafe9aba87c1a93df8c7120a6ae01fe435ad
Gitweb: http://git.kernel.org/tip/c186fafe9aba87c1a93df8c7120a6ae01fe435ad
Author: Peter Zijlstra <a.p.zijlstra@...llo.nl>
AuthorDate: Mon, 21 Feb 2011 18:52:53 +0100
Committer: Ingo Molnar <mingo@...e.hu>
CommitDate: Wed, 23 Feb 2011 11:33:55 +0100
sched: Clean up remnants of sd_idle
With the wholesale removal of the sd_idle SMT logic we can clean up
some more.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@...llo.nl>
Cc: Nikhil Rao <ncrao@...gle.com>
Cc: Venkatesh Pallipadi <venki@...gle.com>
Cc: Suresh Siddha <suresh.b.siddha@...el.com>
Cc: Mike Galbraith <efault@....de>
LKML-Reference: <new-submission>
Signed-off-by: Ingo Molnar <mingo@...e.hu>
---
kernel/sched_fair.c | 23 ++++++++++-------------
1 files changed, 10 insertions(+), 13 deletions(-)
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c
index d384e73..cd18600 100644
--- a/kernel/sched_fair.c
+++ b/kernel/sched_fair.c
@@ -3150,25 +3150,23 @@ find_busiest_group(struct sched_domain *sd, int this_cpu,
if (sds.this_load >= sds.avg_load)
goto out_balanced;
- /*
- * In the CPU_NEWLY_IDLE, use imbalance_pct to be conservative.
- * And to check for busy balance use !idle_cpu instead of
- * CPU_NOT_IDLE. This is because HT siblings will use CPU_NOT_IDLE
- * even when they are idle.
- */
- if (idle == CPU_NEWLY_IDLE || !idle_cpu(this_cpu)) {
- if (100 * sds.max_load <= sd->imbalance_pct * sds.this_load)
- goto out_balanced;
- } else {
+ if (idle == CPU_IDLE) {
/*
* This cpu is idle. If the busiest group load doesn't
* have more tasks than the number of available cpu's and
* there is no imbalance between this and busiest group
* wrt to idle cpu's, it is balanced.
*/
- if ((sds.this_idle_cpus <= sds.busiest_idle_cpus + 1) &&
+ if ((sds.this_idle_cpus <= sds.busiest_idle_cpus + 1) &&
sds.busiest_nr_running <= sds.busiest_group_weight)
goto out_balanced;
+ } else {
+ /*
+ * In the CPU_NEWLY_IDLE, CPU_NOT_IDLE cases, use
+ * imbalance_pct to be conservative.
+ */
+ if (100 * sds.max_load <= sd->imbalance_pct * sds.this_load)
+ goto out_balanced;
}
force_balance:
@@ -3862,8 +3860,7 @@ static void rebalance_domains(int cpu, enum cpu_idle_type idle)
if (load_balance(cpu, rq, sd, idle, &balance)) {
/*
* We've pulled tasks over so either we're no
- * longer idle, or one of our SMT siblings is
- * not idle.
+ * longer idle.
*/
idle = CPU_NOT_IDLE;
}
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists