lists.openwall.net | lists / announce owl-users owl-dev john-users john-dev passwdqc-users yescrypt popa3d-users / oss-security kernel-hardening musl sabotage tlsify passwords / crypt-dev xvendor / Bugtraq Full-Disclosure linux-kernel linux-netdev linux-ext4 linux-hardening linux-cve-announce PHC | |
Open Source and information security mailing list archives
| ||
|
Date: Mon, 12 May 2008 14:14:10 -0400 From: Gregory Haskins <ghaskins@...ell.com> To: Ingo Molnar <mingo@...e.hu>, Peter Zijlstra <peterz@...radead.org> Cc: Srivatsa Vaddagiri <vatsa@...ux.vnet.ibm.com>, Gregory Haskins <ghaskins@...ell.com>, linux-kernel@...r.kernel.org Subject: [RFC PATCH 2/3] sched: pass specific sched_balancer to functions Instead of assuming sd->group_balancer.groups is the intended group. We will use this later in the series to introduce alternate balancers. Signed-off-by: Gregory Haskins <ghaskins@...ell.com> --- kernel/sched.c | 21 ++++++++++++--------- 1 files changed, 12 insertions(+), 9 deletions(-) diff --git a/kernel/sched.c b/kernel/sched.c index a8e0bd3..0bdbfe6 100644 --- a/kernel/sched.c +++ b/kernel/sched.c @@ -3288,11 +3288,12 @@ static int move_one_task(struct rq *this_rq, int this_cpu, struct rq *busiest, * should be moved to restore balance via the imbalance parameter. */ static struct sched_group * -find_busiest_group(struct sched_domain *sd, int this_cpu, +find_busiest_group(struct sched_domain *sd, struct sched_balancer *balancer, + int this_cpu, unsigned long *imbalance, enum cpu_idle_type idle, int *sd_idle, const cpumask_t *cpus, int *balance) { - struct sched_group *first_group = sd->group_balancer.groups; + struct sched_group *first_group = balancer->groups; struct sched_group *busiest = NULL, *this = NULL, *group = first_group; unsigned long max_load, avg_load, total_load, this_load, total_pwr; unsigned long max_pull; @@ -3630,10 +3631,11 @@ find_busiest_queue(struct sched_group *group, enum cpu_idle_type idle, * tasks if there is an imbalance. */ static int load_balance(int this_cpu, struct rq *this_rq, - struct sched_domain *sd, enum cpu_idle_type idle, + struct sched_domain *sd, + struct sched_balancer *balancer, + enum cpu_idle_type idle, int *balance, cpumask_t *cpus) { - struct sched_balancer *balancer = &sd->group_balancer; int ld_moved, all_pinned = 0, active_balance = 0, sd_idle = 0; struct sched_group *group; unsigned long imbalance; @@ -3658,8 +3660,8 @@ static int load_balance(int this_cpu, struct rq *this_rq, schedstat_inc(sd, lb_count[idle]); redo: - group = find_busiest_group(sd, this_cpu, &imbalance, idle, &sd_idle, - cpus, balance); + group = find_busiest_group(sd, balancer, this_cpu, &imbalance, idle, + &sd_idle, cpus, balance); if (*balance == 0) goto out_balanced; @@ -3819,8 +3821,8 @@ load_balance_newidle(int this_cpu, struct rq *this_rq, struct sched_domain *sd, schedstat_inc(sd, lb_count[CPU_NEWLY_IDLE]); redo: - group = find_busiest_group(sd, this_cpu, &imbalance, CPU_NEWLY_IDLE, - &sd_idle, cpus, NULL); + group = find_busiest_group(sd, balancer, this_cpu, &imbalance, + CPU_NEWLY_IDLE, &sd_idle, cpus, NULL); if (!group) { schedstat_inc(sd, lb_nobusyg[CPU_NEWLY_IDLE]); goto out_balanced; @@ -4082,7 +4084,8 @@ static void rebalance_domains(int cpu, enum cpu_idle_type idle) } if (time_after_eq(jiffies, balancer->last_exec + interval)) { - if (load_balance(cpu, rq, sd, idle, &balance, &tmp)) { + if (load_balance(cpu, rq, sd, balancer, + idle, &balance, &tmp)) { /* * We've pulled tasks over so either we're no * longer idle, or one of our SMT siblings is -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@...r.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists