[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <4968D79C.10602@sgi.com>
Date: Sat, 10 Jan 2009 09:15:08 -0800
From: Mike Travis <travis@....com>
To: Ingo Molnar <mingo@...e.hu>
CC: Rusty Russell <rusty@...tcorp.com.au>, Jes Sorensen <jes@....com>,
Jack Steiner <steiner@....com>,
LKML <linux-kernel@...r.kernel.org>
Subject: Re: bad patch in sched.c
Ingo Molnar wrote:
> * Rusty Russell <rusty@...tcorp.com.au> wrote:
>
>> cpumask: fix CONFIG_NUMA=y sched.c
>
> doesnt build:
>
> kernel/sched.c: In function ‘__build_sched_domains’:
> kernel/sched.c:7701: error: ‘struct static_sched_domain’ has no member named ‘sg’
>
> Ingo
Hi,
This one does compile and fix it.
Thanks,
Mike
---
cpumask: fix CONFIG_NUMA=y sched.c
struct sched_domain is now a dangling structure; where we really want
static ones, we need to use static_sched_domain.
(As the FIXME in this file says, cpumask_var_t would be better, but
this code is hairy enough without trying to add initialization code to
the right places).
Signed-off-by: Rusty Russell <rusty@...tcorp.com.au>
---
kernel/sched.c | 10 +++++-----
1 file changed, 5 insertions(+), 5 deletions(-)
--- linux-2.8.orig/kernel/sched.c
+++ linux-2.8/kernel/sched.c
@@ -7278,10 +7278,10 @@ cpu_to_phys_group(int cpu, const struct
* groups, so roll our own. Now each node has its own list of groups which
* gets dynamically allocated.
*/
-static DEFINE_PER_CPU(struct sched_domain, node_domains);
+static DEFINE_PER_CPU(struct static_sched_domain, node_domains);
static struct sched_group ***sched_group_nodes_bycpu;
-static DEFINE_PER_CPU(struct sched_domain, allnodes_domains);
+static DEFINE_PER_CPU(struct static_sched_domain, allnodes_domains);
static DEFINE_PER_CPU(struct static_sched_group, sched_group_allnodes);
static int cpu_to_allnodes_group(int cpu, const struct cpumask *cpu_map,
@@ -7556,7 +7556,7 @@ static int __build_sched_domains(const s
#ifdef CONFIG_NUMA
if (cpumask_weight(cpu_map) >
SD_NODES_PER_DOMAIN*cpumask_weight(nodemask)) {
- sd = &per_cpu(allnodes_domains, i);
+ sd = &per_cpu(allnodes_domains, i).sd;
SD_INIT(sd, ALLNODES);
set_domain_attribute(sd, attr);
cpumask_copy(sched_domain_span(sd), cpu_map);
@@ -7566,7 +7566,7 @@ static int __build_sched_domains(const s
} else
p = NULL;
- sd = &per_cpu(node_domains, i);
+ sd = &per_cpu(node_domains, i).sd;
SD_INIT(sd, NODE);
set_domain_attribute(sd, attr);
sched_domain_node_span(cpu_to_node(i), sched_domain_span(sd));
@@ -7684,7 +7684,7 @@ static int __build_sched_domains(const s
for_each_cpu(j, nodemask) {
struct sched_domain *sd;
- sd = &per_cpu(node_domains, j);
+ sd = &per_cpu(node_domains, j).sd;
sd->groups = sg;
}
sg->__cpu_power = 0;
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists