[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-ID: <4A5C4626.8000000@cn.fujitsu.com>
Date: Tue, 14 Jul 2009 16:47:34 +0800
From: Lai Jiangshan <laijs@...fujitsu.com>
To: Andrew Morton <akpm@...ux-foundation.org>,
Rusty Russell <rusty@...tcorp.com.au>,
LKML <linux-kernel@...r.kernel.org>
Subject: [PATCH] cpu_hotplug: don't affect current task's affinity
_cpu_down() changes current task's affinity and then
recovers it at the end.
It brings two defects:
1) The recovering will failed in some condition.
# grep Cpus_allowed_list /proc/$$/status
Cpus_allowed_list: 0-3
# taskset -pc 2 $$
pid 29075's current affinity list: 0-3
pid 29075's new affinity list: 2
# grep Cpus_allowed_list /proc/$$/status
Cpus_allowed_list: 2
# echo 0 > /sys/devices/system/cpu/cpu2/online
# grep Cpus_allowed_list /proc/$$/status
Cpus_allowed_list: 0
In linux, tasks' "Cpus_allowed_list" which are "2" originally
will become "0-1,3" after the cpu#2 is offlined.
This "Cpus_allowed_list: 0" is suspicionful.
2) current task is a userspace task, the user may change
its cpu-affinity at the same time. The user may get unexpected
result if _cpu_down() changes current task's affinity.
Actually, we don't have to change the affinity.
We create a kernel thread to do the works.
Signed-off-by: Lai Jiangshan <laijs@...fujitsu.com>
---
diff --git a/kernel/cpu.c b/kernel/cpu.c
index 8ce1004..901caeb 100644
--- a/kernel/cpu.c
+++ b/kernel/cpu.c
@@ -162,15 +162,17 @@ static inline void check_for_tasks(int cpu)
write_unlock_irq(&tasklist_lock);
}
-struct take_cpu_down_param {
+struct cpu_down_param {
unsigned long mod;
- void *hcpu;
+ unsigned int cpu;
+ int ret;
+ struct completion done;
};
/* Take this CPU down. */
static int __ref take_cpu_down(void *_param)
{
- struct take_cpu_down_param *param = _param;
+ struct cpu_down_param *param = _param;
int err;
/* Ensure this CPU doesn't handle any more interrupts. */
@@ -179,7 +181,7 @@ static int __ref take_cpu_down(void *_param)
return err;
raw_notifier_call_chain(&cpu_chain, CPU_DYING | param->mod,
- param->hcpu);
+ (void *)(long)param->cpu);
/* Force idle task to run as soon as we yield: it should
immediately notice cpu is offline and die quickly. */
@@ -187,26 +189,13 @@ static int __ref take_cpu_down(void *_param)
return 0;
}
-/* Requires cpu_add_remove_lock to be held */
-static int __ref _cpu_down(unsigned int cpu, int tasks_frozen)
+static int __ref _cpu_down_thread(void *_param)
{
+ struct cpu_down_param *param = _param;
int err, nr_calls = 0;
- cpumask_var_t old_allowed;
+ unsigned long mod = param->mod;
+ unsigned int cpu = param->cpu;
void *hcpu = (void *)(long)cpu;
- unsigned long mod = tasks_frozen ? CPU_TASKS_FROZEN : 0;
- struct take_cpu_down_param tcd_param = {
- .mod = mod,
- .hcpu = hcpu,
- };
-
- if (num_online_cpus() == 1)
- return -EBUSY;
-
- if (!cpu_online(cpu))
- return -EINVAL;
-
- if (!alloc_cpumask_var(&old_allowed, GFP_KERNEL))
- return -ENOMEM;
cpu_hotplug_begin();
err = __raw_notifier_call_chain(&cpu_chain, CPU_DOWN_PREPARE | mod,
@@ -222,18 +211,16 @@ static int __ref _cpu_down(unsigned int cpu, int tasks_frozen)
}
/* Ensure that we are not runnable on dying cpu */
- cpumask_copy(old_allowed, ¤t->cpus_allowed);
- set_cpus_allowed_ptr(current,
- cpumask_of(cpumask_any_but(cpu_online_mask, cpu)));
+ set_cpus_allowed_ptr(current, cpu_active_mask);
- err = __stop_machine(take_cpu_down, &tcd_param, cpumask_of(cpu));
+ err = __stop_machine(take_cpu_down, param, cpumask_of(cpu));
if (err) {
/* CPU didn't die: tell everyone. Can't complain. */
if (raw_notifier_call_chain(&cpu_chain, CPU_DOWN_FAILED | mod,
hcpu) == NOTIFY_BAD)
BUG();
- goto out_allowed;
+ goto out_release;
}
BUG_ON(cpu_online(cpu));
@@ -251,8 +238,6 @@ static int __ref _cpu_down(unsigned int cpu, int tasks_frozen)
check_for_tasks(cpu);
-out_allowed:
- set_cpus_allowed_ptr(current, old_allowed);
out_release:
cpu_hotplug_done();
if (!err) {
@@ -260,8 +245,35 @@ out_release:
hcpu) == NOTIFY_BAD)
BUG();
}
- free_cpumask_var(old_allowed);
- return err;
+ param->ret = err;
+ complete(¶m->done);
+
+ return 0;
+}
+
+/* Requires cpu_add_remove_lock to be held */
+static int __ref _cpu_down(unsigned int cpu, int tasks_frozen)
+{
+ struct task_struct *k;
+ struct cpu_down_param param = {
+ .mod = tasks_frozen ? CPU_TASKS_FROZEN : 0,
+ .cpu = cpu,
+ .ret = 0,
+ };
+
+ if (num_online_cpus() == 1)
+ return -EBUSY;
+
+ if (!cpu_online(cpu))
+ return -EINVAL;
+
+ init_completion(¶m.done);
+ k = kthread_run(_cpu_down_thread, ¶m, "kcpu_down");
+ if (IS_ERR(k))
+ return PTR_ERR(k);
+ wait_for_completion(¶m.done);
+
+ return param.ret;
}
int __ref cpu_down(unsigned int cpu)
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists