The scheduler sports quite a bunch of hotplug notifiers. One reason for multiple notifiers is the fact, that the startup and teardown process are asymetric. Now the scheduler wants to be called early on startup and late on teardown. That requires to install two different notifiers for the same issue. With the state machine implementation we can register a callback pair for startup and teardown at the appropriate spot. This patch converts the notifiers which are setup with special priorities and combines CPU_PRI_SCHED and CPU_PRI_CPUSET notifiers to a single callback. They run back to back anyway and we can make sure in the callbacks that the ordering inside the scheduler is correct. These notifiers are installed in sched_init_smp() as we can't run them during the bringup of the non boot cpus because the smp scheduler is setup after that. It would be nice if we just could compile them in, but that needs a larger surgery to the scheduler code and is beyond the scope of this patch. Signed-off-by: Thomas Gleixner --- include/linux/cpu.h | 16 ---- include/linux/cpuhotplug.h | 6 + kernel/cpu.c | 4 + kernel/sched/core.c | 154 +++++++++++++++++---------------------------- 4 files changed, 69 insertions(+), 111 deletions(-) Index: linux-2.6/include/linux/cpu.h =================================================================== --- linux-2.6.orig/include/linux/cpu.h +++ linux-2.6/include/linux/cpu.h @@ -58,22 +58,6 @@ extern ssize_t arch_print_cpu_modalias(s * CPU notifier priorities. */ enum { - /* - * SCHED_ACTIVE marks a cpu which is coming up active during - * CPU_ONLINE and CPU_DOWN_FAILED and must be the first - * notifier. CPUSET_ACTIVE adjusts cpuset according to - * cpu_active mask right after SCHED_ACTIVE. During - * CPU_DOWN_PREPARE, SCHED_INACTIVE and CPUSET_INACTIVE are - * ordered in the similar way. - * - * This ordering guarantees consistent cpu_active mask and - * migration behavior to all cpu notifiers. - */ - CPU_PRI_SCHED_ACTIVE = INT_MAX, - CPU_PRI_CPUSET_ACTIVE = INT_MAX - 1, - CPU_PRI_SCHED_INACTIVE = INT_MIN + 1, - CPU_PRI_CPUSET_INACTIVE = INT_MIN, - /* migration should happen before other stuff but after perf */ CPU_PRI_PERF = 20, CPU_PRI_MIGRATION = 10, Index: linux-2.6/include/linux/cpuhotplug.h =================================================================== --- linux-2.6.orig/include/linux/cpuhotplug.h +++ linux-2.6/include/linux/cpuhotplug.h @@ -6,13 +6,16 @@ enum cpuhp_states { CPUHP_CREATE_THREADS, CPUHP_NOTIFY_PREPARE, CPUHP_NOTIFY_DEAD, + CPUHP_SCHED_DEAD, CPUHP_BRINGUP_CPU, CPUHP_AP_OFFLINE, + CPUHP_AP_SCHED_STARTING, CPUHP_AP_NOTIFY_STARTING, CPUHP_AP_NOTIFY_DYING, CPUHP_AP_MAX, CPUHP_TEARDOWN_CPU, CPUHP_PERCPU_THREADS, + CPUHP_SCHED_ONLINE, CPUHP_NOTIFY_ONLINE, CPUHP_NOTIFY_DOWN_PREPARE, CPUHP_MAX, @@ -87,4 +90,7 @@ static inline void cpuhp_remove_state_no __cpuhp_remove_state(state, false); } +/* Compiled in scheduler hotplug functions */ +int sched_starting_cpu(unsigned int cpu); + #endif Index: linux-2.6/kernel/cpu.c =================================================================== --- linux-2.6.orig/kernel/cpu.c +++ linux-2.6/kernel/cpu.c @@ -788,6 +788,10 @@ static struct cpuhp_step cpuhp_bp_states /* Application processor state steps */ static struct cpuhp_step cpuhp_ap_states[] = { #ifdef CONFIG_SMP + [CPUHP_AP_SCHED_STARTING] = { + .startup = sched_starting_cpu, + .teardown = NULL, + }, [CPUHP_AP_NOTIFY_STARTING] = { .startup = notify_starting, .teardown = NULL, Index: linux-2.6/kernel/sched/core.c =================================================================== --- linux-2.6.orig/kernel/sched/core.c +++ linux-2.6/kernel/sched/core.c @@ -5167,31 +5167,6 @@ static struct notifier_block __cpuinitda .priority = CPU_PRI_MIGRATION, }; -static int __cpuinit sched_cpu_active(struct notifier_block *nfb, - unsigned long action, void *hcpu) -{ - switch (action & ~CPU_TASKS_FROZEN) { - case CPU_STARTING: - case CPU_DOWN_FAILED: - set_cpu_active((long)hcpu, true); - return NOTIFY_OK; - default: - return NOTIFY_DONE; - } -} - -static int __cpuinit sched_cpu_inactive(struct notifier_block *nfb, - unsigned long action, void *hcpu) -{ - switch (action & ~CPU_TASKS_FROZEN) { - case CPU_DOWN_PREPARE: - set_cpu_active((long)hcpu, false); - return NOTIFY_OK; - default: - return NOTIFY_DONE; - } -} - static int __init migration_init(void) { void *cpu = (void *)(long)smp_processor_id(); @@ -5203,10 +5178,6 @@ static int __init migration_init(void) migration_call(&migration_notifier, CPU_ONLINE, cpu); register_cpu_notifier(&migration_notifier); - /* Register cpu active notifiers */ - cpu_notifier(sched_cpu_active, CPU_PRI_SCHED_ACTIVE); - cpu_notifier(sched_cpu_inactive, CPU_PRI_SCHED_INACTIVE); - return 0; } early_initcall(migration_init); @@ -6292,42 +6263,12 @@ static void sched_domains_numa_masks_cle } } -/* - * Update sched_domains_numa_masks[level][node] array when new cpus - * are onlined. - */ -static int sched_domains_numa_masks_update(struct notifier_block *nfb, - unsigned long action, - void *hcpu) -{ - int cpu = (long)hcpu; - - switch (action & ~CPU_TASKS_FROZEN) { - case CPU_ONLINE: - sched_domains_numa_masks_set(cpu); - break; - - case CPU_DEAD: - sched_domains_numa_masks_clear(cpu); - break; - - default: - return NOTIFY_DONE; - } - - return NOTIFY_OK; -} #else -static inline void sched_init_numa(void) -{ -} - -static int sched_domains_numa_masks_update(struct notifier_block *nfb, - unsigned long action, - void *hcpu) -{ - return 0; -} +static inline void sched_init_numa(void) { } +#ifdef CONFIG_HOTPLUG_CPU +static void sched_domains_numa_masks_set(int cpu) { } +static void sched_domains_numa_masks_clear(int cpu) { } +#endif #endif /* CONFIG_NUMA */ static int __sdt_alloc(const struct cpumask *cpu_map) @@ -6696,6 +6637,7 @@ match2: mutex_unlock(&sched_domains_mutex); } +#ifdef CONFIG_HOTPLUG_CPU static int num_cpus_frozen; /* used to mark begin/end of suspend/resume */ /* @@ -6706,13 +6648,9 @@ static int num_cpus_frozen; /* used to m * If we come here as part of a suspend/resume, don't touch cpusets because we * want to restore it back to its original state upon resume anyway. */ -static int cpuset_cpu_active(struct notifier_block *nfb, unsigned long action, - void *hcpu) +static void cpuset_cpu_active(void) { - switch (action) { - case CPU_ONLINE_FROZEN: - case CPU_DOWN_FAILED_FROZEN: - + if (cpuhp_tasks_frozen) { /* * num_cpus_frozen tracks how many CPUs are involved in suspend * resume sequence. As long as this is not the last online @@ -6722,40 +6660,62 @@ static int cpuset_cpu_active(struct noti num_cpus_frozen--; if (likely(num_cpus_frozen)) { partition_sched_domains(1, NULL, NULL); - break; + return; } - /* * This is the last CPU online operation. So fall through and * restore the original sched domains by considering the * cpuset configurations. */ - - case CPU_ONLINE: - case CPU_DOWN_FAILED: - cpuset_update_active_cpus(true); - break; - default: - return NOTIFY_DONE; } - return NOTIFY_OK; + cpuset_update_active_cpus(true); } -static int cpuset_cpu_inactive(struct notifier_block *nfb, unsigned long action, - void *hcpu) +static void cpuset_cpu_inactive(void) { - switch (action) { - case CPU_DOWN_PREPARE: - cpuset_update_active_cpus(false); - break; - case CPU_DOWN_PREPARE_FROZEN: + if (cpuhp_tasks_frozen) { num_cpus_frozen++; partition_sched_domains(1, NULL, NULL); - break; - default: - return NOTIFY_DONE; - } - return NOTIFY_OK; + } else + cpuset_update_active_cpus(false); +} + +static int sched_dead_cpu(unsigned int cpu) +{ + sched_domains_numa_masks_clear(cpu); + return 0; +} + +static int sched_online_cpu(unsigned int cpu) +{ + /* Looks redundant, but we need it in case of down canceled */ + set_cpu_active(cpu, true); + /* + * Asymetric to sched_dead_cpu, but this just fiddles with + * bits. Sigh + */ + sched_domains_numa_masks_set(cpu); + /* This is actually symetric */ + cpuset_cpu_active(); + return 0; +} + +static int sched_offline_cpu(unsigned int cpu) +{ + set_cpu_active(cpu, false); + cpuset_cpu_inactive(); + return 0; +} +#else +#define sched_dead_cpu NULL +#define sched_online_cpu NULL +#define sched_offline_cpu NULL +#endif + +int __cpuinit sched_starting_cpu(unsigned int cpu) +{ + set_cpu_active(cpu, true); + return 0; } void __init sched_init_smp(void) @@ -6776,9 +6736,13 @@ void __init sched_init_smp(void) mutex_unlock(&sched_domains_mutex); put_online_cpus(); - hotcpu_notifier(sched_domains_numa_masks_update, CPU_PRI_SCHED_ACTIVE); - hotcpu_notifier(cpuset_cpu_active, CPU_PRI_CPUSET_ACTIVE); - hotcpu_notifier(cpuset_cpu_inactive, CPU_PRI_CPUSET_INACTIVE); + /* + * Note: These callbacks are installed late because we init + * numa and sched domains after we brought up the cpus. + */ + cpuhp_setup_state_nocalls(CPUHP_SCHED_DEAD, NULL, sched_dead_cpu); + cpuhp_setup_state_nocalls(CPUHP_SCHED_ONLINE, sched_online_cpu, + sched_offline_cpu); /* RT runtime code needs to handle some hotplug events */ hotcpu_notifier(update_runtime, 0); -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/