[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Message-Id: <1390611565-18709-1-git-send-email-yinghai@kernel.org>
Date: Fri, 24 Jan 2014 16:59:25 -0800
From: Yinghai Lu <yinghai@...nel.org>
To: "H. Peter Anvin" <hpa@...or.com>
Cc: Thomas Gleixner <tglx@...utronix.de>, Ingo Molnar <mingo@...e.hu>,
linux-kernel@...r.kernel.org, Yinghai Lu <yinghai@...nel.org>,
Prarit Bhargava <prarit@...hat.com>
Subject: [PATCH -v2] x86: allocate cpumask during check irq vectors
Fix warning:
arch/x86/kernel/irq.c: In function check_irq_vectors_for_cpu_disable:
arch/x86/kernel/irq.c:337:1: warning: the frame size of 2052 bytes is larger than 2048 bytes
when NR_CPUS=8192
We should use zalloc_cpumask_var() instead.
-v2: update to GFP_ATOMIC instead and free the allocated cpumask at last.
Signed-off-by: Yinghai Lu <yinghai@...nel.org>
Cc: Prarit Bhargava <prarit@...hat.com>
---
arch/x86/kernel/irq.c | 24 +++++++++++++++++-------
1 file changed, 17 insertions(+), 7 deletions(-)
Index: linux-2.6/arch/x86/kernel/irq.c
===================================================================
--- linux-2.6.orig/arch/x86/kernel/irq.c
+++ linux-2.6/arch/x86/kernel/irq.c
@@ -277,11 +277,18 @@ int check_irq_vectors_for_cpu_disable(vo
unsigned int this_cpu, vector, this_count, count;
struct irq_desc *desc;
struct irq_data *data;
- struct cpumask affinity_new, online_new;
+ cpumask_var_t affinity_new, online_new;
+
+ if (!alloc_cpumask_var(&affinity_new, GFP_ATOMIC))
+ return -ENOMEM;
+ if (!alloc_cpumask_var(&online_new, GFP_ATOMIC)) {
+ free_cpumask_var(affinity_new);
+ return -ENOMEM;
+ }
this_cpu = smp_processor_id();
- cpumask_copy(&online_new, cpu_online_mask);
- cpu_clear(this_cpu, online_new);
+ cpumask_copy(online_new, cpu_online_mask);
+ cpumask_clear_cpu(this_cpu, online_new);
this_count = 0;
for (vector = FIRST_EXTERNAL_VECTOR; vector < NR_VECTORS; vector++) {
@@ -289,8 +296,8 @@ int check_irq_vectors_for_cpu_disable(vo
if (irq >= 0) {
desc = irq_to_desc(irq);
data = irq_desc_get_irq_data(desc);
- cpumask_copy(&affinity_new, data->affinity);
- cpu_clear(this_cpu, affinity_new);
+ cpumask_copy(affinity_new, data->affinity);
+ cpumask_clear_cpu(this_cpu, affinity_new);
/* Do not count inactive or per-cpu irqs. */
if (!irq_has_action(irq) || irqd_is_per_cpu(data))
@@ -311,12 +318,15 @@ int check_irq_vectors_for_cpu_disable(vo
* mask is not zero; that is the down'd cpu is the
* last online cpu in a user set affinity mask.
*/
- if (cpumask_empty(&affinity_new) ||
- !cpumask_subset(&affinity_new, &online_new))
+ if (cpumask_empty(affinity_new) ||
+ !cpumask_subset(affinity_new, online_new))
this_count++;
}
}
+ free_cpumask_var(affinity_new);
+ free_cpumask_var(online_new);
+
count = 0;
for_each_online_cpu(cpu) {
if (cpu == this_cpu)
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists