[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20250625191108.1646208-8-sshegde@linux.ibm.com>
Date: Thu, 26 Jun 2025 00:41:06 +0530
From: Shrikanth Hegde <sshegde@...ux.ibm.com>
To: mingo@...hat.com, peterz@...radead.org, juri.lelli@...hat.com,
vincent.guittot@...aro.org, tglx@...utronix.de, yury.norov@...il.com,
maddy@...ux.ibm.com
Cc: sshegde@...ux.ibm.com, vschneid@...hat.com, dietmar.eggemann@....com,
rostedt@...dmis.org, kprateek.nayak@....com, huschle@...ux.ibm.com,
srikar@...ux.ibm.com, linux-kernel@...r.kernel.org,
christophe.leroy@...roup.eu, linuxppc-dev@...ts.ozlabs.org,
gregkh@...uxfoundation.org
Subject: [RFC v2 7/9] sched: Add static key check for cpu_avoid
Checking if a CPU is avoid can add a slight overhead and should be
done only when necessary.
Add a static key check which makes it almost nop when key is false.
Arch needs to set the key when it decides to. Refer to debug patch
for example.
Signed-off-by: Shrikanth Hegde <sshegde@...ux.ibm.com>
---
This method avoids additional ifdefs. So kept it that way instead of
CONFIG_PARAVIRT.
Added a helper function for cpu_avoid, since including sched.h fails in
cpumask.h
kernel/sched/core.c | 8 ++++----
kernel/sched/fair.c | 5 +++--
kernel/sched/rt.c | 8 ++++----
kernel/sched/sched.h | 9 +++++++++
4 files changed, 20 insertions(+), 10 deletions(-)
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index aea4232e3ec4..51426b17ef55 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -148,9 +148,9 @@ __read_mostly int sysctl_resched_latency_warn_once = 1;
* Limited because this is done with IRQs disabled.
*/
__read_mostly unsigned int sysctl_sched_nr_migrate = SCHED_NR_MIGRATE_BREAK;
-
__read_mostly int scheduler_running;
+DEFINE_STATIC_KEY_FALSE(paravirt_cpu_avoid_enabled);
#ifdef CONFIG_SCHED_CORE
DEFINE_STATIC_KEY_FALSE(__sched_core_enabled);
@@ -2438,7 +2438,7 @@ static inline bool is_cpu_allowed(struct task_struct *p, int cpu)
return false;
/* CPU marked as avoid, shouldn't chosen to run any task*/
- if (cpu_avoid(cpu))
+ if (cpu_avoid_check(cpu))
return false;
/* But are allowed during online. */
@@ -5578,7 +5578,7 @@ void sched_tick(void)
sched_clock_tick();
/* push the current task out if cpu is marked as avoid */
- if (cpu_avoid(cpu))
+ if (cpu_avoid_check(cpu))
push_current_task(rq);
rq_lock(rq, &rf);
@@ -8048,7 +8048,7 @@ void push_current_task(struct rq *rq)
unsigned long flags;
/* idle task can't be pused out */
- if (rq->curr == rq->idle || !cpu_avoid(rq->cpu))
+ if (rq->curr == rq->idle || !cpu_avoid_check(rq->cpu))
return;
/* Do for only SCHED_NORMAL AND RT for now */
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index 406288aef535..21370f76d61b 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -8547,7 +8547,7 @@ select_task_rq_fair(struct task_struct *p, int prev_cpu, int wake_flags)
rcu_read_unlock();
/* Don't select a CPU marked as avoid for wakeup */
- if (cpu_avoid(new_cpu))
+ if (cpu_avoid_check(new_cpu))
return cpu;
else
return new_cpu;
@@ -11668,7 +11668,8 @@ static int sched_balance_rq(int this_cpu, struct rq *this_rq,
cpumask_and(cpus, sched_domain_span(sd), cpu_active_mask);
/* Don't spread load into CPUs marked as avoid */
- cpumask_andnot(cpus, cpus, cpu_avoid_mask);
+ if (static_branch_unlikely(¶virt_cpu_avoid_enabled))
+ cpumask_andnot(cpus, cpus, cpu_avoid_mask);
schedstat_inc(sd->lb_count[idle]);
diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c
index fd9df6f46135..0ab3fdf7a637 100644
--- a/kernel/sched/rt.c
+++ b/kernel/sched/rt.c
@@ -1549,7 +1549,7 @@ select_task_rq_rt(struct task_struct *p, int cpu, int flags)
if (!test && target != -1 && !rt_task_fits_capacity(p, target))
goto out_unlock;
- if (cpu_avoid(target))
+ if (cpu_avoid_check(target))
goto out_unlock;
/*
* Don't bother moving it if the destination CPU is
@@ -1873,7 +1873,7 @@ static struct rq *find_lock_lowest_rq(struct task_struct *task, struct rq *rq)
for (tries = 0; tries < RT_MAX_TRIES; tries++) {
cpu = find_lowest_rq(task);
- if ((cpu == -1) || (cpu == rq->cpu) || cpu_avoid(cpu))
+ if ((cpu == -1) || (cpu == rq->cpu) || cpu_avoid_check(cpu))
break;
lowest_rq = cpu_rq(cpu);
@@ -1971,7 +1971,7 @@ static int push_rt_task(struct rq *rq, bool pull)
return 0;
cpu = find_lowest_rq(rq->curr);
- if (cpu == -1 || cpu == rq->cpu || cpu_avoid(cpu))
+ if (cpu == -1 || cpu == rq->cpu || cpu_avoid_check(cpu))
return 0;
/*
@@ -2234,7 +2234,7 @@ static void pull_rt_task(struct rq *this_rq)
if (likely(!rt_overload_count))
return;
- if (cpu_avoid(this_rq->cpu))
+ if (cpu_avoid_check(this_rq->cpu))
return;
/*
diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
index b9614873762e..707fdfa46772 100644
--- a/kernel/sched/sched.h
+++ b/kernel/sched/sched.h
@@ -1710,6 +1710,15 @@ struct rq_flags {
extern struct balance_callback balance_push_callback;
void push_current_task(struct rq *rq);
+DECLARE_STATIC_KEY_FALSE(paravirt_cpu_avoid_enabled);
+
+static inline bool cpu_avoid_check(int cpu)
+{
+ if (static_branch_unlikely(¶virt_cpu_avoid_enabled))
+ return cpu_avoid(cpu);
+
+ return false;
+}
#ifdef CONFIG_SCHED_CLASS_EXT
extern const struct sched_class ext_sched_class;
--
2.43.0
Powered by blists - more mailing lists