[<prev] [next>] [day] [month] [year] [list]
Message-ID: <20250903150824.16960-1-yajun.deng@linux.dev>
Date: Wed, 3 Sep 2025 15:08:24 +0000
From: Yajun Deng <yajun.deng@...ux.dev>
To: mingo@...hat.com,
peterz@...radead.org,
juri.lelli@...hat.com,
vincent.guittot@...aro.org,
dietmar.eggemann@....com,
rostedt@...dmis.org,
bsegall@...gle.com,
mgorman@...e.de,
vschneid@...hat.com
Cc: linux-kernel@...r.kernel.org,
Yajun Deng <yajun.deng@...ux.dev>
Subject: [PATCH] sched/rt: pass rq and task_struct pointers to update_stats functions
These __update_stats functions only require the rq and task_struct
parameters. All update_stats functions convert the rt_rq to rq and the
rt_se to task_struct. And all callers of update_stats will get the
rt_rq from the rq and the rt_se from the task_struct.
The rq_of_rt_rq() always returns the top-level rq, which is unique for
each CPU. The rt_task_of() is only available for the task_struct. They
are doing the convering and being converted. However, these conversions
don't change anything.
Just pass the rq and task_struct pointers to these update_stats functions.
If it don't have a task_struct pointer, call rt_entity_is_task() to ensure
the task isn't NULL.
Signed-off-by: Yajun Deng <yajun.deng@...ux.dev>
---
kernel/sched/rt.c | 91 ++++++++++-------------------------------------
1 file changed, 18 insertions(+), 73 deletions(-)
diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c
index 390f3d08abbe..feef4508efca 100644
--- a/kernel/sched/rt.c
+++ b/kernel/sched/rt.c
@@ -1212,107 +1212,56 @@ static void __delist_rt_entity(struct sched_rt_entity *rt_se, struct rt_prio_arr
rt_se->on_list = 0;
}
-static inline struct sched_statistics *
-__schedstats_from_rt_se(struct sched_rt_entity *rt_se)
-{
- /* schedstats is not supported for rt group. */
- if (!rt_entity_is_task(rt_se))
- return NULL;
-
- return &rt_task_of(rt_se)->stats;
-}
-
static inline void
-update_stats_wait_start_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se)
+update_stats_wait_start_rt(struct rq *rq, struct task_struct *p)
{
- struct sched_statistics *stats;
- struct task_struct *p = NULL;
-
if (!schedstat_enabled())
return;
- if (rt_entity_is_task(rt_se))
- p = rt_task_of(rt_se);
-
- stats = __schedstats_from_rt_se(rt_se);
- if (!stats)
- return;
-
- __update_stats_wait_start(rq_of_rt_rq(rt_rq), p, stats);
+ __update_stats_wait_start(rq, p, &p->stats);
}
static inline void
-update_stats_enqueue_sleeper_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se)
+update_stats_enqueue_sleeper_rt(struct rq *rq, struct task_struct *p)
{
- struct sched_statistics *stats;
- struct task_struct *p = NULL;
-
if (!schedstat_enabled())
return;
- if (rt_entity_is_task(rt_se))
- p = rt_task_of(rt_se);
-
- stats = __schedstats_from_rt_se(rt_se);
- if (!stats)
- return;
-
- __update_stats_enqueue_sleeper(rq_of_rt_rq(rt_rq), p, stats);
+ __update_stats_enqueue_sleeper(rq, p, &p->stats);
}
static inline void
-update_stats_enqueue_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se,
- int flags)
+update_stats_enqueue_rt(struct rq *rq, struct task_struct *p, int flags)
{
- if (!schedstat_enabled())
- return;
-
if (flags & ENQUEUE_WAKEUP)
- update_stats_enqueue_sleeper_rt(rt_rq, rt_se);
+ update_stats_enqueue_sleeper_rt(rq, p);
}
static inline void
-update_stats_wait_end_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se)
+update_stats_wait_end_rt(struct rq *rq, struct task_struct *p)
{
- struct sched_statistics *stats;
- struct task_struct *p = NULL;
-
if (!schedstat_enabled())
return;
- if (rt_entity_is_task(rt_se))
- p = rt_task_of(rt_se);
-
- stats = __schedstats_from_rt_se(rt_se);
- if (!stats)
- return;
-
- __update_stats_wait_end(rq_of_rt_rq(rt_rq), p, stats);
+ __update_stats_wait_end(rq, p, &p->stats);
}
static inline void
-update_stats_dequeue_rt(struct rt_rq *rt_rq, struct sched_rt_entity *rt_se,
- int flags)
+update_stats_dequeue_rt(struct rq *rq, struct task_struct *p, int flags)
{
- struct task_struct *p = NULL;
if (!schedstat_enabled())
return;
- if (rt_entity_is_task(rt_se))
- p = rt_task_of(rt_se);
-
if ((flags & DEQUEUE_SLEEP) && p) {
unsigned int state;
state = READ_ONCE(p->__state);
if (state & TASK_INTERRUPTIBLE)
- __schedstat_set(p->stats.sleep_start,
- rq_clock(rq_of_rt_rq(rt_rq)));
+ __schedstat_set(p->stats.sleep_start, rq_clock(rq));
if (state & TASK_UNINTERRUPTIBLE)
- __schedstat_set(p->stats.block_start,
- rq_clock(rq_of_rt_rq(rt_rq)));
+ __schedstat_set(p->stats.block_start, rq_clock(rq));
}
}
@@ -1392,7 +1341,8 @@ static void enqueue_rt_entity(struct sched_rt_entity *rt_se, unsigned int flags)
{
struct rq *rq = rq_of_rt_se(rt_se);
- update_stats_enqueue_rt(rt_rq_of_se(rt_se), rt_se, flags);
+ if (rt_entity_is_task(rt_se))
+ update_stats_enqueue_rt(rq, rt_task_of(rt_se), flags);
dequeue_rt_stack(rt_se, flags);
for_each_sched_rt_entity(rt_se)
@@ -1404,7 +1354,8 @@ static void dequeue_rt_entity(struct sched_rt_entity *rt_se, unsigned int flags)
{
struct rq *rq = rq_of_rt_se(rt_se);
- update_stats_dequeue_rt(rt_rq_of_se(rt_se), rt_se, flags);
+ if (rt_entity_is_task(rt_se))
+ update_stats_dequeue_rt(rq, rt_task_of(rt_se), flags);
dequeue_rt_stack(rt_se, flags);
@@ -1429,7 +1380,7 @@ enqueue_task_rt(struct rq *rq, struct task_struct *p, int flags)
rt_se->timeout = 0;
check_schedstat_required();
- update_stats_wait_start_rt(rt_rq_of_se(rt_se), rt_se);
+ update_stats_wait_start_rt(rq_of_rt_se(rt_se), p);
enqueue_rt_entity(rt_se, flags);
@@ -1631,12 +1582,9 @@ static void wakeup_preempt_rt(struct rq *rq, struct task_struct *p, int flags)
static inline void set_next_task_rt(struct rq *rq, struct task_struct *p, bool first)
{
- struct sched_rt_entity *rt_se = &p->rt;
- struct rt_rq *rt_rq = &rq->rt;
-
p->se.exec_start = rq_clock_task(rq);
if (on_rt_rq(&p->rt))
- update_stats_wait_end_rt(rt_rq, rt_se);
+ update_stats_wait_end_rt(rq, p);
/* The running task is never eligible for pushing */
dequeue_pushable_task(rq, p);
@@ -1702,11 +1650,8 @@ static struct task_struct *pick_task_rt(struct rq *rq)
static void put_prev_task_rt(struct rq *rq, struct task_struct *p, struct task_struct *next)
{
- struct sched_rt_entity *rt_se = &p->rt;
- struct rt_rq *rt_rq = &rq->rt;
-
if (on_rt_rq(&p->rt))
- update_stats_wait_start_rt(rt_rq, rt_se);
+ update_stats_wait_start_rt(rq, p);
update_curr_rt(rq);
--
2.25.1
Powered by blists - more mailing lists