[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <09ea63b7-7294-a143-c797-95eba87c765e@arm.com>
Date: Fri, 24 Mar 2023 10:46:21 +0000
From: Lukasz Luba <lukasz.luba@....com>
To: Yajun Deng <yajun.deng@...ux.dev>
Cc: linux-pm@...r.kernel.org, linux-kernel@...r.kernel.org,
rostedt@...dmis.org, dietmar.eggemann@....com,
vincent.guittot@...aro.org, mingo@...hat.com, vschneid@...hat.com,
bristot@...hat.com, bsegall@...gle.com, juri.lelli@...hat.com,
peterz@...radead.org, mgorman@...e.de, viresh.kumar@...aro.org,
rafael@...nel.org
Subject: Re: [PATCH] cpufreq: schedutil: Combine two loops into one in
sugov_start()
Hi Yajun,
On 3/24/23 10:00, Yajun Deng wrote:
> The sugov_start() function currently contains two for loops that
> traverse the CPU list and perform some initialization tasks. The first
> loop initializes each sugov_cpu struct and assigns the CPU number and
> sugov_policy pointer. The second loop sets up the update_util hook for
> each CPU based on the policy type.
>
> Since both loops operate on the same CPU list, it is possible to combine
> them into a single for loop. This simplifies the code and reduces the
> number of times the CPU list needs to be traversed, which can improve
> performance.
>
> Signed-off-by: Yajun Deng <yajun.deng@...ux.dev>
> ---
> kernel/sched/cpufreq_schedutil.c | 12 ++++--------
> 1 file changed, 4 insertions(+), 8 deletions(-)
>
> diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c
> index e3211455b203..9a28ebbb9c1e 100644
> --- a/kernel/sched/cpufreq_schedutil.c
> +++ b/kernel/sched/cpufreq_schedutil.c
> @@ -766,14 +766,6 @@ static int sugov_start(struct cpufreq_policy *policy)
>
> sg_policy->need_freq_update = cpufreq_driver_test_flags(CPUFREQ_NEED_UPDATE_LIMITS);
>
> - for_each_cpu(cpu, policy->cpus) {
> - struct sugov_cpu *sg_cpu = &per_cpu(sugov_cpu, cpu);
> -
> - memset(sg_cpu, 0, sizeof(*sg_cpu));
> - sg_cpu->cpu = cpu;
> - sg_cpu->sg_policy = sg_policy;
> - }
> -
> if (policy_is_shared(policy))
> uu = sugov_update_shared;
> else if (policy->fast_switch_enabled && cpufreq_driver_has_adjust_perf())
> @@ -784,6 +776,10 @@ static int sugov_start(struct cpufreq_policy *policy)
> for_each_cpu(cpu, policy->cpus) {
> struct sugov_cpu *sg_cpu = &per_cpu(sugov_cpu, cpu);
>
> + memset(sg_cpu, 0, sizeof(*sg_cpu));
> + sg_cpu->cpu = cpu;
> + sg_cpu->sg_policy = sg_policy;
> +
> cpufreq_add_update_util_hook(cpu, &sg_cpu->update_util, uu);
> }
> return 0;
IMO the change might cause a race.
There is a call to set scheduler hook in the 2nd loop.
If you combine two loops that hook might be used
from other CPU in the meantime, while still the rest CPUs are not
finished.
The first loop makes sure all CPUs in the 'policy->cpus' get a clean
context 'sg_cpu' and proper 'cpu' values first (and 'sg_policy' as
well). When the two loops are combined, there might be fast usage
from scheduler on other CPU the sugov code path.
If the policy is shared for many CPUs and any of them is able to
change the freq, then some CPU can enter this code flow, where
remotely wants to check the other CPUs' utilization:
sugov_next_freq_shared()
for each cpu in policy->cpus:
sugov_get_util()
where the 'sg_cpu->cpu' is used
Therefore, IMO this optimization in a start function (which is
only called once and is not part of the 'hot path') is not
worth the race risk.
Regards
Lukasz
Powered by blists - more mailing lists