[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20230906213831.jclswjplxeim2oy4@airbuntu>
Date: Wed, 6 Sep 2023 22:38:31 +0100
From: Qais Yousef <qyousef@...alina.io>
To: Dietmar Eggemann <dietmar.eggemann@....com>
Cc: Ingo Molnar <mingo@...nel.org>,
Peter Zijlstra <peterz@...radead.org>,
"Rafael J. Wysocki" <rafael@...nel.org>,
Viresh Kumar <viresh.kumar@...aro.org>,
Vincent Guittot <vincent.guittot@...aro.org>,
linux-kernel@...r.kernel.org, linux-pm@...r.kernel.org,
Lukasz Luba <lukasz.luba@....com>
Subject: Re: [RFC PATCH 2/7] sched/pelt: Add a new function to approximate
runtime to reach given util
On 09/06/23 22:44, Dietmar Eggemann wrote:
> On 06/09/2023 14:56, Dietmar Eggemann wrote:
> > On 28/08/2023 01:31, Qais Yousef wrote:
> >> It is basically the ramp-up time from 0 to a given value. Will be used
> >> later to implement new tunable to control response time for schedutil.
> >>
> >> Signed-off-by: Qais Yousef (Google) <qyousef@...alina.io>
> >> ---
> >> kernel/sched/pelt.c | 21 +++++++++++++++++++++
> >> kernel/sched/sched.h | 1 +
> >> 2 files changed, 22 insertions(+)
> >>
> >> diff --git a/kernel/sched/pelt.c b/kernel/sched/pelt.c
> >> index 50322005a0ae..f673b9ab92dc 100644
> >> --- a/kernel/sched/pelt.c
> >> +++ b/kernel/sched/pelt.c
> >> @@ -487,3 +487,24 @@ unsigned long approximate_util_avg(unsigned long util, u64 delta)
> >>
> >> return sa.util_avg;
> >> }
> >> +
> >> +/*
> >> + * Approximate the required amount of runtime in ms required to reach @util.
> >> + */
> >> +u64 approximate_runtime(unsigned long util)
> >> +{
> >> + struct sched_avg sa = {};
> >> + u64 delta = 1024; // period = 1024 = ~1ms
> >> + u64 runtime = 0;
> >> +
> >> + if (unlikely(!util))
> >> + return runtime;
> >> +
> >> + while (sa.util_avg < util) {
> >> + accumulate_sum(delta, &sa, 0, 0, 1);
> >> + ___update_load_avg(&sa, 0);
> >> + runtime++;
> >> + }
> >> +
> >> + return runtime;
> >> +}
> >
> > S_n = S_inv * (1 - 0.5^(t/hl))
> >
> > t = hl * ln(1 - Sn/S_inv)/ln(0.5)
> >
> > (1) for a little CPU (capacity_orig = 446)
> >
> > t = 32ms * ln(1 - 446/1024)/ln(0.5)
> >
> > t = 26ms
> >
> > (2) for a big CPU (capacity = 1023 (*instead of 1024 since ln(0) not
> > defined
> >
> > t = 32ms * ln(1 - 1023/1024)/ln(0.5)
> >
> > t = 320ms
>
> Forgot half of what I wanted to ask:
>
> And you want to be able to have a schedutil interface:
>
> /sys/devices/system/cpu/cpufreq/policy*/schedutil/response_time_ms
>
> in which by default we have 26ms for a CPU with the capacity_orig of 446.
Note that this *is* the default. I'm just exposing it not really changing it :)
It is actually much less than that if you take into account the current 25%
headroom.
>
> I.e. you want to have a time-based interface there? Which the user can
> overwrite, say with 52ms and this then will lower the return value of
> get_next_freq() so the system will respond slower?
>
> And the time based interface is more intuitive than staying in the
> capacity world of [0-1024]?
Yes this is exactly how I am defining the interface :-) I think this is generic
and will give users what they need and hopefully should stand the test of time.
The slow down aspect has a limitation though as I highlight in the cover
letter. I haven't figured out how to resolve it yet, or worth the effort.
If anyone has thoughts on that that'd be useful to learn about. It should be
fixable though.
Generally perf first is not always the desired outcome. Power and thermal play
bigger roles in a lot of systems today and I can see even sever market pays
more attention to them now.
Hence I didn't see why I should limit it to improving perf only and disregard
that there are situations where the system might be more concerned about power
or thermals and this could allow more fine tuning than limiting max frequencies
abruptly. They just get harder to reach so in average we get different
residencies (for same workload), but freqs are still reachable. There will be
a perf hit of course, but that's userspace problem to decide if it's worth it
or not.
Generally I am a big advocate of userspace to take the leap and be smarter
about what it needs and when. Fixing everything automagically has its appeal
but I don't think this is sustainable anymore.
Thanks!
--
Qais Yousef
Powered by blists - more mailing lists