[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20190122144329.ziimv6fejwvky7yb@e110439-lin>
Date: Tue, 22 Jan 2019 14:43:29 +0000
From: Patrick Bellasi <patrick.bellasi@....com>
To: Peter Zijlstra <peterz@...radead.org>
Cc: linux-kernel@...r.kernel.org, linux-pm@...r.kernel.org,
linux-api@...r.kernel.org, Ingo Molnar <mingo@...hat.com>,
Tejun Heo <tj@...nel.org>,
"Rafael J . Wysocki" <rafael.j.wysocki@...el.com>,
Vincent Guittot <vincent.guittot@...aro.org>,
Viresh Kumar <viresh.kumar@...aro.org>,
Paul Turner <pjt@...gle.com>,
Quentin Perret <quentin.perret@....com>,
Dietmar Eggemann <dietmar.eggemann@....com>,
Morten Rasmussen <morten.rasmussen@....com>,
Juri Lelli <juri.lelli@...hat.com>,
Todd Kjos <tkjos@...gle.com>,
Joel Fernandes <joelaf@...gle.com>,
Steve Muckle <smuckle@...gle.com>,
Suren Baghdasaryan <surenb@...gle.com>
Subject: Re: [PATCH v6 07/16] sched/core: uclamp: Add system default clamps
On 22-Jan 14:56, Peter Zijlstra wrote:
> On Tue, Jan 15, 2019 at 10:15:04AM +0000, Patrick Bellasi wrote:
>
> > diff --git a/include/linux/sched.h b/include/linux/sched.h
> > index 84294925d006..c8f391d1cdc5 100644
> > --- a/include/linux/sched.h
> > +++ b/include/linux/sched.h
> > @@ -625,6 +625,11 @@ struct uclamp_se {
> > unsigned int bucket_id : bits_per(UCLAMP_BUCKETS);
> > unsigned int mapped : 1;
> > unsigned int active : 1;
> > + /* Clamp bucket and value actually used by a RUNNABLE task */
> > + struct {
> > + unsigned int value : bits_per(SCHED_CAPACITY_SCALE);
> > + unsigned int bucket_id : bits_per(UCLAMP_BUCKETS);
> > + } effective;
>
> I am confuzled by this thing.. so uclamp_se already has a value,bucket,
> which per the prior code is the effective one.
>
> Now; I think I see why you want another value; you need the second to
> store the original value for when the system limits change and we must
> re-evaluate.
Yes, that's one reason, the other one being to properly support
CGroup when we add them in the following patches.
Effective will always track the value/bucket in which the task has
been refcounted at enqueue time and it depends on the aggregated
value.
> So why are you not adding something like:
>
> unsigned int orig_value : bits_per(SCHED_CAPACITY_SCALE);
Would say that can be enough if we decide to ditch the mapping and use
a linear mapping. In that case the value will always be enough to find
in which bucket a task has been accounted.
> > +unsigned int sysctl_sched_uclamp_util_min;
>
> > +unsigned int sysctl_sched_uclamp_util_max = SCHED_CAPACITY_SCALE;
>
> > +static inline void
> > +uclamp_effective_get(struct task_struct *p, unsigned int clamp_id,
> > + unsigned int *clamp_value, unsigned int *bucket_id)
> > +{
> > + /* Task specific clamp value */
> > + *clamp_value = p->uclamp[clamp_id].value;
> > + *bucket_id = p->uclamp[clamp_id].bucket_id;
> > +
> > + /* System default restriction */
> > + if (unlikely(*clamp_value < uclamp_default[UCLAMP_MIN].value ||
> > + *clamp_value > uclamp_default[UCLAMP_MAX].value)) {
> > + /* Keep it simple: unconditionally enforce system defaults */
> > + *clamp_value = uclamp_default[clamp_id].value;
> > + *bucket_id = uclamp_default[clamp_id].bucket_id;
> > + }
> > +}
>
> That would then turn into something like:
>
> unsigned int high = READ_ONCE(sysctl_sched_uclamp_util_max);
> unsigned int low = READ_ONCE(sysctl_sched_uclamp_util_min);
>
> uclamp_se->orig_value = value;
> uclamp_se->value = clamp(value, low, high);
>
> And then determine bucket_id based on value.
Right... if I ditch the mapping that should work.
> > +int sched_uclamp_handler(struct ctl_table *table, int write,
> > + void __user *buffer, size_t *lenp,
> > + loff_t *ppos)
> > +{
> > + int old_min, old_max;
> > + int result = 0;
> > +
> > + mutex_lock(&uclamp_mutex);
> > +
> > + old_min = sysctl_sched_uclamp_util_min;
> > + old_max = sysctl_sched_uclamp_util_max;
> > +
> > + result = proc_dointvec(table, write, buffer, lenp, ppos);
> > + if (result)
> > + goto undo;
> > + if (!write)
> > + goto done;
> > +
> > + if (sysctl_sched_uclamp_util_min > sysctl_sched_uclamp_util_max ||
> > + sysctl_sched_uclamp_util_max > SCHED_CAPACITY_SCALE) {
> > + result = -EINVAL;
> > + goto undo;
> > + }
> > +
> > + if (old_min != sysctl_sched_uclamp_util_min) {
> > + uclamp_bucket_inc(NULL, &uclamp_default[UCLAMP_MIN],
> > + UCLAMP_MIN, sysctl_sched_uclamp_util_min);
> > + }
> > + if (old_max != sysctl_sched_uclamp_util_max) {
> > + uclamp_bucket_inc(NULL, &uclamp_default[UCLAMP_MAX],
> > + UCLAMP_MAX, sysctl_sched_uclamp_util_max);
> > + }
>
> Should you not update all tasks?
That's true, but that's also an expensive operation, that's why now
I'm doing only lazy updates at next enqueue time.
Do you think that could be acceptable?
Perhaps I can sanity check all the CPU to ensure that they all have a
current clamp value within the new enforced range. This kind-of
anticipate the idea to have an in-kernel API which has higher priority
and allows to set clamp values across all the CPUs...
--
#include <best/regards.h>
Patrick Bellasi
Powered by blists - more mailing lists