[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <tip-d18be45dbfef2e0bb12b9696c21aeae92f83b1ea@git.kernel.org>
Date: Wed, 10 Jan 2018 04:19:05 -0800
From: tip-bot for Juri Lelli <tipbot@...or.com>
To: linux-tip-commits@...r.kernel.org
Cc: tglx@...utronix.de, viresh.kumar@...aro.org,
torvalds@...ux-foundation.org, claudio@...dence.eu.com,
rafael.j.wysocki@...el.com, hpa@...or.com, juri.lelli@....com,
linux-kernel@...r.kernel.org, luca.abeni@...tannapisa.it,
peterz@...radead.org, mingo@...nel.org
Subject: [tip:sched/core] sched/cpufreq: Split utilization signals
Commit-ID: d18be45dbfef2e0bb12b9696c21aeae92f83b1ea
Gitweb: https://git.kernel.org/tip/d18be45dbfef2e0bb12b9696c21aeae92f83b1ea
Author: Juri Lelli <juri.lelli@....com>
AuthorDate: Mon, 4 Dec 2017 11:23:21 +0100
Committer: Ingo Molnar <mingo@...nel.org>
CommitDate: Wed, 10 Jan 2018 12:53:34 +0100
sched/cpufreq: Split utilization signals
To be able to treat utilization signals of different scheduling classes
in different ways (e.g., CFS signal might be stale while DEADLINE signal
is never stale by design) we need to split sugov_cpu::util signal in two:
util_cfs and util_dl.
This patch does that by also changing sugov_get_util() parameter list.
After this change, aggregation of the different signals has to be performed
by sugov_get_util() users (so that they can decide what to do with the
different signals).
Suggested-by: Rafael J. Wysocki <rafael.j.wysocki@...el.com>
Signed-off-by: Juri Lelli <juri.lelli@....com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@...radead.org>
Acked-by: Viresh Kumar <viresh.kumar@...aro.org>
Cc: Claudio Scordino <claudio@...dence.eu.com>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>
Cc: Luca Abeni <luca.abeni@...tannapisa.it>
Cc: Peter Zijlstra <peterz@...radead.org>
Cc: Thomas Gleixner <tglx@...utronix.de>
Cc: alessio.balsini@....com
Cc: bristot@...hat.com
Cc: dietmar.eggemann@....com
Cc: joelaf@...gle.com
Cc: juri.lelli@...hat.com
Cc: mathieu.poirier@...aro.org
Cc: morten.rasmussen@....com
Cc: patrick.bellasi@....com
Cc: rjw@...ysocki.net
Cc: rostedt@...dmis.org
Cc: tkjos@...roid.com
Cc: tommaso.cucinotta@...tannapisa.it
Cc: vincent.guittot@...aro.org
Link: http://lkml.kernel.org/r/20171204102325.5110-5-juri.lelli@redhat.com
Signed-off-by: Ingo Molnar <mingo@...nel.org>
---
kernel/sched/cpufreq_schedutil.c | 30 ++++++++++++++++--------------
1 file changed, 16 insertions(+), 14 deletions(-)
diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c
index bd5f997..e9e0713 100644
--- a/kernel/sched/cpufreq_schedutil.c
+++ b/kernel/sched/cpufreq_schedutil.c
@@ -60,7 +60,8 @@ struct sugov_cpu {
u64 last_update;
/* The fields below are only needed when sharing a policy. */
- unsigned long util;
+ unsigned long util_cfs;
+ unsigned long util_dl;
unsigned long max;
unsigned int flags;
@@ -176,20 +177,23 @@ static unsigned int get_next_freq(struct sugov_policy *sg_policy,
return cpufreq_driver_resolve_freq(policy, freq);
}
-static void sugov_get_util(unsigned long *util, unsigned long *max, int cpu)
+static void sugov_get_util(struct sugov_cpu *sg_cpu)
{
- struct rq *rq = cpu_rq(cpu);
- unsigned long util_cfs = cpu_util_cfs(rq);
- unsigned long util_dl = cpu_util_dl(rq);
+ struct rq *rq = cpu_rq(sg_cpu->cpu);
- *max = arch_scale_cpu_capacity(NULL, cpu);
+ sg_cpu->max = arch_scale_cpu_capacity(NULL, sg_cpu->cpu);
+ sg_cpu->util_cfs = cpu_util_cfs(rq);
+ sg_cpu->util_dl = cpu_util_dl(rq);
+}
+static unsigned long sugov_aggregate_util(struct sugov_cpu *sg_cpu)
+{
/*
* Ideally we would like to set util_dl as min/guaranteed freq and
* util_cfs + util_dl as requested freq. However, cpufreq is not yet
* ready for such an interface. So, we only do the latter for now.
*/
- *util = min(util_cfs + util_dl, *max);
+ return min(sg_cpu->util_cfs + sg_cpu->util_dl, sg_cpu->max);
}
static void sugov_set_iowait_boost(struct sugov_cpu *sg_cpu, u64 time)
@@ -279,7 +283,9 @@ static void sugov_update_single(struct update_util_data *hook, u64 time,
if (flags & SCHED_CPUFREQ_RT) {
next_f = policy->cpuinfo.max_freq;
} else {
- sugov_get_util(&util, &max, sg_cpu->cpu);
+ sugov_get_util(sg_cpu);
+ max = sg_cpu->max;
+ util = sugov_aggregate_util(sg_cpu);
sugov_iowait_boost(sg_cpu, &util, &max);
next_f = get_next_freq(sg_policy, util, max);
/*
@@ -324,8 +330,8 @@ static unsigned int sugov_next_freq_shared(struct sugov_cpu *sg_cpu, u64 time)
if (j_sg_cpu->flags & SCHED_CPUFREQ_RT)
return policy->cpuinfo.max_freq;
- j_util = j_sg_cpu->util;
j_max = j_sg_cpu->max;
+ j_util = sugov_aggregate_util(j_sg_cpu);
if (j_util * max > j_max * util) {
util = j_util;
max = j_max;
@@ -342,15 +348,11 @@ static void sugov_update_shared(struct update_util_data *hook, u64 time,
{
struct sugov_cpu *sg_cpu = container_of(hook, struct sugov_cpu, update_util);
struct sugov_policy *sg_policy = sg_cpu->sg_policy;
- unsigned long util, max;
unsigned int next_f;
- sugov_get_util(&util, &max, sg_cpu->cpu);
-
raw_spin_lock(&sg_policy->update_lock);
- sg_cpu->util = util;
- sg_cpu->max = max;
+ sugov_get_util(sg_cpu);
sg_cpu->flags = flags;
sugov_set_iowait_boost(sg_cpu, time);
Powered by blists - more mailing lists