[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <20180712172942.10094-11-hannes@cmpxchg.org>
Date: Thu, 12 Jul 2018 13:29:42 -0400
From: Johannes Weiner <hannes@...xchg.org>
To: Ingo Molnar <mingo@...hat.com>,
Peter Zijlstra <peterz@...radead.org>,
Andrew Morton <akpm@...ux-foundation.org>,
Linus Torvalds <torvalds@...ux-foundation.org>
Cc: Tejun Heo <tj@...nel.org>, Suren Baghdasaryan <surenb@...gle.com>,
Vinayak Menon <vinmenon@...eaurora.org>,
Christopher Lameter <cl@...ux.com>,
Mike Galbraith <efault@....de>,
Shakeel Butt <shakeelb@...gle.com>, linux-mm@...ck.org,
cgroups@...r.kernel.org, linux-kernel@...r.kernel.org,
kernel-team@...com
Subject: [RFC PATCH 10/10] psi: aggregate ongoing stall events when somebody reads pressure
Right now, psi reports pressure and stall times of already concluded
stall events. For most use cases this is current enough, but certain
highly latency-sensitive applications, like the Android OOM killer,
might want to know about and react to stall states before they have
even concluded (e.g. a prolonged reclaim cycle).
This patches the procfs/cgroupfs interface such that when the pressure
metrics are read, the current per-cpu states, if any, are taken into
account as well.
Any ongoing states are concluded, their time snapshotted, and then
restarted. This requires holding the rq lock to avoid corruption. It
could use some form of rq lock ratelimiting or avoidance.
Requested-by: Suren Baghdasaryan <surenb@...gle.com>
Not-yet-signed-off-by: Johannes Weiner <hannes@...xchg.org>
---
kernel/sched/psi.c | 56 +++++++++++++++++++++++++++++++++++++---------
1 file changed, 46 insertions(+), 10 deletions(-)
diff --git a/kernel/sched/psi.c b/kernel/sched/psi.c
index 53e0b7b83e2e..5a6c6057f775 100644
--- a/kernel/sched/psi.c
+++ b/kernel/sched/psi.c
@@ -190,7 +190,7 @@ static void calc_avgs(unsigned long avg[3], u64 time, int missed_periods)
}
}
-static bool psi_update_stats(struct psi_group *group)
+static bool psi_update_stats(struct psi_group *group, bool ondemand)
{
u64 some[NR_PSI_RESOURCES] = { 0, };
u64 full[NR_PSI_RESOURCES] = { 0, };
@@ -200,8 +200,6 @@ static bool psi_update_stats(struct psi_group *group)
int cpu;
int r;
- mutex_lock(&group->stat_lock);
-
/*
* Collect the per-cpu time buckets and average them into a
* single time sample that is normalized to wallclock time.
@@ -218,10 +216,36 @@ static bool psi_update_stats(struct psi_group *group)
for_each_online_cpu(cpu) {
struct psi_group_cpu *groupc = per_cpu_ptr(group->cpus, cpu);
unsigned long nonidle;
+ struct rq_flags rf;
+ struct rq *rq;
+ u64 now;
- if (!groupc->nonidle_time)
+ if (!groupc->nonidle_time && !groupc->nonidle)
continue;
+ /*
+ * We come here for two things: 1) periodic per-cpu
+ * bucket flushing and averaging and 2) when the user
+ * wants to read a pressure file. For flushing and
+ * averaging, which is relatively infrequent, we can
+ * be lazy and tolerate some raciness with concurrent
+ * updates to the per-cpu counters. However, if a user
+ * polls the pressure state, we want to give them the
+ * most uptodate information we have, including any
+ * currently active state which hasn't been timed yet,
+ * because in case of an iowait or a reclaim run, that
+ * can be significant.
+ */
+ if (ondemand) {
+ rq = cpu_rq(cpu);
+ rq_lock_irq(rq, &rf);
+
+ now = cpu_clock(cpu);
+
+ groupc->nonidle_time += now - groupc->nonidle_start;
+ groupc->nonidle_start = now;
+ }
+
nonidle = nsecs_to_jiffies(groupc->nonidle_time);
groupc->nonidle_time = 0;
nonidle_total += nonidle;
@@ -229,13 +253,27 @@ static bool psi_update_stats(struct psi_group *group)
for (r = 0; r < NR_PSI_RESOURCES; r++) {
struct psi_resource *res = &groupc->res[r];
+ if (ondemand && res->state != PSI_NONE) {
+ bool is_full = res->state == PSI_FULL;
+
+ res->times[is_full] += now - res->state_start;
+ res->state_start = now;
+ }
+
some[r] += (res->times[0] + res->times[1]) * nonidle;
full[r] += res->times[1] * nonidle;
- /* It's racy, but we can tolerate some error */
res->times[0] = 0;
res->times[1] = 0;
}
+
+ if (ondemand)
+ rq_unlock_irq(rq, &rf);
+ }
+
+ for (r = 0; r < NR_PSI_RESOURCES; r++) {
+ do_div(some[r], max(nonidle_total, 1UL));
+ do_div(full[r], max(nonidle_total, 1UL));
}
/*
@@ -249,12 +287,10 @@ static bool psi_update_stats(struct psi_group *group)
* activity, thus no data, and clock ticks are sporadic. The
* below handles both.
*/
+ mutex_lock(&group->stat_lock);
/* total= */
for (r = 0; r < NR_PSI_RESOURCES; r++) {
- do_div(some[r], max(nonidle_total, 1UL));
- do_div(full[r], max(nonidle_total, 1UL));
-
group->some[r] += some[r];
group->full[r] += full[r];
}
@@ -301,7 +337,7 @@ static void psi_clock(struct work_struct *work)
* go - see calc_avgs() and missed_periods.
*/
- nonidle = psi_update_stats(group);
+ nonidle = psi_update_stats(group, false);
if (nonidle) {
unsigned long delay = 0;
@@ -570,7 +606,7 @@ int psi_show(struct seq_file *m, struct psi_group *group, enum psi_res res)
if (psi_disabled)
return -EOPNOTSUPP;
- psi_update_stats(group);
+ psi_update_stats(group, true);
for (w = 0; w < 3; w++) {
avg[0][w] = group->avg_some[res][w];
--
2.18.0
Powered by blists - more mailing lists