[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1501530579.9118.43.camel@gmx.de>
Date: Mon, 31 Jul 2017 21:49:39 +0200
From: Mike Galbraith <efault@....de>
To: Johannes Weiner <hannes@...xchg.org>,
Peter Zijlstra <peterz@...radead.org>
Cc: Ingo Molnar <mingo@...hat.com>,
Andrew Morton <akpm@...ux-foundation.org>,
Rik van Riel <riel@...hat.com>, Mel Gorman <mgorman@...e.de>,
linux-mm@...ck.org, linux-kernel@...r.kernel.org,
kernel-team@...com
Subject: Re: [PATCH 3/3] mm/sched: memdelay: memory health interface for
systems and workloads
On Mon, 2017-07-31 at 14:41 -0400, Johannes Weiner wrote:
>
> Adding an rq counter for tasks inside memdelay sections should be
> straight-forward as well (except for maybe the migration cost of that
> state between CPUs in ttwu that Mike pointed out).
What I pointed out should be easily eliminated (zero use case).
> That leaves the question of how to track these numbers per cgroup at
> an acceptable cost. The idea for a tree of cgroups is that walltime
> impact of delays at each level is reported for all tasks at or below
> that level. E.g. a leave group aggregates the state of its own tasks,
> the root/system aggregates the state of all tasks in the system; hence
> the propagation of the task state counters up the hierarchy.
The crux of the biscuit is where exactly the investment return lies.
Gathering of these numbers ain't gonna be free, no matter how hard you
try, and you're plugging into paths where every cycle added is made of
userspace hide.
-Mike
Powered by blists - more mailing lists