[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAATkVEw22YAfSH4GKY1Y9Qz9chCAz1cgcesz_xg3O2-0XxY_ng@mail.gmail.com>
Date: Thu, 6 Jul 2017 14:00:00 -0400
From: Debabrata Banerjee <dbavatar@...il.com>
To: Mel Gorman <mgorman@...hsingularity.net>
Cc: Roman Gushchin <guro@...com>, linux-mm@...ck.org,
Andrew Morton <akpm@...ux-foundation.org>,
Johannes Weiner <hannes@...xchg.org>,
Michal Hocko <mhocko@...e.com>,
Vladimir Davydov <vdavydov.dev@...il.com>,
Rik van Riel <riel@...hat.com>, kernel-team@...com,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] mm: make allocation counters per-order
On Thu, Jul 6, 2017 at 1:16 PM, Mel Gorman <mgorman@...hsingularity.net> wrote:
>
> I'm still struggling to see how counters help when an agent that monitors
> for high CPU usage could be activated
>
I suspect Roman has the same problem set as us, the CPU usage is
either always high, high and service critical likely when something
interesting is happening. We'd like to collect data on 200k machines,
and study the results statistically and with respect to time based on
kernel versions, build configs, hardware types, process types, load
patterns, etc, etc. Even finding good candidate machines and at the
right time of day to manually debug with ftrace is problematic.
Granted we could be utilizing existing counters like compact_fail
better. Ultimately the data either leads to dealing with certain bad
actors, different vm tunings, or patches to mm.
Powered by blists - more mailing lists