[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <e269f5df3af1157232b01a9b0dae3edf4880d786.1613584277.git.tim.c.chen@linux.intel.com>
Date: Wed, 17 Feb 2021 12:41:36 -0800
From: Tim Chen <tim.c.chen@...ux.intel.com>
To: Andrew Morton <akpm@...ux-foundation.org>,
Johannes Weiner <hannes@...xchg.org>,
Michal Hocko <mhocko@...e.cz>,
Vladimir Davydov <vdavydov.dev@...il.com>
Cc: Tim Chen <tim.c.chen@...ux.intel.com>,
Dave Hansen <dave.hansen@...el.com>,
Ying Huang <ying.huang@...el.com>, linux-mm@...ck.org,
cgroups@...r.kernel.org, linux-kernel@...r.kernel.org
Subject: [PATCH v2 3/3] mm: Fix missing mem cgroup soft limit tree updates
On a per node basis, the mem cgroup soft limit tree on each node tracks
how much a cgroup has exceeded its soft limit memory limit and sorts
the cgroup by its excess usage. On page release, the trees are not
updated right away, until we have gathered a batch of pages belonging to
the same cgroup. This reduces the frequency of updating the soft limit tree
and locking of the tree and associated cgroup.
However, the batch of pages could contain pages from multiple nodes but
only the soft limit tree from one node would get updated. Change the
logic so that we update the tree in batch of pages, with each batch of
pages all in the same mem cgroup and memory node. An update is issued for
the batch of pages of a node collected till now whenever we encounter
a page belonging to a different node. Note that this batching for
the same node logic is only relevant for v1 cgroup that has a memory
soft limit.
Reviewed-by: Ying Huang <ying.huang@...el.com>
Signed-off-by: Tim Chen <tim.c.chen@...ux.intel.com>
---
mm/memcontrol.c | 10 +++++++++-
1 file changed, 9 insertions(+), 1 deletion(-)
diff --git a/mm/memcontrol.c b/mm/memcontrol.c
index d72449eeb85a..8bddee75f5cb 100644
--- a/mm/memcontrol.c
+++ b/mm/memcontrol.c
@@ -6804,6 +6804,7 @@ struct uncharge_gather {
unsigned long pgpgout;
unsigned long nr_kmem;
struct page *dummy_page;
+ int nid;
};
static inline void uncharge_gather_clear(struct uncharge_gather *ug)
@@ -6849,7 +6850,13 @@ static void uncharge_page(struct page *page, struct uncharge_gather *ug)
* exclusive access to the page.
*/
- if (ug->memcg != page_memcg(page)) {
+ if (ug->memcg != page_memcg(page) ||
+ /*
+ * Update soft limit tree used in v1 cgroup in page batch for
+ * the same node. Relevant only to v1 cgroup with a soft limit.
+ */
+ (ug->dummy_page && ug->nid != page_to_nid(page) &&
+ ug->memcg->soft_limit != PAGE_COUNTER_MAX)) {
if (ug->memcg) {
uncharge_batch(ug);
uncharge_gather_clear(ug);
@@ -6869,6 +6876,7 @@ static void uncharge_page(struct page *page, struct uncharge_gather *ug)
ug->pgpgout++;
ug->dummy_page = page;
+ ug->nid = page_to_nid(page);
page->memcg_data = 0;
css_put(&ug->memcg->css);
}
--
2.20.1
Powered by blists - more mailing lists