[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200414154204.GC42877@carbon.lan>
Date: Tue, 14 Apr 2020 08:42:04 -0700
From: Roman Gushchin <guro@...com>
To: Vlastimil Babka <vbabka@...e.cz>
CC: Andrew Morton <akpm@...ux-foundation.org>,
Michal Hocko <mhocko@...nel.org>, <linux-mm@...ck.org>,
<kernel-team@...com>, <linux-kernel@...r.kernel.org>,
Rik van Riel <riel@...riel.com>,
Mel Gorman <mgorman@...hsingularity.net>,
Qian Cai <cai@....pw>, Andrea Arcangeli <aarcange@...hat.com>,
Joonsoo Kim <js1304@...il.com>
Subject: Re: [PATCH RFC] mm: compaction: avoid migrating non-cma pages to a
cma area
On Tue, Apr 14, 2020 at 01:49:45PM +0200, Vlastimil Babka wrote:
> On 4/8/20 9:41 PM, Roman Gushchin wrote:
> > Compaction does treat cma pageblocks on pair with any movable
> > pageblocks. It means it can easily move non-cma pages into a cma zone.
> >
> > It can create problems for the cma allocator.
> >
> > The particular problem I'm looking at is related to btrfs metadata
> > pages, which are allocated without __GFP_MOVABLE, but beside that
> > are generic pagecache pages. In fact, they are sometimes movable
> > and sometimes not, depending on whether they are dirty and also
> > on the extent buffer reference counter.
>
Hello, Vlastimil!
Thank you for looking into it.
> Hm I think I'd rather make such pages really unmovable (by a pin?) than deny the
> whole CMA area to compaction. Would it be feasible?
Well, it's an option too, however I'm not sure it's the best one.
First, because these pages can be moved quite often, making
them completely unmovable will make the compaction less efficient.
Second, because it's not only about these pages, but in general
about migrating pages into a cma area without a clear need.
As I wrote in the commit log, if a cma area is located close to end
of a node (which seems to be default at least on x86 without a movable
zone), compaction can fill it quite aggressively. And it might bring
some hot pages (e.g. executable pagecache pages), which will cause
cma allocation failures. I've seen something like this in our prod.
>
> > Compaction moves them to the hugetlb_cma area, and then sometimes
> > the cma allocator fails to move them back from the cma area. It
> > results in failures of gigantic hugepages allocations.
> >
> > Also in general cma areas are reserved close to the end of a zone,
> > and it's where compaction tries to migrate pages. It means
> > compaction will aggressively fill cma areas, which makes not much
> > sense.
>
> So now the free page scanner will have to skip those areas, which is not much
> effective. But I suspect a worse problem in __compaction_suitable() which will
> now falsely report that there are enough free pages, so compaction will start
> but fail to do anytning. Minimally the __zone_watermark_ok() check there would
> have to lose ALLOC_CMA, but there might be other similar checks that would need
> adjusting.
A really good point! I've looked around for any other checks, but haven't found
anything. Please, find an updated version of the patch below.
>
> And long-term what happens if the "CMA using ZONE_MOVABLE" approach is merged
> and there are not more CMA migratetypes to test? Might this change actually also
> avoid your issue, as said pages without __GFP_MOVABLE won't end up in a
> ZONE_MOVABLE?
Yeah, this is what I was thinking about. Basically I want to mimic this behavior
right now. Once this approach will be implemented and merged, it will happen
automatically: obviously, compaction won't move pages between different zones.
Thank you!
--
>From f4a0cfff41c7acab78116478e8e69ae8773b405c Mon Sep 17 00:00:00 2001
From: Roman Gushchin <guro@...com>
Date: Wed, 8 Apr 2020 11:16:38 -0700
Subject: [PATCH v2] mm: compaction: avoid migrating non-cma pages to a cma
area
Compaction does treat cma pageblocks on pair with any movable
pageblocks. It means it can easily move non-cma pages into a cma zone.
It can create problems for the cma allocator.
The particular problem I'm looking at is related to btrfs metadata
pages, which are allocated without __GFP_MOVABLE, but beside that
are generic pagecache pages. In fact, they are sometimes movable
and sometimes not, depending on whether they are dirty and also
on the extent buffer reference counter.
Compaction moves them to the hugetlb_cma area, and then sometimes
the cma allocator fails to move them back from the cma area. It
results in failures of gigantic hugepages allocations.
Also in general cma areas are reserved close to the end of a zone,
and it's where compaction tries to migrate pages. It means
compaction will aggressively fill cma areas, which makes not much
sense.
So to avoid it, let's preserve non-cma pages from being moved into
a cma area. Because cma areas are usually quite large and the number
of areas is small, it should not significantly affect the memory
fragmentation.
v2:
1) adjusted the __zone_watermark_ok() check, suggested by
Vlastimil Babka
2) fixed a typo in a comment
Signed-off-by: Roman Gushchin <guro@...com>
---
mm/compaction.c | 10 ++++++++--
mm/internal.h | 1 +
2 files changed, 9 insertions(+), 2 deletions(-)
diff --git a/mm/compaction.c b/mm/compaction.c
index 46f0fcc93081..9ca036cb148a 100644
--- a/mm/compaction.c
+++ b/mm/compaction.c
@@ -1159,6 +1159,10 @@ static bool suitable_migration_target(struct compact_control *cc,
return false;
}
+ /* Do not bring non-cma pages into a cma area */
+ if (is_migrate_cma(get_pageblock_migratetype(page)) && !cc->cma)
+ return false;
+
if (cc->ignore_block_suitable)
return true;
@@ -1832,6 +1836,8 @@ static isolate_migrate_t isolate_migratepages(struct compact_control *cc)
if (!low_pfn)
return ISOLATE_ABORT;
+ cc->cma = is_migrate_cma(get_pageblock_migratetype(page));
+
/*
* Either we isolated something and proceed with migration. Or
* we failed and compact_zone should decide if we should
@@ -2000,8 +2006,8 @@ static enum compact_result __compaction_suitable(struct zone *zone, int order,
watermark = (order > PAGE_ALLOC_COSTLY_ORDER) ?
low_wmark_pages(zone) : min_wmark_pages(zone);
watermark += compact_gap(order);
- if (!__zone_watermark_ok(zone, 0, watermark, classzone_idx,
- ALLOC_CMA, wmark_target))
+ if (!__zone_watermark_ok(zone, 0, watermark, classzone_idx, 0,
+ wmark_target))
return COMPACT_SKIPPED;
return COMPACT_CONTINUE;
diff --git a/mm/internal.h b/mm/internal.h
index b5634e78f01d..0ce649da824b 100644
--- a/mm/internal.h
+++ b/mm/internal.h
@@ -232,6 +232,7 @@ struct compact_control {
bool contended; /* Signal lock or sched contention */
bool rescan; /* Rescanning the same pageblock */
bool alloc_contig; /* alloc_contig_range allocation */
+ bool cma; /* migratepages contains cma pages */
};
/*
--
2.25.2
Powered by blists - more mailing lists