[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200323054037.GI2987@MiWiFi-R3L-srv>
Date: Mon, 23 Mar 2020 13:40:37 +0800
From: Baoquan He <bhe@...hat.com>
To: js1304@...il.com
Cc: Andrew Morton <akpm@...ux-foundation.org>, linux-mm@...ck.org,
linux-kernel@...r.kernel.org, Johannes Weiner <hannes@...xchg.org>,
Michal Hocko <mhocko@...nel.org>,
Minchan Kim <minchan@...nel.org>,
Vlastimil Babka <vbabka@...e.cz>,
Mel Gorman <mgorman@...hsingularity.net>, kernel-team@....com,
Ye Xiaolong <xiaolong.ye@...el.com>,
David Rientjes <rientjes@...gle.com>,
Joonsoo Kim <iamjoonsoo.kim@....com>
Subject: Re: [PATCH v4 2/2] mm/page_alloc: integrate classzone_idx and
high_zoneidx
On 03/23/20 at 01:21pm, Baoquan He wrote:
> On 03/23/20 at 01:49pm, js1304@...il.com wrote:
> > From: Joonsoo Kim <iamjoonsoo.kim@....com>
> >
> > classzone_idx is just different name for high_zoneidx now.
> > So, integrate them and add some comment to struct alloc_context
> > in order to reduce future confusion about the meaning of this variable.
> >
> > The accessor, ac_classzone_idx() is also removed since it isn't needed
> > after integration.
> >
> > In addition to integration, this patch also renames high_zoneidx
> > to highest_zoneidx since it represents more precise meaning.
> >
> > Signed-off-by: Joonsoo Kim <iamjoonsoo.kim@....com>
> > ---
> > include/linux/compaction.h | 9 ++--
> > include/linux/mmzone.h | 12 ++---
> > include/trace/events/compaction.h | 22 ++++----
> > include/trace/events/vmscan.h | 14 +++--
> > mm/compaction.c | 64 +++++++++++------------
> > mm/internal.h | 21 +++++---
> > mm/memory_hotplug.c | 6 +--
> > mm/oom_kill.c | 4 +-
> > mm/page_alloc.c | 60 +++++++++++-----------
> > mm/slab.c | 4 +-
> > mm/slub.c | 4 +-
> > mm/vmscan.c | 105 ++++++++++++++++++++------------------
> > 12 files changed, 175 insertions(+), 150 deletions(-)
> >
> > diff --git a/include/linux/compaction.h b/include/linux/compaction.h
> > index 4b898cd..3ed2f22 100644
> > --- a/include/linux/compaction.h
> > +++ b/include/linux/compaction.h
> > @@ -97,7 +97,7 @@ extern enum compact_result try_to_compact_pages(gfp_t gfp_mask,
> > struct page **page);
> > extern void reset_isolation_suitable(pg_data_t *pgdat);
> > extern enum compact_result compaction_suitable(struct zone *zone, int order,
> > - unsigned int alloc_flags, int classzone_idx);
> > + unsigned int alloc_flags, int highest_zoneidx);
> >
> > extern void defer_compaction(struct zone *zone, int order);
> > extern bool compaction_deferred(struct zone *zone, int order);
> > @@ -182,7 +182,7 @@ bool compaction_zonelist_suitable(struct alloc_context *ac, int order,
> >
> > extern int kcompactd_run(int nid);
> > extern void kcompactd_stop(int nid);
> > -extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int classzone_idx);
> > +extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx);
> >
> > #else
> > static inline void reset_isolation_suitable(pg_data_t *pgdat)
> > @@ -190,7 +190,7 @@ static inline void reset_isolation_suitable(pg_data_t *pgdat)
> > }
> >
> > static inline enum compact_result compaction_suitable(struct zone *zone, int order,
> > - int alloc_flags, int classzone_idx)
> > + int alloc_flags, int highest_zoneidx)
> > {
> > return COMPACT_SKIPPED;
> > }
> > @@ -232,7 +232,8 @@ static inline void kcompactd_stop(int nid)
> > {
> > }
> >
> > -static inline void wakeup_kcompactd(pg_data_t *pgdat, int order, int classzone_idx)
> > +static inline void wakeup_kcompactd(pg_data_t *pgdat,
> > + int order, int highest_zoneidx)
> > {
> > }
> >
> > diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h
> > index f3f2648..337b5ec 100644
> > --- a/include/linux/mmzone.h
> > +++ b/include/linux/mmzone.h
> > @@ -698,13 +698,13 @@ typedef struct pglist_data {
> > struct task_struct *kswapd; /* Protected by
> > mem_hotplug_begin/end() */
> > int kswapd_order;
> > - enum zone_type kswapd_classzone_idx;
> > + enum zone_type kswapd_highest_zoneidx;
> >
> > int kswapd_failures; /* Number of 'reclaimed == 0' runs */
> >
> > #ifdef CONFIG_COMPACTION
> > int kcompactd_max_order;
> > - enum zone_type kcompactd_classzone_idx;
> > + enum zone_type kcompactd_highest_zoneidx;
> > wait_queue_head_t kcompactd_wait;
> > struct task_struct *kcompactd;
> > #endif
> > @@ -782,15 +782,15 @@ static inline bool pgdat_is_empty(pg_data_t *pgdat)
> >
> > void build_all_zonelists(pg_data_t *pgdat);
> > void wakeup_kswapd(struct zone *zone, gfp_t gfp_mask, int order,
> > - enum zone_type classzone_idx);
> > + enum zone_type highest_zoneidx);
> > bool __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark,
> > - int classzone_idx, unsigned int alloc_flags,
> > + int highest_zoneidx, unsigned int alloc_flags,
> > long free_pages);
> > bool zone_watermark_ok(struct zone *z, unsigned int order,
> > - unsigned long mark, int classzone_idx,
> > + unsigned long mark, int highest_zoneidx,
> > unsigned int alloc_flags);
> > bool zone_watermark_ok_safe(struct zone *z, unsigned int order,
> > - unsigned long mark, int classzone_idx);
> > + unsigned long mark, int highest_zoneidx);
> > enum memmap_context {
> > MEMMAP_EARLY,
> > MEMMAP_HOTPLUG,
> > diff --git a/include/trace/events/compaction.h b/include/trace/events/compaction.h
> > index e5bf6ee..54e5bf0 100644
> > --- a/include/trace/events/compaction.h
> > +++ b/include/trace/events/compaction.h
> > @@ -314,40 +314,44 @@ TRACE_EVENT(mm_compaction_kcompactd_sleep,
> >
> > DECLARE_EVENT_CLASS(kcompactd_wake_template,
> >
> > - TP_PROTO(int nid, int order, enum zone_type classzone_idx),
> > + TP_PROTO(int nid, int order, enum zone_type highest_zoneidx),
> >
> > - TP_ARGS(nid, order, classzone_idx),
> > + TP_ARGS(nid, order, highest_zoneidx),
> >
> > TP_STRUCT__entry(
> > __field(int, nid)
> > __field(int, order)
> > - __field(enum zone_type, classzone_idx)
> > + __field(enum zone_type, highest_zoneidx)
> > ),
> >
> > TP_fast_assign(
> > __entry->nid = nid;
> > __entry->order = order;
> > - __entry->classzone_idx = classzone_idx;
> > + __entry->highest_zoneidx = highest_zoneidx;
> > ),
> >
> > + /*
> > + * classzone_idx is previous name of the highest_zoneidx.
> > + * Reason not to change it is the ABI requirement of the tracepoint.
> > + */
> > TP_printk("nid=%d order=%d classzone_idx=%-8s",
> ~~~~~~~~~~ this one could be missed.
Oh, I realized above 'classzone_idx=%-8s' being kept is for the old
script compatibility? then it's OK.
>
> And I applied this patch series, use 'git grep classzone_idx' to search
> any relics, found below classzone_idx leftover in perl script.
> [~]$ git grep classzone_idx
> Documentation/trace/postprocess/trace-vmscan-postprocess.pl:my $regex_lru_isolate_default = 'isolate_mode=([0-9]*) classzone_idx=([0-9]*) order=([0-9]*) nr_requested=([0-9]*) nr_scanned=([0-9]*) nr_skipped=([0-9]*) nr_taken=([0-9]*) lru=([a-z_]*)';
> Documentation/trace/postprocess/trace-vmscan-postprocess.pl: "isolate_mode", "classzone_idx", "order",
>
> > __entry->nid,
> > __entry->order,
> > - __print_symbolic(__entry->classzone_idx, ZONE_TYPE))
> > + __print_symbolic(__entry->highest_zoneidx, ZONE_TYPE))
> > );
> >
> > DEFINE_EVENT(kcompactd_wake_template, mm_compaction_wakeup_kcompactd,
> >
> > - TP_PROTO(int nid, int order, enum zone_type classzone_idx),
> > + TP_PROTO(int nid, int order, enum zone_type highest_zoneidx),
> >
> > - TP_ARGS(nid, order, classzone_idx)
> > + TP_ARGS(nid, order, highest_zoneidx)
> > );
> >
> > DEFINE_EVENT(kcompactd_wake_template, mm_compaction_kcompactd_wake,
> >
> > - TP_PROTO(int nid, int order, enum zone_type classzone_idx),
> > + TP_PROTO(int nid, int order, enum zone_type highest_zoneidx),
> >
> > - TP_ARGS(nid, order, classzone_idx)
> > + TP_ARGS(nid, order, highest_zoneidx)
> > );
> > #endif
> >
> > diff --git a/include/trace/events/vmscan.h b/include/trace/events/vmscan.h
> > index a5ab297..f2b3b9c 100644
> > --- a/include/trace/events/vmscan.h
> > +++ b/include/trace/events/vmscan.h
> > @@ -265,7 +265,7 @@ TRACE_EVENT(mm_shrink_slab_end,
> > );
> >
> > TRACE_EVENT(mm_vmscan_lru_isolate,
> > - TP_PROTO(int classzone_idx,
> > + TP_PROTO(int highest_zoneidx,
> > int order,
> > unsigned long nr_requested,
> > unsigned long nr_scanned,
> > @@ -274,10 +274,10 @@ TRACE_EVENT(mm_vmscan_lru_isolate,
> > isolate_mode_t isolate_mode,
> > int lru),
> >
> > - TP_ARGS(classzone_idx, order, nr_requested, nr_scanned, nr_skipped, nr_taken, isolate_mode, lru),
> > + TP_ARGS(highest_zoneidx, order, nr_requested, nr_scanned, nr_skipped, nr_taken, isolate_mode, lru),
> >
> > TP_STRUCT__entry(
> > - __field(int, classzone_idx)
> > + __field(int, highest_zoneidx)
> > __field(int, order)
> > __field(unsigned long, nr_requested)
> > __field(unsigned long, nr_scanned)
> > @@ -288,7 +288,7 @@ TRACE_EVENT(mm_vmscan_lru_isolate,
> > ),
> >
> > TP_fast_assign(
> > - __entry->classzone_idx = classzone_idx;
> > + __entry->highest_zoneidx = highest_zoneidx;
> > __entry->order = order;
> > __entry->nr_requested = nr_requested;
> > __entry->nr_scanned = nr_scanned;
> > @@ -298,9 +298,13 @@ TRACE_EVENT(mm_vmscan_lru_isolate,
> > __entry->lru = lru;
> > ),
> >
> > + /*
> > + * classzone is previous name of the highest_zoneidx.
> > + * Reason not to change it is the ABI requirement of the tracepoint.
> > + */
> > TP_printk("isolate_mode=%d classzone=%d order=%d nr_requested=%lu nr_scanned=%lu nr_skipped=%lu nr_taken=%lu lru=%s",
> > __entry->isolate_mode,
> > - __entry->classzone_idx,
> > + __entry->highest_zoneidx,
> > __entry->order,
> > __entry->nr_requested,
> > __entry->nr_scanned,
> > diff --git a/mm/compaction.c b/mm/compaction.c
> > index 827d8a2..80d6424 100644
> > --- a/mm/compaction.c
> > +++ b/mm/compaction.c
> > @@ -1966,7 +1966,7 @@ static enum compact_result compact_finished(struct compact_control *cc)
> > */
> > static enum compact_result __compaction_suitable(struct zone *zone, int order,
> > unsigned int alloc_flags,
> > - int classzone_idx,
> > + int highest_zoneidx,
> > unsigned long wmark_target)
> > {
> > unsigned long watermark;
> > @@ -1979,7 +1979,7 @@ static enum compact_result __compaction_suitable(struct zone *zone, int order,
> > * If watermarks for high-order allocation are already met, there
> > * should be no need for compaction at all.
> > */
> > - if (zone_watermark_ok(zone, order, watermark, classzone_idx,
> > + if (zone_watermark_ok(zone, order, watermark, highest_zoneidx,
> > alloc_flags))
> > return COMPACT_SUCCESS;
> >
> > @@ -1989,9 +1989,9 @@ static enum compact_result __compaction_suitable(struct zone *zone, int order,
> > * watermark and alloc_flags have to match, or be more pessimistic than
> > * the check in __isolate_free_page(). We don't use the direct
> > * compactor's alloc_flags, as they are not relevant for freepage
> > - * isolation. We however do use the direct compactor's classzone_idx to
> > - * skip over zones where lowmem reserves would prevent allocation even
> > - * if compaction succeeds.
> > + * isolation. We however do use the direct compactor's highest_zoneidx
> > + * to skip over zones where lowmem reserves would prevent allocation
> > + * even if compaction succeeds.
> > * For costly orders, we require low watermark instead of min for
> > * compaction to proceed to increase its chances.
> > * ALLOC_CMA is used, as pages in CMA pageblocks are considered
> > @@ -2000,7 +2000,7 @@ static enum compact_result __compaction_suitable(struct zone *zone, int order,
> > watermark = (order > PAGE_ALLOC_COSTLY_ORDER) ?
> > low_wmark_pages(zone) : min_wmark_pages(zone);
> > watermark += compact_gap(order);
> > - if (!__zone_watermark_ok(zone, 0, watermark, classzone_idx,
> > + if (!__zone_watermark_ok(zone, 0, watermark, highest_zoneidx,
> > ALLOC_CMA, wmark_target))
> > return COMPACT_SKIPPED;
> >
> > @@ -2009,12 +2009,12 @@ static enum compact_result __compaction_suitable(struct zone *zone, int order,
> >
> > enum compact_result compaction_suitable(struct zone *zone, int order,
> > unsigned int alloc_flags,
> > - int classzone_idx)
> > + int highest_zoneidx)
> > {
> > enum compact_result ret;
> > int fragindex;
> >
> > - ret = __compaction_suitable(zone, order, alloc_flags, classzone_idx,
> > + ret = __compaction_suitable(zone, order, alloc_flags, highest_zoneidx,
> > zone_page_state(zone, NR_FREE_PAGES));
> > /*
> > * fragmentation index determines if allocation failures are due to
> > @@ -2055,8 +2055,8 @@ bool compaction_zonelist_suitable(struct alloc_context *ac, int order,
> > * Make sure at least one zone would pass __compaction_suitable if we continue
> > * retrying the reclaim.
> > */
> > - for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx,
> > - ac->nodemask) {
> > + for_each_zone_zonelist_nodemask(zone, z, ac->zonelist,
> > + ac->highest_zoneidx, ac->nodemask) {
> > unsigned long available;
> > enum compact_result compact_result;
> >
> > @@ -2069,7 +2069,7 @@ bool compaction_zonelist_suitable(struct alloc_context *ac, int order,
> > available = zone_reclaimable_pages(zone) / order;
> > available += zone_page_state_snapshot(zone, NR_FREE_PAGES);
> > compact_result = __compaction_suitable(zone, order, alloc_flags,
> > - ac_classzone_idx(ac), available);
> > + ac->highest_zoneidx, available);
> > if (compact_result != COMPACT_SKIPPED)
> > return true;
> > }
> > @@ -2100,7 +2100,7 @@ compact_zone(struct compact_control *cc, struct capture_control *capc)
> >
> > cc->migratetype = gfpflags_to_migratetype(cc->gfp_mask);
> > ret = compaction_suitable(cc->zone, cc->order, cc->alloc_flags,
> > - cc->classzone_idx);
> > + cc->highest_zoneidx);
> > /* Compaction is likely to fail */
> > if (ret == COMPACT_SUCCESS || ret == COMPACT_SKIPPED)
> > return ret;
> > @@ -2296,7 +2296,7 @@ compact_zone(struct compact_control *cc, struct capture_control *capc)
> >
> > static enum compact_result compact_zone_order(struct zone *zone, int order,
> > gfp_t gfp_mask, enum compact_priority prio,
> > - unsigned int alloc_flags, int classzone_idx,
> > + unsigned int alloc_flags, int highest_zoneidx,
> > struct page **capture)
> > {
> > enum compact_result ret;
> > @@ -2308,7 +2308,7 @@ static enum compact_result compact_zone_order(struct zone *zone, int order,
> > .mode = (prio == COMPACT_PRIO_ASYNC) ?
> > MIGRATE_ASYNC : MIGRATE_SYNC_LIGHT,
> > .alloc_flags = alloc_flags,
> > - .classzone_idx = classzone_idx,
> > + .highest_zoneidx = highest_zoneidx,
> > .direct_compaction = true,
> > .whole_zone = (prio == MIN_COMPACT_PRIORITY),
> > .ignore_skip_hint = (prio == MIN_COMPACT_PRIORITY),
> > @@ -2364,8 +2364,8 @@ enum compact_result try_to_compact_pages(gfp_t gfp_mask, unsigned int order,
> > trace_mm_compaction_try_to_compact_pages(order, gfp_mask, prio);
> >
> > /* Compact each zone in the list */
> > - for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx,
> > - ac->nodemask) {
> > + for_each_zone_zonelist_nodemask(zone, z, ac->zonelist,
> > + ac->highest_zoneidx, ac->nodemask) {
> > enum compact_result status;
> >
> > if (prio > MIN_COMPACT_PRIORITY
> > @@ -2375,7 +2375,7 @@ enum compact_result try_to_compact_pages(gfp_t gfp_mask, unsigned int order,
> > }
> >
> > status = compact_zone_order(zone, order, gfp_mask, prio,
> > - alloc_flags, ac_classzone_idx(ac), capture);
> > + alloc_flags, ac->highest_zoneidx, capture);
> > rc = max(status, rc);
> >
> > /* The allocation should succeed, stop compacting */
> > @@ -2510,16 +2510,16 @@ static bool kcompactd_node_suitable(pg_data_t *pgdat)
> > {
> > int zoneid;
> > struct zone *zone;
> > - enum zone_type classzone_idx = pgdat->kcompactd_classzone_idx;
> > + enum zone_type highest_zoneidx = pgdat->kcompactd_highest_zoneidx;
> >
> > - for (zoneid = 0; zoneid <= classzone_idx; zoneid++) {
> > + for (zoneid = 0; zoneid <= highest_zoneidx; zoneid++) {
> > zone = &pgdat->node_zones[zoneid];
> >
> > if (!populated_zone(zone))
> > continue;
> >
> > if (compaction_suitable(zone, pgdat->kcompactd_max_order, 0,
> > - classzone_idx) == COMPACT_CONTINUE)
> > + highest_zoneidx) == COMPACT_CONTINUE)
> > return true;
> > }
> >
> > @@ -2537,16 +2537,16 @@ static void kcompactd_do_work(pg_data_t *pgdat)
> > struct compact_control cc = {
> > .order = pgdat->kcompactd_max_order,
> > .search_order = pgdat->kcompactd_max_order,
> > - .classzone_idx = pgdat->kcompactd_classzone_idx,
> > + .highest_zoneidx = pgdat->kcompactd_highest_zoneidx,
> > .mode = MIGRATE_SYNC_LIGHT,
> > .ignore_skip_hint = false,
> > .gfp_mask = GFP_KERNEL,
> > };
> > trace_mm_compaction_kcompactd_wake(pgdat->node_id, cc.order,
> > - cc.classzone_idx);
> > + cc.highest_zoneidx);
> > count_compact_event(KCOMPACTD_WAKE);
> >
> > - for (zoneid = 0; zoneid <= cc.classzone_idx; zoneid++) {
> > + for (zoneid = 0; zoneid <= cc.highest_zoneidx; zoneid++) {
> > int status;
> >
> > zone = &pgdat->node_zones[zoneid];
> > @@ -2595,16 +2595,16 @@ static void kcompactd_do_work(pg_data_t *pgdat)
> >
> > /*
> > * Regardless of success, we are done until woken up next. But remember
> > - * the requested order/classzone_idx in case it was higher/tighter than
> > - * our current ones
> > + * the requested order/highest_zoneidx in case it was higher/tighter
> > + * than our current ones
> > */
> > if (pgdat->kcompactd_max_order <= cc.order)
> > pgdat->kcompactd_max_order = 0;
> > - if (pgdat->kcompactd_classzone_idx >= cc.classzone_idx)
> > - pgdat->kcompactd_classzone_idx = pgdat->nr_zones - 1;
> > + if (pgdat->kcompactd_highest_zoneidx >= cc.highest_zoneidx)
> > + pgdat->kcompactd_highest_zoneidx = pgdat->nr_zones - 1;
> > }
> >
> > -void wakeup_kcompactd(pg_data_t *pgdat, int order, int classzone_idx)
> > +void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx)
> > {
> > if (!order)
> > return;
> > @@ -2612,8 +2612,8 @@ void wakeup_kcompactd(pg_data_t *pgdat, int order, int classzone_idx)
> > if (pgdat->kcompactd_max_order < order)
> > pgdat->kcompactd_max_order = order;
> >
> > - if (pgdat->kcompactd_classzone_idx > classzone_idx)
> > - pgdat->kcompactd_classzone_idx = classzone_idx;
> > + if (pgdat->kcompactd_highest_zoneidx > highest_zoneidx)
> > + pgdat->kcompactd_highest_zoneidx = highest_zoneidx;
> >
> > /*
> > * Pairs with implicit barrier in wait_event_freezable()
> > @@ -2626,7 +2626,7 @@ void wakeup_kcompactd(pg_data_t *pgdat, int order, int classzone_idx)
> > return;
> >
> > trace_mm_compaction_wakeup_kcompactd(pgdat->node_id, order,
> > - classzone_idx);
> > + highest_zoneidx);
> > wake_up_interruptible(&pgdat->kcompactd_wait);
> > }
> >
> > @@ -2647,7 +2647,7 @@ static int kcompactd(void *p)
> > set_freezable();
> >
> > pgdat->kcompactd_max_order = 0;
> > - pgdat->kcompactd_classzone_idx = pgdat->nr_zones - 1;
> > + pgdat->kcompactd_highest_zoneidx = pgdat->nr_zones - 1;
> >
> > while (!kthread_should_stop()) {
> > unsigned long pflags;
> > diff --git a/mm/internal.h b/mm/internal.h
> > index aebaa33..d84b0d9 100644
> > --- a/mm/internal.h
> > +++ b/mm/internal.h
> > @@ -102,10 +102,10 @@ extern pmd_t *mm_find_pmd(struct mm_struct *mm, unsigned long address);
> > * between functions involved in allocations, including the alloc_pages*
> > * family of functions.
> > *
> > - * nodemask, migratetype and high_zoneidx are initialized only once in
> > + * nodemask, migratetype and highest_zoneidx are initialized only once in
> > * __alloc_pages_nodemask() and then never change.
> > *
> > - * zonelist, preferred_zone and classzone_idx are set first in
> > + * zonelist, preferred_zone and highest_zoneidx are set first in
> > * __alloc_pages_nodemask() for the fast path, and might be later changed
> > * in __alloc_pages_slowpath(). All other functions pass the whole strucure
> > * by a const pointer.
> > @@ -115,12 +115,21 @@ struct alloc_context {
> > nodemask_t *nodemask;
> > struct zoneref *preferred_zoneref;
> > int migratetype;
> > - enum zone_type high_zoneidx;
> > +
> > + /*
> > + * highest_zoneidx represents highest usable zone index of
> > + * the allocation request. Due to the nature of the zone,
> > + * memory on lower zone than the highest_zoneidx will be
> > + * protected by lowmem_reserve[highest_zoneidx].
> > + *
> > + * highest_zoneidx is also used by reclaim/compaction to limit
> > + * the target zone since higher zone than this index cannot be
> > + * usable for this allocation request.
> > + */
> > + enum zone_type highest_zoneidx;
> > bool spread_dirty_pages;
> > };
> >
> > -#define ac_classzone_idx(ac) (ac->high_zoneidx)
> > -
> > /*
> > * Locate the struct page for both the matching buddy in our
> > * pair (buddy1) and the combined O(n+1) page they form (page).
> > @@ -199,7 +208,7 @@ struct compact_control {
> > int order; /* order a direct compactor needs */
> > int migratetype; /* migratetype of direct compactor */
> > const unsigned int alloc_flags; /* alloc flags of a direct compactor */
> > - const int classzone_idx; /* zone index of a direct compactor */
> > + const int highest_zoneidx; /* zone index of a direct compactor */
> > enum migrate_mode mode; /* Async or sync migration mode */
> > bool ignore_skip_hint; /* Scan blocks even if marked skip */
> > bool no_set_skip_hint; /* Don't mark blocks for skipping */
> > diff --git a/mm/memory_hotplug.c b/mm/memory_hotplug.c
> > index 8bdf484..f942969 100644
> > --- a/mm/memory_hotplug.c
> > +++ b/mm/memory_hotplug.c
> > @@ -880,13 +880,13 @@ static pg_data_t __ref *hotadd_new_pgdat(int nid, u64 start)
> > } else {
> > int cpu;
> > /*
> > - * Reset the nr_zones, order and classzone_idx before reuse.
> > - * Note that kswapd will init kswapd_classzone_idx properly
> > + * Reset the nr_zones, order and highest_zoneidx before reuse.
> > + * Note that kswapd will init kswapd_highest_zoneidx properly
> > * when it starts in the near future.
> > */
> > pgdat->nr_zones = 0;
> > pgdat->kswapd_order = 0;
> > - pgdat->kswapd_classzone_idx = 0;
> > + pgdat->kswapd_highest_zoneidx = 0;
> > for_each_online_cpu(cpu) {
> > struct per_cpu_nodestat *p;
> >
> > diff --git a/mm/oom_kill.c b/mm/oom_kill.c
> > index dfc3576..4daedf7 100644
> > --- a/mm/oom_kill.c
> > +++ b/mm/oom_kill.c
> > @@ -254,7 +254,7 @@ static enum oom_constraint constrained_alloc(struct oom_control *oc)
> > {
> > struct zone *zone;
> > struct zoneref *z;
> > - enum zone_type high_zoneidx = gfp_zone(oc->gfp_mask);
> > + enum zone_type highest_zoneidx = gfp_zone(oc->gfp_mask);
> > bool cpuset_limited = false;
> > int nid;
> >
> > @@ -294,7 +294,7 @@ static enum oom_constraint constrained_alloc(struct oom_control *oc)
> >
> > /* Check this allocation failure is caused by cpuset's wall function */
> > for_each_zone_zonelist_nodemask(zone, z, oc->zonelist,
> > - high_zoneidx, oc->nodemask)
> > + highest_zoneidx, oc->nodemask)
> > if (!cpuset_zone_allowed(zone, oc->gfp_mask))
> > cpuset_limited = true;
> >
> > diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> > index cb2f896..27bacee 100644
> > --- a/mm/page_alloc.c
> > +++ b/mm/page_alloc.c
> > @@ -2620,7 +2620,7 @@ static bool unreserve_highatomic_pageblock(const struct alloc_context *ac,
> > int order;
> > bool ret;
> >
> > - for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->high_zoneidx,
> > + for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->highest_zoneidx,
> > ac->nodemask) {
> > /*
> > * Preserve at least one pageblock unless memory pressure
> > @@ -3488,7 +3488,7 @@ ALLOW_ERROR_INJECTION(should_fail_alloc_page, TRUE);
> > * to check in the allocation paths if no pages are free.
> > */
> > bool __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark,
> > - int classzone_idx, unsigned int alloc_flags,
> > + int highest_zoneidx, unsigned int alloc_flags,
> > long free_pages)
> > {
> > long min = mark;
> > @@ -3533,7 +3533,7 @@ bool __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark,
> > * are not met, then a high-order request also cannot go ahead
> > * even if a suitable page happened to be free.
> > */
> > - if (free_pages <= min + z->lowmem_reserve[classzone_idx])
> > + if (free_pages <= min + z->lowmem_reserve[highest_zoneidx])
> > return false;
> >
> > /* If this is an order-0 request then the watermark is fine */
> > @@ -3566,14 +3566,15 @@ bool __zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark,
> > }
> >
> > bool zone_watermark_ok(struct zone *z, unsigned int order, unsigned long mark,
> > - int classzone_idx, unsigned int alloc_flags)
> > + int highest_zoneidx, unsigned int alloc_flags)
> > {
> > - return __zone_watermark_ok(z, order, mark, classzone_idx, alloc_flags,
> > + return __zone_watermark_ok(z, order, mark, highest_zoneidx, alloc_flags,
> > zone_page_state(z, NR_FREE_PAGES));
> > }
> >
> > static inline bool zone_watermark_fast(struct zone *z, unsigned int order,
> > - unsigned long mark, int classzone_idx, unsigned int alloc_flags)
> > + unsigned long mark, int highest_zoneidx,
> > + unsigned int alloc_flags)
> > {
> > long free_pages = zone_page_state(z, NR_FREE_PAGES);
> > long cma_pages = 0;
> > @@ -3591,22 +3592,23 @@ static inline bool zone_watermark_fast(struct zone *z, unsigned int order,
> > * the caller is !atomic then it'll uselessly search the free
> > * list. That corner case is then slower but it is harmless.
> > */
> > - if (!order && (free_pages - cma_pages) > mark + z->lowmem_reserve[classzone_idx])
> > + if (!order && (free_pages - cma_pages) >
> > + mark + z->lowmem_reserve[highest_zoneidx])
> > return true;
> >
> > - return __zone_watermark_ok(z, order, mark, classzone_idx, alloc_flags,
> > + return __zone_watermark_ok(z, order, mark, highest_zoneidx, alloc_flags,
> > free_pages);
> > }
> >
> > bool zone_watermark_ok_safe(struct zone *z, unsigned int order,
> > - unsigned long mark, int classzone_idx)
> > + unsigned long mark, int highest_zoneidx)
> > {
> > long free_pages = zone_page_state(z, NR_FREE_PAGES);
> >
> > if (z->percpu_drift_mark && free_pages < z->percpu_drift_mark)
> > free_pages = zone_page_state_snapshot(z, NR_FREE_PAGES);
> >
> > - return __zone_watermark_ok(z, order, mark, classzone_idx, 0,
> > + return __zone_watermark_ok(z, order, mark, highest_zoneidx, 0,
> > free_pages);
> > }
> >
> > @@ -3683,8 +3685,8 @@ get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags,
> > */
> > no_fallback = alloc_flags & ALLOC_NOFRAGMENT;
> > z = ac->preferred_zoneref;
> > - for_next_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx,
> > - ac->nodemask) {
> > + for_next_zone_zonelist_nodemask(zone, z, ac->zonelist,
> > + ac->highest_zoneidx, ac->nodemask) {
> > struct page *page;
> > unsigned long mark;
> >
> > @@ -3739,7 +3741,7 @@ get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags,
> >
> > mark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK);
> > if (!zone_watermark_fast(zone, order, mark,
> > - ac_classzone_idx(ac), alloc_flags)) {
> > + ac->highest_zoneidx, alloc_flags)) {
> > int ret;
> >
> > #ifdef CONFIG_DEFERRED_STRUCT_PAGE_INIT
> > @@ -3772,7 +3774,7 @@ get_page_from_freelist(gfp_t gfp_mask, unsigned int order, int alloc_flags,
> > default:
> > /* did we reclaim enough */
> > if (zone_watermark_ok(zone, order, mark,
> > - ac_classzone_idx(ac), alloc_flags))
> > + ac->highest_zoneidx, alloc_flags))
> > goto try_this_zone;
> >
> > continue;
> > @@ -3931,7 +3933,7 @@ __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order,
> > if (gfp_mask & __GFP_RETRY_MAYFAIL)
> > goto out;
> > /* The OOM killer does not needlessly kill tasks for lowmem */
> > - if (ac->high_zoneidx < ZONE_NORMAL)
> > + if (ac->highest_zoneidx < ZONE_NORMAL)
> > goto out;
> > if (pm_suspended_storage())
> > goto out;
> > @@ -4134,10 +4136,10 @@ should_compact_retry(struct alloc_context *ac, unsigned int order, int alloc_fla
> > * Let's give them a good hope and keep retrying while the order-0
> > * watermarks are OK.
> > */
> > - for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx,
> > - ac->nodemask) {
> > + for_each_zone_zonelist_nodemask(zone, z, ac->zonelist,
> > + ac->highest_zoneidx, ac->nodemask) {
> > if (zone_watermark_ok(zone, 0, min_wmark_pages(zone),
> > - ac_classzone_idx(ac), alloc_flags))
> > + ac->highest_zoneidx, alloc_flags))
> > return true;
> > }
> > return false;
> > @@ -4261,12 +4263,12 @@ static void wake_all_kswapds(unsigned int order, gfp_t gfp_mask,
> > struct zoneref *z;
> > struct zone *zone;
> > pg_data_t *last_pgdat = NULL;
> > - enum zone_type high_zoneidx = ac->high_zoneidx;
> > + enum zone_type highest_zoneidx = ac->highest_zoneidx;
> >
> > - for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, high_zoneidx,
> > + for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, highest_zoneidx,
> > ac->nodemask) {
> > if (last_pgdat != zone->zone_pgdat)
> > - wakeup_kswapd(zone, gfp_mask, order, high_zoneidx);
> > + wakeup_kswapd(zone, gfp_mask, order, highest_zoneidx);
> > last_pgdat = zone->zone_pgdat;
> > }
> > }
> > @@ -4401,8 +4403,8 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order,
> > * request even if all reclaimable pages are considered then we are
> > * screwed and have to go OOM.
> > */
> > - for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx,
> > - ac->nodemask) {
> > + for_each_zone_zonelist_nodemask(zone, z, ac->zonelist,
> > + ac->highest_zoneidx, ac->nodemask) {
> > unsigned long available;
> > unsigned long reclaimable;
> > unsigned long min_wmark = min_wmark_pages(zone);
> > @@ -4416,7 +4418,7 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order,
> > * reclaimable pages?
> > */
> > wmark = __zone_watermark_ok(zone, order, min_wmark,
> > - ac_classzone_idx(ac), alloc_flags, available);
> > + ac->highest_zoneidx, alloc_flags, available);
> > trace_reclaim_retry_zone(z, order, reclaimable,
> > available, min_wmark, *no_progress_loops, wmark);
> > if (wmark) {
> > @@ -4535,7 +4537,7 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
> > * could end up iterating over non-eligible zones endlessly.
> > */
> > ac->preferred_zoneref = first_zones_zonelist(ac->zonelist,
> > - ac->high_zoneidx, ac->nodemask);
> > + ac->highest_zoneidx, ac->nodemask);
> > if (!ac->preferred_zoneref->zone)
> > goto nopage;
> >
> > @@ -4622,7 +4624,7 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
> > if (!(alloc_flags & ALLOC_CPUSET) || reserve_flags) {
> > ac->nodemask = NULL;
> > ac->preferred_zoneref = first_zones_zonelist(ac->zonelist,
> > - ac->high_zoneidx, ac->nodemask);
> > + ac->highest_zoneidx, ac->nodemask);
> > }
> >
> > /* Attempt with potentially adjusted zonelist and alloc_flags */
> > @@ -4756,7 +4758,7 @@ static inline bool prepare_alloc_pages(gfp_t gfp_mask, unsigned int order,
> > struct alloc_context *ac, gfp_t *alloc_mask,
> > unsigned int *alloc_flags)
> > {
> > - ac->high_zoneidx = gfp_zone(gfp_mask);
> > + ac->highest_zoneidx = gfp_zone(gfp_mask);
> > ac->zonelist = node_zonelist(preferred_nid, gfp_mask);
> > ac->nodemask = nodemask;
> > ac->migratetype = gfpflags_to_migratetype(gfp_mask);
> > @@ -4795,7 +4797,7 @@ static inline void finalise_ac(gfp_t gfp_mask, struct alloc_context *ac)
> > * may get reset for allocations that ignore memory policies.
> > */
> > ac->preferred_zoneref = first_zones_zonelist(ac->zonelist,
> > - ac->high_zoneidx, ac->nodemask);
> > + ac->highest_zoneidx, ac->nodemask);
> > }
> >
> > /*
> > @@ -6992,7 +6994,7 @@ void __init free_area_init_node(int nid, unsigned long *zones_size,
> > unsigned long end_pfn = 0;
> >
> > /* pg_data_t should be reset to zero when it's allocated */
> > - WARN_ON(pgdat->nr_zones || pgdat->kswapd_classzone_idx);
> > + WARN_ON(pgdat->nr_zones || pgdat->kswapd_highest_zoneidx);
> >
> > pgdat->node_id = nid;
> > pgdat->node_start_pfn = node_start_pfn;
> > diff --git a/mm/slab.c b/mm/slab.c
> > index a896336..9350062 100644
> > --- a/mm/slab.c
> > +++ b/mm/slab.c
> > @@ -3106,7 +3106,7 @@ static void *fallback_alloc(struct kmem_cache *cache, gfp_t flags)
> > struct zonelist *zonelist;
> > struct zoneref *z;
> > struct zone *zone;
> > - enum zone_type high_zoneidx = gfp_zone(flags);
> > + enum zone_type highest_zoneidx = gfp_zone(flags);
> > void *obj = NULL;
> > struct page *page;
> > int nid;
> > @@ -3124,7 +3124,7 @@ static void *fallback_alloc(struct kmem_cache *cache, gfp_t flags)
> > * Look through allowed nodes for objects available
> > * from existing per node queues.
> > */
> > - for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) {
> > + for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) {
> > nid = zone_to_nid(zone);
> >
> > if (cpuset_zone_allowed(zone, flags) &&
> > diff --git a/mm/slub.c b/mm/slub.c
> > index 1c55bf7..d220671 100644
> > --- a/mm/slub.c
> > +++ b/mm/slub.c
> > @@ -1909,7 +1909,7 @@ static void *get_any_partial(struct kmem_cache *s, gfp_t flags,
> > struct zonelist *zonelist;
> > struct zoneref *z;
> > struct zone *zone;
> > - enum zone_type high_zoneidx = gfp_zone(flags);
> > + enum zone_type highest_zoneidx = gfp_zone(flags);
> > void *object;
> > unsigned int cpuset_mems_cookie;
> >
> > @@ -1938,7 +1938,7 @@ static void *get_any_partial(struct kmem_cache *s, gfp_t flags,
> > do {
> > cpuset_mems_cookie = read_mems_allowed_begin();
> > zonelist = node_zonelist(mempolicy_slab_node(), flags);
> > - for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) {
> > + for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) {
> > struct kmem_cache_node *n;
> >
> > n = get_node(s, zone_to_nid(zone));
> > diff --git a/mm/vmscan.c b/mm/vmscan.c
> > index dca623d..0616abe 100644
> > --- a/mm/vmscan.c
> > +++ b/mm/vmscan.c
> > @@ -3133,8 +3133,8 @@ static bool allow_direct_reclaim(pg_data_t *pgdat)
> >
> > /* kswapd must be awake if processes are being throttled */
> > if (!wmark_ok && waitqueue_active(&pgdat->kswapd_wait)) {
> > - if (READ_ONCE(pgdat->kswapd_classzone_idx) > ZONE_NORMAL)
> > - WRITE_ONCE(pgdat->kswapd_classzone_idx, ZONE_NORMAL);
> > + if (READ_ONCE(pgdat->kswapd_highest_zoneidx) > ZONE_NORMAL)
> > + WRITE_ONCE(pgdat->kswapd_highest_zoneidx, ZONE_NORMAL);
> >
> > wake_up_interruptible(&pgdat->kswapd_wait);
> > }
> > @@ -3387,7 +3387,7 @@ static void age_active_anon(struct pglist_data *pgdat,
> > } while (memcg);
> > }
> >
> > -static bool pgdat_watermark_boosted(pg_data_t *pgdat, int classzone_idx)
> > +static bool pgdat_watermark_boosted(pg_data_t *pgdat, int highest_zoneidx)
> > {
> > int i;
> > struct zone *zone;
> > @@ -3399,7 +3399,7 @@ static bool pgdat_watermark_boosted(pg_data_t *pgdat, int classzone_idx)
> > * start prematurely when there is no boosting and a lower
> > * zone is balanced.
> > */
> > - for (i = classzone_idx; i >= 0; i--) {
> > + for (i = highest_zoneidx; i >= 0; i--) {
> > zone = pgdat->node_zones + i;
> > if (!managed_zone(zone))
> > continue;
> > @@ -3413,9 +3413,9 @@ static bool pgdat_watermark_boosted(pg_data_t *pgdat, int classzone_idx)
> >
> > /*
> > * Returns true if there is an eligible zone balanced for the request order
> > - * and classzone_idx
> > + * and highest_zoneidx
> > */
> > -static bool pgdat_balanced(pg_data_t *pgdat, int order, int classzone_idx)
> > +static bool pgdat_balanced(pg_data_t *pgdat, int order, int highest_zoneidx)
> > {
> > int i;
> > unsigned long mark = -1;
> > @@ -3425,19 +3425,19 @@ static bool pgdat_balanced(pg_data_t *pgdat, int order, int classzone_idx)
> > * Check watermarks bottom-up as lower zones are more likely to
> > * meet watermarks.
> > */
> > - for (i = 0; i <= classzone_idx; i++) {
> > + for (i = 0; i <= highest_zoneidx; i++) {
> > zone = pgdat->node_zones + i;
> >
> > if (!managed_zone(zone))
> > continue;
> >
> > mark = high_wmark_pages(zone);
> > - if (zone_watermark_ok_safe(zone, order, mark, classzone_idx))
> > + if (zone_watermark_ok_safe(zone, order, mark, highest_zoneidx))
> > return true;
> > }
> >
> > /*
> > - * If a node has no populated zone within classzone_idx, it does not
> > + * If a node has no populated zone within highest_zoneidx, it does not
> > * need balancing by definition. This can happen if a zone-restricted
> > * allocation tries to wake a remote kswapd.
> > */
> > @@ -3463,7 +3463,8 @@ static void clear_pgdat_congested(pg_data_t *pgdat)
> > *
> > * Returns true if kswapd is ready to sleep
> > */
> > -static bool prepare_kswapd_sleep(pg_data_t *pgdat, int order, int classzone_idx)
> > +static bool prepare_kswapd_sleep(pg_data_t *pgdat, int order,
> > + int highest_zoneidx)
> > {
> > /*
> > * The throttled processes are normally woken up in balance_pgdat() as
> > @@ -3485,7 +3486,7 @@ static bool prepare_kswapd_sleep(pg_data_t *pgdat, int order, int classzone_idx)
> > if (pgdat->kswapd_failures >= MAX_RECLAIM_RETRIES)
> > return true;
> >
> > - if (pgdat_balanced(pgdat, order, classzone_idx)) {
> > + if (pgdat_balanced(pgdat, order, highest_zoneidx)) {
> > clear_pgdat_congested(pgdat);
> > return true;
> > }
> > @@ -3549,7 +3550,7 @@ static bool kswapd_shrink_node(pg_data_t *pgdat,
> > * or lower is eligible for reclaim until at least one usable zone is
> > * balanced.
> > */
> > -static int balance_pgdat(pg_data_t *pgdat, int order, int classzone_idx)
> > +static int balance_pgdat(pg_data_t *pgdat, int order, int highest_zoneidx)
> > {
> > int i;
> > unsigned long nr_soft_reclaimed;
> > @@ -3577,7 +3578,7 @@ static int balance_pgdat(pg_data_t *pgdat, int order, int classzone_idx)
> > * stall or direct reclaim until kswapd is finished.
> > */
> > nr_boost_reclaim = 0;
> > - for (i = 0; i <= classzone_idx; i++) {
> > + for (i = 0; i <= highest_zoneidx; i++) {
> > zone = pgdat->node_zones + i;
> > if (!managed_zone(zone))
> > continue;
> > @@ -3595,7 +3596,7 @@ static int balance_pgdat(pg_data_t *pgdat, int order, int classzone_idx)
> > bool balanced;
> > bool ret;
> >
> > - sc.reclaim_idx = classzone_idx;
> > + sc.reclaim_idx = highest_zoneidx;
> >
> > /*
> > * If the number of buffer_heads exceeds the maximum allowed
> > @@ -3625,7 +3626,7 @@ static int balance_pgdat(pg_data_t *pgdat, int order, int classzone_idx)
> > * on the grounds that the normal reclaim should be enough to
> > * re-evaluate if boosting is required when kswapd next wakes.
> > */
> > - balanced = pgdat_balanced(pgdat, sc.order, classzone_idx);
> > + balanced = pgdat_balanced(pgdat, sc.order, highest_zoneidx);
> > if (!balanced && nr_boost_reclaim) {
> > nr_boost_reclaim = 0;
> > goto restart;
> > @@ -3725,7 +3726,7 @@ static int balance_pgdat(pg_data_t *pgdat, int order, int classzone_idx)
> > if (boosted) {
> > unsigned long flags;
> >
> > - for (i = 0; i <= classzone_idx; i++) {
> > + for (i = 0; i <= highest_zoneidx; i++) {
> > if (!zone_boosts[i])
> > continue;
> >
> > @@ -3740,7 +3741,7 @@ static int balance_pgdat(pg_data_t *pgdat, int order, int classzone_idx)
> > * As there is now likely space, wakeup kcompact to defragment
> > * pageblocks.
> > */
> > - wakeup_kcompactd(pgdat, pageblock_order, classzone_idx);
> > + wakeup_kcompactd(pgdat, pageblock_order, highest_zoneidx);
> > }
> >
> > snapshot_refaults(NULL, pgdat);
> > @@ -3758,22 +3759,22 @@ static int balance_pgdat(pg_data_t *pgdat, int order, int classzone_idx)
> > }
> >
> > /*
> > - * The pgdat->kswapd_classzone_idx is used to pass the highest zone index to be
> > - * reclaimed by kswapd from the waker. If the value is MAX_NR_ZONES which is not
> > - * a valid index then either kswapd runs for first time or kswapd couldn't sleep
> > - * after previous reclaim attempt (node is still unbalanced). In that case
> > - * return the zone index of the previous kswapd reclaim cycle.
> > + * The pgdat->kswapd_highest_zoneidx is used to pass the highest zone index to
> > + * be reclaimed by kswapd from the waker. If the value is MAX_NR_ZONES which is
> > + * not a valid index then either kswapd runs for first time or kswapd couldn't
> > + * sleep after previous reclaim attempt (node is still unbalanced). In that
> > + * case return the zone index of the previous kswapd reclaim cycle.
> > */
> > -static enum zone_type kswapd_classzone_idx(pg_data_t *pgdat,
> > - enum zone_type prev_classzone_idx)
> > +static enum zone_type kswapd_highest_zoneidx(pg_data_t *pgdat,
> > + enum zone_type prev_highest_zoneidx)
> > {
> > - enum zone_type curr_idx = READ_ONCE(pgdat->kswapd_classzone_idx);
> > + enum zone_type curr_idx = READ_ONCE(pgdat->kswapd_highest_zoneidx);
> >
> > - return curr_idx == MAX_NR_ZONES ? prev_classzone_idx : curr_idx;
> > + return curr_idx == MAX_NR_ZONES ? prev_highest_zoneidx : curr_idx;
> > }
> >
> > static void kswapd_try_to_sleep(pg_data_t *pgdat, int alloc_order, int reclaim_order,
> > - unsigned int classzone_idx)
> > + unsigned int highest_zoneidx)
> > {
> > long remaining = 0;
> > DEFINE_WAIT(wait);
> > @@ -3790,7 +3791,7 @@ static void kswapd_try_to_sleep(pg_data_t *pgdat, int alloc_order, int reclaim_o
> > * eligible zone balanced that it's also unlikely that compaction will
> > * succeed.
> > */
> > - if (prepare_kswapd_sleep(pgdat, reclaim_order, classzone_idx)) {
> > + if (prepare_kswapd_sleep(pgdat, reclaim_order, highest_zoneidx)) {
> > /*
> > * Compaction records what page blocks it recently failed to
> > * isolate pages from and skips them in the future scanning.
> > @@ -3803,18 +3804,19 @@ static void kswapd_try_to_sleep(pg_data_t *pgdat, int alloc_order, int reclaim_o
> > * We have freed the memory, now we should compact it to make
> > * allocation of the requested order possible.
> > */
> > - wakeup_kcompactd(pgdat, alloc_order, classzone_idx);
> > + wakeup_kcompactd(pgdat, alloc_order, highest_zoneidx);
> >
> > remaining = schedule_timeout(HZ/10);
> >
> > /*
> > - * If woken prematurely then reset kswapd_classzone_idx and
> > + * If woken prematurely then reset kswapd_highest_zoneidx and
> > * order. The values will either be from a wakeup request or
> > * the previous request that slept prematurely.
> > */
> > if (remaining) {
> > - WRITE_ONCE(pgdat->kswapd_classzone_idx,
> > - kswapd_classzone_idx(pgdat, classzone_idx));
> > + WRITE_ONCE(pgdat->kswapd_highest_zoneidx,
> > + kswapd_highest_zoneidx(pgdat,
> > + highest_zoneidx));
> >
> > if (READ_ONCE(pgdat->kswapd_order) < reclaim_order)
> > WRITE_ONCE(pgdat->kswapd_order, reclaim_order);
> > @@ -3829,7 +3831,7 @@ static void kswapd_try_to_sleep(pg_data_t *pgdat, int alloc_order, int reclaim_o
> > * go fully to sleep until explicitly woken up.
> > */
> > if (!remaining &&
> > - prepare_kswapd_sleep(pgdat, reclaim_order, classzone_idx)) {
> > + prepare_kswapd_sleep(pgdat, reclaim_order, highest_zoneidx)) {
> > trace_mm_vmscan_kswapd_sleep(pgdat->node_id);
> >
> > /*
> > @@ -3871,7 +3873,7 @@ static void kswapd_try_to_sleep(pg_data_t *pgdat, int alloc_order, int reclaim_o
> > static int kswapd(void *p)
> > {
> > unsigned int alloc_order, reclaim_order;
> > - unsigned int classzone_idx = MAX_NR_ZONES - 1;
> > + unsigned int highest_zoneidx = MAX_NR_ZONES - 1;
> > pg_data_t *pgdat = (pg_data_t*)p;
> > struct task_struct *tsk = current;
> > const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id);
> > @@ -3895,22 +3897,24 @@ static int kswapd(void *p)
> > set_freezable();
> >
> > WRITE_ONCE(pgdat->kswapd_order, 0);
> > - WRITE_ONCE(pgdat->kswapd_classzone_idx, MAX_NR_ZONES);
> > + WRITE_ONCE(pgdat->kswapd_highest_zoneidx, MAX_NR_ZONES);
> > for ( ; ; ) {
> > bool ret;
> >
> > alloc_order = reclaim_order = READ_ONCE(pgdat->kswapd_order);
> > - classzone_idx = kswapd_classzone_idx(pgdat, classzone_idx);
> > + highest_zoneidx = kswapd_highest_zoneidx(pgdat,
> > + highest_zoneidx);
> >
> > kswapd_try_sleep:
> > kswapd_try_to_sleep(pgdat, alloc_order, reclaim_order,
> > - classzone_idx);
> > + highest_zoneidx);
> >
> > - /* Read the new order and classzone_idx */
> > + /* Read the new order and highest_zoneidx */
> > alloc_order = reclaim_order = READ_ONCE(pgdat->kswapd_order);
> > - classzone_idx = kswapd_classzone_idx(pgdat, classzone_idx);
> > + highest_zoneidx = kswapd_highest_zoneidx(pgdat,
> > + highest_zoneidx);
> > WRITE_ONCE(pgdat->kswapd_order, 0);
> > - WRITE_ONCE(pgdat->kswapd_classzone_idx, MAX_NR_ZONES);
> > + WRITE_ONCE(pgdat->kswapd_highest_zoneidx, MAX_NR_ZONES);
> >
> > ret = try_to_freeze();
> > if (kthread_should_stop())
> > @@ -3931,9 +3935,10 @@ static int kswapd(void *p)
> > * but kcompactd is woken to compact for the original
> > * request (alloc_order).
> > */
> > - trace_mm_vmscan_kswapd_wake(pgdat->node_id, classzone_idx,
> > + trace_mm_vmscan_kswapd_wake(pgdat->node_id, highest_zoneidx,
> > alloc_order);
> > - reclaim_order = balance_pgdat(pgdat, alloc_order, classzone_idx);
> > + reclaim_order = balance_pgdat(pgdat, alloc_order,
> > + highest_zoneidx);
> > if (reclaim_order < alloc_order)
> > goto kswapd_try_sleep;
> > }
> > @@ -3951,7 +3956,7 @@ static int kswapd(void *p)
> > * needed.
> > */
> > void wakeup_kswapd(struct zone *zone, gfp_t gfp_flags, int order,
> > - enum zone_type classzone_idx)
> > + enum zone_type highest_zoneidx)
> > {
> > pg_data_t *pgdat;
> > enum zone_type curr_idx;
> > @@ -3963,10 +3968,10 @@ void wakeup_kswapd(struct zone *zone, gfp_t gfp_flags, int order,
> > return;
> >
> > pgdat = zone->zone_pgdat;
> > - curr_idx = READ_ONCE(pgdat->kswapd_classzone_idx);
> > + curr_idx = READ_ONCE(pgdat->kswapd_highest_zoneidx);
> >
> > - if (curr_idx == MAX_NR_ZONES || curr_idx < classzone_idx)
> > - WRITE_ONCE(pgdat->kswapd_classzone_idx, classzone_idx);
> > + if (curr_idx == MAX_NR_ZONES || curr_idx < highest_zoneidx)
> > + WRITE_ONCE(pgdat->kswapd_highest_zoneidx, highest_zoneidx);
> >
> > if (READ_ONCE(pgdat->kswapd_order) < order)
> > WRITE_ONCE(pgdat->kswapd_order, order);
> > @@ -3976,8 +3981,8 @@ void wakeup_kswapd(struct zone *zone, gfp_t gfp_flags, int order,
> >
> > /* Hopeless node, leave it to direct reclaim if possible */
> > if (pgdat->kswapd_failures >= MAX_RECLAIM_RETRIES ||
> > - (pgdat_balanced(pgdat, order, classzone_idx) &&
> > - !pgdat_watermark_boosted(pgdat, classzone_idx))) {
> > + (pgdat_balanced(pgdat, order, highest_zoneidx) &&
> > + !pgdat_watermark_boosted(pgdat, highest_zoneidx))) {
> > /*
> > * There may be plenty of free memory available, but it's too
> > * fragmented for high-order allocations. Wake up kcompactd
> > @@ -3986,11 +3991,11 @@ void wakeup_kswapd(struct zone *zone, gfp_t gfp_flags, int order,
> > * ratelimit its work.
> > */
> > if (!(gfp_flags & __GFP_DIRECT_RECLAIM))
> > - wakeup_kcompactd(pgdat, order, classzone_idx);
> > + wakeup_kcompactd(pgdat, order, highest_zoneidx);
> > return;
> > }
> >
> > - trace_mm_vmscan_wakeup_kswapd(pgdat->node_id, classzone_idx, order,
> > + trace_mm_vmscan_wakeup_kswapd(pgdat->node_id, highest_zoneidx, order,
> > gfp_flags);
> > wake_up_interruptible(&pgdat->kswapd_wait);
> > }
> > --
> > 2.7.4
> >
> >
Powered by blists - more mailing lists