[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ff1add3a-446b-1e30-c4c2-cfab035f11f2@linux.intel.com>
Date: Wed, 17 Oct 2018 08:17:05 -0700
From: Alexander Duyck <alexander.h.duyck@...ux.intel.com>
To: Michal Hocko <mhocko@...nel.org>
Cc: linux-mm@...ck.org, akpm@...ux-foundation.org,
pavel.tatashin@...rosoft.com, dave.jiang@...el.com,
linux-kernel@...r.kernel.org, willy@...radead.org,
davem@...emloft.net, yi.z.zhang@...ux.intel.com,
khalid.aziz@...cle.com, rppt@...ux.vnet.ibm.com, vbabka@...e.cz,
sparclinux@...r.kernel.org, dan.j.williams@...el.com,
ldufour@...ux.vnet.ibm.com, mgorman@...hsingularity.net,
mingo@...nel.org, kirill.shutemov@...ux.intel.com
Subject: Re: [mm PATCH v3 3/6] mm: Use memblock/zone specific iterator for
handling deferred page init
On 10/17/2018 2:11 AM, Michal Hocko wrote:
> On Mon 15-10-18 13:27:09, Alexander Duyck wrote:
>> This patch introduces a new iterator for_each_free_mem_pfn_range_in_zone.
>>
>> This iterator will take care of making sure a given memory range provided
>> is in fact contained within a zone. It takes are of all the bounds checking
>> we were doing in deferred_grow_zone, and deferred_init_memmap. In addition
>> it should help to speed up the search a bit by iterating until the end of a
>> range is greater than the start of the zone pfn range, and will exit
>> completely if the start is beyond the end of the zone.
>>
>> This patch adds yet another iterator called
>> for_each_free_mem_range_in_zone_from and then uses it to support
>> initializing and freeing pages in groups no larger than MAX_ORDER_NR_PAGES.
>> By doing this we can greatly improve the cache locality of the pages while
>> we do several loops over them in the init and freeing process.
>>
>> We are able to tighten the loops as a result since we only really need the
>> checks for first_init_pfn in our first iteration and after that we can
>> assume that all future values will be greater than this. So I have added a
>> function called deferred_init_mem_pfn_range_in_zone that primes the
>> iterators and if it fails we can just exit.
>
> Numbers please.
>
> Besides that, this adds a lot of code and I am not convinced the result
> is so much better to justify that.
If I recall most of the gains are due to better cache locality. Instead
of running through all of memory once for init, and once for freeing
this patch has us doing it in MAX_ORDER_NR_PAGES sized chunks. So the
advantage is that we can keep most of the pages structs in the L2 cache
at least on x86 processors to avoid having to go to memory as much.
I'll run performance numbers per patch today and try to make certain I
have a line mentioning the delta for each patch in the v4 patch set.
- Alex
Powered by blists - more mailing lists