[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <c13c0c78-55d1-b2e2-c24b-897ce2469410@suse.cz>
Date: Tue, 6 Aug 2019 10:03:42 +0200
From: Vlastimil Babka <vbabka@...e.cz>
To: Mike Kravetz <mike.kravetz@...cle.com>, linux-mm@...ck.org,
linux-kernel@...r.kernel.org
Cc: Hillf Danton <hdanton@...a.com>, Michal Hocko <mhocko@...nel.org>,
Mel Gorman <mgorman@...e.de>,
Johannes Weiner <hannes@...xchg.org>,
Andrea Arcangeli <aarcange@...hat.com>,
David Rientjes <rientjes@...gle.com>,
Andrew Morton <akpm@...ux-foundation.org>
Subject: Re: [PATCH v2 4/4] hugetlbfs: don't retry when pool page allocations
start to fail
On 8/6/19 3:47 AM, Mike Kravetz wrote:
> When allocating hugetlbfs pool pages via /proc/sys/vm/nr_hugepages,
> the pages will be interleaved between all nodes of the system. If
> nodes are not equal, it is quite possible for one node to fill up
> before the others. When this happens, the code still attempts to
> allocate pages from the full node. This results in calls to direct
> reclaim and compaction which slow things down considerably.
>
> When allocating pool pages, note the state of the previous allocation
> for each node. If previous allocation failed, do not use the
> aggressive retry algorithm on successive attempts. The allocation
> will still succeed if there is memory available, but it will not try
> as hard to free up memory.
>
> Signed-off-by: Mike Kravetz <mike.kravetz@...cle.com>
Acked-by: Vlastimil Babka <vbabka@...e.cz>
Thanks.
Powered by blists - more mailing lists