lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Tue, 05 Dec 2006 15:48:06 +0000
From:	Andy Whitcroft <apw@...dowen.org>
To:	Mel Gorman <mel@....ul.ie>
CC:	Andrew Morton <akpm@...l.org>, clameter@....com,
	Linux Memory Management List <linux-mm@...ck.org>,
	Linux Kernel Mailing List <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] Add __GFP_MOVABLE for callers to flag allocations that
 may be migrated

Mel Gorman wrote:
> On Mon, 4 Dec 2006, Andrew Morton wrote:

>> , but I would of course prefer to avoid
>> merging the anti-frag patches simply based on their stupendous size. 
>> It seems to me that lumpy-reclaim is suitable for the e1000 problem
>> , but perhaps not for the hugetlbpage problem.
> 
> I believe you'll hit similar problems even with lumpy-reclaim for the 
> e1000 (I've added Andy to the cc so he can comment more). Lumpy provides 
> a much smarter way of freeing higher-order contiguous blocks without 
> having to reclaim 95%+ of memory - this is good. However, if you are 
> currently seeing situations where the allocations fails even after you 
> page out everything possible, smarter reclaim that eventually pages out 
> everything anyway will not help you (chances are it's something like 
> page tables that are in your way).

The pre-lumpy algorithm is capable of producing reasonable numbers of 
very low order pages.  Lumpy should improve success rates producing 
sucessful reclaim at higher order than that.  Its success is limited 
however by the percentage of non-reclaimable pages and their distribution.

The e1000 problem is that it wants order=3 pages ie. 8 pages in size. 
For lumpy to have a high chance of success we would need the average 
unmovable page count to be significantly less than 1 in 8 pages 
(assuming a random distribution) (<12% pinned).  In stress testing we 
find we can reclaim of the order of 70% of memory, this tends to 
indicates that the pinned memory is more like 25% than 10%.  It would 
suggest that we are going to find reclaim rates above order=2 are poor 
without explicit placement control.

Obviously this all depends on the workload.  Our test workloads are 
known to be fairly hostile in terms of fragmentation.  So I would love 
to see lumpy tested in the problem scenario to get some data on that setup.

> This is where anti-frag comes in. It clusters pages together based on 
> their type - unmovable, reapable (inode caches, short-lived kernel 
> allocations, skbuffs etc) and movable. When kswapd kicks in, the slab 
> caches will be reaped. As reapable pages are clustered together, that 
> will free some contiguous areas - probably enough for the e1000 
> allocations to succeed!
> 
> If that doesn't work, kswapd and direct reclaim will start reclaiming 
> the "movable" pages. Without lumpy reclaim, 95%+ of memory could be 
> paged out which is bad. Lumpy finds the contiguous pages faster and with 
> less IO, that's why it's important.
> 
> Tests I am aware of show that lumpy-reclaim on it's own makes little or 
> no difference to the hugetlb page problem. However, with anti-frag, 
> hugetlb-sized allocations succeed much more often even when under memory 
> pressure.

At high order both traditional and lumpy reclaim are next to useless 
without placement controls.

> 
>> Whereas anti-fragmentation adds
>> vastly more code, but can address both problems?  Or something.
>>
> 
> Anti-frag goes a long way to addressing both problems. Lumpy-reclaim 
> increases it's success rates under memory pressure and reduces the 
> amount of reclaim that occurs.
> 
>> IOW: big-picture where-do-we-go-from-here stuff.
>>
> 
> Start with lumpy reclaim, then I'd like to merge page clustering piece 
> by piece, ideally with one of the people with e1000 problems testing to 
> see does it make a difference.
> 
> Assuming they are shown to help, where we'd go from there would be stuff 
> like;
> 
> 1. Keep non-movable and reapable allocations at the lower PFNs as much as
>    possible. This is so DIMMS for higher PFNs can be removed (doesn't
>    exist)
> 2. Use page migration to compact memory rather than depending solely on
>    reclaim (doesn't exist)
> 3. Introduce a mechanism for marking a group of pages as being offlined so
>    that they are not reallocated (code that does something like this
>    exists)
> 4. Resurrect the hotplug-remove code (exists, but probably very stale)
> 5. Allow allocations for hugepages outside of the pool as long as the
>    process remains with it's locked_vm limits (patches were posted to
>    libhugetlbfs last Friday. will post to linux-mm tomorrow).

-apw

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ