lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Fri, 8 Jul 2016 10:52:03 +0100
From:	Mel Gorman <mgorman@...hsingularity.net>
To:	Dave Chinner <david@...morbit.com>
Cc:	Andrew Morton <akpm@...ux-foundation.org>,
	Linux-MM <linux-mm@...ck.org>, Rik van Riel <riel@...riel.com>,
	Vlastimil Babka <vbabka@...e.cz>,
	Johannes Weiner <hannes@...xchg.org>,
	LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH 00/31] Move LRU page reclaim from zones to nodes v8

On Fri, Jul 08, 2016 at 09:27:13AM +1000, Dave Chinner wrote:
> .....
> > This series is not without its hazards. There are at least three areas
> > that I'm concerned with even though I could not reproduce any problems in
> > that area.
> > 
> > 1. Reclaim/compaction is going to be affected because the amount of reclaim is
> >    no longer targetted at a specific zone. Compaction works on a per-zone basis
> >    so there is no guarantee that reclaiming a few THP's worth page pages will
> >    have a positive impact on compaction success rates.
> > 
> > 2. The Slab/LRU reclaim ratio is affected because the frequency the shrinkers
> >    are called is now different. This may or may not be a problem but if it
> >    is, it'll be because shrinkers are not called enough and some balancing
> >    is required.
> 
> Given that XFS has a much more complex set of shrinkers and has a
> much more finely tuned balancing between LRU and shrinker reclaim,
> I'd be interested to see if you get the same results on XFS for the
> tests you ran on ext4. It might also be worth running some highly
> concurrent inode cache benchmarks (e.g. the 50-million inode, 16-way
> concurrent fsmark tests) to see what impact heavy slab cache
> pressure has on shrinker behaviour and system balance...
> 

I had tested XFS with earlier releases and noticed no major problems
so later releases tested only one filesystem.  Given the changes since,
a retest is desirable. I've posted the current version of the series but
I'll queue the tests to run over the weekend. They are quite time consuming
to run unfortunately.

On the fsmark configuration, I configured the test to use 4K files
instead of 0-sized files that normally would be used to stress inode
creation/deletion. This is to have a mix of page cache and slab
allocations. Shout if this does not suit your expectations.

Finally, not all the machines I'm using can store 50 million inodes
of this size. The benchmark has been configured to use as many inodes
as it estimates will fit in the disk. In all cases, it'll exert memory
pressure. Unfortunately, the storage is simple so there is no guarantee
it'll find all problems but that's standard unfortunately.

Thanks.

-- 
Mel Gorman
SUSE Labs

Powered by blists - more mailing lists