[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <alpine.DEB.2.00.1105171220400.5438@chino.kir.corp.google.com>
Date: Tue, 17 May 2011 12:25:39 -0700 (PDT)
From: David Rientjes <rientjes@...gle.com>
To: Mel Gorman <mgorman@...e.de>
cc: Andrea Arcangeli <aarcange@...hat.com>,
Andrew Morton <akpm@...ux-foundation.org>,
James Bottomley <James.Bottomley@...senpartnership.com>,
Colin King <colin.king@...onical.com>,
Raghavendra D Prabhu <raghu.prabhu13@...il.com>,
Jan Kara <jack@...e.cz>, Chris Mason <chris.mason@...cle.com>,
Christoph Lameter <cl@...ux.com>,
Pekka Enberg <penberg@...nel.org>,
Rik van Riel <riel@...hat.com>,
Johannes Weiner <hannes@...xchg.org>,
linux-fsdevel <linux-fsdevel@...r.kernel.org>,
linux-mm <linux-mm@...ck.org>,
linux-kernel <linux-kernel@...r.kernel.org>,
linux-ext4 <linux-ext4@...r.kernel.org>
Subject: Re: [PATCH 3/3] mm: slub: Default slub_max_order to 0
On Tue, 17 May 2011, Mel Gorman wrote:
> > The fragmentation isn't the only issue with the netperf TCP_RR benchmark,
> > the problem is that the slub slowpath is being used >95% of the time on
> > every allocation and free for the very large number of kmalloc-256 and
> > kmalloc-2K caches.
>
> Ok, that makes sense as I'd full expect that benchmark to exhaust
> the per-cpu page (high order or otherwise) of slab objects routinely
> during default and I'd also expect the freeing on the other side to
> be releasing slabs frequently to the partial or empty lists.
>
That's most of the problem, but it's compounded on this benchmark because
the slab pulled from the partial list to replace the per-cpu page
typically only has a very minimal number (2 or 3) of free objects, so it
can only serve one allocation and then require the allocation slowpath to
pull yet another slab from the partial list the next time around. I had a
patchset that addressed that, which I called "slab thrashing", by only
pulling a slab from the partial list when it had a pre-defined proportion
of available objects and otherwise skipping it, and that ended up helping
the benchmark by 5-7%. Smaller orders will make this worse, as well,
since if there were only 2 or 3 free objects on an order-3 slab before,
there's no chance that's going to be equivalent on an order-0 slab.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists