[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <1199378818.8274.25.camel@cinder.waste.org>
Date: Thu, 03 Jan 2008 10:46:58 -0600
From: Matt Mackall <mpm@...enic.com>
To: Ingo Molnar <mingo@...e.hu>
Cc: Linus Torvalds <torvalds@...ux-foundation.org>,
Pekka Enberg <penberg@...helsinki.fi>,
Hugh Dickins <hugh@...itas.com>,
Andi Kleen <andi@...stfloor.org>,
Christoph Lameter <clameter@....com>,
Peter Zijlstra <a.p.zijlstra@...llo.nl>,
Linux Kernel Mailing List <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] procfs: provide slub's /proc/slabinfo
On Thu, 2008-01-03 at 09:52 +0100, Ingo Molnar wrote:
> * Matt Mackall <mpm@...enic.com> wrote:
>
> > > Which means that SLOB could also trivially implement the same thing,
> > > with no new #ifdef'fery or other crud.
> >
> > Except SLOB's emulation of slabs is so thin, it doesn't have the
> > relevant information. We have a very small struct kmem_cache, which I
> > suppose could contain a counter. But we don't have anything like the
> > kmalloc slabs, so you'd only be getting half the picture anyway. The
> > output of slabtop would simply be misleading because there are no
> > underlying "slabs" in the first place.
>
> i think SLOB/embedded is sufficiently special that a "no /proc/slabinfo"
> restriction is perfectly supportable. (for instance it's only selectable
> if CONFIG_EMBEDDED=y) If a SLOB user has any memory allocation problems
> it's worth going to the bigger allocators anyway, to get all the
> debugging goodies.
>
> btw., do you think it would be worth/possible to have build mode for
> SLUB that is acceptably close to the memory efficiency of SLOB? (and
> hence work towards unifying all the 3 allocators into SLUB in essence)
There are three downsides with the slab-like approach: internal
fragmentation, under-utilized slabs, and pinning.
The first is the situation where we ask for a kmalloc of 33 bytes and
get 64. I think the average kmalloc wastes about 30% trying to fit into
power-of-two buckets. We can tune our buckets a bit, but I think in
general trying to back kmalloc with slabs is problematic. SLOB has a
2-byte granularity up to the point where it just hands things off to the
page allocator.
If we tried to add more slabs to fill the gaps, we'd exacerbate the
second problem: because only one type of object can go on a slab, a lot
of slabs are half-full. SLUB's automerging of slabs helps some here, but
is still restricted to objects of the same size.
And finally, there's the whole pinning problem: we can have a cache like
the dcache grow very large and then contract, but still have most of its
slabs used by pinned dentries. Christoph has some rather hairy patches
to address this, but SLOB doesn't have much of a problem here - those
pages are still available to allocate other objects on.
By comparison, SLOB's big downsides are that it's not O(1) and it has a
single lock. But it's currently fast enough to keep up with SLUB on
kernel compiles on my 2G box and Nick had an allocator benchmark where
scalability didn't fall off until beyond 4 CPUs.
> right now we are far away from it - SLUB has an order of magnitude
> larger .o than SLOB, even on UP. I'm wondering why that is so - SLUB's
> data structures _are_ quite compact and could in theory be used in a
> SLOB-alike way. Perhaps one problem is that much of SLUB's debugging
> code is always built in?
I think we should probably just accept that it makes sense to have more
than one allocator. A 64MB single CPU machine is very, very different
than a 64TB 4096-CPU machine. On one of those, it probably makes some
sense to burn some memory for maximum scalability.
--
Mathematics is the supreme nostalgia of our time.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists