[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20150907101610.44504597@redhat.com>
Date: Mon, 7 Sep 2015 10:16:10 +0200
From: Jesper Dangaard Brouer <brouer@...hat.com>
To: Alexander Duyck <alexander.duyck@...il.com>
Cc: netdev@...r.kernel.org, akpm@...ux-foundation.org,
linux-mm@...ck.org, aravinda@...ux.vnet.ibm.com,
Christoph Lameter <cl@...ux.com>,
"Paul E. McKenney" <paulmck@...ux.vnet.ibm.com>,
iamjoonsoo.kim@....com, brouer@...hat.com
Subject: Re: [RFC PATCH 0/3] Network stack, first user of SLAB/kmem_cache
bulk free API.
On Fri, 4 Sep 2015 11:09:21 -0700
Alexander Duyck <alexander.duyck@...il.com> wrote:
> This is an interesting start. However I feel like it might work better
> if you were to create a per-cpu pool for skbs that could be freed and
> allocated in NAPI context. So for example we already have
> napi_alloc_skb, why not just add a napi_free_skb
I do like the idea...
> and then make the array
> of objects to be freed part of a pool that could be used for either
> allocation or freeing? If the pool runs empty you just allocate
> something like 8 or 16 new skb heads, and if you fill it you just free
> half of the list?
But I worry that this algorithm will "randomize" the (skb) objects.
And the SLUB bulk optimization only works if we have many objects
belonging to the same page.
It would likely be fastest to implement a simple stack (for these
per-cpu pools), but I again worry that it would randomize the
object-pages. A simple queue might be better, but slightly slower.
Guess I could just reuse part of qmempool / alf_queue as a quick test.
Having a per-cpu pool in networking would solve the problem of the slub
per-cpu pool isn't large enough for our use-case. On the other hand,
maybe we should fix slub to dynamically adjust the size of it's per-cpu
resources?
A pre-req knowledge (for people not knowing slub's internal details):
Slub alloc path will pickup a page, and empty all objects for that page
before proceeding to the next page. Thus, slub bulk alloc will give
many objects belonging to the page. I'm trying to keep these objects
grouped together until they can be free'ed in a bulk.
--
Best regards,
Jesper Dangaard Brouer
MSc.CS, Sr. Network Kernel Developer at Red Hat
Author of http://www.iptv-analyzer.org
LinkedIn: http://www.linkedin.com/in/brouer
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists