lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:   Thu, 14 Jan 2021 11:41:14 +0000
From:   Alexander Lobakin <alobakin@...me>
To:     Eric Dumazet <edumazet@...gle.com>
Cc:     Alexander Lobakin <alobakin@...me>,
        "David S. Miller" <davem@...emloft.net>,
        Jakub Kicinski <kuba@...nel.org>,
        Edward Cree <ecree.xilinx@...il.com>,
        Jonathan Lemon <jonathan.lemon@...il.com>,
        Willem de Bruijn <willemb@...gle.com>,
        Miaohe Lin <linmiaohe@...wei.com>,
        Steffen Klassert <steffen.klassert@...unet.com>,
        Guillaume Nault <gnault@...hat.com>,
        Yadu Kishore <kyk.segfault@...il.com>,
        Al Viro <viro@...iv.linux.org.uk>,
        netdev <netdev@...r.kernel.org>,
        LKML <linux-kernel@...r.kernel.org>,
        Dmitry Vyukov <dvyukov@...gle.com>
Subject: Re: [PATCH v2 net-next 2/3] skbuff: (re)use NAPI skb cache on allocation path

From: Eric Dumazet <edumazet@...gle.com>
Date: Wed, 13 Jan 2021 15:36:05 +0100

> On Wed, Jan 13, 2021 at 2:37 PM Alexander Lobakin <alobakin@...me> wrote:
>>
>> Instead of calling kmem_cache_alloc() every time when building a NAPI
>> skb, (re)use skbuff_heads from napi_alloc_cache.skb_cache. Previously
>> this cache was only used for bulk-freeing skbuff_heads consumed via
>> napi_consume_skb() or __kfree_skb_defer().
>>
>> Typical path is:
>>  - skb is queued for freeing from driver or stack, its skbuff_head
>>    goes into the cache instead of immediate freeing;
>>  - driver or stack requests NAPI skb allocation, an skbuff_head is
>>    taken from the cache instead of allocation.
>>
>> Corner cases:
>>  - if it's empty on skb allocation, bulk-allocate the first half;
>>  - if it's full on skb consuming, bulk-wipe the second half.
>>
>> Also try to balance its size after completing network softirqs
>> (__kfree_skb_flush()).
>
> I do not see the point of doing this rebalance (especially if we do not change
> its name describing its purpose more accurately).
>
> For moderate load, we will have a reduced bulk size (typically one or two).
> Number of skbs in the cache is in [0, 64[ , there is really no risk of
> letting skbs there for a long period of time.
> (32 * sizeof(sk_buff) = 8192)
> I would personally get rid of this function completely.

When I had a cache of 128 entries, I had worse results without this
function. But seems like I forgot to retest when I switched to the
original size of 64.
I also thought about removing this function entirely, will test.

> Also it seems you missed my KASAN support request ?
> I guess this is a matter of using kasan_unpoison_range(), we can ask for help.

I saw your request, but don't see a reason for doing this.
We are not caching already freed skbuff_heads. They don't get
kmem_cache_freed before getting into local cache. KASAN poisons
them no earlier than at kmem_cache_free() (or did I miss someting?).
heads being cached just get rid of all references and at the moment
of dropping to the cache they are pretty the same as if they were
allocated.

I also remind that only skbs that are caught by napi_consume_skb() or
__kfree_skb_defer() are getting into skb_cache, not every single one.

Regarding other emails:

1. NUMA awareness.

napi_alloc_cache is percpu, we're partly protected. The only thing
that might happen is that napi_consume_skb() can be called for skb
that was allocated at a distant node, and then it's requested by
napi_alloc_skb() (and there were no bulk-wipes between).
This can occur only if a NAPI polling cycle for cleaning up the
completion/send queue(s) is scheduled on a CPU that is far away
from the one(s) that clean(s) up the receive queue(s).
That is really very unlikely to be caught, but...

One of the ways to handle this is like (inside napi_skb_cache_get()):

	skb = nc->skb_cache[--nc->skb_count];
	if (unlikely(pfn_to_nid(virt_to_pfn(skb)) != numa_mem_id())) {
		kmem_cache_free(skbuff_head_cache, skb);
		skb = kmem_cache_alloc(skbuff_head_cache, GFP_ATOMIC);
	}

	return skb;

This whole condition will be optimized out on !CONFIG_NUMA, as
pfn_to_nid() and numa_mem_id() are compile-time 0 in this case.
This won't break currently present bulk-freeing.

2. Where do optimizations come from.

Not only from bulk allocations, but also from the shortcut:

napi_consume_skb()/__kfree_skb_defer() -> skb_cache -> napi_alloc_skb();

napi_alloc_skb() will get a new head directly without calling for MM
functions.
I'm aware that kmem_cache has its own cache, but this also applies to
page allocators etc. which doesn't prevent from having things like
page_frag_cache or page_pool to recycle pages and fragments directly,
not through MM layer.

>> prefetchw() on CONFIG_SLUB is dropped since it makes no sense anymore.
>>
>> Suggested-by: Edward Cree <ecree.xilinx@...il.com>
>> Signed-off-by: Alexander Lobakin <alobakin@...me>
>> ---
>>  net/core/skbuff.c | 54 ++++++++++++++++++++++++++++++-----------------
>>  1 file changed, 35 insertions(+), 19 deletions(-)
>>
>> diff --git a/net/core/skbuff.c b/net/core/skbuff.c
>> index dc3300dc2ac4..f42a3a04b918 100644
>> --- a/net/core/skbuff.c
>> +++ b/net/core/skbuff.c
>> @@ -364,6 +364,7 @@ struct sk_buff *build_skb_around(struct sk_buff *skb,
>>  EXPORT_SYMBOL(build_skb_around);
>>
>>  #define NAPI_SKB_CACHE_SIZE    64
>> +#define NAPI_SKB_CACHE_HALF    (NAPI_SKB_CACHE_SIZE / 2)
>>
>>  struct napi_alloc_cache {
>>         struct page_frag_cache page;
>> @@ -487,7 +488,15 @@ EXPORT_SYMBOL(__netdev_alloc_skb);
>>
>>  static struct sk_buff *napi_skb_cache_get(struct napi_alloc_cache *nc)
>>  {
>> -       return kmem_cache_alloc(skbuff_head_cache, GFP_ATOMIC);
>> +       if (unlikely(!nc->skb_count))
>> +               nc->skb_count = kmem_cache_alloc_bulk(skbuff_head_cache,
>> +                                                     GFP_ATOMIC,
>> +                                                     NAPI_SKB_CACHE_HALF,
>> +                                                     nc->skb_cache);
>> +       if (unlikely(!nc->skb_count))
>> +               return NULL;
>> +
>> +       return nc->skb_cache[--nc->skb_count];
>>  }
>>
>>  /**
>> @@ -867,40 +876,47 @@ void __consume_stateless_skb(struct sk_buff *skb)
>>  void __kfree_skb_flush(void)
>>  {
>>         struct napi_alloc_cache *nc = this_cpu_ptr(&napi_alloc_cache);
>> +       size_t count;
>> +       void **ptr;
>> +
>> +       if (unlikely(nc->skb_count == NAPI_SKB_CACHE_HALF))
>> +               return;
>> +
>> +       if (nc->skb_count > NAPI_SKB_CACHE_HALF) {
>> +               count = nc->skb_count - NAPI_SKB_CACHE_HALF;
>> +               ptr = nc->skb_cache + NAPI_SKB_CACHE_HALF;
>>
>> -       /* flush skb_cache if containing objects */
>> -       if (nc->skb_count) {
>> -               kmem_cache_free_bulk(skbuff_head_cache, nc->skb_count,
>> -                                    nc->skb_cache);
>> -               nc->skb_count = 0;
>> +               kmem_cache_free_bulk(skbuff_head_cache, count, ptr);
>> +               nc->skb_count = NAPI_SKB_CACHE_HALF;
>> +       } else {
>> +               count = NAPI_SKB_CACHE_HALF - nc->skb_count;
>> +               ptr = nc->skb_cache + nc->skb_count;
>> +
>> +               nc->skb_count += kmem_cache_alloc_bulk(skbuff_head_cache,
>> +                                                      GFP_ATOMIC, count,
>> +                                                      ptr);
>>         }
>>  }
>>

Al

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ