[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAEf4BzZaJaYw0tB0R+q3qoQX7=qy3T9jvzf5q=TH++t66wNd-w@mail.gmail.com>
Date: Mon, 26 Oct 2020 15:47:51 -0700
From: Andrii Nakryiko <andrii.nakryiko@...il.com>
To: David Verbeiren <david.verbeiren@...sares.net>
Cc: bpf <bpf@...r.kernel.org>, Networking <netdev@...r.kernel.org>,
Matthieu Baerts <matthieu.baerts@...sares.net>
Subject: Re: [PATCH bpf] bpf: zero-fill re-used per-cpu map element
On Fri, Oct 23, 2020 at 8:48 AM David Verbeiren
<david.verbeiren@...sares.net> wrote:
>
> Zero-fill element values for all cpus, just as when not using
> prealloc. This is the only way the bpf program can ensure known
> initial values for cpus other than the current one ('onallcpus'
> cannot be set when coming from the bpf program).
>
> The scenario is: bpf program inserts some elements in a per-cpu
> map, then deletes some (or userspace does). When later adding
> new elements using bpf_map_update_elem(), the bpf program can
> only set the value of the new elements for the current cpu.
> When prealloc is enabled, previously deleted elements are re-used.
> Without the fix, values for other cpus remain whatever they were
> when the re-used entry was previously freed.
>
> Fixes: 6c9059817432 ("bpf: pre-allocate hash map elements")
> Acked-by: Matthieu Baerts <matthieu.baerts@...sares.net>
> Signed-off-by: David Verbeiren <david.verbeiren@...sares.net>
> ---
> kernel/bpf/hashtab.c | 12 ++++++++++++
> 1 file changed, 12 insertions(+)
>
> diff --git a/kernel/bpf/hashtab.c b/kernel/bpf/hashtab.c
> index 1815e97d4c9c..667553cce65a 100644
> --- a/kernel/bpf/hashtab.c
> +++ b/kernel/bpf/hashtab.c
> @@ -836,6 +836,7 @@ static struct htab_elem *alloc_htab_elem(struct bpf_htab *htab, void *key,
> bool prealloc = htab_is_prealloc(htab);
> struct htab_elem *l_new, **pl_new;
> void __percpu *pptr;
> + int cpu;
>
> if (prealloc) {
> if (old_elem) {
> @@ -880,6 +881,17 @@ static struct htab_elem *alloc_htab_elem(struct bpf_htab *htab, void *key,
> size = round_up(size, 8);
> if (prealloc) {
> pptr = htab_elem_get_ptr(l_new, key_size);
> +
> + /* zero-fill element values for all cpus, just as when
> + * not using prealloc. Only way for bpf program to
> + * ensure known initial values for cpus other than
> + * current one (onallcpus=false when coming from bpf
> + * prog).
> + */
> + if (!onallcpus)
> + for_each_possible_cpu(cpu)
> + memset((void *)per_cpu_ptr(pptr, cpu),
> + 0, size);
Technically, you don't have to memset() for the current CPU, right?
Don't know if extra check is cheaper than avoiding one memset() call,
though.
But regardless, this 6 level nesting looks pretty bad, maybe move the
for_each_possible_cpu() loop into a helper function?
Also, does the per-CPU LRU hashmap need the same treatment?
> } else {
> /* alloc_percpu zero-fills */
> pptr = __alloc_percpu_gfp(size, 8,
> --
> 2.29.0
>
Powered by blists - more mailing lists