lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAPhsuW7j9SK5OJxK-ZnDYgwVhoPJGP8K+1Q+pUJ8mGUA41ZvHQ@mail.gmail.com>
Date:   Sun, 20 Feb 2022 17:28:06 -0800
From:   Song Liu <song@...nel.org>
To:     trix@...hat.com
Cc:     Alexei Starovoitov <ast@...nel.org>,
        Daniel Borkmann <daniel@...earbox.net>,
        Andrii Nakryiko <andrii@...nel.org>,
        Martin KaFai Lau <kafai@...com>,
        Song Liu <songliubraving@...com>, Yonghong Song <yhs@...com>,
        John Fastabend <john.fastabend@...il.com>,
        KP Singh <kpsingh@...nel.org>,
        Networking <netdev@...r.kernel.org>, bpf <bpf@...r.kernel.org>,
        open list <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] bpf: cleanup comments

On Sun, Feb 20, 2022 at 10:41 AM <trix@...hat.com> wrote:
>
> From: Tom Rix <trix@...hat.com>
>
> Add leading space to spdx tag
> Use // for spdx c file comment
>
> Replacements
> resereved to reserved
> inbetween to in between
> everytime to every time

I think everytime could be a single word? Other than that,

Acked-by: Song Liu <songliubraving@...com>

> intutivie to intuitive
> currenct to current
> encontered to encountered
> referenceing to referencing
> upto to up to
> exectuted to executed
>
> Signed-off-by: Tom Rix <trix@...hat.com>
> ---
>  kernel/bpf/bpf_local_storage.c | 2 +-
>  kernel/bpf/btf.c               | 6 +++---
>  kernel/bpf/cgroup.c            | 8 ++++----
>  kernel/bpf/hashtab.c           | 2 +-
>  kernel/bpf/helpers.c           | 2 +-
>  kernel/bpf/local_storage.c     | 2 +-
>  kernel/bpf/reuseport_array.c   | 2 +-
>  kernel/bpf/syscall.c           | 2 +-
>  kernel/bpf/trampoline.c        | 2 +-
>  9 files changed, 14 insertions(+), 14 deletions(-)
>
> diff --git a/kernel/bpf/bpf_local_storage.c b/kernel/bpf/bpf_local_storage.c
> index 71de2a89869c..092a1ac772d7 100644
> --- a/kernel/bpf/bpf_local_storage.c
> +++ b/kernel/bpf/bpf_local_storage.c
> @@ -136,7 +136,7 @@ bool bpf_selem_unlink_storage_nolock(struct bpf_local_storage *local_storage,
>                  * will be done by the caller.
>                  *
>                  * Although the unlock will be done under
> -                * rcu_read_lock(),  it is more intutivie to
> +                * rcu_read_lock(),  it is more intuitive to
>                  * read if the freeing of the storage is done
>                  * after the raw_spin_unlock_bh(&local_storage->lock).
>                  *
> diff --git a/kernel/bpf/btf.c b/kernel/bpf/btf.c
> index 02d7014417a0..8b11d1a9bee1 100644
> --- a/kernel/bpf/btf.c
> +++ b/kernel/bpf/btf.c
> @@ -1,4 +1,4 @@
> -/* SPDX-License-Identifier: GPL-2.0 */
> +// SPDX-License-Identifier: GPL-2.0
>  /* Copyright (c) 2018 Facebook */
>
>  #include <uapi/linux/btf.h>
> @@ -2547,7 +2547,7 @@ static int btf_ptr_resolve(struct btf_verifier_env *env,
>          *
>          * We now need to continue from the last-resolved-ptr to
>          * ensure the last-resolved-ptr will not referring back to
> -        * the currenct ptr (t).
> +        * the current ptr (t).
>          */
>         if (btf_type_is_modifier(next_type)) {
>                 const struct btf_type *resolved_type;
> @@ -6148,7 +6148,7 @@ int btf_type_snprintf_show(const struct btf *btf, u32 type_id, void *obj,
>
>         btf_type_show(btf, type_id, obj, (struct btf_show *)&ssnprintf);
>
> -       /* If we encontered an error, return it. */
> +       /* If we encountered an error, return it. */
>         if (ssnprintf.show.state.status)
>                 return ssnprintf.show.state.status;
>
> diff --git a/kernel/bpf/cgroup.c b/kernel/bpf/cgroup.c
> index 098632fdbc45..128028efda64 100644
> --- a/kernel/bpf/cgroup.c
> +++ b/kernel/bpf/cgroup.c
> @@ -1031,7 +1031,7 @@ int cgroup_bpf_prog_query(const union bpf_attr *attr,
>   * __cgroup_bpf_run_filter_skb() - Run a program for packet filtering
>   * @sk: The socket sending or receiving traffic
>   * @skb: The skb that is being sent or received
> - * @type: The type of program to be exectuted
> + * @type: The type of program to be executed
>   *
>   * If no socket is passed, or the socket is not of type INET or INET6,
>   * this function does nothing and returns 0.
> @@ -1094,7 +1094,7 @@ EXPORT_SYMBOL(__cgroup_bpf_run_filter_skb);
>  /**
>   * __cgroup_bpf_run_filter_sk() - Run a program on a sock
>   * @sk: sock structure to manipulate
> - * @type: The type of program to be exectuted
> + * @type: The type of program to be executed
>   *
>   * socket is passed is expected to be of type INET or INET6.
>   *
> @@ -1119,7 +1119,7 @@ EXPORT_SYMBOL(__cgroup_bpf_run_filter_sk);
>   *                                       provided by user sockaddr
>   * @sk: sock struct that will use sockaddr
>   * @uaddr: sockaddr struct provided by user
> - * @type: The type of program to be exectuted
> + * @type: The type of program to be executed
>   * @t_ctx: Pointer to attach type specific context
>   * @flags: Pointer to u32 which contains higher bits of BPF program
>   *         return value (OR'ed together).
> @@ -1166,7 +1166,7 @@ EXPORT_SYMBOL(__cgroup_bpf_run_filter_sock_addr);
>   * @sock_ops: bpf_sock_ops_kern struct to pass to program. Contains
>   * sk with connection information (IP addresses, etc.) May not contain
>   * cgroup info if it is a req sock.
> - * @type: The type of program to be exectuted
> + * @type: The type of program to be executed
>   *
>   * socket passed is expected to be of type INET or INET6.
>   *
> diff --git a/kernel/bpf/hashtab.c b/kernel/bpf/hashtab.c
> index d29af9988f37..65877967f414 100644
> --- a/kernel/bpf/hashtab.c
> +++ b/kernel/bpf/hashtab.c
> @@ -1636,7 +1636,7 @@ __htab_map_lookup_and_delete_batch(struct bpf_map *map,
>                 value_size = size * num_possible_cpus();
>         total = 0;
>         /* while experimenting with hash tables with sizes ranging from 10 to
> -        * 1000, it was observed that a bucket can have upto 5 entries.
> +        * 1000, it was observed that a bucket can have up to 5 entries.
>          */
>         bucket_size = 5;
>
> diff --git a/kernel/bpf/helpers.c b/kernel/bpf/helpers.c
> index 49817755b8c3..ae64110a98b5 100644
> --- a/kernel/bpf/helpers.c
> +++ b/kernel/bpf/helpers.c
> @@ -1093,7 +1093,7 @@ struct bpf_hrtimer {
>  struct bpf_timer_kern {
>         struct bpf_hrtimer *timer;
>         /* bpf_spin_lock is used here instead of spinlock_t to make
> -        * sure that it always fits into space resereved by struct bpf_timer
> +        * sure that it always fits into space reserved by struct bpf_timer
>          * regardless of LOCKDEP and spinlock debug flags.
>          */
>         struct bpf_spin_lock lock;
> diff --git a/kernel/bpf/local_storage.c b/kernel/bpf/local_storage.c
> index 23f7f9d08a62..497916060ac7 100644
> --- a/kernel/bpf/local_storage.c
> +++ b/kernel/bpf/local_storage.c
> @@ -1,4 +1,4 @@
> -//SPDX-License-Identifier: GPL-2.0
> +// SPDX-License-Identifier: GPL-2.0
>  #include <linux/bpf-cgroup.h>
>  #include <linux/bpf.h>
>  #include <linux/bpf_local_storage.h>
> diff --git a/kernel/bpf/reuseport_array.c b/kernel/bpf/reuseport_array.c
> index 556a769b5b80..962556917c4d 100644
> --- a/kernel/bpf/reuseport_array.c
> +++ b/kernel/bpf/reuseport_array.c
> @@ -143,7 +143,7 @@ static void reuseport_array_free(struct bpf_map *map)
>
>         /*
>          * Once reaching here, all sk->sk_user_data is not
> -        * referenceing this "array".  "array" can be freed now.
> +        * referencing this "array".  "array" can be freed now.
>          */
>         bpf_map_area_free(array);
>  }
> diff --git a/kernel/bpf/syscall.c b/kernel/bpf/syscall.c
> index 35646db3d950..ce4657a00dae 100644
> --- a/kernel/bpf/syscall.c
> +++ b/kernel/bpf/syscall.c
> @@ -2562,7 +2562,7 @@ static int bpf_link_alloc_id(struct bpf_link *link)
>   * pre-allocated resources are to be freed with bpf_cleanup() call. All the
>   * transient state is passed around in struct bpf_link_primer.
>   * This is preferred way to create and initialize bpf_link, especially when
> - * there are complicated and expensive operations inbetween creating bpf_link
> + * there are complicated and expensive operations in between creating bpf_link
>   * itself and attaching it to BPF hook. By using bpf_link_prime() and
>   * bpf_link_settle() kernel code using bpf_link doesn't have to perform
>   * expensive (and potentially failing) roll back operations in a rare case
> diff --git a/kernel/bpf/trampoline.c b/kernel/bpf/trampoline.c
> index 7224691df2ec..0b41fa993825 100644
> --- a/kernel/bpf/trampoline.c
> +++ b/kernel/bpf/trampoline.c
> @@ -45,7 +45,7 @@ void *bpf_jit_alloc_exec_page(void)
>
>         set_vm_flush_reset_perms(image);
>         /* Keep image as writeable. The alternative is to keep flipping ro/rw
> -        * everytime new program is attached or detached.
> +        * every time new program is attached or detached.
>          */
>         set_memory_x((long)image, 1);
>         return image;
> --
> 2.26.3
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ