lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <20221026173846.GO5600@paulmck-ThinkPad-P17-Gen-1>
Date:   Wed, 26 Oct 2022 10:38:46 -0700
From:   "Paul E. McKenney" <paulmck@...nel.org>
To:     "Uladzislau Rezki (Sony)" <urezki@...il.com>
Cc:     LKML <linux-kernel@...r.kernel.org>, RCU <rcu@...r.kernel.org>,
        Frederic Weisbecker <frederic@...nel.org>,
        Neeraj Upadhyay <neeraj.iitr10@...il.com>,
        Joel Fernandes <joel@...lfernandes.org>,
        Oleksiy Avramchenko <oleksiy.avramchenko@...y.com>
Subject: Re: [PATCH 1/1] rcu: Refactor kvfree_call_rcu() and high-level
 helpers

On Tue, Oct 25, 2022 at 04:46:12PM +0200, Uladzislau Rezki (Sony) wrote:
> Currently a kvfree_call_rcu() takes an offset within a
> structure as a second parameter, so a helper such as a
> kvfree_rcu_arg_2() has to convert rcu_head and a freed
> ptr to an offset in order to pass it. That leads to an
> extra conversion on macro entry.
> 
> Instead of converting, refactor the code in way that a
> pointer that has to be freed is passed directly to the
> kvfree_call_rcu().
> 
> This patch does not make any functional change and is
> transparent to all kvfree_rcu() users.
> 
> Signed-off-by: Uladzislau Rezki (Sony) <urezki@...il.com>

Nice simplification!

Hearing no immediate protests, I have pulled this in for testing and
for further review.

							Thanx, Paul

> ---
>  include/linux/rcupdate.h |  5 ++---
>  include/linux/rcutiny.h  | 12 ++++++------
>  include/linux/rcutree.h  |  2 +-
>  kernel/rcu/tiny.c        |  9 +++------
>  kernel/rcu/tree.c        | 29 ++++++++++++-----------------
>  5 files changed, 24 insertions(+), 33 deletions(-)
> 
> diff --git a/include/linux/rcupdate.h b/include/linux/rcupdate.h
> index 08605ce7379d..cbd34058fff1 100644
> --- a/include/linux/rcupdate.h
> +++ b/include/linux/rcupdate.h
> @@ -985,8 +985,7 @@ do {									\
>  									\
>  	if (___p) {									\
>  		BUILD_BUG_ON(!__is_kvfree_rcu_offset(offsetof(typeof(*(ptr)), rhf)));	\
> -		kvfree_call_rcu(&((___p)->rhf), (rcu_callback_t)(unsigned long)		\
> -			(offsetof(typeof(*(ptr)), rhf)));				\
> +		kvfree_call_rcu(&((___p)->rhf), (void *) (___p));			\
>  	}										\
>  } while (0)
>  
> @@ -995,7 +994,7 @@ do {								\
>  	typeof(ptr) ___p = (ptr);				\
>  								\
>  	if (___p)						\
> -		kvfree_call_rcu(NULL, (rcu_callback_t) (___p));	\
> +		kvfree_call_rcu(NULL, (void *) (___p));		\
>  } while (0)
>  
>  /*
> diff --git a/include/linux/rcutiny.h b/include/linux/rcutiny.h
> index 768196a5f39d..9623c039964c 100644
> --- a/include/linux/rcutiny.h
> +++ b/include/linux/rcutiny.h
> @@ -98,25 +98,25 @@ static inline void synchronize_rcu_expedited(void)
>   */
>  extern void kvfree(const void *addr);
>  
> -static inline void __kvfree_call_rcu(struct rcu_head *head, rcu_callback_t func)
> +static inline void __kvfree_call_rcu(struct rcu_head *head, void *ptr)
>  {
>  	if (head) {
> -		call_rcu(head, func);
> +		call_rcu(head, (rcu_callback_t) ((void *) head - ptr));
>  		return;
>  	}
>  
>  	// kvfree_rcu(one_arg) call.
>  	might_sleep();
>  	synchronize_rcu();
> -	kvfree((void *) func);
> +	kvfree(ptr);
>  }
>  
>  #ifdef CONFIG_KASAN_GENERIC
> -void kvfree_call_rcu(struct rcu_head *head, rcu_callback_t func);
> +void kvfree_call_rcu(struct rcu_head *head, void *ptr);
>  #else
> -static inline void kvfree_call_rcu(struct rcu_head *head, rcu_callback_t func)
> +static inline void kvfree_call_rcu(struct rcu_head *head, void *ptr)
>  {
> -	__kvfree_call_rcu(head, func);
> +	__kvfree_call_rcu(head, ptr);
>  }
>  #endif
>  
> diff --git a/include/linux/rcutree.h b/include/linux/rcutree.h
> index 5efb51486e8a..e37a0747a8b2 100644
> --- a/include/linux/rcutree.h
> +++ b/include/linux/rcutree.h
> @@ -33,7 +33,7 @@ static inline void rcu_virt_note_context_switch(int cpu)
>  }
>  
>  void synchronize_rcu_expedited(void);
> -void kvfree_call_rcu(struct rcu_head *head, rcu_callback_t func);
> +void kvfree_call_rcu(struct rcu_head *head, void *ptr);
>  
>  void rcu_barrier(void);
>  bool rcu_eqs_special_set(int cpu);
> diff --git a/kernel/rcu/tiny.c b/kernel/rcu/tiny.c
> index a33a8d4942c3..6c9496d08669 100644
> --- a/kernel/rcu/tiny.c
> +++ b/kernel/rcu/tiny.c
> @@ -246,15 +246,12 @@ bool poll_state_synchronize_rcu(unsigned long oldstate)
>  EXPORT_SYMBOL_GPL(poll_state_synchronize_rcu);
>  
>  #ifdef CONFIG_KASAN_GENERIC
> -void kvfree_call_rcu(struct rcu_head *head, rcu_callback_t func)
> +void kvfree_call_rcu(struct rcu_head *head, void *ptr)
>  {
> -	if (head) {
> -		void *ptr = (void *) head - (unsigned long) func;
> -
> +	if (head)
>  		kasan_record_aux_stack_noalloc(ptr);
> -	}
>  
> -	__kvfree_call_rcu(head, func);
> +	__kvfree_call_rcu(head, ptr);
>  }
>  EXPORT_SYMBOL_GPL(kvfree_call_rcu);
>  #endif
> diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
> index 0ca21ac0f064..9fea2aff87a1 100644
> --- a/kernel/rcu/tree.c
> +++ b/kernel/rcu/tree.c
> @@ -3068,8 +3068,8 @@ static void kfree_rcu_work(struct work_struct *work)
>  	 * This list is named "Channel 3".
>  	 */
>  	for (; head; head = next) {
> -		unsigned long offset = (unsigned long)head->func;
> -		void *ptr = (void *)head - offset;
> +		void *ptr = (void *) head->func;
> +		unsigned long offset = (void *) head - ptr;
>  
>  		next = head->next;
>  		debug_rcu_head_unqueue((struct rcu_head *)ptr);
> @@ -3307,26 +3307,21 @@ add_ptr_to_bulk_krc_lock(struct kfree_rcu_cpu **krcp,
>   * be free'd in workqueue context. This allows us to: batch requests together to
>   * reduce the number of grace periods during heavy kfree_rcu()/kvfree_rcu() load.
>   */
> -void kvfree_call_rcu(struct rcu_head *head, rcu_callback_t func)
> +void kvfree_call_rcu(struct rcu_head *head, void *ptr)
>  {
>  	unsigned long flags;
>  	struct kfree_rcu_cpu *krcp;
>  	bool success;
> -	void *ptr;
>  
> -	if (head) {
> -		ptr = (void *) head - (unsigned long) func;
> -	} else {
> -		/*
> -		 * Please note there is a limitation for the head-less
> -		 * variant, that is why there is a clear rule for such
> -		 * objects: it can be used from might_sleep() context
> -		 * only. For other places please embed an rcu_head to
> -		 * your data.
> -		 */
> +	/*
> +	 * Please note there is a limitation for the head-less
> +	 * variant, that is why there is a clear rule for such
> +	 * objects: it can be used from might_sleep() context
> +	 * only. For other places please embed an rcu_head to
> +	 * your data.
> +	 */
> +	if (!head)
>  		might_sleep();
> -		ptr = (unsigned long *) func;
> -	}
>  
>  	// Queue the object but don't yet schedule the batch.
>  	if (debug_rcu_head_queue(ptr)) {
> @@ -3347,7 +3342,7 @@ void kvfree_call_rcu(struct rcu_head *head, rcu_callback_t func)
>  			// Inline if kvfree_rcu(one_arg) call.
>  			goto unlock_return;
>  
> -		head->func = func;
> +		head->func = ptr;
>  		head->next = krcp->head;
>  		krcp->head = head;
>  		success = true;
> -- 
> 2.30.2
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ