lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <64bebc29-a007-4ebc-bf86-8c2c0a7e6bf6@suse.cz>
Date: Wed, 28 Aug 2024 16:58:48 +0200
From: Vlastimil Babka <vbabka@...e.cz>
To: "Uladzislau Rezki (Sony)" <urezki@...il.com>,
 "Paul E . McKenney" <paulmck@...nel.org>
Cc: RCU <rcu@...r.kernel.org>, LKML <linux-kernel@...r.kernel.org>,
 Neeraj upadhyay <Neeraj.Upadhyay@....com>, Boqun Feng
 <boqun.feng@...il.com>, Joel Fernandes <joel@...lfernandes.org>,
 Frederic Weisbecker <frederic@...nel.org>,
 Oleksiy Avramchenko <oleksiy.avramchenko@...y.com>
Subject: Re: [PATCH 1/4] rcu/kvfree: Support dynamic rcu_head for single
 argument objects

On 8/28/24 13:09, Uladzislau Rezki (Sony) wrote:
> Add a support of dynamically attaching an rcu_head to an object
> which gets freed via the single argument of kvfree_rcu(). This is
> used in the path, when a page allocation fails due to a high memory
> pressure.
> 
> The basic idea behind of this is to minimize a hit of slow path
> which requires a caller to wait until a grace period is passed.
> 
> Signed-off-by: Uladzislau Rezki (Sony) <urezki@...il.com>

So IIUC it's a situation where we can't allocate a page, but we hope the
kmalloc-32 slab has still free objects to give us dyn_rcu_head's before it
would have to also make a page allocation?

So that may really be possible and there might potentially be many such
objects, but I wonder if there's really a benefit. The system is struggling
for memory and the single-argument caller specifically is _mightsleep so it
could e.g. instead go direct reclaim a page rather than start depleting the
kmalloc-32 slab, no?

> ---
>  kernel/rcu/tree.c | 53 +++++++++++++++++++++++++++++++++++++++++++----
>  1 file changed, 49 insertions(+), 4 deletions(-)
> 
> diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
> index be00aac5f4e7..0124411fecfb 100644
> --- a/kernel/rcu/tree.c
> +++ b/kernel/rcu/tree.c
> @@ -3425,6 +3425,11 @@ kvfree_rcu_bulk(struct kfree_rcu_cpu *krcp,
>  	cond_resched_tasks_rcu_qs();
>  }
>  
> +struct dyn_rcu_head {
> +	unsigned long *ptr;
> +	struct rcu_head rh;
> +};
> +
>  static void
>  kvfree_rcu_list(struct rcu_head *head)
>  {
> @@ -3433,15 +3438,32 @@ kvfree_rcu_list(struct rcu_head *head)
>  	for (; head; head = next) {
>  		void *ptr = (void *) head->func;
>  		unsigned long offset = (void *) head - ptr;
> +		struct dyn_rcu_head *drhp = NULL;
> +
> +		/*
> +		 * For dynamically attached rcu_head, a ->func field
> +		 * points to _offset_, i.e. not to a pointer which has
> +		 * to be freed. For such objects, adjust an offset and
> +		 * pointer.
> +		 */
> +		if (__is_kvfree_rcu_offset((unsigned long) ptr)) {
> +			drhp = container_of(head, struct dyn_rcu_head, rh);
> +			offset = (unsigned long) drhp->rh.func;
> +			ptr = drhp->ptr;
> +		}
>  
>  		next = head->next;
>  		debug_rcu_head_unqueue((struct rcu_head *)ptr);
>  		rcu_lock_acquire(&rcu_callback_map);
>  		trace_rcu_invoke_kvfree_callback(rcu_state.name, head, offset);
>  
> -		if (!WARN_ON_ONCE(!__is_kvfree_rcu_offset(offset)))
> +		if (!WARN_ON_ONCE(!__is_kvfree_rcu_offset(offset))) {
>  			kvfree(ptr);
>  
> +			if (drhp)
> +				kvfree(drhp);
> +		}
> +
>  		rcu_lock_release(&rcu_callback_map);
>  		cond_resched_tasks_rcu_qs();
>  	}
> @@ -3787,6 +3809,21 @@ add_ptr_to_bulk_krc_lock(struct kfree_rcu_cpu **krcp,
>  	return true;
>  }
>  
> +static struct rcu_head *
> +attach_rcu_head_to_object(void *obj)
> +{
> +	struct dyn_rcu_head *rhp;
> +
> +	rhp = kmalloc(sizeof(struct dyn_rcu_head), GFP_KERNEL |
> +		__GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN);
> +
> +	if (!rhp)
> +		return NULL;
> +
> +	rhp->ptr = obj;
> +	return &rhp->rh;
> +}
> +
>  /*
>   * Queue a request for lazy invocation of the appropriate free routine
>   * after a grace period.  Please note that three paths are maintained,
> @@ -3830,9 +3867,17 @@ void kvfree_call_rcu(struct rcu_head *head, void *ptr)
>  	if (!success) {
>  		run_page_cache_worker(krcp);
>  
> -		if (head == NULL)
> -			// Inline if kvfree_rcu(one_arg) call.
> -			goto unlock_return;
> +		if (!head) {
> +			krc_this_cpu_unlock(krcp, flags);
> +			head = attach_rcu_head_to_object(ptr);
> +			krcp = krc_this_cpu_lock(&flags);
> +
> +			if (!head)
> +				// Inline if kvfree_rcu(one_arg) call.
> +				goto unlock_return;
> +
> +			ptr = (rcu_callback_t) offsetof(struct dyn_rcu_head, rh);
> +		}
>  
>  		head->func = ptr;
>  		head->next = krcp->head;


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ