lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date: Fri, 15 Mar 2024 14:21:28 +0100
From: Vlastimil Babka <vbabka@...e.cz>
To: Oscar Salvador <osalvador@...e.de>,
 Andrew Morton <akpm@...ux-foundation.org>
Cc: linux-kernel@...r.kernel.org, linux-mm@...ck.org,
 Michal Hocko <mhocko@...e.com>, Marco Elver <elver@...gle.com>,
 Andrey Konovalov <andreyknvl@...il.com>,
 Alexander Potapenko <glider@...gle.com>
Subject: Re: [PATCH] mm,page_owner: Fix refcount imbalance

On 3/14/24 15:47, Oscar Salvador wrote:
> Current code does not contemplate scenarios were an allocation and
> free operation on the same pages do not handle it in the same amount
> at once.
> To give an example, page_alloc_exact(), where we will allocate a page
> of enough order to stafisfy the size request, but we will free the
> remainings right away.
> 
> In the above example, we will increment the stack_record refcount
> only once, but we will decrease it the same number of times as number
> of unused pages we have to free.
> This will lead to a warning because of refcount imbalance.
> 
> Fix this by recording the number of base pages every stack_record holds,
> and only let the last decrementing of refcount succeed if the number of
> base pages equals 0, which means we freed all the pages.
> 
> As a bonus, show the aggregate of stack_count + base_pages as this gives
> a much better picture of the memory usage.
> 
> Signed-off-by: Oscar Salvador <osalvador@...e.de>
> Fixes: 217b2119b9e2 ("mm,page_owner: implement the tracking of the stacks count")
> ---
>  include/linux/stackdepot.h |  3 ++
>  mm/page_owner.c            | 57 +++++++++++++++++++++++++++++++-------
>  2 files changed, 50 insertions(+), 10 deletions(-)
> 
> diff --git a/include/linux/stackdepot.h b/include/linux/stackdepot.h
> index 3c6caa5abc7c..261472807c32 100644
> --- a/include/linux/stackdepot.h
> +++ b/include/linux/stackdepot.h
> @@ -57,6 +57,9 @@ struct stack_record {
>  	u32 size;			/* Number of stored frames */
>  	union handle_parts handle;	/* Constant after initialization */
>  	refcount_t count;
> +#ifdef CONFIG_PAGE_OWNER
> +	unsigned long nr_base_pages;
> +#endif

The stackdepot guys probably won't be thrilled about this addition to
stack_record. Can't we instead make the refcount itself reflect the number
of base pages?

..

> -static void dec_stack_record_count(depot_stack_handle_t handle)
> +static void dec_stack_record_count(depot_stack_handle_t handle,
> +				   unsigned long nr_base_pages)
>  {
>  	struct stack_record *stack_record = __stack_depot_get_stack_record(handle);
> +	unsigned long curr_nr_pages;
> +
> +	if (!stack_record)
> +		return;
> +
> +	curr_nr_pages = smp_load_acquire(&stack_record->nr_base_pages);
> +	smp_store_release(&stack_record->nr_base_pages,
> +			  curr_nr_pages - nr_base_pages);

If the intent of this is to have stack_record->nr_base_pages updated
atomically with respect to parallel updates, these smp_ operations won't
help I'm afraid.

> +	curr_nr_pages = smp_load_acquire(&stack_record->nr_base_pages);
> +
> +	/*
> +	 * If this stack_record is going to reach a refcount == 1, which means
> +	 * free, only do it if all the base pages it allocated were freed.
> +	 * E.g: scenarios like THP splitting, or alloc_pages_exact() can have
> +	 * an alloc/free operation with different amount of pages
> +	 */
> +	if (refcount_read(&stack_record->count) == 2 &&
> +	    curr_nr_pages)
> +		return;

This is very suspicious. We shouldn't manipulate refcount based on the other
counter. This suggest the refcount will eventually stop reflecting reality
as we could withold some legitimate decreases and not retry them afterwards?
Another reason to try making refcount itself represent nr_base_pages.


>  
> -	if (stack_record)
> -		refcount_dec(&stack_record->count);
> +	refcount_dec(&stack_record->count);

This refcount_read() followed by refcount_dec() is also potentially racy.

>  }
>  
>  void __reset_page_owner(struct page *page, unsigned short order)
> @@ -260,7 +294,7 @@ void __reset_page_owner(struct page *page, unsigned short order)
>  		 * the machinery is not ready yet, we cannot decrement
>  		 * their refcount either.
>  		 */
> -		dec_stack_record_count(alloc_handle);
> +		dec_stack_record_count(alloc_handle, 1UL << order);
>  }
>  
>  static inline void __set_page_owner_handle(struct page_ext *page_ext,
> @@ -303,7 +337,7 @@ noinline void __set_page_owner(struct page *page, unsigned short order,
>  	__set_page_owner_handle(page_ext, handle, order, gfp_mask);
>  	page_ext_put(page_ext);
>  	set_current_in_page_owner();
> -	inc_stack_record_count(handle, gfp_mask);
> +	inc_stack_record_count(handle, gfp_mask, 1UL << order);
>  	unset_current_in_page_owner();
>  }
>  
> @@ -868,6 +902,7 @@ static int stack_print(struct seq_file *m, void *v)
>  	struct stack *stack = v;
>  	unsigned long *entries;
>  	unsigned long nr_entries;
> +	unsigned long nr_base_pages;
>  	struct stack_record *stack_record = stack->stack_record;
>  
>  	if (!stack->stack_record)
> @@ -875,6 +910,7 @@ static int stack_print(struct seq_file *m, void *v)
>  
>  	nr_entries = stack_record->size;
>  	entries = stack_record->entries;
> +	nr_base_pages = stack_record->nr_base_pages;
>  	stack_count = refcount_read(&stack_record->count) - 1;
>  
>  	if (stack_count < 1 || stack_count < page_owner_stack_threshold)
> @@ -882,7 +918,8 @@ static int stack_print(struct seq_file *m, void *v)
>  
>  	for (i = 0; i < nr_entries; i++)
>  		seq_printf(m, " %pS\n", (void *)entries[i]);
> -	seq_printf(m, "stack_count: %d\n\n", stack_count);
> +	seq_printf(m, "stack_count: %d curr_nr_base_pages: %lu\n\n",
> +		   stack_count, nr_base_pages);
>  
>  	return 0;
>  }


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ