lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <714e77d3-f629-bc06-4808-f7f33ac6872d@google.com>
Date: Sat, 3 Aug 2024 23:55:51 -0700 (PDT)
From: Hugh Dickins <hughd@...gle.com>
To: Andrew Morton <akpm@...ux-foundation.org>
cc: Yu Zhao <yuzhao@...gle.com>, Barry Song <21cnbao@...il.com>, 
    Hugh Dickins <hughd@...gle.com>, linux-mm@...ck.org, 
    linux-kernel@...r.kernel.org
Subject: Re: [PATCH mm-unstable v1 5/5] mm/swap: remove boilerplate

On Wed, 10 Jul 2024, Yu Zhao wrote:

> Remove boilerplate by using a macro to choose the corresponding lock
> and handler for each folio_batch in cpu_fbatches.
> 
> Signed-off-by: Yu Zhao <yuzhao@...gle.com>

Andrew, please revert this "remove boilerplate" patch (and of course its
followup fix) from mm-unstable. From the title I presume it was intended
to make no functional change, but that's far from so.

Under tmpfs swapping load, on different runs I see various badnesses:
"Bad page" in __free_one_page(), Oops in __run_timer_base(),
WARNING at kernel/workqueue.c:790 in set_work_data(), PageBuddy BUG
at page-flags.h:1009 from __del_page_from_freelist(), something (I'd
given up taking better notes by this time) in __queue_work(), others.

All those were including the fix to Barry's report: without that fix,
the boot is drowned in warnings scrolling past too fast to be read.

(All the above were on the HP workstation, swapping to SSD; whereas on
this ThinkPad, swapping to NVMe, no problem seen at all - I mention the
swapping medium, but have no idea whether that's a relevant difference.
In each case, MGLRU compiled in but not enabled. THPs and 64kTHPs active.)

Sorry, but I've put no effort whatsoever into debugging this: "remove
boilerplate" didn't seem worth the effort, and my personal preference
is for readable boilerplate over hiding in a macro.  If you prefer the
macro, I expect Yu can soon come up with a fix (which I could test here):
but for now please revert "remove boilerplate", since its issues get in
the way of further mm testing.

Thanks,
Hugh

> ---
>  mm/swap.c | 107 +++++++++++++++++++-----------------------------------
>  1 file changed, 37 insertions(+), 70 deletions(-)
> 
> diff --git a/mm/swap.c b/mm/swap.c
> index 4a66d2f87f26..342ff4e39ba4 100644
> --- a/mm/swap.c
> +++ b/mm/swap.c
> @@ -220,16 +220,45 @@ static void folio_batch_move_lru(struct folio_batch *fbatch, move_fn_t move_fn)
>  	folios_put(fbatch);
>  }
>  
> -static void folio_batch_add_and_move(struct folio_batch *fbatch,
> -		struct folio *folio, move_fn_t move_fn)
> +static void __folio_batch_add_and_move(struct folio_batch *fbatch,
> +		struct folio *folio, move_fn_t move_fn,
> +		bool on_lru, bool disable_irq)
>  {
> +	unsigned long flags;
> +
> +	folio_get(folio);
> +
> +	if (on_lru && !folio_test_clear_lru(folio)) {
> +		folio_put(folio);
> +		return;
> +	}
> +
>  	if (folio_batch_add(fbatch, folio) && !folio_test_large(folio) &&
>  	    !lru_cache_disabled())
>  		return;
>  
> +	if (disable_irq)
> +		local_lock_irqsave(&cpu_fbatches.lock_irq, flags);
> +	else
> +		local_lock(&cpu_fbatches.lock);
> +
>  	folio_batch_move_lru(fbatch, move_fn);
> +
> +	if (disable_irq)
> +		local_unlock_irqrestore(&cpu_fbatches.lock_irq, flags);
> +	else
> +		local_unlock(&cpu_fbatches.lock);
>  }
>  
> +#define folio_batch_add_and_move(folio, op, on_lru)						\
> +	__folio_batch_add_and_move(								\
> +		this_cpu_ptr(&cpu_fbatches.op),							\
> +		folio,										\
> +		op,										\
> +		on_lru,										\
> +		offsetof(struct cpu_fbatches, op) > offsetof(struct cpu_fbatches, lock_irq)	\
> +	)
> +
>  static void lru_move_tail(struct lruvec *lruvec, struct folio *folio)
>  {
>  	if (folio_test_unevictable(folio))
> @@ -250,23 +279,11 @@ static void lru_move_tail(struct lruvec *lruvec, struct folio *folio)
>   */
>  void folio_rotate_reclaimable(struct folio *folio)
>  {
> -	struct folio_batch *fbatch;
> -	unsigned long flags;
> -
>  	if (folio_test_locked(folio) || folio_test_dirty(folio) ||
>  	    folio_test_unevictable(folio))
>  		return;
>  
> -	folio_get(folio);
> -	if (!folio_test_clear_lru(folio)) {
> -		folio_put(folio);
> -		return;
> -	}
> -
> -	local_lock_irqsave(&cpu_fbatches.lock_irq, flags);
> -	fbatch = this_cpu_ptr(&cpu_fbatches.lru_move_tail);
> -	folio_batch_add_and_move(fbatch, folio, lru_move_tail);
> -	local_unlock_irqrestore(&cpu_fbatches.lock_irq, flags);
> +	folio_batch_add_and_move(folio, lru_move_tail, true);
>  }
>  
>  void lru_note_cost(struct lruvec *lruvec, bool file,
> @@ -355,21 +372,10 @@ static void folio_activate_drain(int cpu)
>  
>  void folio_activate(struct folio *folio)
>  {
> -	struct folio_batch *fbatch;
> -
>  	if (folio_test_active(folio) || folio_test_unevictable(folio))
>  		return;
>  
> -	folio_get(folio);
> -	if (!folio_test_clear_lru(folio)) {
> -		folio_put(folio);
> -		return;
> -	}
> -
> -	local_lock(&cpu_fbatches.lock);
> -	fbatch = this_cpu_ptr(&cpu_fbatches.lru_activate);
> -	folio_batch_add_and_move(fbatch, folio, lru_activate);
> -	local_unlock(&cpu_fbatches.lock);
> +	folio_batch_add_and_move(folio, lru_activate, true);
>  }
>  
>  #else
> @@ -513,8 +519,6 @@ EXPORT_SYMBOL(folio_mark_accessed);
>   */
>  void folio_add_lru(struct folio *folio)
>  {
> -	struct folio_batch *fbatch;
> -
>  	VM_BUG_ON_FOLIO(folio_test_active(folio) &&
>  			folio_test_unevictable(folio), folio);
>  	VM_BUG_ON_FOLIO(folio_test_lru(folio), folio);
> @@ -524,11 +528,7 @@ void folio_add_lru(struct folio *folio)
>  	    lru_gen_in_fault() && !(current->flags & PF_MEMALLOC))
>  		folio_set_active(folio);
>  
> -	folio_get(folio);
> -	local_lock(&cpu_fbatches.lock);
> -	fbatch = this_cpu_ptr(&cpu_fbatches.lru_add);
> -	folio_batch_add_and_move(fbatch, folio, lru_add);
> -	local_unlock(&cpu_fbatches.lock);
> +	folio_batch_add_and_move(folio, lru_add, false);
>  }
>  EXPORT_SYMBOL(folio_add_lru);
>  
> @@ -702,22 +702,11 @@ void lru_add_drain_cpu(int cpu)
>   */
>  void deactivate_file_folio(struct folio *folio)
>  {
> -	struct folio_batch *fbatch;
> -
>  	/* Deactivating an unevictable folio will not accelerate reclaim */
>  	if (folio_test_unevictable(folio))
>  		return;
>  
> -	folio_get(folio);
> -	if (!folio_test_clear_lru(folio)) {
> -		folio_put(folio);
> -		return;
> -	}
> -
> -	local_lock(&cpu_fbatches.lock);
> -	fbatch = this_cpu_ptr(&cpu_fbatches.lru_deactivate_file);
> -	folio_batch_add_and_move(fbatch, folio, lru_deactivate_file);
> -	local_unlock(&cpu_fbatches.lock);
> +	folio_batch_add_and_move(folio, lru_deactivate_file, true);
>  }
>  
>  /*
> @@ -730,21 +719,10 @@ void deactivate_file_folio(struct folio *folio)
>   */
>  void folio_deactivate(struct folio *folio)
>  {
> -	struct folio_batch *fbatch;
> -
>  	if (folio_test_unevictable(folio) || !(folio_test_active(folio) || lru_gen_enabled()))
>  		return;
>  
> -	folio_get(folio);
> -	if (!folio_test_clear_lru(folio)) {
> -		folio_put(folio);
> -		return;
> -	}
> -
> -	local_lock(&cpu_fbatches.lock);
> -	fbatch = this_cpu_ptr(&cpu_fbatches.lru_deactivate);
> -	folio_batch_add_and_move(fbatch, folio, lru_deactivate);
> -	local_unlock(&cpu_fbatches.lock);
> +	folio_batch_add_and_move(folio, lru_deactivate, true);
>  }
>  
>  /**
> @@ -756,22 +734,11 @@ void folio_deactivate(struct folio *folio)
>   */
>  void folio_mark_lazyfree(struct folio *folio)
>  {
> -	struct folio_batch *fbatch;
> -
>  	if (!folio_test_anon(folio) || !folio_test_swapbacked(folio) ||
>  	    folio_test_swapcache(folio) || folio_test_unevictable(folio))
>  		return;
>  
> -	folio_get(folio);
> -	if (!folio_test_clear_lru(folio)) {
> -		folio_put(folio);
> -		return;
> -	}
> -
> -	local_lock(&cpu_fbatches.lock);
> -	fbatch = this_cpu_ptr(&cpu_fbatches.lru_lazyfree);
> -	folio_batch_add_and_move(fbatch, folio, lru_lazyfree);
> -	local_unlock(&cpu_fbatches.lock);
> +	folio_batch_add_and_move(folio, lru_lazyfree, true);
>  }
>  
>  void lru_add_drain(void)
> -- 
> 2.45.2.803.g4e1b14247a-goog

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ