lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <fa9d72ac-1b46-4a09-8f29-af97f2ca6e2e@redhat.com>
Date: Wed, 4 Jun 2025 12:07:21 +0200
From: David Hildenbrand <david@...hat.com>
To: Hyesoo Yu <hyesoo.yu@...sung.com>
Cc: janghyuck.kim@...sung.com, zhaoyang.huang@...soc.com,
 jaewon31.kim@...il.com, Andrew Morton <akpm@...ux-foundation.org>,
 Jason Gunthorpe <jgg@...pe.ca>, John Hubbard <jhubbard@...dia.com>,
 Peter Xu <peterx@...hat.com>, linux-mm@...ck.org,
 linux-kernel@...r.kernel.org
Subject: Re: [PATCH] mm: gup: fail migration when no migratable page to
 prevent CMA pinning

On 04.06.25 11:50, Hyesoo Yu wrote:
> Commit 1aaf8c122918 ("mm: gup: fix infinite loop within __get_longterm_locked")
> caused CMA pages to become pinned in some cases when handling longterm GUP.
> This happened because migration would return success immediately if no pages
> were in the movable_page_list, without retrying.
> 
> However, CMA pages can be temporarily off the LRU (e.g., in pagevecs), and

A better example might be concurrent isolation. Just imagine two of 
these longterm pinnings racing.

> therefore not appear in movable_page_list, even though they can be migrated
> later. Before commit 1aaf8c, the kernel would retry migration in such cases,
> which helped avoid accidental CMA pinning.
> 
> The commit 1aaf8c aimed to support an out-of-tree use case (like pKVM), where
> longterm GUP was applied to non-LRU CMA pages. But allowing CMA pinning
> in general for this corner case could lead to more fragmentation and
> reliability issues. So this patch prevents that.
> 
> Instead of retrying, this patch explicitly fails the migration attempt
> (-EBUSY) if no movable pages are found and unpinnable pages remain.
> This avoids infinite loops and gives user a clear signal to retry,
> rather then spinning inside kernel.

Hmmm, that means we will return EBUSY to the caller. Are all users 
actually prepared to deal with that?

So far we only returned EBUSY in this corner-case 
migrate_device_coherent_folio() that most callers never actually trigger.

Maybe we should do EAGAIN for now (old way of doing it?), and look into 
doing EBUSY separately.

> 
> Fixes: 1aaf8c122918 ("mm: gup: fix infinite loop within __get_longterm_locked")
> Signed-off-by: Hyesoo Yu <hyesoo.yu@...sung.com>
> ---
>   mm/gup.c | 49 ++++++++++++++++++++++++++-----------------------
>   1 file changed, 26 insertions(+), 23 deletions(-)
> 
> diff --git a/mm/gup.c b/mm/gup.c
> index e065a49842a8..446938aedcc9 100644
> --- a/mm/gup.c
> +++ b/mm/gup.c
> @@ -2303,12 +2303,13 @@ static void pofs_unpin(struct pages_or_folios *pofs)
>   /*
>    * Returns the number of collected folios. Return value is always >= 0.
>    */

Comment should be removed.

> -static void collect_longterm_unpinnable_folios(
> +static bool collect_longterm_unpinnable_folios(
>   		struct list_head *movable_folio_list,
>   		struct pages_or_folios *pofs)
>   {
>   	struct folio *prev_folio = NULL;
>   	bool drain_allow = true;
> +	bool any_unpinnable = false;
>   	unsigned long i;
>   
>   	for (i = 0; i < pofs->nr_entries; i++) {
> @@ -2321,6 +2322,8 @@ static void collect_longterm_unpinnable_folios(
>   		if (folio_is_longterm_pinnable(folio))
>   			continue;
>   
> +		any_unpinnable = true;
> +
>   		if (folio_is_device_coherent(folio))
>   			continue;
>   
> @@ -2342,6 +2345,8 @@ static void collect_longterm_unpinnable_folios(
>   				    NR_ISOLATED_ANON + folio_is_file_lru(folio),
>   				    folio_nr_pages(folio));
>   	}
> +
> +	return any_unpinnable;
>   }
>   
>   /*
> @@ -2353,8 +2358,13 @@ static int
>   migrate_longterm_unpinnable_folios(struct list_head *movable_folio_list,
>   				   struct pages_or_folios *pofs)
>   {
> -	int ret;
> +	int ret = -EAGAIN;
>   	unsigned long i;
> +	struct migration_target_control mtc = {
> +		.nid = NUMA_NO_NODE,
> +		.gfp_mask = GFP_USER | __GFP_NOWARN,
> +		.reason = MR_LONGTERM_PIN,
> +	};

Reverse xmas tree while we're at it.

But, can we do this cleanup here separately, and not as part of the fix?

>   
>   	for (i = 0; i < pofs->nr_entries; i++) {
>   		struct folio *folio = pofs_get_folio(pofs, i);
> @@ -2370,6 +2380,7 @@ migrate_longterm_unpinnable_folios(struct list_head *movable_folio_list,
>   			gup_put_folio(folio, 1, FOLL_PIN);
>   
>   			if (migrate_device_coherent_folio(folio)) {
> +				pofs_unpin(pofs);
>   				ret = -EBUSY;
>   				goto err;
>   			}
> @@ -2388,27 +2399,11 @@ migrate_longterm_unpinnable_folios(struct list_head *movable_folio_list,
>   		pofs_clear_entry(pofs, i);
>   	}
>   
> -	if (!list_empty(movable_folio_list)) {
> -		struct migration_target_control mtc = {
> -			.nid = NUMA_NO_NODE,
> -			.gfp_mask = GFP_USER | __GFP_NOWARN,
> -			.reason = MR_LONGTERM_PIN,
> -		};
> -
> -		if (migrate_pages(movable_folio_list, alloc_migration_target,
> -				  NULL, (unsigned long)&mtc, MIGRATE_SYNC,
> -				  MR_LONGTERM_PIN, NULL)) {
> -			ret = -ENOMEM;
> -			goto err;
> -		}
> -	}
> -
> -	putback_movable_pages(movable_folio_list);
> -
> -	return -EAGAIN;
> +	if (migrate_pages(movable_folio_list, alloc_migration_target, NULL,
> +			  (unsigned long)&mtc, MIGRATE_SYNC, MR_LONGTERM_PIN, NULL))
> +		ret = -ENOMEM;
>   
>   err:
> -	pofs_unpin(pofs);
>   	putback_movable_pages(movable_folio_list);
>   
>   	return ret;
> @@ -2417,11 +2412,19 @@ migrate_longterm_unpinnable_folios(struct list_head *movable_folio_list,
>   static long
>   check_and_migrate_movable_pages_or_folios(struct pages_or_folios *pofs)
>   {
> +	bool any_unpinnable;
> +
>   	LIST_HEAD(movable_folio_list);
>   
> -	collect_longterm_unpinnable_folios(&movable_folio_list, pofs);
> -	if (list_empty(&movable_folio_list))
> +	any_unpinnable = collect_longterm_unpinnable_folios(&movable_folio_list, pofs);
> +
> +	if (list_empty(&movable_folio_list)) {
> +		if (any_unpinnable) {

/*
  * If we find any longterm unpinnable page that we failed to
  * isolated for migration, it might be because someone else
  * concurrently isolated it. Make the caller retry until it
  * succeeds.
  */


> +			pofs_unpin(pofs);
> +			return -EBUSY;
> +		}
>   		return 0;
> +	}
>   
>   	return migrate_longterm_unpinnable_folios(&movable_folio_list, pofs);
>   }


-- 
Cheers,

David / dhildenb


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ