lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <20255174537-aBMmobhpYTFtoONI-arkamar@atlas.cz>
Date: Thu, 1 May 2025 09:45:37 +0200
From: Petr Vaněk <arkamar@...as.cz>
To: David Hildenbrand <david@...hat.com>
Cc: linux-kernel@...r.kernel.org, Andrew Morton <akpm@...ux-foundation.org>,
	Ryan Roberts <ryan.roberts@....com>, linux-mm@...ck.org,
	stable@...r.kernel.org
Subject: Re: [PATCH 1/1] mm: Fix folio_pte_batch() overcount with zero PTEs

On Wed, Apr 30, 2025 at 11:25:56PM +0200, David Hildenbrand wrote:
> On 30.04.25 18:00, Petr Vaněk wrote:
> > On Wed, Apr 30, 2025 at 04:37:21PM +0200, David Hildenbrand wrote:
> >> On 30.04.25 13:52, Petr Vaněk wrote:
> >>> On Tue, Apr 29, 2025 at 08:56:03PM +0200, David Hildenbrand wrote:
> >>>> On 29.04.25 20:33, Petr Vaněk wrote:
> >>>>> On Tue, Apr 29, 2025 at 05:45:53PM +0200, David Hildenbrand wrote:
> >>>>>> On 29.04.25 16:52, David Hildenbrand wrote:
> >>>>>>> On 29.04.25 16:45, Petr Vaněk wrote:
> >>>>>>>> On Tue, Apr 29, 2025 at 04:29:30PM +0200, David Hildenbrand wrote:
> >>>>>>>>> On 29.04.25 16:22, Petr Vaněk wrote:
> >>>>>>>>>> folio_pte_batch() could overcount the number of contiguous PTEs when
> >>>>>>>>>> pte_advance_pfn() returns a zero-valued PTE and the following PTE in
> >>>>>>>>>> memory also happens to be zero. The loop doesn't break in such a case
> >>>>>>>>>> because pte_same() returns true, and the batch size is advanced by one
> >>>>>>>>>> more than it should be.
> >>>>>>>>>>
> >>>>>>>>>> To fix this, bail out early if a non-present PTE is encountered,
> >>>>>>>>>> preventing the invalid comparison.
> >>>>>>>>>>
> >>>>>>>>>> This issue started to appear after commit 10ebac4f95e7 ("mm/memory:
> >>>>>>>>>> optimize unmap/zap with PTE-mapped THP") and was discovered via git
> >>>>>>>>>> bisect.
> >>>>>>>>>>
> >>>>>>>>>> Fixes: 10ebac4f95e7 ("mm/memory: optimize unmap/zap with PTE-mapped THP")
> >>>>>>>>>> Cc: stable@...r.kernel.org
> >>>>>>>>>> Signed-off-by: Petr Vaněk <arkamar@...as.cz>
> >>>>>>>>>> ---
> >>>>>>>>>>        mm/internal.h | 2 ++
> >>>>>>>>>>        1 file changed, 2 insertions(+)
> >>>>>>>>>>
> >>>>>>>>>> diff --git a/mm/internal.h b/mm/internal.h
> >>>>>>>>>> index e9695baa5922..c181fe2bac9d 100644
> >>>>>>>>>> --- a/mm/internal.h
> >>>>>>>>>> +++ b/mm/internal.h
> >>>>>>>>>> @@ -279,6 +279,8 @@ static inline int folio_pte_batch(struct folio *folio, unsigned long addr,
> >>>>>>>>>>        			dirty = !!pte_dirty(pte);
> >>>>>>>>>>        		pte = __pte_batch_clear_ignored(pte, flags);
> >>>>>>>>>>        
> >>>>>>>>>> +		if (!pte_present(pte))
> >>>>>>>>>> +			break;
> >>>>>>>>>>        		if (!pte_same(pte, expected_pte))
> >>>>>>>>>>        			break;
> >>>>>>>>>
> >>>>>>>>> How could pte_same() suddenly match on a present and non-present PTE.
> >>>>>>>>
> >>>>>>>> In the problematic case pte.pte == 0 and expected_pte.pte == 0 as well.
> >>>>>>>> pte_same() returns a.pte == b.pte -> 0 == 0. Both are non-present PTEs.
> >>>>>>>
> >>>>>>> Observe that folio_pte_batch() was called *with a present pte*.
> >>>>>>>
> >>>>>>> do_zap_pte_range()
> >>>>>>> 	if (pte_present(ptent))
> >>>>>>> 		zap_present_ptes()
> >>>>>>> 			folio_pte_batch()
> >>>>>>>
> >>>>>>> How can we end up with an expected_pte that is !present, if it is based
> >>>>>>> on the provided pte that *is present* and we only used pte_advance_pfn()
> >>>>>>> to advance the pfn?
> >>>>>>
> >>>>>> I've been staring at the code for too long and don't see the issue.
> >>>>>>
> >>>>>> We even have
> >>>>>>
> >>>>>> VM_WARN_ON_FOLIO(!pte_present(pte), folio);
> >>>>>>
> >>>>>> So the initial pteval we got is present.
> >>>>>>
> >>>>>> I don't see how
> >>>>>>
> >>>>>> 	nr = pte_batch_hint(start_ptep, pte);
> >>>>>> 	expected_pte = __pte_batch_clear_ignored(pte_advance_pfn(pte, nr), flags);
> >>>>>>
> >>>>>> would suddenly result in !pte_present(expected_pte).
> >>>>>
> >>>>> The issue is not happening in __pte_batch_clear_ignored but later in
> >>>>> following line:
> >>>>>
> >>>>>      expected_pte = pte_advance_pfn(expected_pte, nr);
> >>>>>
> >>>>> The issue seems to be in __pte function which converts PTE value to
> >>>>> pte_t in pte_advance_pfn, because warnings disappears when I change the
> >>>>> line to
> >>>>>
> >>>>>      expected_pte = (pte_t){ .pte = pte_val(expected_pte) + (nr << PFN_PTE_SHIFT) };
> >>>>>
> >>>>> The kernel probably uses __pte function from
> >>>>> arch/x86/include/asm/paravirt.h because it is configured with
> >>>>> CONFIG_PARAVIRT=y:
> >>>>>
> >>>>>      static inline pte_t __pte(pteval_t val)
> >>>>>      {
> >>>>>      	return (pte_t) { PVOP_ALT_CALLEE1(pteval_t, mmu.make_pte, val,
> >>>>>      					  "mov %%rdi, %%rax", ALT_NOT_XEN) };
> >>>>>      }
> >>>>>
> >>>>> I guess it might cause this weird magic, but I need more time to
> >>>>> understand what it does :)
> >>>
> >>> I understand it slightly more. __pte() uses xen_make_pte(), which calls
> >>> pte_pfn_to_mfn(), however, mfn for this pfn contains INVALID_P2M_ENTRY
> >>> value, therefore the pte_pfn_to_mfn() returns 0, see [1].
> >>>
> >>> I guess that the mfn was invalidated by xen-balloon driver?
> >>>
> >>> [1] https://web.git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/arch/x86/xen/mmu_pv.c?h=v6.15-rc4#n408
> >>>
> >>>> What XEN does with basic primitives that convert between pteval and
> >>>> pte_t is beyond horrible.
> >>>>
> >>>> How come set_ptes() that uses pte_next_pfn()->pte_advance_pfn() does not
> >>>> run into this?
> >>>
> >>> I don't know, but I guess it is somehow related to pfn->mfn translation.
> >>>
> >>>> Is it only a problem if we exceed a certain pfn?
> >>>
> >>> No, it is a problem if the corresponding mft to given pfn is invalid.
> >>>
> >>> I am not sure if my original patch is a good fix.
> >>
> >> No :)
> >>
> >> Maybe it would be
> >>> better to have some sort of native_pte_advance_pfn() which will use
> >>> native_make_pte() rather than __pte(). Or do you think the issue is in
> >>> Xen part?
> >>
> >> I think what's happening is that -- under XEN only -- we might get garbage when
> >> calling pte_advance_pfn() and the next PFN would no longer fall into the folio. And
> >> the current code cannot deal with that XEN garbage.
> >>
> >> But still not 100% sure.
> >>
> >> The following is completely untested, could you give that a try?
> > 
> > Yes, it solves the issue for me.
> 
> Cool!
> 
> > 
> > However, maybe it would be better to solve it with the following patch.
> > The pte_pfn_to_mfn() actually returns the same value for non-present
> > PTEs. I suggest to return original PTE if the mfn is INVALID_P2M_ENTRY,
> > rather than empty non-present PTE, but the _PAGE_PRESENT bit will be set
> > to zero. Thus, we will not loose information about original pfn but it
> > will be clear that the page is not present.
> > 
> >  From e84781f9ec4fb7275d5e7629cf7e222466caf759 Mon Sep 17 00:00:00 2001
> > From: =?UTF-8?q?Petr=20Van=C4=9Bk?= <arkamar@...as.cz>
> > Date: Wed, 30 Apr 2025 17:08:41 +0200
> > Subject: [PATCH] x86/mm: Reset pte _PAGE_PRESENT bit for invalid mft
> > MIME-Version: 1.0
> > Content-Type: text/plain; charset=UTF-8
> > Content-Transfer-Encoding: 8bit
> > 
> > Signed-off-by: Petr Vaněk <arkamar@...as.cz>
> > ---
> >   arch/x86/xen/mmu_pv.c | 9 +++------
> >   1 file changed, 3 insertions(+), 6 deletions(-)
> > 
> > diff --git a/arch/x86/xen/mmu_pv.c b/arch/x86/xen/mmu_pv.c
> > index 38971c6dcd4b..92a6a9af0c65 100644
> > --- a/arch/x86/xen/mmu_pv.c
> > +++ b/arch/x86/xen/mmu_pv.c
> > @@ -392,28 +392,25 @@ static pteval_t pte_mfn_to_pfn(pteval_t val)
> >   static pteval_t pte_pfn_to_mfn(pteval_t val)
> >   {
> >   	if (val & _PAGE_PRESENT) {
> >   		unsigned long pfn = (val & PTE_PFN_MASK) >> PAGE_SHIFT;
> >   		pteval_t flags = val & PTE_FLAGS_MASK;
> >   		unsigned long mfn;
> >   
> >   		mfn = __pfn_to_mfn(pfn);
> >   
> >   		/*
> > -		 * If there's no mfn for the pfn, then just create an
> > -		 * empty non-present pte.  Unfortunately this loses
> > -		 * information about the original pfn, so
> > -		 * pte_mfn_to_pfn is asymmetric.
> > +		 * If there's no mfn for the pfn, then just reset present pte bit.
> >   		 */
> >   		if (unlikely(mfn == INVALID_P2M_ENTRY)) {
> > -			mfn = 0;
> > -			flags = 0;
> > +			mfn = pfn;
> > +			flags &= ~_PAGE_PRESENT;
> >   		} else
> >   			mfn &= ~(FOREIGN_FRAME_BIT | IDENTITY_FRAME_BIT);
> >   		val = ((pteval_t)mfn << PAGE_SHIFT) | flags;
> >   	}
> >   
> >   	return val;
> >   }
> >   
> >   __visible pteval_t xen_pte_val(pte_t pte)
> >   {
> 
> That might do as well.
>
>
> I assume the following would also work? (removing the early ==1 check)

Yes, it also works in my case and the removal makes sense to me.

> It has the general benefit of removing the pte_pfn() call from the
> loop body, which is why I like that fix. (almost looks like a cleanup)

Indeed, it looks like a cleanup to me as well :)

I am still considering if it would make sense to send both patches, I am
not sure if reseting _PAGE_PRESENT flag is enough, because of swapping
or other areas which I am not aware of.

>  From 75948778b586d4759a480bf412fd4682067b12ea Mon Sep 17 00:00:00 2001
> From: David Hildenbrand <david@...hat.com>
> Date: Wed, 30 Apr 2025 16:35:12 +0200
> Subject: [PATCH] tmp
> 
> Signed-off-by: David Hildenbrand <david@...hat.com>
> ---
>   mm/internal.h | 27 +++++++++++----------------
>   1 file changed, 11 insertions(+), 16 deletions(-)
> 
> diff --git a/mm/internal.h b/mm/internal.h
> index e9695baa59226..25a29872c634b 100644
> --- a/mm/internal.h
> +++ b/mm/internal.h
> @@ -248,11 +248,9 @@ static inline int folio_pte_batch(struct folio *folio, unsigned long addr,
>   		pte_t *start_ptep, pte_t pte, int max_nr, fpb_t flags,
>   		bool *any_writable, bool *any_young, bool *any_dirty)
>   {
> -	unsigned long folio_end_pfn = folio_pfn(folio) + folio_nr_pages(folio);
> -	const pte_t *end_ptep = start_ptep + max_nr;
>   	pte_t expected_pte, *ptep;
>   	bool writable, young, dirty;
> -	int nr;
> +	int nr, cur_nr;
>   
>   	if (any_writable)
>   		*any_writable = false;
> @@ -265,11 +263,15 @@ static inline int folio_pte_batch(struct folio *folio, unsigned long addr,
>   	VM_WARN_ON_FOLIO(!folio_test_large(folio) || max_nr < 1, folio);
>   	VM_WARN_ON_FOLIO(page_folio(pfn_to_page(pte_pfn(pte))) != folio, folio);
>   
> +	/* Limit max_nr to the actual remaining PFNs in the folio we could batch. */
> +	max_nr = min_t(unsigned long, max_nr,
> +		       folio_pfn(folio) + folio_nr_pages(folio) - pte_pfn(pte));
> +
>   	nr = pte_batch_hint(start_ptep, pte);
>   	expected_pte = __pte_batch_clear_ignored(pte_advance_pfn(pte, nr), flags);
>   	ptep = start_ptep + nr;
>   
> -	while (ptep < end_ptep) {
> +	while (nr < max_nr) {
>   		pte = ptep_get(ptep);
>   		if (any_writable)
>   			writable = !!pte_write(pte);
> @@ -282,14 +284,6 @@ static inline int folio_pte_batch(struct folio *folio, unsigned long addr,
>   		if (!pte_same(pte, expected_pte))
>   			break;
>   
> -		/*
> -		 * Stop immediately once we reached the end of the folio. In
> -		 * corner cases the next PFN might fall into a different
> -		 * folio.
> -		 */
> -		if (pte_pfn(pte) >= folio_end_pfn)
> -			break;
> -
>   		if (any_writable)
>   			*any_writable |= writable;
>   		if (any_young)
> @@ -297,12 +291,13 @@ static inline int folio_pte_batch(struct folio *folio, unsigned long addr,
>   		if (any_dirty)
>   			*any_dirty |= dirty;
>   
> -		nr = pte_batch_hint(ptep, pte);
> -		expected_pte = pte_advance_pfn(expected_pte, nr);
> -		ptep += nr;
> +		cur_nr = pte_batch_hint(ptep, pte);
> +		expected_pte = pte_advance_pfn(expected_pte, cur_nr);
> +		ptep += cur_nr;
> +		nr += cur_nr;
>   	}
>   
> -	return min(ptep - start_ptep, max_nr);
> +	return min(nr, max_nr);
>   }
>   
>   /**
> -- 
> 2.49.0
> 
> 
> -- 
> Cheers,
> 
> David / dhildenb
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ