lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <26569718-050f-fc90-e3ac-79edfaae9ac7@suse.cz>
Date:   Mon, 15 Feb 2021 19:37:20 +0100
From:   Vlastimil Babka <vbabka@...e.cz>
To:     Andrea Arcangeli <aarcange@...hat.com>
Cc:     stable@...r.kernel.org, Jann Horn <jannh@...gle.com>,
        "Kirill A. Shutemov" <kirill.shutemov@...ux.intel.com>,
        Linus Torvalds <torvalds@...ux-foundation.org>,
        Greg Kroah-Hartman <gregkh@...uxfoundation.org>,
        linux-kernel@...r.kernel.org,
        "linux-mm@...ck.org" <linux-mm@...ck.org>,
        Jann Horn <jannh@...gle.com>, Nicolai Stange <nstange@...e.de>,
        Michal Hocko <mhocko@...nel.org>
Subject: Re: [PATCH 4.9 098/128] mm: thp: make the THP mapcount atomic against
 __split_huge_pmd_locked()

On 6/19/20 4:33 PM, Greg Kroah-Hartman wrote:
> From: Andrea Arcangeli <aarcange@...hat.com>
> 
> commit c444eb564fb16645c172d550359cb3d75fe8a040 upstream.
> 
> Write protect anon page faults require an accurate mapcount to decide
> if to break the COW or not. This is implemented in the THP path with
> reuse_swap_page() ->
> page_trans_huge_map_swapcount()/page_trans_huge_mapcount().
> 
> If the COW triggers while the other processes sharing the page are
> under a huge pmd split, to do an accurate reading, we must ensure the
> mapcount isn't computed while it's being transferred from the head
> page to the tail pages.
> 
> reuse_swap_cache() already runs serialized by the page lock, so it's
> enough to add the page lock around __split_huge_pmd_locked too, in
> order to add the missing serialization.
> 
> Note: the commit in "Fixes" is just to facilitate the backporting,
> because the code before such commit didn't try to do an accurate THP
> mapcount calculation and it instead used the page_count() to decide if
> to COW or not. Both the page_count and the pin_count are THP-wide
> refcounts, so they're inaccurate if used in
> reuse_swap_page(). Reverting such commit (besides the unrelated fix to
> the local anon_vma assignment) would have also opened the window for
> memory corruption side effects to certain workloads as documented in
> such commit header.
> 
> Signed-off-by: Andrea Arcangeli <aarcange@...hat.com>
> Suggested-by: Jann Horn <jannh@...gle.com>
> Reported-by: Jann Horn <jannh@...gle.com>
> Acked-by: Kirill A. Shutemov <kirill.shutemov@...ux.intel.com>
> Fixes: 6d0a07edd17c ("mm: thp: calculate the mapcount correctly for THP pages during WP faults")
> Cc: stable@...r.kernel.org
> Signed-off-by: Linus Torvalds <torvalds@...ux-foundation.org>
> Signed-off-by: Greg Kroah-Hartman <gregkh@...uxfoundation.org>

Hi, when evaluating this backport for our 4.12 based kernel, Nicolai found out
that Jann's POC still triggers, AFAICS because do_huge_pmd_wp_page() doesn't
take the page lock, which was only added by ba3c4ce6def4 ("mm, THP, swap: make
reuse_swap_page() works for THP swapped out") in 4.14. The upstream stable 4.9
is thus in the same situation (didn't actually test the POC there, but should be
obvious), so this is a heads up.

Now just backporting ba3c4ce6def4 to 4.9 stable isn't that simple, as that's
part of a larger series (maybe with even more prerequisities, didn't check). I'm
considering just taking the part of ba3c4ce6def4 that's wrapping
page_trans_huge_mapcount() in the page lock (without changing it to
reuse_swap_page() and changing the latter to deal with swapped out THP) and will
look at it tomorrow. But suggestions (and/or later review) from Andrea/Kirill
are welcome.

Thanks,
Vlastimil

> ---
>  mm/huge_memory.c |   31 ++++++++++++++++++++++++++++---
>  1 file changed, 28 insertions(+), 3 deletions(-)
> 
> --- a/mm/huge_memory.c
> +++ b/mm/huge_memory.c
> @@ -1755,6 +1755,8 @@ void __split_huge_pmd(struct vm_area_str
>  	spinlock_t *ptl;
>  	struct mm_struct *mm = vma->vm_mm;
>  	unsigned long haddr = address & HPAGE_PMD_MASK;
> +	bool was_locked = false;
> +	pmd_t _pmd;
>  
>  	mmu_notifier_invalidate_range_start(mm, haddr, haddr + HPAGE_PMD_SIZE);
>  	ptl = pmd_lock(mm, pmd);
> @@ -1764,11 +1766,32 @@ void __split_huge_pmd(struct vm_area_str
>  	 * pmd against. Otherwise we can end up replacing wrong page.
>  	 */
>  	VM_BUG_ON(freeze && !page);
> -	if (page && page != pmd_page(*pmd))
> -	        goto out;
> +	if (page) {
> +		VM_WARN_ON_ONCE(!PageLocked(page));
> +		was_locked = true;
> +		if (page != pmd_page(*pmd))
> +			goto out;
> +	}
>  
> +repeat:
>  	if (pmd_trans_huge(*pmd)) {
> -		page = pmd_page(*pmd);
> +		if (!page) {
> +			page = pmd_page(*pmd);
> +			if (unlikely(!trylock_page(page))) {
> +				get_page(page);
> +				_pmd = *pmd;
> +				spin_unlock(ptl);
> +				lock_page(page);
> +				spin_lock(ptl);
> +				if (unlikely(!pmd_same(*pmd, _pmd))) {
> +					unlock_page(page);
> +					put_page(page);
> +					page = NULL;
> +					goto repeat;
> +				}
> +				put_page(page);
> +			}
> +		}
>  		if (PageMlocked(page))
>  			clear_page_mlock(page);
>  	} else if (!pmd_devmap(*pmd))
> @@ -1776,6 +1799,8 @@ void __split_huge_pmd(struct vm_area_str
>  	__split_huge_pmd_locked(vma, pmd, haddr, freeze);
>  out:
>  	spin_unlock(ptl);
> +	if (!was_locked && page)
> +		unlock_page(page);
>  	mmu_notifier_invalidate_range_end(mm, haddr, haddr + HPAGE_PMD_SIZE);
>  }
>  
> 
> 
> 

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ