lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date: Fri, 24 May 2024 15:20:47 +1200
From: Barry Song <21cnbao@...il.com>
To: Lance Yang <ioworker0@...il.com>
Cc: akpm@...ux-foundation.org, ryan.roberts@....com, david@...hat.com, 
	mhocko@...e.com, fengwei.yin@...el.com, zokeefe@...gle.com, 
	shy828301@...il.com, xiehuan09@...il.com, wangkefeng.wang@...wei.com, 
	songmuchun@...edance.com, peterx@...hat.com, minchan@...nel.org, 
	linux-mm@...ck.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH v10 4/4] mm/madvise: optimize lazyfreeing with mTHP in madvise_free

On Fri, May 24, 2024 at 2:41 PM Lance Yang <ioworker0@...il.com> wrote:
>
> Hi Barry,
>
> Thanks a lot for reporting!
>
> On Fri, May 24, 2024 at 6:20 AM Barry Song <21cnbao@...il.com> wrote:
> >
> > On Fri, Apr 19, 2024 at 1:44 AM Lance Yang <ioworker0@...il.com> wrote:
> > >
> > > This patch optimizes lazyfreeing with PTE-mapped mTHP[1]
> > > (Inspired by David Hildenbrand[2]). We aim to avoid unnecessary folio
> > > splitting if the large folio is fully mapped within the target range.
> > >
> > > If a large folio is locked or shared, or if we fail to split it, we just
> > > leave it in place and advance to the next PTE in the range. But note that
> > > the behavior is changed; previously, any failure of this sort would cause
> > > the entire operation to give up. As large folios become more common,
> > > sticking to the old way could result in wasted opportunities.
> > >
> > > On an Intel I5 CPU, lazyfreeing a 1GiB VMA backed by PTE-mapped folios of
> > > the same size results in the following runtimes for madvise(MADV_FREE) in
> > > seconds (shorter is better):
> > >
> > > Folio Size |   Old    |   New    | Change
> > > ------------------------------------------
> > >       4KiB | 0.590251 | 0.590259 |    0%
> > >      16KiB | 2.990447 | 0.185655 |  -94%
> > >      32KiB | 2.547831 | 0.104870 |  -95%
> > >      64KiB | 2.457796 | 0.052812 |  -97%
> > >     128KiB | 2.281034 | 0.032777 |  -99%
> > >     256KiB | 2.230387 | 0.017496 |  -99%
> > >     512KiB | 2.189106 | 0.010781 |  -99%
> > >    1024KiB | 2.183949 | 0.007753 |  -99%
> > >    2048KiB | 0.002799 | 0.002804 |    0%
> > >
> > > [1] https://lkml.kernel.org/r/20231207161211.2374093-5-ryan.roberts@arm.com
> > > [2] https://lore.kernel.org/linux-mm/20240214204435.167852-1-david@redhat.com
> > >
> > > Reviewed-by: Ryan Roberts <ryan.roberts@....com>
> > > Acked-by: David Hildenbrand <david@...hat.com>
> > > Signed-off-by: Lance Yang <ioworker0@...il.com>
> > > ---
> >
> > Hi Lance,
> > I am getting kernel panic with this patch,
>
> Good spot!
>
> I just noticed that you posted a patch[1] to fix the bug introduced by the
> commit 89e86854fb0a (mm/arm64: override clear_young_dirty_ptes() batch helper).
>
> Could you please try your patch and see if the kernel panic issue still occurs?
>
> [1] https://lore.kernel.org/all/20240524005444.135417-1-21cnbao@gmail.com/


Indeed, this is all attributed to the overflow in clear_young_dirty_ptes().
I'm reaching out to update you that the issue I reported earlier has been
resolved by the patch.

>
> >
> > / # /home/barry/develop/linux/madvfree
> > [   78.345305] watchdog: BUG: soft lockup - CPU#3 stuck for 22s! [madvfree:101]
> > [   78.345992] Modules linked in:
> > [   78.346942] irq event stamp: 0
> > [   78.347311] hardirqs last  enabled at (0): [<0000000000000000>] 0x0
> > [   78.348407] hardirqs last disabled at (0): [<ffff8000800add04>]
> > copy_process+0x654/0x19a8
> > [   78.349291] softirqs last  enabled at (0): [<ffff8000800add04>]
> > copy_process+0x654/0x19a8
> > [   78.349851] softirqs last disabled at (0): [<0000000000000000>] 0x0
> > [   78.350544] CPU: 3 PID: 101 Comm: madvfree Not tainted
> > 6.9.0-ge51ae633c861 #253
> > [   78.351200] Hardware name: linux,dummy-virt (DT)
> > [   78.351747] pstate: 01400005 (nzcv daif +PAN -UAO -TCO +DIT -SSBS BTYPE=--)
> > [   78.352314] pc : queued_spin_lock_slowpath+0x5c/0x528
> > [   78.352772] lr : do_raw_spin_lock+0xc8/0x120
> > [   78.353245] sp : ffff8000863d3720
> > [   78.353657] x29: ffff8000863d3720 x28: ffff0000c45a8ff8 x27: 0800000103b24003
> > [   78.354632] x26: ffff0000c3b26080 x25: fffffdffc0000000 x24: ffff8000822c2d10
> > [   78.355446] x23: ffff80008403018f x22: ffff8000863d38e0 x21: 0000ffff7f000000
> > [   78.356259] x20: ffff800082fbe008 x19: ffff0000c3b26080 x18: 0000000000000000
> > [   78.357120] x17: 0000000000000000 x16: 0000000000000000 x15: 0000000000000000
> > [   78.357967] x14: 0000000000000000 x13: 0000000000000000 x12: 0000000000000000
> > [   78.358878] x11: ff7ffffffffffbff x10: 0040000000000041 x9 : ffff800080143750
> > [   78.359818] x8 : ffff8000863d3708 x7 : 0000000000000000 x6 : ffff8000803b34e8
> > [   78.360688] x5 : 0000000000000000 x4 : 0000000000000001 x3 : ffff800082fbe008
> > [   78.361602] x2 : ffff80012ac1f000 x1 : 0000000000000000 x0 : 0000000000000080
> > [   78.362755] Call trace:
> > [   78.363132]  queued_spin_lock_slowpath+0x5c/0x528
> > [   78.363614]  do_raw_spin_lock+0xc8/0x120
> > [   78.364041]  _raw_spin_lock+0x58/0x70
> > [   78.364455]  __pte_offset_map_lock+0x98/0x210
> > [   78.364891]  madvise_free_pte_range+0x88/0xa58
> > [   78.365406]  walk_pgd_range+0x390/0x808
> > [   78.365829]  __walk_page_range+0x1e0/0x1f0
> > [   78.366256]  walk_page_range+0x1f0/0x2c8
> > [   78.366676]  madvise_free_single_vma+0x16c/0x308
> > [   78.367115]  madvise_vma_behavior+0x504/0xa20
> > [   78.367549]  madvise_walk_vmas+0xc0/0x128
> > [   78.367972]  do_madvise.part.0+0x110/0x558
> > [   78.368398]  __arm64_sys_madvise+0x68/0x88
> > [   78.368826]  invoke_syscall+0x50/0x128
> > [   78.369332]  el0_svc_common.constprop.0+0x48/0xf8
> > [   78.369778]  do_el0_svc+0x28/0x40
> > [   78.370184]  el0_svc+0x50/0x150
> > [   78.370583]  el0t_64_sync_handler+0x13c/0x158
> > [   78.371017]  el0t_64_sync+0x1a4/0x1a8
> > [  102.345217] watchdog: BUG: soft lockup - CPU#3 stuck for 45s! [madvfree:101]
> > [  102.345835] Modules linked in:
> > [  102.346290] irq event stamp: 0
> > [  102.346715] hardirqs last  enabled at (0): [<0000000000000000>] 0x0
> > [  102.347252] hardirqs last disabled at (0): [<ffff8000800add04>]
> > copy_process+0x654/0x19a8
> > [  102.347796] softirqs last  enabled at (0): [<ffff8000800add04>]
> > copy_process+0x654/0x19a8
> > [  102.348333] softirqs last disabled at (0): [<0000000000000000>] 0x0
> > [  102.348925] CPU: 3 PID: 101 Comm: madvfree Tainted: G             L
> >     6.9.0-ge51ae633c861 #253
> > [  102.349549] Hardware name: linux,dummy-virt (DT)
> > [  102.349988] pstate: 01400005 (nzcv daif +PAN -UAO -TCO +DIT -SSBS BTYPE=--)
> > [  102.350535] pc : queued_spin_lock_slowpath+0x5c/0x528
> > [  102.351010] lr : do_raw_spin_lock+0xc8/0x120
> > [  102.351508] sp : ffff8000863d3720
> > [  102.351939] x29: ffff8000863d3720 x28: ffff0000c45a8ff8 x27: 0800000103b24003
> > [  102.352811] x26: ffff0000c3b26080 x25: fffffdffc0000000 x24: ffff8000822c2d10
> > [  102.353772] x23: ffff80008403018f x22: ffff8000863d38e0 x21: 0000ffff7f000000
> > [  102.354625] x20: ffff800082fbe008 x19: ffff0000c3b26080 x18: 0000000000000000
> > [  102.355495] x17: 0000000000000000 x16: 0000000000000000 x15: 0000000000000000
> > [  102.356370] x14: 0000000000000000 x13: 0000000000000000 x12: 0000000000000000
> > [  102.357333] x11: ff7ffffffffffbff x10: 0040000000000041 x9 : ffff800080143750
> > [  102.358273] x8 : ffff8000863d3708 x7 : 0000000000000000 x6 : ffff8000803b34e8
> > [  102.359112] x5 : 0000000000000000 x4 : 0000000000000001 x3 : ffff800082fbe008
> > [  102.360001] x2 : ffff80012ac1f000 x1 : 0000000000000000 x0 : 0000000000000080
> > [  102.360887] Call trace:
> > [  102.361289]  queued_spin_lock_slowpath+0x5c/0x528
> > [  102.361768]  do_raw_spin_lock+0xc8/0x120
> > [  102.362294]  _raw_spin_lock+0x58/0x70
> > [  102.362825]  __pte_offset_map_lock+0x98/0x210
> > [  102.363299]  madvise_free_pte_range+0x88/0xa58
> > [  102.363771]  walk_pgd_range+0x390/0x808
> > [  102.364268]  __walk_page_range+0x1e0/0x1f0
> > [  102.364729]  walk_page_range+0x1f0/0x2c8
> > [  102.365263]  madvise_free_single_vma+0x16c/0x308
> > [  102.365786]  madvise_vma_behavior+0x504/0xa20
> > [  102.366315]  madvise_walk_vmas+0xc0/0x128
> > [  102.366779]  do_madvise.part.0+0x110/0x558
> > [  102.367269]  __arm64_sys_madvise+0x68/0x88
> > [  102.367714]  invoke_syscall+0x50/0x128
> > [  102.368110]  el0_svc_common.constprop.0+0x48/0xf8
> > [  102.368574]  do_el0_svc+0x28/0x40
> > [  102.369001]  el0_svc+0x50/0x150
> > [  102.369464]  el0t_64_sync_handler+0x13c/0x158
> > [  102.369916]  el0t_64_sync+0x1a4/0x1a8
> > [  126.345236] watchdog: BUG: soft lockup - CPU#3 stuck for 67s! [madvfree:101]
> >
> > the test code is as belows,
> >
> > #define MADV_PAGEOUT    21      /* reclaim these pages */
> >
> > #define DATA_SIZE (128UL * 1024 * 1024)
> > #define PAGE_SIZE (4UL * 1024)
> > #define LARGE_FOLIO_SIZE (64UL * 1024)
> >
> > static void *read_data(void *addr)
> > {
> >         unsigned long i;
> >
> >         for (i = 0; i < DATA_SIZE * 2; i += PAGE_SIZE) {
> >                 if (*((char *)addr + i) != (char)i) {
> >                 }
> >         }
> > }
> >
> > static void *lazyfree(void *addr)
> > {
> >         unsigned long i;
> >
> >         madvise(addr, DATA_SIZE * 2, MADV_FREE);
> > }
> >
> > int main(int argc, char **argv)
> > {
> >         void *addr = mmap(NULL, DATA_SIZE * 2, PROT_READ | PROT_WRITE,
> >                         MAP_ANONYMOUS | MAP_PRIVATE, -1, 0);
>
> Could you please check the /sys/kernel/mm/transparent_hugepage/enabled?
>
> Is it set to 'always'?
>
> Also, what size of mTHP are you using here?

That is 64KiB with "always" policy.

>
> >         memset(addr, 0x11, DATA_SIZE * 2);
> >         lazyfree(addr);
> >
> >         while(1) {
> >                 sleep(1);
> >                 read_data(addr);
> >         }
> >         return 0;
> > }
>
> Thanks again for reaching out!

No worries. please test and ack the patch which fixed clear_young_dirty_ptes()
so that it can be merged as soon as possible.

> Lance
> >
> > >  mm/madvise.c | 85 +++++++++++++++++++++++++++-------------------------
> > >  1 file changed, 44 insertions(+), 41 deletions(-)
> > >
> > > diff --git a/mm/madvise.c b/mm/madvise.c
> > > index 4597a3568e7e..ed125ad8a21e 100644
> > > --- a/mm/madvise.c
> > > +++ b/mm/madvise.c
> > > @@ -643,6 +643,7 @@ static int madvise_free_pte_range(pmd_t *pmd, unsigned long addr,
> > >                                 unsigned long end, struct mm_walk *walk)
> > >
> > >  {
> > > +       const cydp_t cydp_flags = CYDP_CLEAR_YOUNG | CYDP_CLEAR_DIRTY;
> > >         struct mmu_gather *tlb = walk->private;
> > >         struct mm_struct *mm = tlb->mm;
> > >         struct vm_area_struct *vma = walk->vma;
> > > @@ -697,44 +698,57 @@ static int madvise_free_pte_range(pmd_t *pmd, unsigned long addr,
> > >                         continue;
> > >
> > >                 /*
> > > -                * If pmd isn't transhuge but the folio is large and
> > > -                * is owned by only this process, split it and
> > > -                * deactivate all pages.
> > > +                * If we encounter a large folio, only split it if it is not
> > > +                * fully mapped within the range we are operating on. Otherwise
> > > +                * leave it as is so that it can be marked as lazyfree. If we
> > > +                * fail to split a folio, leave it in place and advance to the
> > > +                * next pte in the range.
> > >                  */
> > >                 if (folio_test_large(folio)) {
> > > -                       int err;
> > > +                       bool any_young, any_dirty;
> > >
> > > -                       if (folio_likely_mapped_shared(folio))
> > > -                               break;
> > > -                       if (!folio_trylock(folio))
> > > -                               break;
> > > -                       folio_get(folio);
> > > -                       arch_leave_lazy_mmu_mode();
> > > -                       pte_unmap_unlock(start_pte, ptl);
> > > -                       start_pte = NULL;
> > > -                       err = split_folio(folio);
> > > -                       folio_unlock(folio);
> > > -                       folio_put(folio);
> > > -                       if (err)
> > > -                               break;
> > > -                       start_pte = pte =
> > > -                               pte_offset_map_lock(mm, pmd, addr, &ptl);
> > > -                       if (!start_pte)
> > > -                               break;
> > > -                       arch_enter_lazy_mmu_mode();
> > > -                       pte--;
> > > -                       addr -= PAGE_SIZE;
> > > -                       continue;
> > > +                       nr = madvise_folio_pte_batch(addr, end, folio, pte,
> > > +                                                    ptent, &any_young, &any_dirty);
> > > +
> > > +                       if (nr < folio_nr_pages(folio)) {
> > > +                               int err;
> > > +
> > > +                               if (folio_likely_mapped_shared(folio))
> > > +                                       continue;
> > > +                               if (!folio_trylock(folio))
> > > +                                       continue;
> > > +                               folio_get(folio);
> > > +                               arch_leave_lazy_mmu_mode();
> > > +                               pte_unmap_unlock(start_pte, ptl);
> > > +                               start_pte = NULL;
> > > +                               err = split_folio(folio);
> > > +                               folio_unlock(folio);
> > > +                               folio_put(folio);
> > > +                               pte = pte_offset_map_lock(mm, pmd, addr, &ptl);
> > > +                               start_pte = pte;
> > > +                               if (!start_pte)
> > > +                                       break;
> > > +                               arch_enter_lazy_mmu_mode();
> > > +                               if (!err)
> > > +                                       nr = 0;
> > > +                               continue;
> > > +                       }
> > > +
> > > +                       if (any_young)
> > > +                               ptent = pte_mkyoung(ptent);
> > > +                       if (any_dirty)
> > > +                               ptent = pte_mkdirty(ptent);
> > >                 }
> > >
> > >                 if (folio_test_swapcache(folio) || folio_test_dirty(folio)) {
> > >                         if (!folio_trylock(folio))
> > >                                 continue;
> > >                         /*
> > > -                        * If folio is shared with others, we mustn't clear
> > > -                        * the folio's dirty flag.
> > > +                        * If we have a large folio at this point, we know it is
> > > +                        * fully mapped so if its mapcount is the same as its
> > > +                        * number of pages, it must be exclusive.
> > >                          */
> > > -                       if (folio_mapcount(folio) != 1) {
> > > +                       if (folio_mapcount(folio) != folio_nr_pages(folio)) {
> > >                                 folio_unlock(folio);
> > >                                 continue;
> > >                         }
> > > @@ -750,19 +764,8 @@ static int madvise_free_pte_range(pmd_t *pmd, unsigned long addr,
> > >                 }
> > >
> > >                 if (pte_young(ptent) || pte_dirty(ptent)) {
> > > -                       /*
> > > -                        * Some of architecture(ex, PPC) don't update TLB
> > > -                        * with set_pte_at and tlb_remove_tlb_entry so for
> > > -                        * the portability, remap the pte with old|clean
> > > -                        * after pte clearing.
> > > -                        */
> > > -                       ptent = ptep_get_and_clear_full(mm, addr, pte,
> > > -                                                       tlb->fullmm);
> > > -
> > > -                       ptent = pte_mkold(ptent);
> > > -                       ptent = pte_mkclean(ptent);
> > > -                       set_pte_at(mm, addr, pte, ptent);
> > > -                       tlb_remove_tlb_entry(tlb, pte, addr);
> > > +                       clear_young_dirty_ptes(vma, addr, pte, nr, cydp_flags);
> > > +                       tlb_remove_tlb_entries(tlb, pte, nr, addr);
> > >                 }
> > >                 folio_mark_lazyfree(folio);
> > >         }
> > > --
> > > 2.33.1
> > >

Thanks
Barry

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ