[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20210305093656.6c262b19@alex-virtual-machine>
Date: Fri, 5 Mar 2021 09:36:56 +0800
From: Aili Yao <yaoaili@...gsoft.com>
To: "Luck, Tony" <tony.luck@...el.com>
CC: "HORIGUCHI NAOYA堀口 直也)"
<naoya.horiguchi@....com>, Oscar Salvador <osalvador@...e.de>,
"david@...hat.com" <david@...hat.com>,
"akpm@...ux-foundation.org" <akpm@...ux-foundation.org>,
"bp@...en8.de" <bp@...en8.de>,
"tglx@...utronix.de" <tglx@...utronix.de>,
"mingo@...hat.com" <mingo@...hat.com>,
"hpa@...or.com" <hpa@...or.com>, "x86@...nel.org" <x86@...nel.org>,
"linux-edac@...r.kernel.org" <linux-edac@...r.kernel.org>,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"linux-mm@...ck.org" <linux-mm@...ck.org>,
"yangfeng1@...gsoft.com" <yangfeng1@...gsoft.com>,
<yaoaili@...gsoft.com>
Subject: Re: [PATCH] mm,hwpoison: return -EBUSY when page already poisoned
On Fri, 5 Mar 2021 09:30:16 +0800
Aili Yao <yaoaili@...gsoft.com> wrote:
> On Thu, 4 Mar 2021 15:57:20 -0800
> "Luck, Tony" <tony.luck@...el.com> wrote:
>
> > On Thu, Mar 04, 2021 at 02:45:24PM +0800, Aili Yao wrote:
> > > > > if your methods works, should it be like this?
> > > > >
> > > > > 1582 pteval = swp_entry_to_pte(make_hwpoison_entry(subpage));
> > > > > 1583 if (PageHuge(page)) {
> > > > > 1584 hugetlb_count_sub(compound_nr(page), mm);
> > > > > 1585 set_huge_swap_pte_at(mm, address,
> > > > > 1586 pvmw.pte, pteval,
> > > > > 1587 vma_mmu_pagesize(vma));
> > > > > 1588 } else {
> > > > > 1589 dec_mm_counter(mm, mm_counter(page));
> > > > > 1590 set_pte_at(mm, address, pvmw.pte, pteval);
> > > > > 1591 }
> > > > >
> > > > > the page fault check if it's a poison page using is_hwpoison_entry(),
> > > > >
> > > >
> > > > And if it works, does we need some locking mechanism before we call walk_page_range();
> > > > if we lock, does we need to process the blocking interrupted error as other places will do?
> > > >
> > >
> > > And another thing:
> > > Do we need a call to flush_tlb_page(vma, address) to make the pte changes into effect?
> >
> > Thanks for all the pointers. I added them to the patch (see below).
> > [The pmd/pud cases may need some tweaking ... but just trying to get
> > the 4K page case working first]
> >
> > I tried testing by skipping the call to memory_failure() and just
> > using this new code to search the page tables for current page and
> > marking it hwpoison (to simulate the case where 2nd process gets the
> > early return from memory_failure(). Something is still missing because I get:
> >
> > [ 481.911298] mce: pte_entry: matched pfn - mark poison & zap pte
> > [ 481.917935] MCE: Killing einj_mem_uc:5555 due to hardware memory corruption fault at 7fe64b33b400
> > [ 482.933775] BUG: Bad page cache in process einj_mem_uc pfn:408b6d6
> > [ 482.940777] page:0000000013ea6e96 refcount:3 mapcount:1 mapping:00000000e3a069d9 index:0x0 pfn:0x408b6d6
> > [ 482.951355] memcg:ffff94a809834000
> > [ 482.955153] aops:shmem_aops ino:3c04
> > [ 482.959142] flags: 0x97ffffc0880015(locked|uptodate|lru|swapbacked|hwpoison)
> > [ 482.967018] raw: 0097ffffc0880015 ffff94c80e93ec00 ffff94c80e93ec00 ffff94c80a9b25a8
> > [ 482.975666] raw: 0000000000000000 0000000000000000 0000000300000000 ffff94a809834000
> > [ 482.984310] page dumped because: still mapped when deleted
>
> From the walk, it seems we have got the virtual address, can we just send a SIGBUS with it?
Does this walk proper for other memory-failure error cases, can it be applied to if (p->mce_vaddr != (void __user *)-1l) branch
in kill_me_maybe()?
> > commit e5de44560b33e2d407704243566253a70f858a59
> > Author: Tony Luck <tony.luck@...el.com>
> > Date: Tue Mar 2 15:06:33 2021 -0800
> >
> > x86/mce: Handle races between machine checks
> >
> > When multiple CPUs hit the same poison memory there is a race. The
> > first CPU into memory_failure() atomically marks the page as poison
> > and continues processing to hunt down all the tasks that map this page
> > so that the virtual addresses can be marked not-present and SIGBUS
> > sent to the task that did the access.
> >
> > Later CPUs get an early return from memory_failure() and may return
> > to user mode and access the poison again.
> >
> > Add a new argument to memory_failure() so that it can indicate when
> > the race has been lost. Fix kill_me_maybe() to scan page tables in
> > this case to unmap pages.
> >
> > diff --git a/arch/x86/kernel/cpu/mce/core.c b/arch/x86/kernel/cpu/mce/core.c
> > index 7962355436da..a52c6a772de2 100644
> > --- a/arch/x86/kernel/cpu/mce/core.c
> > +++ b/arch/x86/kernel/cpu/mce/core.c
> > @@ -28,8 +28,12 @@
> > #include <linux/sysfs.h>
> > #include <linux/types.h>
> > #include <linux/slab.h>
> > +#include <linux/hugetlb.h>
> > +#include <linux/swap.h>
> > +#include <linux/swapops.h>
> > #include <linux/init.h>
> > #include <linux/kmod.h>
> > +#include <linux/pagewalk.h>
> > #include <linux/poll.h>
> > #include <linux/nmi.h>
> > #include <linux/cpu.h>
> > @@ -637,6 +641,7 @@ static int uc_decode_notifier(struct notifier_block *nb, unsigned long val,
> > {
> > struct mce *mce = (struct mce *)data;
> > unsigned long pfn;
> > + int already = 0;
> >
> > if (!mce || !mce_usable_address(mce))
> > return NOTIFY_DONE;
> > @@ -646,8 +651,9 @@ static int uc_decode_notifier(struct notifier_block *nb, unsigned long val,
> > return NOTIFY_DONE;
> >
> > pfn = mce->addr >> PAGE_SHIFT;
> > - if (!memory_failure(pfn, 0)) {
> > - set_mce_nospec(pfn, whole_page(mce));
> > + if (!memory_failure(pfn, 0, &already)) {
> > + if (!already)
> > + set_mce_nospec(pfn, whole_page(mce));
> > mce->kflags |= MCE_HANDLED_UC;
> > }
> >
> > @@ -1248,6 +1254,79 @@ static void __mc_scan_banks(struct mce *m, struct pt_regs *regs, struct mce *fin
> > *m = *final;
> > }
> >
> > +static int pte_entry(pte_t *pte, unsigned long addr, unsigned long next, struct mm_walk *walk)
> > +{
> > + u64 pfn = (u64)walk->private;
> > + struct page *page;
> > + pte_t pteval;
> > +
> > + if (pte_pfn(*pte) == pfn) {
> > +pr_info("pte_entry: matched pfn - mark poison & zap pte\n");
> > + page = pfn_to_page(pfn);
> > + lock_page(page);
> > +SetPageHWPoison(page);
> > + pteval = swp_entry_to_pte(make_hwpoison_entry(page));
> > + dec_mm_counter(walk->mm, mm_counter(page));
> > + set_pte_at(current->mm, addr, pte, pteval);
> > + unlock_page(page);
> > + flush_tlb_page(walk->vma, addr);
> > + }
> > +
> > + return 0;
> > +}
> > +
> > +static int pmd_entry(pmd_t *pmd, unsigned long addr, unsigned long next, struct mm_walk *walk)
> > +{
> > + int shift = PMD_SHIFT - PAGE_SHIFT;
> > + u64 pfn = (u64)walk->private;
> > + struct page *page;
> > + pte_t pteval;
> > +
> > + if (!pmd_large(*pmd))
> > + return 0;
> > +
> > + if (pmd_pfn(*pmd) >> shift == pfn >> shift) {
> > + page = pfn_to_page(pfn);
> > + lock_page(page);
> > + pteval = swp_entry_to_pte(make_hwpoison_entry(page));
> > + hugetlb_count_sub(compound_nr(page), walk->mm);
> > + set_huge_swap_pte_at(walk->mm, addr, (pte_t *)pmd, pteval, vma_mmu_pagesize(walk->vma));
> > + unlock_page(page);
> > + flush_tlb_page(walk->vma, addr);
> > + }
> > +
> > + return 0;
> > +}
> > +
> > +static int pud_entry(pud_t *pud, unsigned long addr, unsigned long next, struct mm_walk *walk)
> > +{
> > + int shift = PUD_SHIFT - PAGE_SHIFT;
> > + u64 pfn = (u64)walk->private;
> > + struct page *page;
> > + pte_t pteval;
> > +
> > + if (!pud_large(*pud))
> > + return 0;
> > +
> > + if (pud_pfn(*pud) >> shift == pfn >> shift) {
> > + page = pfn_to_page(pfn);
> > + lock_page(page);
> > + pteval = swp_entry_to_pte(make_hwpoison_entry(page));
> > + hugetlb_count_sub(compound_nr(page), walk->mm);
> > + set_huge_swap_pte_at(walk->mm, addr, (pte_t *)pud, pteval, vma_mmu_pagesize(walk->vma));
> > + unlock_page(page);
> > + flush_tlb_page(walk->vma, addr);
> > + }
> > +
> > + return 0;
> > +}
> > +
> > +static struct mm_walk_ops walk = {
> > + .pte_entry = pte_entry,
> > + .pmd_entry = pmd_entry,
> > + .pud_entry = pud_entry
> > +};
> > +
> > static void kill_me_now(struct callback_head *ch)
> > {
> > force_sig(SIGBUS);
> > @@ -1257,15 +1336,22 @@ static void kill_me_maybe(struct callback_head *cb)
> > {
> > struct task_struct *p = container_of(cb, struct task_struct, mce_kill_me);
> > int flags = MF_ACTION_REQUIRED;
> > + int already = 0;
> >
> > pr_err("Uncorrected hardware memory error in user-access at %llx", p->mce_addr);
> >
> > if (!p->mce_ripv)
> > flags |= MF_MUST_KILL;
> >
> > - if (!memory_failure(p->mce_addr >> PAGE_SHIFT, flags) &&
> > + if (!memory_failure(p->mce_addr >> PAGE_SHIFT, flags, &already) &&
> > !(p->mce_kflags & MCE_IN_KERNEL_COPYIN)) {
> > - set_mce_nospec(p->mce_addr >> PAGE_SHIFT, p->mce_whole_page);
> > + if (already) {
> > + mmap_read_lock(current->mm);
> > + walk_page_range(current->mm, 0, TASK_SIZE_MAX, &walk, (void *)(p->mce_addr >> PAGE_SHIFT));
> > + mmap_read_unlock(current->mm);
> > + } else {
> > + set_mce_nospec(p->mce_addr >> PAGE_SHIFT, p->mce_whole_page);
> > + }
> > sync_core();
> > return;
> > }
> > @@ -1452,7 +1538,7 @@ noinstr void do_machine_check(struct pt_regs *regs)
> > EXPORT_SYMBOL_GPL(do_machine_check);
> >
> > #ifndef CONFIG_MEMORY_FAILURE
> > -int memory_failure(unsigned long pfn, int flags)
> > +int memory_failure(unsigned long pfn, int flags, int *already)
> > {
> > /* mce_severity() should not hand us an ACTION_REQUIRED error */
> > BUG_ON(flags & MF_ACTION_REQUIRED);
> > diff --git a/drivers/base/memory.c b/drivers/base/memory.c
> > index f35298425575..144500983656 100644
> > --- a/drivers/base/memory.c
> > +++ b/drivers/base/memory.c
> > @@ -480,7 +480,7 @@ static ssize_t hard_offline_page_store(struct device *dev,
> > if (kstrtoull(buf, 0, &pfn) < 0)
> > return -EINVAL;
> > pfn >>= PAGE_SHIFT;
> > - ret = memory_failure(pfn, 0);
> > + ret = memory_failure(pfn, 0, NULL);
> > return ret ? ret : count;
> > }
> >
> > diff --git a/include/linux/mm.h b/include/linux/mm.h
> > index 77e64e3eac80..beaa6e871cbe 100644
> > --- a/include/linux/mm.h
> > +++ b/include/linux/mm.h
> > @@ -3003,7 +3003,7 @@ enum mf_flags {
> > MF_MUST_KILL = 1 << 2,
> > MF_SOFT_OFFLINE = 1 << 3,
> > };
> > -extern int memory_failure(unsigned long pfn, int flags);
> > +extern int memory_failure(unsigned long pfn, int flags, int *already);
> > extern void memory_failure_queue(unsigned long pfn, int flags);
> > extern void memory_failure_queue_kick(int cpu);
> > extern int unpoison_memory(unsigned long pfn);
> > diff --git a/mm/hwpoison-inject.c b/mm/hwpoison-inject.c
> > index 1ae1ebc2b9b1..bfd5151dcd3f 100644
> > --- a/mm/hwpoison-inject.c
> > +++ b/mm/hwpoison-inject.c
> > @@ -48,7 +48,7 @@ static int hwpoison_inject(void *data, u64 val)
> >
> > inject:
> > pr_info("Injecting memory failure at pfn %#lx\n", pfn);
> > - return memory_failure(pfn, 0);
> > + return memory_failure(pfn, 0, NULL);
> > }
> >
> > static int hwpoison_unpoison(void *data, u64 val)
> > diff --git a/mm/madvise.c b/mm/madvise.c
> > index df692d2e35d4..09f569fed68d 100644
> > --- a/mm/madvise.c
> > +++ b/mm/madvise.c
> > @@ -908,7 +908,7 @@ static int madvise_inject_error(int behavior,
> > } else {
> > pr_info("Injecting memory failure for pfn %#lx at process virtual address %#lx\n",
> > pfn, start);
> > - ret = memory_failure(pfn, MF_COUNT_INCREASED);
> > + ret = memory_failure(pfn, MF_COUNT_INCREASED, NULL);
> > }
> >
> > if (ret)
> > diff --git a/mm/memory-failure.c b/mm/memory-failure.c
> > index 24210c9bd843..9a8911aa5fc9 100644
> > --- a/mm/memory-failure.c
> > +++ b/mm/memory-failure.c
> > @@ -1398,7 +1398,7 @@ static int memory_failure_dev_pagemap(unsigned long pfn, int flags,
> > * Must run in process context (e.g. a work queue) with interrupts
> > * enabled and no spinlocks hold.
> > */
> > -int memory_failure(unsigned long pfn, int flags)
> > +int memory_failure(unsigned long pfn, int flags, int *already)
> > {
> > struct page *p;
> > struct page *hpage;
> > @@ -1428,6 +1428,8 @@ int memory_failure(unsigned long pfn, int flags)
> > if (PageHuge(p))
> > return memory_failure_hugetlb(pfn, flags);
> > if (TestSetPageHWPoison(p)) {
> > + if (already)
> > + *already = 1;
> > pr_err("Memory failure: %#lx: already hardware poisoned\n",
> > pfn);
> > return 0;
> > @@ -1634,7 +1636,7 @@ static void memory_failure_work_func(struct work_struct *work)
> > if (entry.flags & MF_SOFT_OFFLINE)
> > soft_offline_page(entry.pfn, entry.flags);
> > else
> > - memory_failure(entry.pfn, entry.flags);
> > + memory_failure(entry.pfn, entry.flags, NULL);
> > }
> > }
> >
>
--
Thanks!
Aili Yao
Powered by blists - more mailing lists