[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1593426565.3504.6.camel@suse.de>
Date: Mon, 29 Jun 2020 12:29:25 +0200
From: Oscar Salvador <osalvador@...e.de>
To: nao.horiguchi@...il.com, linux-mm@...ck.org
Cc: mhocko@...nel.org, akpm@...ux-foundation.org,
mike.kravetz@...cle.com, tony.luck@...el.com, david@...hat.com,
aneesh.kumar@...ux.vnet.ibm.com, zeil@...dex-team.ru,
naoya.horiguchi@....com, linux-kernel@...r.kernel.org
Subject: Re: [PATCH v3 00/15] HWPOISON: soft offline rework
On Wed, 2020-06-24 at 15:01 +0000, nao.horiguchi@...il.com wrote:
> I rebased soft-offline rework patchset [1][2] onto the latest
> mmotm. The
> rebasing required some non-trivial changes to adjust, but mainly that
> was
> straightforward. I confirmed that the reported problem doesn't
> reproduce on
> compaction after soft offline. For more precise description of the
> problem
> and the motivation of this patchset, please see [2].
Hi Naoya,
Thanks for dusting this off.
To be honest, I got stuck with the hard offline mode so this delayed
the resubmission, along other problems.
> I think that the following two patches in v2 are better to be done
> with
> separate work of hard-offline rework, so it's not included in this
> series.
>
> - mm,hwpoison: Take pages off the buddy when hard-offlining
> - mm/hwpoison-inject: Rip off duplicated checks
>
> These two are not directly related to the reported problem, so they
> seems
> not urgent. And the first one breaks num_poisoned_pages counting in
> some
> testcases, and The second patch needs more consideration about
> commented point.
I fully agree.
> Any comment/suggestion/help would be appreciated.
My "new" version included a patch to make sure we give a chance to
pages that possibly are in a pcplist.
Current behavior is that if someone tries to soft-offline such a page,
we return an error because page count is 0 but page is not in the buddy
system.
Since this patchset already landed in the mm tree, I could send it as a
standalone patch on top if you agree with it.
My patch looked something like:
From: Oscar Salvador <osalvador@...e.de>
Date: Mon, 29 Jun 2020 12:25:11 +0200
Subject: [PATCH] mm,hwpoison: Drain pcplists before bailing out for
non-buddy
zero-refcount page
A page with 0-refcount and !PageBuddy could perfectly be a pcppage.
Currently, we bail out with an error if we encounter such a page,
meaning that we do not give a chance to handle pcppages.
Fix this by draining pcplists whenever we find this kind of page
and retry the check again.
It might be that pcplists have been spilled into the buddy allocator
and so we can handle it.
Signed-off-by: Oscar Salvador <osalvador@...e.de>
---
mm/memory-failure.c | 24 +++++++++++++++++++++++-
1 file changed, 23 insertions(+), 1 deletion(-)
diff --git a/mm/memory-failure.c b/mm/memory-failure.c
index e90ddddab397..3aac3f1eeed0 100644
--- a/mm/memory-failure.c
+++ b/mm/memory-failure.c
@@ -958,7 +958,7 @@ static int page_action(struct page_state *ps,
struct page *p,
* Return: return 0 if failed to grab the refcount, otherwise true
(some
* non-zero value.)
*/
-static int get_hwpoison_page(struct page *page)
+static int __get_hwpoison_page(struct page *page)
{
struct page *head = compound_head(page);
@@ -988,6 +988,28 @@ static int get_hwpoison_page(struct page *page)
return 0;
}
+static int get_hwpoison_page(struct page *p)
+{
+ int ret;
+ bool drained = false;
+
+retry:
+ ret = __get_hwpoison_page(p);
+ if (!ret) {
+ if (!is_free_buddy_page(p) && !page_count(p) &&
!drained) {
+ /*
+ * The page might be in a pcplist, so try to
drain
+ * those and see if we are lucky.
+ */
+ drain_all_pages(page_zone(p));
+ drained = true;
+ goto retry;
+ }
+ }
+
+ return ret;
+}
+
/*
* Do all that is necessary to remove user space mappings. Unmap
* the pages and send SIGBUS to the processes if the data was dirty.
--
2.26.2
--
Oscar Salvador
SUSE L3
Powered by blists - more mailing lists