[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20171130195743.52vc6enr3rnivtdx@dhcp22.suse.cz>
Date: Thu, 30 Nov 2017 20:57:43 +0100
From: Michal Hocko <mhocko@...nel.org>
To: Mike Kravetz <mike.kravetz@...cle.com>
Cc: linux-mm@...ck.org, Naoya Horiguchi <n-horiguchi@...jp.nec.com>,
LKML <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH RFC 2/2] mm, hugetlb: do not rely on overcommit limit
during migration
On Thu 30-11-17 11:35:11, Mike Kravetz wrote:
> On 11/29/2017 11:57 PM, Michal Hocko wrote:
> > On Wed 29-11-17 11:52:53, Mike Kravetz wrote:
> >> On 11/29/2017 01:22 AM, Michal Hocko wrote:
> >>> What about this on top. I haven't tested this yet though.
> >>
> >> Yes, this would work.
> >>
> >> However, I think a simple modification to your previous free_huge_page
> >> changes would make this unnecessary. I was confused in your previous
> >> patch because you decremented the per-node surplus page count, but not
> >> the global count. I think it would have been correct (and made this
> >> patch unnecessary) if you decremented the global counter there as well.
> >
> > We cannot really increment the global counter because the over number of
> > surplus pages during migration doesn't increase.
>
> I was not suggesting we increment the global surplus count. Rather,
> your previous patch should have decremented the global surplus count in
> free_huge_page. Something like:
sorry I meant to say decrement. The point is that overal suprlus count
doesn't change after the migration. The only thing that _might_ change
is the per node distribution of surplus pages. That is why I think we
should handle that during the migration.
> @@ -1283,7 +1283,13 @@ void free_huge_page(struct page *page)
> if (restore_reserve)
> h->resv_huge_pages++;
>
> - if (h->surplus_huge_pages_node[nid]) {
> + if (PageHugeTemporary(page)) {
> + list_del(&page->lru);
> + ClearPageHugeTemporary(page);
> + update_and_free_page(h, page);
> + if (h->surplus_huge_pages_node[nid])
> + h->surplus_huge_pages--;
> + h->surplus_huge_pages_node[nid]--;
> + }
> + } else if (h->surplus_huge_pages_node[nid]) {
> /* remove the page from active list */
> list_del(&page->lru);
> update_and_free_page(h, page);
>
> When we allocate one of these 'PageHugeTemporary' pages, we only increment
> the global and node specific nr_huge_pages counters. To me, this makes all
> the huge page counters be the same as if there were simply one additional
> pre-allocated huge page. This 'extra' (PageHugeTemporary) page will go
> away when free_huge_page is called. So, my thought is that it is not
> necessary to transfer per-node counts from the original to target node.
> Of course, I may be missing something.
The thing is that we do not know whether the original page is surplus
until the deallocation time.
> When thinking about transfering per-node counts as is done in your latest
> patch, I took another look at all the per-node counts. This may show my
> ignorance of huge page migration, but do we need to handle the case where
> the page being migrated is 'free'? Is that possible? If so, there will
> be a count for free_huge_pages_node and the page will be on the per node
> hugepage_freelists that must be handled
I do not understand. What do you mean by free? Sitting on the pool? I do
not think we ever try to migrate those. They simply do not have any
state to migrate. We could very well just allocate fresh pages on the
remote node and dissolve free ones. I am not sure we do that during the
memory hotplug to preserve the pool size and I am too tired to check
that now. This would be a different topic I guess.
--
Michal Hocko
SUSE Labs
Powered by blists - more mailing lists