[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <Yqy8EaibIH1k7/qR@monkey>
Date: Fri, 17 Jun 2022 10:38:25 -0700
From: Mike Kravetz <mike.kravetz@...cle.com>
To: Muchun Song <songmuchun@...edance.com>
Cc: Oscar Salvador <osalvador@...e.de>, akpm@...ux-foundation.org,
linux-mm@...ck.org, linux-kernel@...r.kernel.org
Subject: Re: [PATCH] mm: hugetlb: remove minimum_order variable
On 06/17/22 15:41, Muchun Song wrote:
> On Thu, Jun 16, 2022 at 11:03:34AM -0700, Mike Kravetz wrote:
> > On 06/16/22 11:38, Muchun Song wrote:
> > > The following commit:
> > >
> > > commit 641844f5616d ("mm/hugetlb: introduce minimum hugepage order")
> > >
> > > fixed a static checker warning and introduced a global variable minimum_order
> > > to fix the warning. However, the local variable in dissolve_free_huge_pages()
> > > can be initialized to huge_page_order(&default_hstate) to fix the warning.
> > > So remove minimum_order to simplify the code.
> > >
> > > Signed-off-by: Muchun Song <songmuchun@...edance.com>
> > > ---
> > > mm/hugetlb.c | 18 +++++++-----------
> > > 1 file changed, 7 insertions(+), 11 deletions(-)
> > >
> > > diff --git a/mm/hugetlb.c b/mm/hugetlb.c
> > > index 8ea4e51d8186..405d1c7441c9 100644
> > > --- a/mm/hugetlb.c
> > > +++ b/mm/hugetlb.c
> > > @@ -66,12 +66,6 @@ static bool hugetlb_cma_page(struct page *page, unsigned int order)
> > > #endif
> > > static unsigned long hugetlb_cma_size __initdata;
> > >
> > > -/*
> > > - * Minimum page order among possible hugepage sizes, set to a proper value
> > > - * at boot time.
> > > - */
> > > -static unsigned int minimum_order __read_mostly = UINT_MAX;
> > > -
> > > __initdata LIST_HEAD(huge_boot_pages);
> > >
> > > /* for command line parsing */
> > > @@ -2161,11 +2155,17 @@ int dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn)
> > > unsigned long pfn;
> > > struct page *page;
> > > int rc = 0;
> > > + unsigned int order;
> > > + struct hstate *h;
> > >
> > > if (!hugepages_supported())
> > > return rc;
> > >
> > > - for (pfn = start_pfn; pfn < end_pfn; pfn += 1 << minimum_order) {
> > > + order = huge_page_order(&default_hstate);
> > > + for_each_hstate(h)
> > > + order = min(order, huge_page_order(h));
> >
> > Since we will be traversing the array of hstates, I wonder if we should
> > optimize this further? We could:
> > - Pass the node into dissolve_free_huge_pages
> > - When traversing the hstate array, check free_huge_pages_node[node] in
> > each hstate.
> > - If no free huge pages, no need to do the pfn scan.
> >
> > Yes, the above is racy. However, the code is already racy as hugetlb
> > page state can change while performing this scan. We only hold the hugetlb
> > lock when checking an individual hugetlb page. The change above may
> > make the code a bit more racy.
> >
>
> Agree.
>
> > If we think that is too racy, they we could at least check
> > nr_huge_pages_node[node]. If there are no hugetlb pages on the node
> > there is no need to scan. And, I think we have isolated this pfn range
> > so no new hugetlb pages can be created.
> >
> > Not sure if the above optimizations are worth the effort. IIUC, the
> > pfn range is at most a memory block size which is not huge.
> >
>
> Right. It is not huge.
>
> I have no strong opinion. dissolve_free_huge_pages() is only called in
> memory offline path and it is not a hot path. If we think the optimization
> is necessary, I think it should be a separate patch.
>
Don't really think there is a need for such an optimization. It was
mostly just a thought I had while looking at your changes.
Not suggesting we do this, but if we wanted to optimize I would think
about adding a global variable (or something) that indicates whether or
not hugetlb pages have been created. If no hugetlb pages exist, there
there is no need to execute ANY of this code.
As mentioned above, the size of a memory block is not huge. However, I
can 'imagine' some cloud environments where dynamic memory resizing may
be common and there could be operations that want to offline TB's of
memory.
Again, just more to think about.
For this patch,
Reviewed-by: Mike Kravetz <mike.kravetz@...cle.com>
--
Mike Kravetz
Powered by blists - more mailing lists