lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <d97a8744-6d49-4582-82fc-a38ee2ae06fc@intel.com>
Date: Thu, 15 Jan 2026 23:09:11 +0800
From: "Li, Tianyou" <tianyou.li@...el.com>
To: "David Hildenbrand (Red Hat)" <david@...nel.org>, Oscar Salvador
	<osalvador@...e.de>, Mike Rapoport <rppt@...nel.org>, Wei Yang
	<richard.weiyang@...il.com>
CC: <linux-mm@...ck.org>, Yong Hu <yong.hu@...el.com>, Nanhai Zou
	<nanhai.zou@...el.com>, Yuan Liu <yuan1.liu@...el.com>, Tim Chen
	<tim.c.chen@...ux.intel.com>, Qiuxu Zhuo <qiuxu.zhuo@...el.com>, Yu C Chen
	<yu.c.chen@...el.com>, Pan Deng <pan.deng@...el.com>, Chen Zhang
	<zhangchen.kidd@...com>, <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH v7 2/2] mm/memory hotplug/unplug: Optimize
 zone->contiguous update when changes pfn range

Sorry for my delayed response. Appreciated for all your suggestions 
David. My thoughts inlined for your kind consideration.

On 1/14/2026 7:13 PM, David Hildenbrand (Red Hat) wrote:
>>>
>>> This is nasty. I would wish we could just leave that code path alone.
>>>
>>> In particular: I am 99% sure that we never ever run into this case in
>>> practice.
>>>
>>> E.g., on x86, we can have up to 2 GiB memory blocks. But the memmap of
>>> that is 64/4096*2GiB == 32 MB ... and a memory section is 128 MiB.
>>>
>>>
>>> As commented on patch #1, we should drop the set_zone_contiguous() in
>>> this function either way and let online_pages() deal with it.
>>>
>>> We just have to make sure that we don't create some inconsistencies by
>>> doing that.
>>>
>>> Can you double-check?
>
> I thought about this some more, and it's all a bit nasty. We have to 
> get this right.
>
> Losing the optimization for memmap_on_memory users indicates that we 
> are doing the wrong thing.
>
> You could introduce the set_zone_contiguous() in this patch here. But 
> then, I think instead of
>
> +    /*
> +     * If the allocated memmap pages are not in a full section, keep the
> +     * contiguous state as ZONE_CONTIG_NO.
> +     */
> +    if (IS_ALIGNED(end_pfn, PAGES_PER_SECTION))
> +        new_contiguous_state = zone_contig_state_after_growing(zone,
> +                                pfn, nr_pages);
> +
>
> We'd actually unconditionally have to do that, no?
>
That's a great idea! Probably we need to move 
the adjust_present_page_count right after mhp_init_memmap_on_memory to 
keep the present_pages in sync.

When the zone is contiguous previously, there probably 2 situations:

1. If new added range is at the start of the zone, then 
after mhp_init_memmap_on_memory, the zone contiguous will be false at 
fast path. It should be expected as long as the 
adjust_present_page_count was called before the online_pages, 
so zone_contig_state_after_growing in online_pages will return 
ZONE_CONTIG_MAYBE, which is expected. Then the set_zone_contiguous in 
online_pages will get correct result.

2. If new added range is at the end of the zone, then 
the zone_contig_state_after_growing will return ZONE_CONTIG_YES or 
ZONE_CONTIG_NO, regardless of the memory section online or not. When the 
contiguous check comes into the online_pages, it will follow the fast 
path and get the correct contiguous state.

When the zone is not contiguous previously, in any case the new zone 
contiguous state will be false in mhp_init_memmap_on_memory, because 
either the nr_vmemmap_pages can not fill the hole or the memory section 
is not online. If we update the present_pages correctly, then in 
online_pages, the zone_contig_state_after_growing could have the chance 
to return ZONE_CONTIG_MAYBE, and since all memory sections are onlined, 
the set_zone_contiguous will get the correct result.


I am not sure if you'd like to consider another option: could we 
encapsulate the mhp_init_memmap_on_memory and online_pages into one 
function eg. online_memory_block_pages, and offline_memory_block_pages 
correspondingly as well, in the memory_hotplug.c. So we can check the 
zone contiguous state as the whole for the new added range.

int online_memory_block_pages(unsigned long start_pfn, unsigned long 
nr_pages,
             unsigned long nr_vmemmap_pages, struct zone *zone,
             struct memory_group *group)
{
     bool contiguous = zone->contiguous;
     enum zone_contig_state new_contiguous_state;
     int ret;

     /*
      * Calculate the new zone contig state before move_pfn_range_to_zone()
      * sets the zone temporarily to non-contiguous.
      */
     new_contiguous_state = zone_contig_state_after_growing(zone, start_pfn,
                                 nr_pages);

     if (nr_vmemmap_pages) {
         ret = mhp_init_memmap_on_memory(start_pfn, nr_vmemmap_pages,
                         zone);
         if (ret)
             goto restore_zone_contig;
     }

     ret = online_pages(start_pfn + nr_vmemmap_pages,
                nr_pages - nr_vmemmap_pages, zone, group);
     if (ret) {
         if (nr_vmemmap_pages)
             mhp_deinit_memmap_on_memory(start_pfn, nr_vmemmap_pages);
         goto restore_zone_contig;
     }

     /*
      * Account once onlining succeeded. If the zone was unpopulated, it is
      * now already properly populated.
      */
     if (nr_vmemmap_pages)
         adjust_present_page_count(pfn_to_page(start_pfn), mem->group,
                       nr_vmemmap_pages);

     /*
      * Now that the ranges are indicated as online, check whether the whole
      * zone is contiguous.
      */
     set_zone_contiguous(zone, new_contiguous_state);
     return 0;

restore_zone_contig:
     zone->contiguous = contiguous;
     return ret;
}


I am OK with either way, please let me know your preference. Thanks.



Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ