lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:   Fri, 23 Sep 2022 17:18:54 +0530
From:   Anshuman Khandual <anshuman.khandual@....com>
To:     Kefeng Wang <wangkefeng.wang@...wei.com>,
        Catalin Marinas <catalin.marinas@....com>,
        Will Deacon <will@...nel.org>,
        linux-arm-kernel@...ts.infradead.org
Cc:     linux-kernel@...r.kernel.org
Subject: Re: [PATCH] arm64: mm: handle ARM64_KERNEL_USES_PMD_MAPS in
 vmemmap_populate()



On 9/20/22 07:19, Kefeng Wang wrote:
> Directly check ARM64_SWAPPER_USES_SECTION_MAPS to choose base page
> or PMD level huge page mapping in vmemmap_populate() to simplify
> code a bit.
> 
> Signed-off-by: Kefeng Wang <wangkefeng.wang@...wei.com>

Reviewed-by: Anshuman Khandual <anshuman.khandual@....com>

> ---
>  arch/arm64/mm/mmu.c | 13 ++++---------
>  1 file changed, 4 insertions(+), 9 deletions(-)
> 
> diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c
> index 5810eddfb48e..784afa9c60ac 100644
> --- a/arch/arm64/mm/mmu.c
> +++ b/arch/arm64/mm/mmu.c
> @@ -1180,14 +1180,6 @@ static void free_empty_tables(unsigned long addr, unsigned long end,
>  }
>  #endif
>  
> -#if !ARM64_KERNEL_USES_PMD_MAPS
> -int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
> -		struct vmem_altmap *altmap)
> -{
> -	WARN_ON((start < VMEMMAP_START) || (end > VMEMMAP_END));
> -	return vmemmap_populate_basepages(start, end, node, altmap);
> -}
> -#else	/* !ARM64_KERNEL_USES_PMD_MAPS */
>  int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
>  		struct vmem_altmap *altmap)
>  {
> @@ -1199,6 +1191,10 @@ int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
>  	pmd_t *pmdp;
>  
>  	WARN_ON((start < VMEMMAP_START) || (end > VMEMMAP_END));
> +
> +	if (!ARM64_KERNEL_USES_PMD_MAPS)
> +		return vmemmap_populate_basepages(start, end, node, altmap);
> +
>  	do {
>  		next = pmd_addr_end(addr, end);
>  
> @@ -1232,7 +1228,6 @@ int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,
>  
>  	return 0;
>  }
> -#endif	/* !ARM64_KERNEL_USES_PMD_MAPS */
>  
>  #ifdef CONFIG_MEMORY_HOTPLUG
>  void vmemmap_free(unsigned long start, unsigned long end,

Powered by blists - more mailing lists