lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Date:	Sun, 07 Oct 2012 07:59:51 +0800
From:	Ni zhan Chen <nizhan.chen@...il.com>
To:	Mike Yoknis <mike.yoknis@...com>
CC:	mgorman@...e.de, mingo@...hat.com, akpm@...ux-foundation.org,
	linux-arch@...r.kernel.org, mmarek@...e.cz, tglx@...utronix.de,
	hpa@...or.com, arnd@...db.de, sam@...nborg.org, minchan@...nel.org,
	kamezawa.hiroyu@...fujitsu.com, mhocko@...e.cz,
	linux-kbuild@...r.kernel.org, linux-kernel@...r.kernel.org,
	linux-mm@...ck.org
Subject: Re: [PATCH] mm: memmap_init_zone() performance improvement

On 10/03/2012 10:56 PM, Mike Yoknis wrote:
> memmap_init_zone() loops through every Page Frame Number (pfn),
> including pfn values that are within the gaps between existing
> memory sections.  The unneeded looping will become a boot
> performance issue when machines configure larger memory ranges
> that will contain larger and more numerous gaps.
>
> The code will skip across invalid sections to reduce the
> number of loops executed.

looks reasonable to me.

>
> Signed-off-by: Mike Yoknis <mike.yoknis@...com>
> ---
>   arch/x86/include/asm/mmzone_32.h     |    2 ++
>   arch/x86/include/asm/page_32.h       |    1 +
>   arch/x86/include/asm/page_64_types.h |    3 ++-
>   include/asm-generic/page.h           |    1 +
>   include/linux/mmzone.h               |    6 ++++++
>   mm/page_alloc.c                      |    5 ++++-
>   6 files changed, 16 insertions(+), 2 deletions(-)
>
> diff --git a/arch/x86/include/asm/mmzone_32.h b/arch/x86/include/asm/mmzone_32.h
> index eb05fb3..73c5c74 100644
> --- a/arch/x86/include/asm/mmzone_32.h
> +++ b/arch/x86/include/asm/mmzone_32.h
> @@ -48,6 +48,8 @@ static inline int pfn_to_nid(unsigned long pfn)
>   #endif
>   }
>   
> +#define next_pfn_try(pfn)	((pfn)+1)
> +
>   static inline int pfn_valid(int pfn)
>   {
>   	int nid = pfn_to_nid(pfn);
> diff --git a/arch/x86/include/asm/page_32.h b/arch/x86/include/asm/page_32.h
> index da4e762..e2c4cfc 100644
> --- a/arch/x86/include/asm/page_32.h
> +++ b/arch/x86/include/asm/page_32.h
> @@ -19,6 +19,7 @@ extern unsigned long __phys_addr(unsigned long);
>   
>   #ifdef CONFIG_FLATMEM
>   #define pfn_valid(pfn)		((pfn) < max_mapnr)
> +#define next_pfn_try(pfn)	((pfn)+1)
>   #endif /* CONFIG_FLATMEM */
>   
>   #ifdef CONFIG_X86_USE_3DNOW
> diff --git a/arch/x86/include/asm/page_64_types.h b/arch/x86/include/asm/page_64_types.h
> index 320f7bb..02d82e5 100644
> --- a/arch/x86/include/asm/page_64_types.h
> +++ b/arch/x86/include/asm/page_64_types.h
> @@ -69,7 +69,8 @@ extern void init_extra_mapping_wb(unsigned long phys, unsigned long size);
>   #endif	/* !__ASSEMBLY__ */
>   
>   #ifdef CONFIG_FLATMEM
> -#define pfn_valid(pfn)          ((pfn) < max_pfn)
> +#define pfn_valid(pfn)		((pfn) < max_pfn)
> +#define next_pfn_try(pfn)	((pfn)+1)
>   #endif
>   
>   #endif /* _ASM_X86_PAGE_64_DEFS_H */
> diff --git a/include/asm-generic/page.h b/include/asm-generic/page.h
> index 37d1fe2..316200d 100644
> --- a/include/asm-generic/page.h
> +++ b/include/asm-generic/page.h
> @@ -91,6 +91,7 @@ extern unsigned long memory_end;
>   #endif
>   
>   #define pfn_valid(pfn)		((pfn) >= ARCH_PFN_OFFSET && ((pfn) - ARCH_PFN_OFFSET) < max_mapnr)
> +#define next_pfn_try(pfn)	((pfn)+1)
>   
>   #define	virt_addr_valid(kaddr)	(((void *)(kaddr) >= (void *)PAGE_OFFSET) && \
>   				((void *)(kaddr) < (void *)memory_end))
> diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h
> index f7d88ba..04d3c39 100644
> --- a/include/linux/mmzone.h
> +++ b/include/linux/mmzone.h
> @@ -1166,6 +1166,12 @@ static inline int pfn_valid(unsigned long pfn)
>   		return 0;
>   	return valid_section(__nr_to_section(pfn_to_section_nr(pfn)));
>   }
> +
> +static inline unsigned long next_pfn_try(unsigned long pfn)
> +{
> +	/* Skip entire section, because all of it is invalid. */
> +	return section_nr_to_pfn(pfn_to_section_nr(pfn) + 1);
> +}
>   #endif
>   
>   static inline int pfn_present(unsigned long pfn)
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index 5b6b6b1..dd2af8b 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -3798,8 +3798,11 @@ void __meminit memmap_init_zone(unsigned long size, int nid, unsigned long zone,
>   		 * exist on hotplugged memory.
>   		 */
>   		if (context == MEMMAP_EARLY) {
> -			if (!early_pfn_valid(pfn))
> +			if (!early_pfn_valid(pfn)) {
> +				pfn = next_pfn_try(pfn);
> +				pfn--;
>   				continue;
> +			}
>   			if (!early_pfn_in_nid(pfn, nid))
>   				continue;
>   		}

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists