lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <YntRlrwJeP40q6Hg@kernel.org>
Date:   Wed, 11 May 2022 09:03:02 +0300
From:   Mike Rapoport <rppt@...nel.org>
To:     Andrew Morton <akpm@...ux-foundation.org>
Cc:     Zhou Guanghui <zhouguanghui1@...wei.com>,
        linux-kernel@...r.kernel.org, linux-mm@...ck.org,
        xuqiang36@...wei.com
Subject: Re: [PATCH] memblock: config the number of init memblock regions

On Tue, May 10, 2022 at 06:55:23PM -0700, Andrew Morton wrote:
> On Wed, 11 May 2022 01:05:30 +0000 Zhou Guanghui <zhouguanghui1@...wei.com> wrote:
> 
> > During early boot, the number of memblocks may exceed 128(some memory
> > areas are not reported to the kernel due to test failures. As a result,
> > contiguous memory is divided into multiple parts for reporting). If
> > the size of the init memblock regions is exceeded before the array size
> > can be resized, the excess memory will be lost.

I'd like to see more details about how firmware creates that sparse memory
map in the changelog.

> > 
> > ...
> >
> > --- a/mm/Kconfig
> > +++ b/mm/Kconfig
> > @@ -89,6 +89,14 @@ config SPARSEMEM_VMEMMAP
> >  	  pfn_to_page and page_to_pfn operations.  This is the most
> >  	  efficient option when sufficient kernel resources are available.
> >  
> > +config MEMBLOCK_INIT_REGIONS
> > +	int "Number of init memblock regions"
> > +	range 128 1024
> > +	default 128
> > +	help
> > +	  The number of init memblock regions which used to track "memory" and
> > +	  "reserved" memblocks during early boot.
> > +
> >  config HAVE_MEMBLOCK_PHYS_MAP
> >  	bool
> >  
> > diff --git a/mm/memblock.c b/mm/memblock.c
> > index e4f03a6e8e56..6893d26b750e 100644
> > --- a/mm/memblock.c
> > +++ b/mm/memblock.c
> > @@ -22,7 +22,7 @@
> >  
> >  #include "internal.h"
> >  
> > -#define INIT_MEMBLOCK_REGIONS			128
> > +#define INIT_MEMBLOCK_REGIONS			CONFIG_MEMBLOCK_INIT_REGIONS
> 
> Consistent naming would be nice - MEMBLOCK_INIT versus INIT_MEMBLOCK.
> 
> Can we simply increase INIT_MEMBLOCK_REGIONS to 1024 and avoid the
> config option?  It appears that the overhead from this would be 60kB or
> so.

60k is not big, but using 1024 entries array for 2-4 memory banks on
systems that don't report that fragmented memory map is really a waste.

We can make this per platform opt-in, like INIT_MEMBLOCK_RESERVED_REGIONS ...

> Or zero if CONFIG_ARCH_KEEP_MEMBLOCK and CONFIG_MEMORY_HOTPLUG
> are cooperating.

... or add code that will discard unused parts of memblock arrays even if
CONFIG_ARCH_KEEP_MEMBLOCK=y.

-- 
Sincerely yours,
Mike.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ