lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-Id: <20190221111525.GB32004@rapoport-lnx>
Date:   Thu, 21 Feb 2019 13:15:26 +0200
From:   Mike Rapoport <rppt@...ux.ibm.com>
To:     Helge Deller <deller@....de>
Cc:     "James E.J. Bottomley" <James.Bottomley@...senPartnership.com>,
        Matthew Wilcox <willy@...radead.org>,
        linux-parisc@...r.kernel.org, linux-mm@...ck.org,
        linux-kernel@...r.kernel.org
Subject: Re: [PATCH v2] parisc: use memblock_alloc() instead of custom
 get_memblock()

On Thu, Feb 21, 2019 at 11:00:05AM +0100, Helge Deller wrote:
> On 21.02.19 10:07, Mike Rapoport wrote:
> > On Tue, Feb 12, 2019 at 05:16:12PM +0200, Mike Rapoport wrote:
> >> The get_memblock() function implements custom bottom-up memblock allocator.
> >> Setting 'memblock_bottom_up = true' before any memblock allocation is done
> >> allows replacing get_memblock() calls with memblock_alloc().
> 
> >> Signed-off-by: Mike Rapoport <rppt@...ux.ibm.com>
> 
> Acked-by: Helge Deller <deller@....de>
> Tested-by: Helge Deller <deller@....de>
> 
> Thanks!
> Shall I push the patch upstream with the parisc tree?

Yes, please.
 
> Helge
> 
> 
> 
> >> ---
> >> v2: fix allocation alignment
> >>
> >>  arch/parisc/mm/init.c | 52 +++++++++++++++++++--------------------------------
> >>  1 file changed, 19 insertions(+), 33 deletions(-)
> >>
> >> diff --git a/arch/parisc/mm/init.c b/arch/parisc/mm/init.c
> >> index 059187a..d0b1662 100644
> >> --- a/arch/parisc/mm/init.c
> >> +++ b/arch/parisc/mm/init.c
> >> @@ -79,36 +79,6 @@ static struct resource sysram_resources[MAX_PHYSMEM_RANGES] __read_mostly;
> >>  physmem_range_t pmem_ranges[MAX_PHYSMEM_RANGES] __read_mostly;
> >>  int npmem_ranges __read_mostly;
> >>  
> >> -/*
> >> - * get_memblock() allocates pages via memblock.
> >> - * We can't use memblock_find_in_range(0, KERNEL_INITIAL_SIZE) here since it
> >> - * doesn't allocate from bottom to top which is needed because we only created
> >> - * the initial mapping up to KERNEL_INITIAL_SIZE in the assembly bootup code.
> >> - */
> >> -static void * __init get_memblock(unsigned long size)
> >> -{
> >> -	static phys_addr_t search_addr __initdata;
> >> -	phys_addr_t phys;
> >> -
> >> -	if (!search_addr)
> >> -		search_addr = PAGE_ALIGN(__pa((unsigned long) &_end));
> >> -	search_addr = ALIGN(search_addr, size);
> >> -	while (!memblock_is_region_memory(search_addr, size) ||
> >> -		memblock_is_region_reserved(search_addr, size)) {
> >> -		search_addr += size;
> >> -	}
> >> -	phys = search_addr;
> >> -
> >> -	if (phys)
> >> -		memblock_reserve(phys, size);
> >> -	else
> >> -		panic("get_memblock() failed.\n");
> >> -
> >> -	memset(__va(phys), 0, size);
> >> -
> >> -	return __va(phys);
> >> -}
> >> -
> >>  #ifdef CONFIG_64BIT
> >>  #define MAX_MEM         (~0UL)
> >>  #else /* !CONFIG_64BIT */
> >> @@ -321,6 +291,13 @@ static void __init setup_bootmem(void)
> >>  			max_pfn = start_pfn + npages;
> >>  	}
> >>  
> >> +	/*
> >> +	 * We can't use memblock top-down allocations because we only
> >> +	 * created the initial mapping up to KERNEL_INITIAL_SIZE in
> >> +	 * the assembly bootup code.
> >> +	 */
> >> +	memblock_set_bottom_up(true);
> >> +
> >>  	/* IOMMU is always used to access "high mem" on those boxes
> >>  	 * that can support enough mem that a PCI device couldn't
> >>  	 * directly DMA to any physical addresses.
> >> @@ -442,7 +419,10 @@ static void __init map_pages(unsigned long start_vaddr,
> >>  		 */
> >>  
> >>  		if (!pmd) {
> >> -			pmd = (pmd_t *) get_memblock(PAGE_SIZE << PMD_ORDER);
> >> +			pmd = memblock_alloc(PAGE_SIZE << PMD_ORDER,
> >> +					     PAGE_SIZE << PMD_ORDER);
> >> +			if (!pmd)
> >> +				panic("pmd allocation failed.\n");
> >>  			pmd = (pmd_t *) __pa(pmd);
> >>  		}
> >>  
> >> @@ -461,7 +441,10 @@ static void __init map_pages(unsigned long start_vaddr,
> >>  
> >>  			pg_table = (pte_t *)pmd_address(*pmd);
> >>  			if (!pg_table) {
> >> -				pg_table = (pte_t *) get_memblock(PAGE_SIZE);
> >> +				pg_table = memblock_alloc(PAGE_SIZE,
> >> +							  PAGE_SIZE);
> >> +				if (!pg_table)
> >> +					panic("page table allocation failed\n");
> >>  				pg_table = (pte_t *) __pa(pg_table);
> >>  			}
> >>  
> >> @@ -700,7 +683,10 @@ static void __init pagetable_init(void)
> >>  	}
> >>  #endif
> >>  
> >> -	empty_zero_page = get_memblock(PAGE_SIZE);
> >> +	empty_zero_page = memblock_alloc(PAGE_SIZE, PAGE_SIZE);
> >> +	if (!empty_zero_page)
> >> +		panic("zero page allocation failed.\n");
> >> +
> >>  }
> >>  
> >>  static void __init gateway_init(void)
> >> -- 
> >> 2.7.4
> >>
> > 
> 

-- 
Sincerely yours,
Mike.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ