lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <ZBGY4Im3AQC9GGTu@kernel.org>
Date:   Wed, 15 Mar 2023 12:07:28 +0200
From:   Mike Rapoport <rppt@...nel.org>
To:     "Matthew Wilcox (Oracle)" <willy@...radead.org>
Cc:     linux-arch@...r.kernel.org, linux-mm@...ck.org,
        linux-kernel@...r.kernel.org,
        Geert Uytterhoeven <geert@...ux-m68k.org>,
        linux-m68k@...ts.linux-m68k.org
Subject: Re: [PATCH v4 14/36] m68k: Implement the new page table range API

On Wed, Mar 15, 2023 at 05:14:22AM +0000, Matthew Wilcox (Oracle) wrote:
> Add PFN_PTE_SHIFT, update_mmu_cache_range(), flush_icache_pages() and
> flush_dcache_folio().
> 
> Signed-off-by: Matthew Wilcox (Oracle) <willy@...radead.org>
> Cc: Geert Uytterhoeven <geert@...ux-m68k.org>
> Cc: linux-m68k@...ts.linux-m68k.org

Acked-by: Mike Rapoport (IBM) <rppt@...nel.org>

> ---
>  arch/m68k/include/asm/cacheflush_mm.h    | 27 ++++++++++++++++--------
>  arch/m68k/include/asm/mcf_pgtable.h      |  1 +
>  arch/m68k/include/asm/motorola_pgtable.h |  1 +
>  arch/m68k/include/asm/pgtable_mm.h       |  9 ++++----
>  arch/m68k/include/asm/sun3_pgtable.h     |  1 +
>  arch/m68k/mm/motorola.c                  |  2 +-
>  6 files changed, 27 insertions(+), 14 deletions(-)
> 
> diff --git a/arch/m68k/include/asm/cacheflush_mm.h b/arch/m68k/include/asm/cacheflush_mm.h
> index 1ac55e7b47f0..88eb85e81ef6 100644
> --- a/arch/m68k/include/asm/cacheflush_mm.h
> +++ b/arch/m68k/include/asm/cacheflush_mm.h
> @@ -220,24 +220,29 @@ static inline void flush_cache_page(struct vm_area_struct *vma, unsigned long vm
>  
>  /* Push the page at kernel virtual address and clear the icache */
>  /* RZ: use cpush %bc instead of cpush %dc, cinv %ic */
> -static inline void __flush_page_to_ram(void *vaddr)
> +static inline void __flush_pages_to_ram(void *vaddr, unsigned int nr)
>  {
>  	if (CPU_IS_COLDFIRE) {
>  		unsigned long addr, start, end;
>  		addr = ((unsigned long) vaddr) & ~(PAGE_SIZE - 1);
>  		start = addr & ICACHE_SET_MASK;
> -		end = (addr + PAGE_SIZE - 1) & ICACHE_SET_MASK;
> +		end = (addr + nr * PAGE_SIZE - 1) & ICACHE_SET_MASK;
>  		if (start > end) {
>  			flush_cf_bcache(0, end);
>  			end = ICACHE_MAX_ADDR;
>  		}
>  		flush_cf_bcache(start, end);
>  	} else if (CPU_IS_040_OR_060) {
> -		__asm__ __volatile__("nop\n\t"
> -				     ".chip 68040\n\t"
> -				     "cpushp %%bc,(%0)\n\t"
> -				     ".chip 68k"
> -				     : : "a" (__pa(vaddr)));
> +		unsigned long paddr = __pa(vaddr);
> +
> +		do {
> +			__asm__ __volatile__("nop\n\t"
> +					     ".chip 68040\n\t"
> +					     "cpushp %%bc,(%0)\n\t"
> +					     ".chip 68k"
> +					     : : "a" (paddr));
> +			paddr += PAGE_SIZE;
> +		} while (--nr);
>  	} else {
>  		unsigned long _tmp;
>  		__asm__ __volatile__("movec %%cacr,%0\n\t"
> @@ -249,10 +254,14 @@ static inline void __flush_page_to_ram(void *vaddr)
>  }
>  
>  #define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 1
> -#define flush_dcache_page(page)		__flush_page_to_ram(page_address(page))
> +#define flush_dcache_page(page)	__flush_pages_to_ram(page_address(page), 1)
> +#define flush_dcache_folio(folio)		\
> +	__flush_pages_to_ram(folio_address(folio), folio_nr_pages(folio))
>  #define flush_dcache_mmap_lock(mapping)		do { } while (0)
>  #define flush_dcache_mmap_unlock(mapping)	do { } while (0)
> -#define flush_icache_page(vma, page)	__flush_page_to_ram(page_address(page))
> +#define flush_icache_pages(vma, page, nr)	\
> +	__flush_pages_to_ram(page_address(page), nr)
> +#define flush_icache_page(vma, page) flush_icache_pages(vma, page, 1)
>  
>  extern void flush_icache_user_page(struct vm_area_struct *vma, struct page *page,
>  				    unsigned long addr, int len);
> diff --git a/arch/m68k/include/asm/mcf_pgtable.h b/arch/m68k/include/asm/mcf_pgtable.h
> index 13741c1245e1..1414b607eff4 100644
> --- a/arch/m68k/include/asm/mcf_pgtable.h
> +++ b/arch/m68k/include/asm/mcf_pgtable.h
> @@ -292,6 +292,7 @@ static inline pte_t pte_swp_clear_exclusive(pte_t pte)
>  	return pte;
>  }
>  
> +#define PFN_PTE_SHIFT		PAGE_SHIFT
>  #define pmd_pfn(pmd)		(pmd_val(pmd) >> PAGE_SHIFT)
>  #define pmd_page(pmd)		(pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT))
>  
> diff --git a/arch/m68k/include/asm/motorola_pgtable.h b/arch/m68k/include/asm/motorola_pgtable.h
> index ec0dc19ab834..38d5e5edc3e1 100644
> --- a/arch/m68k/include/asm/motorola_pgtable.h
> +++ b/arch/m68k/include/asm/motorola_pgtable.h
> @@ -112,6 +112,7 @@ static inline void pud_set(pud_t *pudp, pmd_t *pmdp)
>  #define pte_present(pte)	(pte_val(pte) & (_PAGE_PRESENT | _PAGE_PROTNONE))
>  #define pte_clear(mm,addr,ptep)		({ pte_val(*(ptep)) = 0; })
>  
> +#define PFN_PTE_SHIFT		PAGE_SHIFT
>  #define pte_page(pte)		virt_to_page(__va(pte_val(pte)))
>  #define pte_pfn(pte)		(pte_val(pte) >> PAGE_SHIFT)
>  #define pfn_pte(pfn, prot)	__pte(((pfn) << PAGE_SHIFT) | pgprot_val(prot))
> diff --git a/arch/m68k/include/asm/pgtable_mm.h b/arch/m68k/include/asm/pgtable_mm.h
> index b93c41fe2067..8c2db20abdb6 100644
> --- a/arch/m68k/include/asm/pgtable_mm.h
> +++ b/arch/m68k/include/asm/pgtable_mm.h
> @@ -31,8 +31,6 @@
>  	do{							\
>  		*(pteptr) = (pteval);				\
>  	} while(0)
> -#define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval)
> -
>  
>  /* PMD_SHIFT determines the size of the area a second-level page table can map */
>  #if CONFIG_PGTABLE_LEVELS == 3
> @@ -138,11 +136,14 @@ extern void kernel_set_cachemode(void *addr, unsigned long size, int cmode);
>   * tables contain all the necessary information.  The Sun3 does, but
>   * they are updated on demand.
>   */
> -static inline void update_mmu_cache(struct vm_area_struct *vma,
> -				    unsigned long address, pte_t *ptep)
> +static inline void update_mmu_cache_range(struct vm_area_struct *vma,
> +		unsigned long address, pte_t *ptep, unsigned int nr)
>  {
>  }
>  
> +#define update_mmu_cache(vma, addr, ptep) \
> +	update_mmu_cache_range(vma, addr, ptep, 1)
> +
>  #endif /* !__ASSEMBLY__ */
>  
>  /* MMU-specific headers */
> diff --git a/arch/m68k/include/asm/sun3_pgtable.h b/arch/m68k/include/asm/sun3_pgtable.h
> index e582b0484a55..feae73b3b342 100644
> --- a/arch/m68k/include/asm/sun3_pgtable.h
> +++ b/arch/m68k/include/asm/sun3_pgtable.h
> @@ -105,6 +105,7 @@ static inline void pte_clear (struct mm_struct *mm, unsigned long addr, pte_t *p
>  	pte_val (*ptep) = 0;
>  }
>  
> +#define PFN_PTE_SHIFT		0
>  #define pte_pfn(pte)            (pte_val(pte) & SUN3_PAGE_PGNUM_MASK)
>  #define pfn_pte(pfn, pgprot) \
>  ({ pte_t __pte; pte_val(__pte) = pfn | pgprot_val(pgprot); __pte; })
> diff --git a/arch/m68k/mm/motorola.c b/arch/m68k/mm/motorola.c
> index 911301224078..790666c6d146 100644
> --- a/arch/m68k/mm/motorola.c
> +++ b/arch/m68k/mm/motorola.c
> @@ -81,7 +81,7 @@ static inline void cache_page(void *vaddr)
>  
>  void mmu_page_ctor(void *page)
>  {
> -	__flush_page_to_ram(page);
> +	__flush_pages_to_ram(page, 1);
>  	flush_tlb_kernel_page(page);
>  	nocache_page(page);
>  }
> -- 
> 2.39.2
> 

-- 
Sincerely yours,
Mike.

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ