lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Date:	Mon, 02 May 2016 16:47:27 +0530
From:	"Aneesh Kumar K.V" <aneesh.kumar@...ux.vnet.ibm.com>
To:	Stephen Rothwell <sfr@...b.auug.org.au>,
	Andrew Morton <akpm@...ux-foundation.org>
Cc:	linux-next@...r.kernel.org, linux-kernel@...r.kernel.org,
	Hugh Dickins <hughd@...gle.com>
Subject: Re: linux-next: manual merge of the akpm-current tree with the powerpc tree

Stephen Rothwell <sfr@...b.auug.org.au> writes:

> Hi Andrew,
>
> Today's linux-next merge of the akpm-current tree got a conflict in:
>
>   arch/powerpc/include/asm/book3s/64/pgtable.h
>
> between commit:
>
>   dbaba7a16b7b ("powerpc/mm: THP is only available on hash64 as of now")
>
> from the powerpc tree and commit:
>
>   383b50a4e356 ("arch: fix has_transparent_hugepage()")
>
> from the akpm-current tree.
>
> I fixed it up (see below - the code moved) and can carry the fix as
> necessary. This is now fixed as far as linux-next is concerned, but any
> non trivial conflicts should be mentioned to your upstream maintainer
> when your tree is submitted for merging.  You may also want to consider
> cooperating with the maintainer of the conflicting tree to minimise any
> particularly complex conflicts.
>


looks good.

> -- 
> Cheers,
> Stephen Rothwell
>
> diff --cc arch/powerpc/include/asm/book3s/64/pgtable.h
> index 48dc76c13094,8fe6f6b48aa5..000000000000
> --- a/arch/powerpc/include/asm/book3s/64/pgtable.h
> +++ b/arch/powerpc/include/asm/book3s/64/pgtable.h
> @@@ -820,86 -265,9 +820,87 @@@ static inline int pmd_protnone(pmd_t pm
>   #define __HAVE_ARCH_PMD_WRITE
>   #define pmd_write(pmd)		pte_write(pmd_pte(pmd))
>   
>  +#ifdef CONFIG_TRANSPARENT_HUGEPAGE
>  +extern pmd_t pfn_pmd(unsigned long pfn, pgprot_t pgprot);
>  +extern pmd_t mk_pmd(struct page *page, pgprot_t pgprot);
>  +extern pmd_t pmd_modify(pmd_t pmd, pgprot_t newprot);
>  +extern void set_pmd_at(struct mm_struct *mm, unsigned long addr,
>  +		       pmd_t *pmdp, pmd_t pmd);
>  +extern void update_mmu_cache_pmd(struct vm_area_struct *vma, unsigned long addr,
>  +				 pmd_t *pmd);
>  +extern int hash__has_transparent_hugepage(void);
> ++#define has_transparent_hugepage has_transparent_hugepage
>  +static inline int has_transparent_hugepage(void)
>  +{
>  +	if (radix_enabled())
>  +		return radix__has_transparent_hugepage();
>  +	return hash__has_transparent_hugepage();
>  +}
>  +
>  +static inline unsigned long
>  +pmd_hugepage_update(struct mm_struct *mm, unsigned long addr, pmd_t *pmdp,
>  +		    unsigned long clr, unsigned long set)
>  +{
>  +	if (radix_enabled())
>  +		return radix__pmd_hugepage_update(mm, addr, pmdp, clr, set);
>  +	return hash__pmd_hugepage_update(mm, addr, pmdp, clr, set);
>  +}
>  +
>  +static inline int pmd_large(pmd_t pmd)
>  +{
>  +	return !!(pmd_val(pmd) & _PAGE_PTE);
>  +}
>  +
>  +static inline pmd_t pmd_mknotpresent(pmd_t pmd)
>  +{
>  +	return __pmd(pmd_val(pmd) & ~_PAGE_PRESENT);
>  +}
>  +/*
>  + * For radix we should always find H_PAGE_HASHPTE zero. Hence
>  + * the below will work for radix too
>  + */
>  +static inline int __pmdp_test_and_clear_young(struct mm_struct *mm,
>  +					      unsigned long addr, pmd_t *pmdp)
>  +{
>  +	unsigned long old;
>  +
>  +	if ((pmd_val(*pmdp) & (_PAGE_ACCESSED | H_PAGE_HASHPTE)) == 0)
>  +		return 0;
>  +	old = pmd_hugepage_update(mm, addr, pmdp, _PAGE_ACCESSED, 0);
>  +	return ((old & _PAGE_ACCESSED) != 0);
>  +}
>  +
>  +#define __HAVE_ARCH_PMDP_SET_WRPROTECT
>  +static inline void pmdp_set_wrprotect(struct mm_struct *mm, unsigned long addr,
>  +				      pmd_t *pmdp)
>  +{
>  +
>  +	if ((pmd_val(*pmdp) & _PAGE_WRITE) == 0)
>  +		return;
>  +
>  +	pmd_hugepage_update(mm, addr, pmdp, _PAGE_WRITE, 0);
>  +}
>  +
>  +static inline int pmd_trans_huge(pmd_t pmd)
>  +{
>  +	if (radix_enabled())
>  +		return radix__pmd_trans_huge(pmd);
>  +	return hash__pmd_trans_huge(pmd);
>  +}
>  +
>  +#define __HAVE_ARCH_PMD_SAME
>  +static inline int pmd_same(pmd_t pmd_a, pmd_t pmd_b)
>  +{
>  +	if (radix_enabled())
>  +		return radix__pmd_same(pmd_a, pmd_b);
>  +	return hash__pmd_same(pmd_a, pmd_b);
>  +}
>  +
>   static inline pmd_t pmd_mkhuge(pmd_t pmd)
>   {
>  -	return __pmd(pmd_val(pmd) | (_PAGE_PTE | _PAGE_THP_HUGE));
>  +	if (radix_enabled())
>  +		return radix__pmd_mkhuge(pmd);
>  +	return hash__pmd_mkhuge(pmd);
>   }
>   
>   #define __HAVE_ARCH_PMDP_SET_ACCESS_FLAGS

Powered by blists - more mailing lists