lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <b48e5b75-5403-223e-d963-da97a09992be@synopsys.com>
Date:   Fri, 16 Jun 2017 11:28:04 -0700
From:   Vineet Gupta <Vineet.Gupta1@...opsys.com>
To:     Christoph Hellwig <hch@....de>,
        <linux-snps-arc@...ts.infradead.org>
CC:     <linux-kernel@...r.kernel.org>
Subject: Re: [PATCH] arc: implement DMA_ATTR_NO_KERNEL_MAPPING

On 06/16/2017 12:11 AM, Christoph Hellwig wrote:
> This way allocations like the NVMe HMB don't consume iomap space
> 
> Signed-off-by: Christoph Hellwig <hch@....de>
> ---
> 
> Note: compile tested only, I stumbled over this when researching dma api
> quirks for HMB support.
> 
>   arch/arc/mm/dma.c | 42 +++++++++++++++++++++++++++++-------------
>   1 file changed, 29 insertions(+), 13 deletions(-)
> 
> diff --git a/arch/arc/mm/dma.c b/arch/arc/mm/dma.c
> index 2a07e6ecafbd..d8999ac88879 100644
> --- a/arch/arc/mm/dma.c
> +++ b/arch/arc/mm/dma.c
> @@ -28,13 +28,22 @@ static void *arc_dma_alloc(struct device *dev, size_t size,
>   	struct page *page;
>   	phys_addr_t paddr;
>   	void *kvaddr;
> -	int need_coh = 1, need_kvaddr = 0;
> +	bool need_cache_sync = true, need_kvaddr = false;
>   
>   	page = alloc_pages(gfp, order);
>   	if (!page)
>   		return NULL;
>   
>   	/*
> +	 * No-kernel mapping memoery doesn't need a kernel virtual address.
> +	 * But we do the initial cache flush to make sure we don't write back
> +	 * to from a previous mapping into the now device owned memory.
> +	 */
> +	if (attrs & DMA_ATTR_NO_KERNEL_MAPPING) {
> +		need_cache_sync = true;
> +		need_kvaddr = false;

This is re-setting to init values. I do understand that a reasonable compiler will 
elide those away. And maybe keeping them here just clarifies the semantics more. 
So ok !


> +
> +	/*
>   	 * IOC relies on all data (even coherent DMA data) being in cache
>   	 * Thus allocate normal cached memory
>   	 *
> @@ -45,17 +54,19 @@ static void *arc_dma_alloc(struct device *dev, size_t size,
>   	 *   -For coherent data, Read/Write to buffers terminate early in cache
>   	 *   (vs. always going to memory - thus are faster)
>   	 */
> -	if ((is_isa_arcv2() && ioc_enable) ||
> -	    (attrs & DMA_ATTR_NON_CONSISTENT))
> -		need_coh = 0;
> +	} else if ((is_isa_arcv2() && ioc_enable) ||
> +		   (attrs & DMA_ATTR_NON_CONSISTENT)) {
> +		need_cache_sync = false;
> +		need_kvaddr = true;

The conditions here can't really help decide need_kvaddr so best to leave it out 
and not use the "else if" construct. Let the various conditions set and reset 
these 2 knobs based on what is over-riding. e.g. DMA_ATTR_NO_KERNEL_MAPPING seems 
like an optimization hint from a subsys, but might be needed after all given the 
constraints of the architecture.


>   
>   	/*
>   	 * - A coherent buffer needs MMU mapping to enforce non-cachability
>   	 * - A highmem page needs a virtual handle (hence MMU mapping)
>   	 *   independent of cachability
>   	 */
> -	if (PageHighMem(page) || need_coh)
> -		need_kvaddr = 1;
> +	} else if (PageHighMem(page)) {
> +		need_kvaddr = true;
> +	}

Again why "else if".

Also need_coh mandates a kvaddr on ARC (despite DMA_ATTR_NO_KERNEL_MAPPING) since 
the uncached kernel mmu mapping is how you get the coherent semantics.

Also now I think about it, I don't like the name change from need_coh to 
need_cache_sync. conceptually we have dma_alloc_coherent() -> arc_dma_alloc()
to get coherent mem and those semantics are only guaranteed with a kernel mapping.

>   
>   	/* This is linear addr (0x8000_0000 based) */
>   	paddr = page_to_phys(page);
> @@ -83,7 +94,7 @@ static void *arc_dma_alloc(struct device *dev, size_t size,
>   	 * Currently flush_cache_vmap nukes the L1 cache completely which
>   	 * will be optimized as a separate commit
>   	 */
> -	if (need_coh)
> +	if (need_cache_sync)
>   		dma_cache_wback_inv(paddr, size);
>   
>   	return kvaddr;
> @@ -93,14 +104,19 @@ static void arc_dma_free(struct device *dev, size_t size, void *vaddr,
>   		dma_addr_t dma_handle, unsigned long attrs)
>   {
>   	phys_addr_t paddr = plat_dma_to_phys(dev, dma_handle);
> -	struct page *page = virt_to_page(paddr);
> -	int is_non_coh = 1;
>   
> -	is_non_coh = (attrs & DMA_ATTR_NON_CONSISTENT) ||
> -			(is_isa_arcv2() && ioc_enable);
> +	if (!(attrs & DMA_ATTR_NO_KERNEL_MAPPING)) {

Again by similar reasoning as above, arch can be forced to ignore 
DMA_ATTR_NO_KERNEL_MAPPING so it alone can't be used to decide whether the mapping 
exists or not.


> +		struct page *page = virt_to_page(paddr);
> +		bool need_iounmap = true;
> +
> +		if (!PageHighMem(page) &&
> +		    ((is_isa_arcv2() && ioc_enable) ||
> +		     (attrs & DMA_ATTR_NON_CONSISTENT)))
> +			need_iounmap = false;
>   
> -	if (PageHighMem(page) || !is_non_coh)
> -		iounmap((void __force __iomem *)vaddr);
> +		if (need_iounmap)
> +			iounmap((void __force __iomem *)vaddr);
> +	}
>   
>   	__free_pages(page, get_order(size));
>   }


Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ