[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-Id: <201108121700.30967.arnd@arndb.de>
Date: Fri, 12 Aug 2011 17:00:30 +0200
From: Arnd Bergmann <arnd@...db.de>
To: linux-arm-kernel@...ts.infradead.org
Cc: Marek Szyprowski <m.szyprowski@...sung.com>,
linux-kernel@...r.kernel.org, linux-media@...r.kernel.org,
linux-mm@...ck.org, linaro-mm-sig@...ts.linaro.org,
Daniel Walker <dwalker@...eaurora.org>,
Russell King <linux@....linux.org.uk>,
Jonathan Corbet <corbet@....net>, Mel Gorman <mel@....ul.ie>,
Chunsang Jeong <chunsang.jeong@...aro.org>,
Michal Nazarewicz <mina86@...a86.com>,
Jesse Barker <jesse.barker@...aro.org>,
Kyungmin Park <kyungmin.park@...sung.com>,
Ankita Garg <ankita@...ibm.com>,
Shariq Hasnain <shariq.hasnain@...aro.org>,
Andrew Morton <akpm@...ux-foundation.org>,
KAMEZAWA Hiroyuki <kamezawa.hiroyu@...fujitsu.com>
Subject: Re: [PATCH 8/9] ARM: integrate CMA with DMA-mapping subsystem
On Friday 12 August 2011, Marek Szyprowski wrote:
> @@ -82,16 +103,16 @@ static struct page *__dma_alloc_buffer(struct device *dev, size_t size, gfp_t gf
> if (mask < 0xffffffffULL)
> gfp |= GFP_DMA;
>
> - page = alloc_pages(gfp, order);
> - if (!page)
> - return NULL;
> -
> /*
> - * Now split the huge page and free the excess pages
> + * Allocate contiguous memory
> */
> - split_page(page, order);
> - for (p = page + (size >> PAGE_SHIFT), e = page + (1 << order); p < e; p++)
> - __free_page(p);
> + if (cma_available())
> + page = dma_alloc_from_contiguous(dev, count, order);
> + else
> + page = __dma_alloc_system_pages(count, gfp, order);
> +
> + if (!page)
> + return NULL;
Why do you need the fallback here? I would assume that CMA now has to be available
on ARMv6 and up to work at all. When you allocate from __dma_alloc_system_pages(),
wouldn't that necessarily fail in the dma_remap_area() stage?
>
> - if (arch_is_coherent() || nommu()) {
> + if (arch_is_coherent() || nommu() ||
> + (cma_available() && !(gfp & GFP_ATOMIC))) {
> + /*
> + * Allocate from system or CMA pages
> + */
> struct page *page = __dma_alloc_buffer(dev, size, gfp);
> if (!page)
> return NULL;
> + dma_remap_area(page, size, area->prot);
> pfn = page_to_pfn(page);
> ret = page_address(page);
Similarly with coherent and nommu. It seems to me that lumping too
many cases together creates extra complexity here.
How about something like
if (arch_is_coherent() || nommu())
ret = alloc_simple_buffer();
else if (arch_is_v4_v5())
ret = alloc_remap();
else if (gfp & GFP_ATOMIC)
ret = alloc_from_pool();
else
ret = alloc_from_contiguous();
This also allows a natural conversion to dma_map_ops when we get there.
> /* reserve any platform specific memblock areas */
> if (mdesc->reserve)
> mdesc->reserve();
>
> + dma_coherent_reserve();
> + dma_contiguous_reserve();
> +
> memblock_analyze();
> memblock_dump_all();
> }
Since we can handle most allocations using CMA on ARMv6+, I would think
that we can have a much smaller reserved area. Have you tried changing
dma_coherent_reserve() to allocate out of the contiguous area instead of
wasting a full 2MB section of memory?
Arnd
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists