lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Date:	Mon, 20 Oct 2008 12:18:46 -0700
From:	Dan Williams <dan.j.williams@...el.com>
To:	Nicolas Ferre <nicolas.ferre@...el.com>
Cc:	Linux Kernel list <linux-kernel@...r.kernel.org>,
	ARM Linux Mailing List 
	<linux-arm-kernel@...ts.arm.linux.org.uk>,
	"Sosnowski, Maciej" <maciej.sosnowski@...el.com>,
	Haavard Skinnemoen <hskinnemoen@...el.com>,
	Andrew Victor <linux@...im.org.za>
Subject: Re: [PATCH] dmaengine: at_hdmac: new driver for the Atmel AHB DMA
	Controller


On Fri, 2008-10-17 at 08:43 -0700, Nicolas Ferre wrote:
> This AHB DMA Controller (aka HDMA or DMAC on AT91 systems) is availlable on
> at91sam9rl chip. It will be used on other products in the future.
> 
> This first release covers only the memory-to-memory tranfer type. This is the
> only tranfer type supported by this chip.
> On other products, it will be used also for peripheral DMA transfer (slave API
> support to come).
> 
> I used dmatest client without problem in different configurations to test
> it.
> 
> Full documentation for this controller can be found in the SAM9RL datasheet :
> http://www.atmel.com/dyn/products/product_card.asp?part_id=4243
> 
> Signed-off-by: Nicolas Ferre <nicolas.ferre@...el.com>
> ---

Hi Nicolas,

A few comments below.

Also, checkpatch reported:

total: 4 errors, 45 warnings, 1475 lines checked

...mostly 80 column warnings (some you may want to take a look at).


Regards,
Dan

>  arch/arm/mach-at91/at91sam9rl_devices.c |   47 ++
>  drivers/dma/Kconfig                     |    8 +
>  drivers/dma/Makefile                    |    1 +
>  drivers/dma/at_hdmac.c                  |  989 +++++++++++++++++++++++++++++++
>  drivers/dma/at_hdmac_regs.h             |  377 ++++++++++++
>  include/linux/at_hdmac.h                |   26 +

...this header should be moved somewhere under arch/arm/include.

>  6 files changed, 1448 insertions(+), 0 deletions(-)
> 
> diff --git a/arch/arm/mach-at91/at91sam9rl_devices.c b/arch/arm/mach-at91/at91sam9rl_devices.c
> index 87deb1e..ad596ff 100644
> --- a/arch/arm/mach-at91/at91sam9rl_devices.c
> +++ b/arch/arm/mach-at91/at91sam9rl_devices.c
> @@ -9,6 +9,7 @@
>  #include <asm/mach/arch.h>
>  #include <asm/mach/map.h>
> 
> +#include <linux/at_hdmac.h>
>  #include <linux/dma-mapping.h>
>  #include <linux/platform_device.h>
>  #include <linux/i2c-gpio.h>
> @@ -26,6 +27,51 @@
> 
> 
>  /* --------------------------------------------------------------------
> + *  HDMAC - AHB DMA Controller
> + * -------------------------------------------------------------------- */
> +
> +#if defined(CONFIG_AT_HDMAC) || defined(CONFIG_AT_HDMAC_MODULE)
> +static u64 hdmac_dmamask = DMA_BIT_MASK(32);
> +
> +static struct at_dma_platform_data atdma_pdata = {
> +       .nr_channels    = 2,
> +};
> +
> +static struct resource hdmac_resources[] = {
> +       [0] = {
> +               .start  = AT91_BASE_SYS + AT91_DMA,
> +               .end    = AT91_BASE_SYS + AT91_DMA + SZ_512 - 1,
> +               .flags  = IORESOURCE_MEM,
> +       },
> +       [2] = {
> +               .start  = AT91SAM9RL_ID_DMA,
> +               .end    = AT91SAM9RL_ID_DMA,
> +               .flags  = IORESOURCE_IRQ,
> +       },
> +};
> +
> +static struct platform_device at_hdmac_device = {
> +       .name           = "at_hdmac",
> +       .id             = -1,
> +       .dev            = {
> +                               .dma_mask               = &hdmac_dmamask,
> +                               .coherent_dma_mask      = DMA_BIT_MASK(32),
> +                               .platform_data          = &atdma_pdata,
> +       },
> +       .resource       = hdmac_resources,
> +       .num_resources  = ARRAY_SIZE(hdmac_resources),
> +};
> +
> +void __init at91_add_device_hdmac(void)
> +{
> +       dma_cap_set(DMA_MEMCPY, atdma_pdata.cap_mask);
> +       platform_device_register(&at_hdmac_device);
> +}
> +#else
> +void __init at91_add_device_hdmac(void) {}
> +#endif
> +
> +/* --------------------------------------------------------------------
>   *  USB HS Device (Gadget)
>   * -------------------------------------------------------------------- */
> 
> @@ -1114,6 +1160,7 @@ void __init at91_add_device_serial(void) {}
>   */
>  static int __init at91_add_standard_devices(void)
>  {
> +       at91_add_device_hdmac();
>         at91_add_device_rtc();
>         at91_add_device_rtt();
>         at91_add_device_watchdog();
> diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig
> index cd30390..03bfd8f 100644
> --- a/drivers/dma/Kconfig
> +++ b/drivers/dma/Kconfig
> @@ -47,6 +47,14 @@ config DW_DMAC
>           Support the Synopsys DesignWare AHB DMA controller.  This
>           can be integrated in chips such as the Atmel AT32ap7000.
> 
> +config AT_HDMAC
> +       tristate "Atmel AHB DMA support"
> +       depends on ARCH_AT91 && ARCH_AT91SAM9RL
> +       select DMA_ENGINE
> +       help
> +         Support the Atmel AHB DMA controller.  This can be integrated in
> +         chips such as the Atmel AT91SAM9RL.
> +
>  config FSL_DMA
>         bool "Freescale MPC85xx/MPC83xx DMA support"
>         depends on PPC
> diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile
> index 14f5952..9c8ce35 100644
> --- a/drivers/dma/Makefile
> +++ b/drivers/dma/Makefile
> @@ -7,3 +7,4 @@ obj-$(CONFIG_INTEL_IOP_ADMA) += iop-adma.o
>  obj-$(CONFIG_FSL_DMA) += fsldma.o
>  obj-$(CONFIG_MV_XOR) += mv_xor.o
>  obj-$(CONFIG_DW_DMAC) += dw_dmac.o
> +obj-$(CONFIG_AT_HDMAC) += at_hdmac.o
> diff --git a/drivers/dma/at_hdmac.c b/drivers/dma/at_hdmac.c
> new file mode 100644
> index 0000000..2f0a386
> --- /dev/null
> +++ b/drivers/dma/at_hdmac.c
> @@ -0,0 +1,989 @@
> +/*
> + * Driver for the Atmel AHB DMA Controller (aka HDMA or DMAC on AT91 systems)
> + *
> + * Copyright (C) 2008 Atmel Corporation
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by
> + * the Free Software Foundation; either version 2 of the License, or
> + * (at your option) any later version.
> + *
> + *
> + * This supports the Atmel AHB DMA Controller,
> + *
> + * The driver has currently been tested with the Atmel AT91SAM9RL
> + * and AT91SAM9M10.
> + */
> +
> +#include <linux/clk.h>
> +#include <linux/dmaengine.h>
> +#include <linux/dma-mapping.h>
> +#include <linux/dmapool.h>
> +#include <linux/interrupt.h>
> +#include <linux/module.h>
> +#include <linux/platform_device.h>
> +
> +#include "at_hdmac_regs.h"
> +
> +/*
> + * Glossary
> + * --------
> + *
> + * at_hdmac            : Name of the ATmel AHB DMA Controller
> + * at_dma_ / atdma     : ATmel DMA controller entity related
> + * atc_        / atchan        : ATmel DMA Channel entity related
> + */
> +
> +#define        ATC_DEFAULT_CFG         (ATC_FIFOCFG_HALFFIFO)
> +#define        ATC_DEFAULT_CTRLA       (0)
> +#define        ATC_DEFAULT_CTRLB       ( ATC_SIF(0)    \
> +                               | ATC_DIF(1))
> +
> +/*
> + * Initial number of descriptors to allocate for each channel. This could
> + * be increased of this amount during dma usage.
> + */
> +#define INIT_NR_DESCS_PER_CHANNEL      16
> +
> +/* prototypes */
> +static dma_cookie_t atc_tx_submit(struct dma_async_tx_descriptor *tx);
> +
> +
> +/*----------------------------------------------------------------------*/
> +
> +static struct at_desc *atc_first_active(struct at_dma_chan *atchan)
> +{
> +       return list_first_entry(&atchan->active_list, struct at_desc, desc_node);
> +}
> +
> +static struct at_desc *atc_first_queued(struct at_dma_chan *atchan)
> +{
> +       return list_first_entry(&atchan->queue, struct at_desc, desc_node);
> +}
> +
> +/**
> + * atc_alloc_descriptor - allocate and return an initilized descriptor
> + * @chan: the channel to allocate descriptors for
> + * @gfp_flags: GFP allocation flags
> + */
> +static struct at_desc *atc_alloc_descriptor(struct dma_chan *chan,
> +                                           gfp_t gfp_flags)
> +{
> +       struct at_desc  *desc = NULL;
> +       struct at_dma   *atdma = to_at_dma(chan->device);
> +       dma_addr_t phys;
> +
> +       desc = dma_pool_alloc(atdma->dma_desc_pool, gfp_flags, &phys);
> +       if (desc) {
> +               BUG_ON(phys & 0x3UL); /* descriptors have to be word aligned */

hmm, yes this is a bug but can't we trust that dma_pool_alloc does its
job correctly?

> +               memset(desc, 0, sizeof(struct at_desc));
> +               dma_async_tx_descriptor_init(&desc->txd, chan);
> +               async_tx_ack(&desc->txd);

the DMA_CTRL_ACK bit is under control of the client.  It should be
read-only to the driver (except for extra descriptors that the driver
creates on behalf of the client).

> +               desc->txd.tx_submit = atc_tx_submit;
> +               INIT_LIST_HEAD(&desc->txd.tx_list);
> +               desc->txd.phys = phys;
> +       }
> +
> +       return desc;
> +}
> +
> +/**
> + * atc_desc_get - get a unsused descriptor from free_list
> + * @atchan: channel we want a new descriptor for
> + */
> +static struct at_desc *atc_desc_get(struct at_dma_chan *atchan)
> +{
> +       struct at_desc *desc, *_desc;
> +       struct at_desc *ret = NULL;
> +       unsigned int i = 0;
> +       LIST_HEAD(tmp_list);
> +
> +       spin_lock_bh(&atchan->lock);
> +       list_for_each_entry_safe(desc, _desc, &atchan->free_list, desc_node) {
> +               if (async_tx_test_ack(&desc->txd)) {
> +                       list_del(&desc->desc_node);
> +                       ret = desc;
> +                       break;
> +               }
> +               dev_dbg(&atchan->chan_common.dev, "desc %p not ACKed\n", desc);
> +               i++;
> +       }
> +       spin_unlock_bh(&atchan->lock);
> +
> +       dev_vdbg(&atchan->chan_common.dev, "scanned %u descriptors on freelist\n", i);
> +
> +       /* no more descriptor available in initial pool : create some more */
> +       if (!ret) {
> +               for (i = 0; i < INIT_NR_DESCS_PER_CHANNEL; i++) {
> +                       desc = atc_alloc_descriptor(&atchan->chan_common, GFP_KERNEL);
> +                       if (!desc)
> +                               break;
> +                       /* return first descripor, queue others in free_list */

sp: descriptor

> +                       if (i)
> +                               list_add_tail(&desc->desc_node, &tmp_list);
> +                       else
> +                               ret = desc;
> +               }
> +
> +               spin_lock_bh(&atchan->lock);
> +               atchan->descs_allocated += i;
> +               list_splice(&tmp_list, &atchan->free_list);
> +               spin_unlock_bh(&atchan->lock);
> +       }
> +
> +       return ret;
> +}
> +
> +/**
> + * atc_desc_put - move a descriptor, including any children, to the free list
> + * @atchan: channel we work on
> + * @desc: descriptor, at the head of a chain, to move to free list
> + */
> +static void atc_desc_put(struct at_dma_chan *atchan, struct at_desc *desc)
> +{
> +       if (desc) {
> +               struct at_desc *child;
> +
> +               spin_lock_bh(&atchan->lock);
> +               list_for_each_entry(child, &desc->txd.tx_list, desc_node)
> +                       dev_vdbg(&atchan->chan_common.dev,
> +                                       "moving child desc %p to freelist\n",
> +                                       child);
> +               list_splice_init(&desc->txd.tx_list, &atchan->free_list);
> +               dev_vdbg(&atchan->chan_common.dev, "moving desc %p to freelist\n", desc);
> +               list_add(&desc->desc_node, &atchan->free_list);
> +               spin_unlock_bh(&atchan->lock);
> +       }
> +}
> +
> +/**
> + * atc_assign_cookie - compute and assign new cookie
> + * @atchan: channel we work on
> + * @desc: descriptor to asign cookie for
> + *
> + * Called with atchan->lock held and bh disabled
> + */
> +static dma_cookie_t
> +atc_assign_cookie(struct at_dma_chan *atchan, struct at_desc *desc)
> +{
> +       dma_cookie_t cookie = atchan->chan_common.cookie;
> +
> +       if (++cookie < 0)
> +               cookie = 1;
> +
> +       atchan->chan_common.cookie = cookie;
> +       desc->txd.cookie = cookie;
> +
> +       return cookie;
> +}
> +
> +/**
> + * atc_dostart - starts the DMA engine for real
> + * @atchan: the channel we want to start
> + * @first: first descriptor in the list we want to begin with
> + *
> + * Called with atchan->lock held and bh disabled
> + */
> +static void atc_dostart(struct at_dma_chan *atchan, struct at_desc *first)
> +{
> +       struct at_dma   *atdma = to_at_dma(atchan->chan_common.device);
> +
> +       /* ASSERT:  channel is idle */
> +       if (atc_chan_is_enabled(atchan)) {
> +               dev_err(&atchan->chan_common.dev,
> +                       "BUG: Attempted to start non-idle channel\n");
> +               dev_err(&atchan->chan_common.dev,
> +                       "  channel: s0x%x d0x%x ctrl0x%x:0x%x l0x%x\n",
> +                       channel_readl(atchan, SADDR),
> +                       channel_readl(atchan, DADDR),
> +                       channel_readl(atchan, CTRLA),
> +                       channel_readl(atchan, CTRLB),
> +                       channel_readl(atchan, DSCR));
> +
> +               /* The tasklet will hopefully advance the queue... */
> +               return;
> +       }
> +
> +       vdbg_dump_regs(atchan);
> +
> +       /* clear any pending interrupt */
> +       while (dma_readl(atdma, EBCISR))
> +               cpu_relax();
> +
> +       channel_writel(atchan, SADDR, 0);
> +       channel_writel(atchan, DADDR, 0);
> +       channel_writel(atchan, CTRLA, 0);
> +       channel_writel(atchan, CTRLB, 0);
> +       channel_writel(atchan, DSCR, first->txd.phys);
> +       dma_writel(atdma, CHER, atchan->mask);
> +
> +       vdbg_dump_regs(atchan);
> +}
> +
> +/**
> + * atc_chain_complete - finish work for one transaction chain
> + * @atchan: channel we work on
> + * @desc: descriptor at the head of the chain we want do complete
> + *
> + * Called with atchan->lock held and bh disabled */
> +static void
> +atc_chain_complete(struct at_dma_chan *atchan, struct at_desc *desc)
> +{
> +       dma_async_tx_callback           callback;
> +       void                            *param;
> +       struct dma_async_tx_descriptor  *txd = &desc->txd;
> +
> +       dev_vdbg(&atchan->chan_common.dev, "descriptor %u complete\n", txd->cookie);
> +
> +       atchan->completed_cookie = txd->cookie;
> +       callback = txd->callback;
> +       param = txd->callback_param;
> +
> +       /* move children to free_list */
> +       list_splice_init(&txd->tx_list, &atchan->free_list);
> +       /* move myself to free_list */
> +       list_move(&desc->desc_node, &atchan->free_list);
> +
> +       /*
> +        * We use dma_unmap_page() regardless of how the buffers were
> +        * mapped before they were submitted...
> +        */
> +       if (!(txd->flags & DMA_COMPL_SKIP_DEST_UNMAP))
> +               dma_unmap_page(atchan->chan_common.dev.parent, desc->lli.daddr, desc->len,
> +                               DMA_FROM_DEVICE);
> +       if (!(txd->flags & DMA_COMPL_SKIP_SRC_UNMAP))
> +               dma_unmap_page(atchan->chan_common.dev.parent, desc->lli.saddr, desc->len,
> +                               DMA_TO_DEVICE);
> +
> +       /*
> +        * The API requires that no submissions are done from a
> +        * callback, so we don't need to drop the lock here
> +        */
> +       if (callback)
> +               callback(param);
> +}
> +
> +/**
> + * atc_complete_all - finish work for all transactions
> + * @atchan: channel to complete transactions for
> + *
> + * Eventually submit queued descriptors if any
> + *
> + * Assume channel is idle while calling this function
> + * Called with atchan->lock held and bh disabled
> + */
> +static void atc_complete_all(struct at_dma_chan *atchan)
> +{
> +       struct at_desc *desc, *_desc;
> +       LIST_HEAD(list);
> +
> +       dev_vdbg(&atchan->chan_common.dev, "complete all\n");
> +
> +       BUG_ON(atc_chan_is_enabled(atchan));
> +
> +       /*
> +        * Submit queued descriptors ASAP, i.e. before we go through
> +        * the completed ones.
> +        */
> +       if (!list_empty(&atchan->queue))
> +               atc_dostart(atchan, atc_first_queued(atchan));
> +       /* empty active_list now it is completed */
> +       list_splice_init(&atchan->active_list, &list);
> +       /* empty queue list by moving descriptors (if any) to active_list */
> +       list_splice_init(&atchan->queue, &atchan->active_list);
> +
> +       list_for_each_entry_safe(desc, _desc, &list, desc_node)
> +               atc_chain_complete(atchan, desc);
> +}
> +
> +/**
> + * atc_cleanup_descriptors - cleanup up finished descriptors in active_list
> + * @atchan: channel to be cleaned up
> + *
> + * Called with atchan->lock held and bh disabled
> + */
> +static void atc_cleanup_descriptors(struct at_dma_chan *atchan)
> +{
> +       struct at_desc  *desc, *_desc;
> +       struct at_desc  *child;
> +
> +       dev_vdbg(&atchan->chan_common.dev, "cleanup descriptors\n");
> +
> +       list_for_each_entry_safe(desc, _desc, &atchan->active_list, desc_node) {
> +               if (!(desc->lli.ctrla & ATC_DONE))
> +                       /* This one is currently in progress */
> +                       return;
> +
> +               list_for_each_entry(child, &desc->txd.tx_list, desc_node)
> +                       if (!(child->lli.ctrla & ATC_DONE))
> +                               /* Currently in progress */
> +                               return;
> +
> +               /*
> +                * No descriptors so far seem to be in progress, i.e.
> +                * this chain must be done.
> +                */
> +               atc_chain_complete(atchan, desc);
> +       }
> +}
> +
> +/**
> + * atc_advance_work - at the end of a transaction, move forward
> + * @atchan: channel where the transaction ended
> + *
> + * Called with atchan->lock held and bh disabled
> + */
> +static void atc_advance_work(struct at_dma_chan *atchan)
> +{
> +       dev_vdbg(&atchan->chan_common.dev, "advance_work\n");
> +
> +       if (   list_empty(&atchan->active_list)
> +           || list_is_singular(&atchan->active_list)) {
> +               atc_complete_all(atchan);
> +       } else {
> +               atc_chain_complete(atchan, atc_first_active(atchan));
> +               /* advance work */
> +               atc_dostart(atchan, atc_first_active(atchan));
> +       }
> +}
> +
> +
> +/**
> + * atc_handle_error - handle errors reported by DMA controller
> + * @atchan: channel where error occurs
> + *
> + * Called with atchan->lock held and bh disabled
> + */
> +static void atc_handle_error(struct at_dma_chan *atchan)
> +{
> +       struct at_desc *bad_desc;
> +       struct at_desc *child;
> +
> +       /*
> +        * The descriptor currently at the head of the active list is
> +        * broked. Since we don't have any way to report errors, we'll
> +        * just have to scream loudly and try to carry on.
> +        */
> +       bad_desc = atc_first_active(atchan);
> +       list_del_init(&bad_desc->desc_node);
> +
> +       /* As we are stopped, take advantage to push queued descriptors
> +        * in active_list */
> +       list_splice_init(&atchan->queue, atchan->active_list.prev);
> +
> +       /* Try to restart the controller */
> +       if (!list_empty(&atchan->active_list))
> +               atc_dostart(atchan, atc_first_active(atchan));
> +
> +       /*
> +        * KERN_CRITICAL may seem harsh, but since this only happens
> +        * when someone submits a bad physical address in a
> +        * descriptor, we should consider ourselves lucky that the
> +        * controller flagged an error instead of scribbling over
> +        * random memory locations.
> +        */
> +       dev_crit(&atchan->chan_common.dev,
> +                       "Bad descriptor submitted for DMA!\n");
> +       dev_crit(&atchan->chan_common.dev,
> +                       "  cookie: %d\n", bad_desc->txd.cookie);
> +       atc_dump_lli(atchan, &bad_desc->lli);
> +       list_for_each_entry(child, &bad_desc->txd.tx_list, desc_node)
> +               atc_dump_lli(atchan, &child->lli);
> +
> +       /* Pretend the descriptor completed successfully */
> +       atc_chain_complete(atchan, bad_desc);
> +}
> +
> +
> +/*--  IRQ & Tasklet  ---------------------------------------------------*/
> +
> +static void atc_tasklet(unsigned long data)
> +{
> +       struct at_dma_chan *atchan = (struct at_dma_chan *)data;
> +
> +       /* Channel cannot be enabled here */
> +       if (atc_chan_is_enabled(atchan)) {
> +               dev_err(&atchan->chan_common.dev,
> +                       "BUG: channel enabled in tasklet\n");
> +               return;
> +       }
> +
> +       spin_lock(&atchan->lock);
> +       if (test_and_clear_bit(0, &atchan->error_status))
> +               atc_handle_error(atchan);
> +       else
> +               atc_advance_work(atchan);
> +
> +       spin_unlock(&atchan->lock);
> +}
> +
> +static irqreturn_t at_dma_interrupt(int irq, void *dev_id)
> +{
> +       struct at_dma           *atdma = (struct at_dma *)dev_id;
> +       struct at_dma_chan      *atchan;
> +       int                     i;
> +       u32                     status, pending, imr;
> +       int                     ret = IRQ_NONE;
> +
> +       do {
> +               imr = dma_readl(atdma, EBCIMR);
> +               status = dma_readl(atdma, EBCISR);
> +               pending = status & imr;
> +
> +               if (!pending)
> +                       break;
> +
> +               dev_vdbg(atdma->dma_common.dev,
> +                       "interrupt: status = 0x%08x, 0x%08x, 0x%08x\n",
> +                        status, imr, pending);
> +
> +               for (i = 0; i < atdma->dma_common.chancnt; i++) {
> +                       atchan = &atdma->chan[i];
> +                       if (pending & (AT_DMA_CBTC(i) | AT_DMA_ERR(i))) {
> +                               if (pending & AT_DMA_ERR(i)) {
> +                                       /*
> +                                       spin_lock(atchan->lock);
> +                                       atchan->error_status = 1;
> +                                       spin_unlock(atchan->lock);

writing to an unsigned long should already be atomic, no?

> +                                       */
> +                                       /* Disable channel on AHB error */
> +                                       dma_writel(atdma, CHDR, atchan->mask);
> +                                       /* Give information to tasklet */
> +                                       set_bit(0, &atchan->error_status);
> +                               }
> +                               tasklet_schedule(&atchan->tasklet);
> +                               ret = IRQ_HANDLED;
> +                       }
> +               }
> +
> +       } while (pending);
> +
> +       return ret;
> +}
> +
> +
> +/*--  DMA Engine API  --------------------------------------------------*/
> +
> +/**
> + * atc_tx_submit - set the prepared descriptor(s) to be executed by the engine
> + * @desc: descriptor at the head of the transaction chain
> + *
> + * Queue chain if DMA engine is working already
> + *
> + * Cookie increment and adding to active_list or queue must be atomic
> + */
> +static dma_cookie_t atc_tx_submit(struct dma_async_tx_descriptor *tx)
> +{
> +       struct at_desc          *desc = txd_to_at_desc(tx);
> +       struct at_dma_chan      *atchan = to_at_dma_chan(tx->chan);
> +       dma_cookie_t            cookie;
> +
> +       spin_lock_bh(&atchan->lock);
> +       cookie = atc_assign_cookie(atchan, desc);
> +
> +       if (list_empty(&atchan->active_list)) {
> +               dev_vdbg(&tx->chan->dev, "tx_submit: started %u\n",
> +                               desc->txd.cookie);
> +               atc_dostart(atchan, desc);
> +               list_add_tail(&desc->desc_node, &atchan->active_list);
> +       } else {
> +               dev_vdbg(&tx->chan->dev, "tx_submit: queued %u\n",
> +                               desc->txd.cookie);
> +               list_add_tail(&desc->desc_node, &atchan->queue);
> +       }
> +
> +       spin_unlock_bh(&atchan->lock);
> +
> +       return cookie;
> +}
> +
> +/**
> + * atc_prep_dma_memcpy - prepare a memcpy operation
> + * @chan: the channel to prepare operation on
> + * @dest: operation virtual destination address
> + * @src: operation virtual source address
> + * @len: operation length
> + * @flags: tx descriptor status flags
> + */
> +static struct dma_async_tx_descriptor *
> +atc_prep_dma_memcpy(struct dma_chan *chan, dma_addr_t dest, dma_addr_t src,
> +               size_t len, unsigned long flags)
> +{
> +       struct at_dma_chan      *atchan = to_at_dma_chan(chan);
> +       struct at_desc          *desc = NULL;
> +       struct at_desc          *first = NULL;
> +       struct at_desc          *prev = NULL;
> +       size_t                  xfer_count;
> +       size_t                  offset;
> +       unsigned int            src_width;
> +       unsigned int            dst_width;
> +       u32                     ctrla;
> +       u32                     ctrlb;
> +
> +       dev_vdbg(&chan->dev, "prep_dma_memcpy: d0x%x s0x%x l0x%zx f0x%lx\n",
> +                       dest, src, len, flags);
> +
> +       if (unlikely(!len)) {
> +               dev_dbg(&chan->dev, "prep_dma_memcpy: length is zero!\n");
> +               return NULL;
> +       }
> +
> +       ctrla =   ATC_DEFAULT_CTRLA;
> +       ctrlb =   ATC_DEFAULT_CTRLB
> +               | ATC_SRC_ADDR_MODE_INCR
> +               | ATC_DST_ADDR_MODE_INCR
> +               | ATC_FC_MEM2MEM;
> +
> +       /*
> +        * We can be a lot more clever here, but this should take care
> +        * of the most common optimization.
> +        */
> +       if (!((src | dest  | len) & 3)) {
> +               ctrla |= ATC_SRC_WIDTH_WORD | ATC_DST_WIDTH_WORD;
> +               src_width = dst_width = 2;
> +       } else if (!((src | dest | len) & 1)) {
> +               ctrla |= ATC_SRC_WIDTH_HALFWORD | ATC_DST_WIDTH_HALFWORD;
> +               src_width = dst_width = 1;
> +       } else {
> +               ctrla |= ATC_SRC_WIDTH_BYTE | ATC_DST_WIDTH_BYTE;
> +               src_width = dst_width = 0;
> +       }
> +
> +       for (offset = 0; offset < len; offset += xfer_count << src_width) {
> +               xfer_count = min_t(size_t, (len - offset) >> src_width,
> +                               ATC_BTSIZE_MAX);
> +
> +               desc = atc_desc_get(atchan);
> +               if (!desc)
> +                       goto err_desc_get;
> +
> +               desc->lli.saddr = src + offset;
> +               desc->lli.daddr = dest + offset;
> +               desc->lli.ctrla = ctrla | xfer_count;
> +               desc->lli.ctrlb = ctrlb;
> +
> +               desc->txd.cookie = 0;
> +               async_tx_ack(&desc->txd);
> +
> +               if (!first) {
> +                       first = desc;
> +               } else {
> +                       /* inform the HW lli about chaining */
> +                       prev->lli.dscr = desc->txd.phys;
> +                       /* insert the link descriptor to the LD ring */
> +                       list_add_tail(&desc->desc_node,
> +                                       &first->txd.tx_list);
> +               }
> +               prev = desc;
> +       }
> +
> +       /* First descriptor of the chain embedds additional information */
> +       first->txd.flags = flags; /* client is in control of this ack */
> +       first->txd.cookie = -EBUSY;
> +       first->len = len;
> +
> +       /* set end-of-link to the last link descriptor of list*/
> +       set_desc_eol(desc);
> +
> +       return &first->txd;
> +
> +err_desc_get:
> +       atc_desc_put(atchan, first);
> +       return NULL;
> +}
> +
> +/**
> + * atc_is_tx_complete - poll for transaction completion
> + * @chan: DMA channel
> + * @cookie: transaction identifier to check status of
> + * @done: if not %NULL, updated with last completed transaction
> + * @used: if not %NULL, updated with last used transaction
> + *
> + * If @done and @used are passed in, upon return they reflect the driver
> + * internal state and can be used with dma_async_is_complete() to check
> + * the status of multiple cookies without re-checking hardware state.
> + */
> +static enum dma_status
> +atc_is_tx_complete(struct dma_chan *chan,
> +               dma_cookie_t cookie,
> +               dma_cookie_t *done, dma_cookie_t *used)
> +{
> +       struct at_dma_chan      *atchan = to_at_dma_chan(chan);
> +       dma_cookie_t            last_used;
> +       dma_cookie_t            last_complete;
> +       enum dma_status         ret;
> +
> +       dev_vdbg(&chan->dev, "is_tx_complete: %d (d%d, u%d)\n",
> +                       cookie, done ? *done : 0, used ? *used : 0);
> +
> +       spin_lock_bh(atchan->lock);
> +
> +       last_complete = atchan->completed_cookie;
> +       last_used = chan->cookie;
> +
> +       ret = dma_async_is_complete(cookie, last_complete, last_used);
> +       if (ret != DMA_SUCCESS) {
> +               atc_cleanup_descriptors(atchan);
> +
> +               last_complete = atchan->completed_cookie;
> +               last_used = chan->cookie;
> +
> +               ret = dma_async_is_complete(cookie, last_complete, last_used);
> +       }
> +
> +       spin_unlock_bh(atchan->lock);
> +
> +       if (done)
> +               *done = last_complete;
> +       if (used)
> +               *used = last_used;
> +
> +       return ret;
> +}
> +
> +/**
> + * atc_issue_pending - try to finish work
> + * @chan: target DMA channel
> + */
> +static void atc_issue_pending(struct dma_chan *chan)
> +{
> +       struct at_dma_chan      *atchan = to_at_dma_chan(chan);
> +
> +       dev_vdbg(&chan->dev, "issue_pending\n");
> +
> +       if (!atc_chan_is_enabled(atchan)) {
> +               spin_lock_bh(&atchan->lock);
> +               atc_advance_work(atchan);
> +               spin_unlock_bh(&atchan->lock);
> +       }
> +}
> +
> +/**
> + * atc_alloc_chan_resources - allocate resources for DMA channel
> + * @chan: allocate descriptor resources for this channel
> + * @client: current client requesting the channel be ready for requests
> + *
> + * return - the number of allocated descriptors
> + */
> +static int atc_alloc_chan_resources(struct dma_chan *chan,
> +                                   struct dma_client *client)
> +{
> +       struct at_dma_chan      *atchan = to_at_dma_chan(chan);
> +       struct at_dma           *atdma = to_at_dma(chan->device);
> +       struct at_desc          *desc;
> +       int                     i;
> +       LIST_HEAD(tmp_list);
> +
> +       dev_vdbg(&chan->dev, "alloc_chan_resources\n");
> +
> +       /* ASSERT:  channel is idle */
> +       if (atc_chan_is_enabled(atchan)) {
> +               dev_dbg(&chan->dev, "DMA channel not idle ?\n");
> +               return -EIO;
> +       }
> +
> +       /* have we already been set up? */
> +       if (!list_empty(&atchan->free_list))
> +               return atchan->descs_allocated;
> +
> +       /* Allocate initial pool of descriptors */
> +       for (i = 0; i < INIT_NR_DESCS_PER_CHANNEL; i++) {
> +               desc = atc_alloc_descriptor(chan, GFP_KERNEL);
> +               if (!desc) {
> +                       dev_err(atdma->dma_common.dev,
> +                               "Only %d initial descriptors\n", i);
> +                       break;
> +               }
> +               list_add_tail(&desc->desc_node, &tmp_list);
> +       }
> +
> +       spin_lock_bh(&atchan->lock);
> +       atchan->descs_allocated = i;
> +       list_splice(&tmp_list, &atchan->free_list);
> +       atchan->completed_cookie = chan->cookie = 1;
> +       spin_unlock_bh(&atchan->lock);
> +
> +       /* channel parameters */
> +       channel_writel(atchan, CFG, ATC_DEFAULT_CFG);
> +
> +       tasklet_init(&atchan->tasklet, atc_tasklet, (unsigned long)atchan);

This routine may be called while the channel is already active,
potentially causing tasklet_init() to be called while a tasklet is
pending.  Can this move to at_dma_probe()?

> +       /* clear any pending interrupt */
> +       while (dma_readl(atdma, EBCISR))
> +               cpu_relax();
> +       atc_enable_irq(atchan);

ditto.

> +
> +       dev_dbg(&chan->dev,
> +               "alloc_chan_resources: allocated %d descriptors\n",
> +               atchan->descs_allocated);
> +
> +       return atchan->descs_allocated;
> +}
> +
> +/**
> + * atc_free_chan_resources - free all channel resources
> + * @chan: DMA channel
> + */
> +static void atc_free_chan_resources(struct dma_chan *chan)
> +{
> +       struct at_dma_chan      *atchan = to_at_dma_chan(chan);
> +       struct at_dma           *atdma = to_at_dma(chan->device);
> +       struct at_desc          *desc, *_desc;
> +       LIST_HEAD(list);
> +
> +       dev_dbg(&chan->dev, "free_chan_resources: (descs allocated=%u)\n",
> +                       atchan->descs_allocated);
> +
> +       /* ASSERT:  channel is idle */
> +       BUG_ON(!list_empty(&atchan->active_list));
> +       BUG_ON(!list_empty(&atchan->queue));
> +       BUG_ON(atc_chan_is_enabled(atchan));
> +
> +       /* Disable interrupts */
> +       atc_disable_irq(atchan);
> +       tasklet_disable(&atchan->tasklet);
> +
> +       list_for_each_entry_safe(desc, _desc, &atchan->free_list, desc_node) {
> +               dev_vdbg(&chan->dev, "  freeing descriptor %p\n", desc);
> +               list_del(&desc->desc_node);
> +               /* free link descriptor */
> +               dma_pool_free(atdma->dma_desc_pool, desc, desc->txd.phys);
> +       }
> +       list_splice_init(&atchan->free_list, &list);
> +       atchan->descs_allocated = 0;
> +
> +       dev_vdbg(&chan->dev, "free_chan_resources: done\n");
> +}
> +
> +
> +/*--  Module Management  -----------------------------------------------*/
> +
> +/**
> + * at_dma_off - disable DMA controller
> + * @atdma: the Atmel HDAMC device
> + */
> +static void at_dma_off(struct at_dma *atdma)
> +{
> +       dma_writel(atdma, EN, 0);
> +
> +       /* disable all interrupts */
> +       dma_writel(atdma, EBCIDR, -1L);
> +
> +       /* confirm that all channels are disabled */
> +       while (dma_readl(atdma, CHSR) & atdma->all_chan_mask)
> +               cpu_relax();
> +}
> +
> +static int __init at_dma_probe(struct platform_device *pdev)
> +{
> +       struct at_dma_platform_data *pdata;
> +       struct resource         *io;
> +       struct at_dma           *atdma;
> +       size_t                  size;
> +       int                     irq;
> +       int                     err;
> +       int                     i;
> +
> +       /* get DMA Controller parameters from platform */
> +       pdata = pdev->dev.platform_data;
> +       if (!pdata || pdata->nr_channels > AT_DMA_MAX_NR_CHANNELS)
> +               return -EINVAL;
> +
> +       io = platform_get_resource(pdev, IORESOURCE_MEM, 0);
> +       if (!io)
> +               return -EINVAL;
> +
> +       irq = platform_get_irq(pdev, 0);
> +       if (irq < 0)
> +               return irq;
> +
> +       size = sizeof(struct at_dma);
> +       size += pdata->nr_channels * sizeof(struct at_dma_chan);
> +       atdma = kzalloc(size, GFP_KERNEL);
> +       if (!atdma)
> +               return -ENOMEM;
> +
> +       /* discover transaction capabilites from the platform data */
> +       atdma->dma_common.cap_mask = pdata->cap_mask;
> +       atdma->all_chan_mask = (1 << pdata->nr_channels) - 1;
> +
> +       size = io->end - io->start + 1;
> +       if (!request_mem_region(io->start, size, pdev->dev.driver->name)) {
> +               err = -EBUSY;
> +               goto err_kfree;
> +       }
> +
> +       atdma->regs = ioremap(io->start, size);
> +       if (!atdma->regs) {
> +               err = -ENOMEM;
> +               goto err_release_r;
> +       }
> +
> +       atdma->clk = clk_get(&pdev->dev, "dma_clk");
> +       if (IS_ERR(atdma->clk)) {
> +               err = PTR_ERR(atdma->clk);
> +               goto err_clk;
> +       }
> +       clk_enable(atdma->clk);
> +
> +       /* force dma off, just in case */
> +       at_dma_off(atdma);
> +
> +       err = request_irq(irq, at_dma_interrupt, 0, "at_hdmac", atdma);
> +       if (err)
> +               goto err_irq;
> +
> +       platform_set_drvdata(pdev, atdma);
> +
> +       /* creates a pool of consistent memory blocks for hardware descriptors */
> +       atdma->dma_desc_pool = dma_pool_create("at_hdmac_desc_pool",
> +                       &pdev->dev, sizeof(struct at_desc),
> +                       4 /* word alignment */, 0);
> +       if (!atdma->dma_desc_pool) {
> +               dev_err(&pdev->dev, "No memory for descriptors dma pool\n");
> +               err = -ENOMEM;
> +               goto err_pool_create;
> +       }
> +
> +       /* initialize channels related values */
> +       INIT_LIST_HEAD(&atdma->dma_common.channels);
> +       for (i = 0; i < pdata->nr_channels; i++, atdma->dma_common.chancnt++) {
> +               struct at_dma_chan      *atchan = &atdma->chan[i];
> +
> +               atchan->chan_common.device = &atdma->dma_common;
> +               atchan->chan_common.cookie = atchan->completed_cookie = 1;
> +               atchan->chan_common.chan_id = i;
> +               list_add_tail(&atchan->chan_common.device_node, &atdma->dma_common.channels);
> +
> +               atchan->ch_regs = atdma->regs + ch_regs(i);
> +               spin_lock_init(&atchan->lock);
> +               atchan->mask = 1 << i;
> +
> +               INIT_LIST_HEAD(&atchan->active_list);
> +               INIT_LIST_HEAD(&atchan->queue);
> +               INIT_LIST_HEAD(&atchan->free_list);
> +       }
> +
> +       /* set base routines */
> +       atdma->dma_common.device_alloc_chan_resources = atc_alloc_chan_resources;
> +       atdma->dma_common.device_free_chan_resources = atc_free_chan_resources;
> +       atdma->dma_common.device_is_tx_complete = atc_is_tx_complete;
> +       atdma->dma_common.device_issue_pending = atc_issue_pending;
> +       atdma->dma_common.dev = &pdev->dev;
> +
> +       /* set prep routines based on capability */
> +       if (dma_has_cap(DMA_MEMCPY, atdma->dma_common.cap_mask))
> +               atdma->dma_common.device_prep_dma_memcpy = atc_prep_dma_memcpy;
> +
> +       dma_writel(atdma, EN, AT_DMA_ENABLE);
> +
> +       dev_info(&pdev->dev, "Atmel AHB DMA Controller ( %s%s), %d channels\n",
> +         dma_has_cap(DMA_MEMCPY, atdma->dma_common.cap_mask) ? "cpy " : "",
> +         dma_has_cap(DMA_SLAVE, atdma->dma_common.cap_mask)  ? "slave " : "",
> +         atdma->dma_common.chancnt);
> +
> +       dma_async_device_register(&atdma->dma_common);
> +
> +       return 0;
> +
> +err_pool_create:
> +       platform_set_drvdata(pdev, NULL);
> +       free_irq(platform_get_irq(pdev, 0), atdma);
> +err_irq:
> +       clk_disable(atdma->clk);
> +       clk_put(atdma->clk);
> +err_clk:
> +       iounmap(atdma->regs);
> +       atdma->regs = NULL;
> +err_release_r:
> +       release_mem_region(io->start, size);
> +err_kfree:
> +       kfree(atdma);
> +       return err;
> +}
> +
> +static int __exit at_dma_remove(struct platform_device *pdev)
> +{
> +       struct at_dma           *atdma = platform_get_drvdata(pdev);
> +       struct dma_chan         *chan, *_chan;
> +       struct resource         *io;
> +
> +       at_dma_off(atdma);
> +       dma_async_device_unregister(&atdma->dma_common);
> +
> +       dma_pool_destroy(atdma->dma_desc_pool);
> +       platform_set_drvdata(pdev, NULL);
> +       free_irq(platform_get_irq(pdev, 0), atdma);
> +
> +       list_for_each_entry_safe(chan, _chan, &atdma->dma_common.channels,
> +                       device_node) {
> +               tasklet_kill(&(to_at_dma_chan(chan)->tasklet));
> +               list_del(&chan->device_node);
> +       }
> +
> +       clk_disable(atdma->clk);
> +       clk_put(atdma->clk);
> +
> +       iounmap(atdma->regs);
> +       atdma->regs = NULL;
> +
> +       io = platform_get_resource(pdev, IORESOURCE_MEM, 0);
> +       release_mem_region(io->start, io->end - io->start + 1);
> +
> +       kfree(atdma);
> +
> +       return 0;
> +}
> +
> +static void at_dma_shutdown(struct platform_device *pdev)
> +{
> +       struct at_dma   *atdma = platform_get_drvdata(pdev);
> +
> +       at_dma_off(platform_get_drvdata(pdev));
> +       clk_disable(atdma->clk);
> +}
> +
> +static int at_dma_suspend_late(struct platform_device *pdev, pm_message_t mesg)
> +{
> +       struct at_dma   *atdma = platform_get_drvdata(pdev);
> +
> +       at_dma_off(platform_get_drvdata(pdev));
> +       clk_disable(atdma->clk);
> +       return 0;
> +}
> +
> +static int at_dma_resume_early(struct platform_device *pdev)
> +{
> +       struct at_dma   *atdma = platform_get_drvdata(pdev);
> +
> +       clk_enable(atdma->clk);
> +       dma_writel(atdma, EN, AT_DMA_ENABLE);
> +       return 0;
> +
> +}
> +
> +static struct platform_driver at_dma_driver = {
> +       .remove         = __exit_p(at_dma_remove),
> +       .shutdown       = at_dma_shutdown,
> +       .suspend_late   = at_dma_suspend_late,
> +       .resume_early   = at_dma_resume_early,
> +       .driver = {
> +               .name   = "at_hdmac",
> +       },
> +};
> +
> +static int __init at_dma_init(void)
> +{
> +       return platform_driver_probe(&at_dma_driver, at_dma_probe);
> +}
> +module_init(at_dma_init);
> +
> +static void __exit at_dma_exit(void)
> +{
> +       platform_driver_unregister(&at_dma_driver);
> +}
> +module_exit(at_dma_exit);
> +
> +MODULE_DESCRIPTION("Atmel AHB DMA Controller driver");
> +MODULE_AUTHOR("Nicolas Ferre <nicolas.ferre@...el.com>");
> +MODULE_LICENSE("GPL");
> +MODULE_ALIAS("platform:at_hdmac");
> diff --git a/drivers/dma/at_hdmac_regs.h b/drivers/dma/at_hdmac_regs.h
> new file mode 100644
> index 0000000..294f625
> --- /dev/null
> +++ b/drivers/dma/at_hdmac_regs.h
> @@ -0,0 +1,377 @@
> +/*
> + * Header file for the Atmel AHB DMA Controller driver
> + *
> + * Copyright (C) 2008 Atmel Corporation
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by
> + * the Free Software Foundation; either version 2 of the License, or
> + * (at your option) any later version.
> + */
> +#ifndef AT_HDMAC_REGS_H
> +#define        AT_HDMAC_REGS_H
> +
> +#include <linux/at_hdmac.h>
> +
> +#define        AT_DMA_MAX_NR_CHANNELS  8
> +
> +
> +#define        AT_DMA_GCFG     0x00    /* Global Configuration Register */
> +#define                AT_DMA_IF_BIGEND(i)     (0x1 << (i))    /* AHB-Lite Interface i in Big-endian mode */
> +#define                AT_DMA_ARB_CFG  (0x1 << 4)      /* Arbiter mode. */
> +#define                        AT_DMA_ARB_CFG_FIXED            (0x0 << 4)
> +#define                        AT_DMA_ARB_CFG_ROUND_ROBIN      (0x1 << 4)
> +
> +#define        AT_DMA_EN       0x04    /* Controller Enable Register */
> +#define                AT_DMA_ENABLE   (0x1 << 0)
> +
> +#define        AT_DMA_SREQ     0x08    /* Software Single Request Register */
> +#define                AT_DMA_SSREQ(x) (0x1 << ((x) << 1))             /* Request a source single transfer on channel x */
> +#define                AT_DMA_DSREQ(x) (0x1 << (1 + ((x) << 1)))       /* Request a destination single transfer on channel x */
> +
> +#define        AT_DMA_CREQ     0x0C    /* Software Chunk Transfer Request Register */
> +#define                AT_DMA_SCREQ(x) (0x1 << ((x) << 1))             /* Request a source chunk transfer on channel x */
> +#define                AT_DMA_DCREQ(x) (0x1 << (1 + ((x) << 1)))       /* Request a destination chunk transfer on channel x */
> +
> +#define        AT_DMA_LAST     0x10    /* Software Last Transfer Flag Register */
> +#define                AT_DMA_SLAST(x) (0x1 << ((x) << 1))             /* This src rq is last tx of buffer on channel x */
> +#define                AT_DMA_DLAST(x) (0x1 << (1 + ((x) << 1)))       /* This dst rq is last tx of buffer on channel x */
> +
> +#define        AT_DMA_SYNC     0x14    /* Request Synchronization Register */
> +#define                AT_DMA_SYR(h)   (0x1 << (h))                    /* Synchronize handshake line h */
> +
> +/* Error, Chained Buffer transfer completed and Buffer transfer completed Interrupt registers */
> +#define        AT_DMA_EBCIER   0x18    /* Enable register */
> +#define        AT_DMA_EBCIDR   0x1C    /* Disable register */
> +#define        AT_DMA_EBCIMR   0x20    /* Mask Register */
> +#define        AT_DMA_EBCISR   0x24    /* Status Register */
> +#define                AT_DMA_CBTC_OFFSET      8
> +#define                AT_DMA_ERR_OFFSET       16
> +#define                AT_DMA_BTC(x)   (0x1 << (x))
> +#define                AT_DMA_CBTC(x)  (0x1 << (AT_DMA_CBTC_OFFSET + (x)))
> +#define                AT_DMA_ERR(x)   (0x1 << (AT_DMA_ERR_OFFSET + (x)))
> +
> +#define        AT_DMA_CHER     0x28    /* Channel Handler Enable Register */
> +#define                AT_DMA_ENA(x)   (0x1 << (x))
> +#define                AT_DMA_SUSP(x)  (0x1 << ( 8 + (x)))
> +#define                AT_DMA_KEEP(x)  (0x1 << (24 + (x)))
> +
> +#define        AT_DMA_CHDR     0x2C    /* Channel Handler Disable Register */
> +#define                AT_DMA_DIS(x)   (0x1 << (x))
> +#define                AT_DMA_RES(x)   (0x1 << ( 8 + (x)))
> +
> +#define        AT_DMA_CHSR     0x30    /* Channel Handler Status Register */
> +#define                AT_DMA_EMPT(x)  (0x1 << (16 + (x)))
> +#define                AT_DMA_STAL(x)  (0x1 << (24 + (x)))
> +
> +
> +#define        AT_DMA_CH_REGS_BASE     0x3C    /* Channel registers base address */
> +#define        ch_regs(x)      (AT_DMA_CH_REGS_BASE + (x) * 0x28) /* Channel x base addr */
> +
> +/* Hardware register offset for each channel */
> +#define        ATC_SADDR_OFFSET        0x00    /* Source Address Register */
> +#define        ATC_DADDR_OFFSET        0x04    /* Destination Address Register */
> +#define        ATC_DSCR_OFFSET         0x08    /* Descriptor Address Register */
> +#define        ATC_CTRLA_OFFSET        0x0C    /* Control A Register */
> +#define        ATC_CTRLB_OFFSET        0x10    /* Control B Register */
> +#define        ATC_CFG_OFFSET          0x14    /* Configuration Register */
> +#define        ATC_SPIP_OFFSET         0x18    /* Src PIP Configuration Register */
> +#define        ATC_DPIP_OFFSET         0x1C    /* Dst PIP Configuration Register */
> +
> +
> +/* Bitfield definitions */
> +
> +/* Bitfields in DSCR */
> +#define        ATC_DSCR_IF(i)          (0x3 & (i))     /* Dsc feched via AHB-Lite Interface i */
> +
> +/* Bitfields in CTRLA */
> +#define        ATC_BTSIZE_MAX          0xFFFFUL        /* Maximum Buffer Transfer Size */
> +#define        ATC_BTSIZE(x)           (ATC_BTSIZE_MAX & (x)) /* Buffer Transfer Size */
> +#define        ATC_SCSIZE_MASK         (0x7 << 16)     /* Source Chunk Transfer Size */
> +#define                ATC_SCSIZE_1            (0x0 << 16)
> +#define                ATC_SCSIZE_4            (0x1 << 16)
> +#define                ATC_SCSIZE_8            (0x2 << 16)
> +#define                ATC_SCSIZE_16           (0x3 << 16)
> +#define                ATC_SCSIZE_32           (0x4 << 16)
> +#define                ATC_SCSIZE_64           (0x5 << 16)
> +#define                ATC_SCSIZE_128          (0x6 << 16)
> +#define                ATC_SCSIZE_256          (0x7 << 16)
> +#define        ATC_DCSIZE_MASK         (0x7 << 20)     /* Destination Chunk Transfer Size */
> +#define                ATC_DCSIZE_1            (0x0 << 20)
> +#define                ATC_DCSIZE_4            (0x1 << 20)
> +#define                ATC_DCSIZE_8            (0x2 << 20)
> +#define                ATC_DCSIZE_16           (0x3 << 20)
> +#define                ATC_DCSIZE_32           (0x4 << 20)
> +#define                ATC_DCSIZE_64           (0x5 << 20)
> +#define                ATC_DCSIZE_128          (0x6 << 20)
> +#define                ATC_DCSIZE_256          (0x7 << 20)
> +#define        ATC_SRC_WIDTH_MASK      (0x3 << 24)     /* Source Single Transfer Size */
> +#define                ATC_SRC_WIDTH_BYTE      (0x0 << 24)
> +#define                ATC_SRC_WIDTH_HALFWORD  (0x1 << 24)
> +#define                ATC_SRC_WIDTH_WORD      (0x2 << 24)
> +#define        ATC_DST_WIDTH_MASK      (0x3 << 28)     /* Destination Single Transfer Size */
> +#define                ATC_DST_WIDTH_BYTE      (0x0 << 28)
> +#define                ATC_DST_WIDTH_HALFWORD  (0x1 << 28)
> +#define                ATC_DST_WIDTH_WORD      (0x2 << 28)
> +#define        ATC_DONE                (0x1 << 31)     /* Tx Done (only written back in descriptor) */
> +
> +/* Bitfields in CTRLB */
> +#define        ATC_SIF(i)              (0x3 & (i))     /* Src tx done via AHB-Lite Interface i */
> +#define        ATC_DIF(i)              ((0x3 & (i)) <<  4)     /* Dst tx done via AHB-Lite Interface i */
> +#define        ATC_SRC_PIP             (0x1 <<  8)     /* Source Picture-in-Picture enabled */
> +#define        ATC_DST_PIP             (0x1 << 12)     /* Destination Picture-in-Picture enabled */
> +#define        ATC_SRC_DSCR_DIS        (0x1 << 16)     /* Src Descriptor fetch disable */
> +#define        ATC_DST_DSCR_DIS        (0x1 << 20)     /* Dst Descriptor fetch disable */
> +#define        ATC_FC_MASK             (0x7 << 21)     /* Choose Flow Controller */
> +#define                ATC_FC_MEM2MEM          (0x0 << 21)     /* Mem-to-Mem (DMA) */
> +#define                ATC_FC_MEM2PER          (0x1 << 21)     /* Mem-to-Periph (DMA) */
> +#define                ATC_FC_PER2MEM          (0x2 << 21)     /* Periph-to-Mem (DMA) */
> +#define                ATC_FC_PER2PER          (0x3 << 21)     /* Periph-to-Periph (DMA) */
> +#define                ATC_FC_PER2MEM_PER      (0x4 << 21)     /* Periph-to-Mem (Peripheral) */
> +#define                ATC_FC_MEM2PER_PER      (0x5 << 21)     /* Mem-to-Periph (Peripheral) */
> +#define                ATC_FC_PER2PER_PER      (0x6 << 21)     /* Periph-to-Periph (Src Peripheral) */
> +#define        ATC_SRC_ADDR_MODE_MASK  (0x3 << 24)
> +#define                ATC_SRC_ADDR_MODE_INCR  (0x0 << 24)     /* Incrementing Mode */
> +#define                ATC_SRC_ADDR_MODE_DECR  (0x1 << 24)     /* Decrementing Mode */
> +#define                ATC_SRC_ADDR_MODE_FIXED (0x2 << 24)     /* Fixed Mode */
> +#define        ATC_DST_ADDR_MODE_MASK  (0x3 << 28)
> +#define                ATC_DST_ADDR_MODE_INCR  (0x0 << 28)     /* Incrementing Mode */
> +#define                ATC_DST_ADDR_MODE_DECR  (0x1 << 28)     /* Decrementing Mode */
> +#define                ATC_DST_ADDR_MODE_FIXED (0x2 << 28)     /* Fixed Mode */
> +#define        ATC_IEN                 (0x1 << 30)     /* BTC interrupt enable (active low) */
> +#define        ATC_AUTO                (0x1 << 31)     /* Auto multiple buffer tx enable */
> +
> +/* Bitfields in CFG */
> +#define        ATC_SRC_PER(h)          (0xFU & (h))    /* Channel src rq associated with periph handshaking ifc h */
> +#define        ATC_DST_PER(h)          ((0xFU & (h)) <<  4)    /* Channel dst rq associated with periph handshaking ifc h */
> +#define        ATC_SRC_REP             (0x1 <<  8)     /* Source Replay Mod */
> +#define        ATC_SRC_H2SEL           (0x1 <<  9)     /* Source Handshaking Mod */
> +#define                ATC_SRC_H2SEL_SW        (0x0 <<  9)
> +#define                ATC_SRC_H2SEL_HW        (0x1 <<  9)
> +#define        ATC_DST_REP             (0x1 << 12)     /* Destination Replay Mod */
> +#define        ATC_DST_H2SEL           (0x1 << 13)     /* Destination Handshaking Mod */
> +#define                ATC_DST_H2SEL_SW        (0x0 << 13)
> +#define                ATC_DST_H2SEL_HW        (0x1 << 13)
> +#define        ATC_SOD                 (0x1 << 16)     /* Stop On Done */
> +#define        ATC_LOCK_IF             (0x1 << 20)     /* Interface Lock */
> +#define        ATC_LOCK_B              (0x1 << 21)     /* AHB Bus Lock */
> +#define        ATC_LOCK_IF_L           (0x1 << 22)     /* Master Interface Arbiter Lock */
> +#define                ATC_LOCK_IF_L_CHUNK     (0x0 << 22)
> +#define                ATC_LOCK_IF_L_BUFFER    (0x1 << 22)
> +#define        ATC_AHB_PROT_MASK       (0x7 << 24)     /* AHB Protection */
> +#define        ATC_FIFOCFG_MASK        (0x3 << 28)     /* FIFO Request Configuration */
> +#define                ATC_FIFOCFG_LARGESTBURST        (0x0 << 28)
> +#define                ATC_FIFOCFG_HALFFIFO            (0x1 << 28)
> +#define                ATC_FIFOCFG_ENOUGHSPACE         (0x2 << 28)
> +
> +/* Bitfields in SPIP */
> +#define        ATC_SPIP_HOLE(x)        (0xFFFFU & (x))
> +#define        ATC_SPIP_BOUNDARY(x)    ((0x3FF & (x)) << 16)
> +
> +/* Bitfields in DPIP */
> +#define        ATC_DPIP_HOLE(x)        (0xFFFFU & (x))
> +#define        ATC_DPIP_BOUNDARY(x)    ((0x3FF & (x)) << 16)
> +
> +
> +/*--  descriptors  -----------------------------------------------------*/
> +
> +/* LLI == Linked List Item; aka DMA buffer descriptor */
> +struct at_lli {
> +       /* values that are not changed by hardware */
> +       dma_addr_t      saddr;
> +       dma_addr_t      daddr;
> +       /* value that may get written back: */
> +       u32             ctrla;
> +       /* more values that are not changed by hardware */
> +       u32             ctrlb;
> +       dma_addr_t      dscr;   /* chain to next lli */
> +};
> +
> +/**
> + * struct at_desc - software descriptor
> + * @at_lli: hardware lli structure
> + * @txd: support for the async_tx api
> + * @desc_node: node on the channed descriptors list
> + * @len: total transaction bytecount
> + */
> +struct at_desc {
> +       /* FIRST values the hardware uses */
> +       struct at_lli                   lli;
> +
> +       /* THEN values for driver housekeeping */
> +       struct dma_async_tx_descriptor  txd;
> +       struct list_head                desc_node;
> +       size_t                          len;
> +};
> +
> +static inline struct at_desc *
> +txd_to_at_desc(struct dma_async_tx_descriptor *txd)
> +{
> +       return container_of(txd, struct at_desc, txd);
> +}
> +
> +
> +/*--  Channels  --------------------------------------------------------*/
> +
> +/**
> + * struct at_dma_chan - internal representation of an Atmel HDMAC channel
> + * @chan_common: common dmaengine channel object members
> + * @device: parent device
> + * @ch_regs: memory mapped register base
> + * @mask: channel index in a mask
> + * @error_status: transmit error status information from irq handler
> + *                to tasklet (use atomic operations)
> + * @tasklet: bottom half to finish transaction work
> + * @lock: serializes enqueue/dequeue operations to descriptors lists
> + * @completed_cookie: identifier for the most recently completed operation
> + * @active_list: list of descriptors dmaengine is being running on
> + * @queue: list of descriptors ready to be submitted to engine
> + * @free_list: list of descriptors usable by the channel
> + * @descs_allocated: records the actual size of the descriptor pool
> + */
> +struct at_dma_chan {
> +       struct dma_chan         chan_common;
> +       struct at_dma           *device;
> +       void __iomem            *ch_regs;
> +       u8                      mask;
> +       unsigned long           error_status;
> +       struct tasklet_struct   tasklet;
> +
> +       spinlock_t              lock;
> +
> +       /* these other elements are all protected by lock */
> +       dma_cookie_t            completed_cookie;
> +       struct list_head        active_list;
> +       struct list_head        queue;
> +       struct list_head        free_list;
> +       unsigned int            descs_allocated;
> +};
> +
> +#define        channel_readl(atchan, name) \
> +       __raw_readl((atchan)->ch_regs + ATC_##name##_OFFSET)
> +
> +#define        channel_writel(atchan, name, val) \
> +       __raw_writel((val), (atchan)->ch_regs + ATC_##name##_OFFSET)
> +
> +static inline struct at_dma_chan *to_at_dma_chan(struct dma_chan *dchan)
> +{
> +       return container_of(dchan, struct at_dma_chan, chan_common);
> +}
> +
> +
> +/*--  Controller  ------------------------------------------------------*/
> +
> +/**
> + * struct at_dma - internal representation of an Atmel HDMA Controller
> + * @chan_common: common dmaengine dma_device object members
> + * @ch_regs: memory mapped register base
> + * @clk: dma controller clock
> + * @all_chan_mask: all channels availlable in a mask
> + * @dma_desc_pool: base of DMA descriptor region (DMA address)
> + * @chan: channels table to store at_dma_chan structures
> + */
> +struct at_dma {
> +       struct dma_device       dma_common;
> +       void __iomem            *regs;
> +       struct clk              *clk;
> +
> +       u8                      all_chan_mask;
> +
> +       struct dma_pool         *dma_desc_pool;
> +       /* AT THE END channels table */
> +       struct at_dma_chan      chan[0];
> +};
> +
> +#define        dma_readl(atdma, name) \
> +       __raw_readl((atdma)->regs + AT_DMA_##name)
> +#define        dma_writel(atdma, name, val) \
> +       __raw_writel((val), (atdma)->regs + AT_DMA_##name)
> +
> +static inline struct at_dma *to_at_dma(struct dma_device *ddev)
> +{
> +       return container_of(ddev, struct at_dma, dma_common);
> +}
> +
> +
> +/*--  Helper functions  ------------------------------------------------*/
> +
> +#if defined(VERBOSE_DEBUG)
> +static void vdbg_dump_regs(struct at_dma_chan *atchan)
> +{
> +       struct at_dma   *atdma = to_at_dma(atchan->chan_common.device);
> +
> +       dev_err(&atchan->chan_common.dev,
> +               "  channel %d : imr = 0x%x, chsr = 0x%x\n",
> +               atchan->chan_common.chan_id,
> +               dma_readl(atdma, EBCIMR),
> +               dma_readl(atdma, CHSR));
> +
> +       dev_err(&atchan->chan_common.dev,
> +               "  channel: s0x%x d0x%x ctrl0x%x:0x%x l0x%x\n",
> +               channel_readl(atchan, SADDR),
> +               channel_readl(atchan, DADDR),
> +               channel_readl(atchan, CTRLA),
> +               channel_readl(atchan, CTRLB),
> +               channel_readl(atchan, DSCR));
> +}
> +#else
> +static void vdbg_dump_regs(struct at_dma_chan *atchan) {}
> +#endif
> +
> +static void atc_dump_lli(struct at_dma_chan *atchan, struct at_lli *lli)
> +{
> +       dev_printk(KERN_CRIT, &atchan->chan_common.dev,
> +                       "  desc: s0x%x d0x%x ctrl0x%x:0x%x l0x%x\n",
> +                       lli->saddr, lli->daddr,
> +                       lli->ctrla, lli->ctrlb, lli->dscr);
> +}
> +
> +
> +static void atc_setup_irq(struct at_dma_chan *atchan, int on)
> +{
> +       struct at_dma   *atdma = to_at_dma(atchan->chan_common.device);
> +       u32             ebci;
> +
> +       /* enable interrupts on buffer chain completion & error */
> +       ebci =    AT_DMA_CBTC(atchan->chan_common.chan_id)
> +               | AT_DMA_ERR(atchan->chan_common.chan_id);
> +       if (on)
> +               dma_writel(atdma, EBCIER, ebci);
> +       else
> +               dma_writel(atdma, EBCIDR, ebci);
> +}
> +
> +static inline void atc_enable_irq(struct at_dma_chan *atchan)
> +{
> +       atc_setup_irq(atchan, 1);
> +}
> +
> +static inline void atc_disable_irq(struct at_dma_chan *atchan)
> +{
> +       atc_setup_irq(atchan, 0);
> +}
> +
> +
> +/**
> + * atc_chan_is_enabled - test if given channel is enabled
> + * @atchan: channel we want to test status
> + */
> +static inline int atc_chan_is_enabled(struct at_dma_chan *atchan)
> +{
> +       struct at_dma   *atdma = to_at_dma(atchan->chan_common.device);
> +
> +       return !!(dma_readl(atdma, CHSR) & atchan->mask);
> +}
> +
> +
> +/**
> + * set_desc_eol - set end-of-link to descriptor so it will end transfer
> + * @desc: descriptor, signle or at the end of a chain, to end chain on
> + */
> +static void set_desc_eol(struct at_desc *desc)
> +{
> +       desc->lli.ctrlb |= ATC_SRC_DSCR_DIS | ATC_DST_DSCR_DIS;
> +       desc->lli.dscr = 0;
> +}
> +
> +#endif /* AT_HDMAC_REGS_H */
> diff --git a/include/linux/at_hdmac.h b/include/linux/at_hdmac.h
> new file mode 100644
> index 0000000..21a5554
> --- /dev/null
> +++ b/include/linux/at_hdmac.h
> @@ -0,0 +1,26 @@
> +/*
> + * Header file for the Atmel AHB DMA Controller driver
> + *
> + * Copyright (C) 2008 Atmel Corporation
> + *
> + * This program is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by
> + * the Free Software Foundation; either version 2 of the License, or
> + * (at your option) any later version.
> + */
> +#ifndef AT_HDMAC_H
> +#define AT_HDMAC_H
> +
> +#include <linux/dmaengine.h>
> +
> +/**
> + * struct at_dma_platform_data - Controller configuration parameters
> + * @nr_channels: Number of channels supported by hardware (max 8)
> + * @cap_mask: dma_capability flags supported by the platform
> + */
> +struct at_dma_platform_data {
> +       unsigned int    nr_channels;
> +       dma_cap_mask_t  cap_mask;
> +};
> +
> +#endif /* AT_HDMAC_H */
> --
> 1.5.3.7
> 
> 
> 
> 
> 

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ