lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <fa686aa41003261500k65470257j7069bc14bbe705a@mail.gmail.com>
Date:	Fri, 26 Mar 2010 16:00:21 -0600
From:	Grant Likely <grant.likely@...retlab.ca>
To:	"Steven J. Magnani" <steve@...idescorp.com>
Cc:	Dan Williams <dan.j.williams@...el.com>,
	linux-kernel@...r.kernel.org, microblaze-uclinux@...e.uq.edu.au,
	monstr@...str.eu
Subject: Re: [PATCH 1/4] dmaengine: xlldma core driver

On Wed, Mar 17, 2010 at 10:26 AM, Steven J. Magnani
<steve@...idescorp.com> wrote:
> Core DMA engine for Xilinx Multi-Port Memory Controller (MPMC) SDMA channels.
> Requires a bus driver to actually connect to a device.
>
> Signed-off-by: Steven J. Magnani <steve@...idescorp.com>
> ---
> diff -uprN a/drivers/dma/xlldma.h b/drivers/dma/xlldma.h
> --- a/drivers/dma/xlldma.h      1969-12-31 18:00:00.000000000 -0600
> +++ b/drivers/dma/xlldma.h      2010-03-17 11:11:16.000000000 -0500

Only put stuff that is actually used by multiple .c files into a .h
file.  Most of this stuff is only used by xlldma.c, so it should be
defined there.

> @@ -0,0 +1,259 @@
> +/*
> + * Xilinx Multi-Port Memory Controller (MPMC) DMA Engine support
> + *
> + * Copyright (C) 2010 Digital Design Corporation. All rights reserved.
> + *
> + * Based on fsldma code that is:
> + * Copyright (C) 2007 Freescale Semiconductor, Inc. All rights reserved.
> + *
> + * Authors:
> + *   Steve Magnani <steve@...idescorp.com>, Mar 2010
> + *
> + * fsldma authors:
> + *   Zhang Wei     <wei.zhang@...escale.com>, Jul 2007
> + *   Ebony Zhu     <ebony.zhu@...escale.com>, May 2007
> + *
> + * Description:
> + *   DMA engine driver for Xilinx MPMC Soft Direct Memory Access Controller.
> + *   The driver can be attached to "loopback" MPMC SDMA ports, i.e., those
> + *   whose LL_TX signals are connected to the LL_RX signals.
> + *
> + * This is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by
> + * the Free Software Foundation; either version 2 of the License, or
> + * (at your option) any later version.
> + *
> + */
> +
> +#ifndef __DMA_XLLDMA_H
> +#define __DMA_XLLDMA_H
> +
> +#include <linux/device.h>
> +#include <linux/dmapool.h>
> +#include <linux/dmaengine.h>
> +
> +/* MPMC SDMA registers definition */
> +
> +#define TX_NXTDESC_PTR      0x00            /* r */
> +#define TX_CURBUF_ADDR      0x01            /* r */
> +#define TX_CURBUF_LENGTH    0x02            /* r */
> +#define TX_CURDESC_PTR      0x03            /* rw */
> +#define TX_TAILDESC_PTR     0x04            /* rw */
> +#define TX_CHNL_CTRL        0x05            /* rw */
> +/*
> + 0:7      24:31       IRQTimeout
> + 8:15     16:23       IRQCount
> + 16:20    11:15       Reserved
> + 21       10          0
> + 22       9           UseIntOnEnd
> + 23       8           LdIRQCnt
> + 24       7           IRQEn
> + 25:28    3:6         Reserved
> + 29       2           IrqErrEn
> + 30       1           IrqDlyEn
> + 31       0           IrqCoalEn
> +*/
> +#define CHNL_CTRL_IRQ_IOE       (1 << 9)
> +#define CHNL_CTRL_IRQ_LD_CNT   (1 << 8)
> +#define CHNL_CTRL_IRQ_EN        (1 << 7)
> +#define CHNL_CTRL_IRQ_ERR_EN    (1 << 2)
> +#define CHNL_CTRL_IRQ_DLY_EN    (1 << 1)
> +#define CHNL_CTRL_IRQ_COAL_EN   (1 << 0)
> +#define TX_IRQ_REG          0x06            /* rw */
> +/*
> +  0:7      24:31       DltTmrValue
> + 8:15     16:23       ClscCntrValue
> + 16:17    14:15       Reserved
> + 18:21    10:13       ClscCnt
> + 22:23    8:9         DlyCnt
> + 24:28    3::7        Reserved
> + 29       2           ErrIrq
> + 30       1           DlyIrq
> + 31       0           CoalIrq
> + */
> +#define TX_CHNL_STS         0x07            /* r */
> +/*
> +   0:9      22:31   Reserved
> + 10       21      TailPErr
> + 11       20      CmpErr
> + 12       19      AddrErr
> + 13       18      NxtPErr
> + 14       17      CurPErr
> + 15       16      BsyWr
> + 16:23    8:15    Reserved
> + 24       7       Error
> + 25       6       IOE
> + 26       5       SOE
> + 27       4       Cmplt
> + 28       3       SOP
> + 29       2       EOP
> + 30       1       EngBusy
> + 31       0       Reserved
> +*/
> +
> +#define RX_NXTDESC_PTR      0x08            /* r */
> +#define RX_CURBUF_ADDR      0x09            /* r */
> +#define RX_CURBUF_LENGTH    0x0a            /* r */
> +#define RX_CURDESC_PTR      0x0b            /* rw */
> +#define RX_TAILDESC_PTR     0x0c            /* rw */
> +#define RX_CHNL_CTRL        0x0d            /* rw */
> +/*
> + 0:7      24:31       IRQTimeout
> + 8:15     16:23       IRQCount
> + 16:20    11:15       Reserved
> + 21       10          0
> + 22       9           UseIntOnEnd
> + 23       8           LdIRQCnt
> + 24       7           IRQEn
> + 25:28    3:6         Reserved
> + 29       2           IrqErrEn
> + 30       1           IrqDlyEn
> + 31       0           IrqCoalEn
> + */
> +#define RX_IRQ_REG          0x0e            /* rw */
> +#define IRQ_COAL        (1 << 0)
> +#define IRQ_DLY         (1 << 1)
> +#define IRQ_ERR         (1 << 2)
> +#define IRQ_DMAERR      (1 << 7)            /* this is not documented ??? */
> +/*
> + 0:7      24:31       DltTmrValue
> + 8:15     16:23       ClscCntrValue
> + 16:17    14:15       Reserved
> + 18:21    10:13       ClscCnt
> + 22:23    8:9         DlyCnt
> + 24:28    3::7        Reserved
> +*/
> +#define RX_CHNL_STS         0x0f        /* r */
> +#define CHNL_STS_ENGBUSY    (1 << 1)
> +#define CHNL_STS_EOP        (1 << 2)
> +#define CHNL_STS_SOP        (1 << 3)
> +#define CHNL_STS_CMPLT      (1 << 4)
> +#define CHNL_STS_SOE        (1 << 5)
> +#define CHNL_STS_IOE        (1 << 6)
> +#define CHNL_STS_ERR        (1 << 7)
> +
> +#define CHNL_STS_BSYWR      (1 << 16)
> +#define CHNL_STS_CURPERR    (1 << 17)
> +#define CHNL_STS_NXTPERR    (1 << 18)
> +#define CHNL_STS_ADDRERR    (1 << 19)
> +#define CHNL_STS_CMPERR     (1 << 20)
> +#define CHNL_STS_TAILERR    (1 << 21)
> +/*
> + 0:9      22:31   Reserved
> + 10       21      TailPErr
> + 11       20      CmpErr
> + 12       19      AddrErr
> + 13       18      NxtPErr
> + 14       17      CurPErr
> + 15       16      BsyWr
> + 16:23    8:15    Reserved
> + 24       7       Error
> + 25       6       IOE
> + 26       5       SOE
> + 27       4       Cmplt
> + 28       3       SOP
> + 29       2       EOP
> + 30       1       EngBusy
> + 31       0       Reserved
> +*/
> +
> +#define DMA_CONTROL_REG             0x10            /* rw */
> +#define DMA_CONTROL_RST                 (1 << 0)
> +#define DMA_TAIL_ENABLE                 (1 << 2)
> +
> +/* CDMAC descriptor status bit definitions */
> +
> +#define STS_CTRL_APP0_ERR         (1 << 31)
> +#define STS_CTRL_APP0_IRQONEND    (1 << 30)
> +/* undoccumented */
> +#define STS_CTRL_APP0_STOPONEND   (1 << 29)
> +#define STS_CTRL_APP0_CMPLT       (1 << 28)
> +#define STS_CTRL_APP0_SOP         (1 << 27)
> +#define STS_CTRL_APP0_EOP         (1 << 26)
> +#define STS_CTRL_APP0_ENGBUSY     (1 << 25)
> +/* undocumented */
> +#define STS_CTRL_APP0_ENGRST      (1 << 24)
> +
> +/**
> + * struct cdmac_bd - LocalLink buffer descriptor format
> + *
> + * app0 bits:
> + *     0    Error
> + *     1    IrqOnEnd    generate an interrupt at completion of DMA  op
> + *     2    reserved
> + *     3    completed   Current descriptor completed
> + *     4    SOP         TX - marks first desc/ RX marks first desct
> + *     5    EOP         TX marks last desc/RX marks last desc
> + *     6    EngBusy     DMA is processing
> + *     7    reserved
> + *     8:31 application specific
> + */
> +struct cdmac_bd {
> +       u32 next;       /* Physical address of next buffer descriptor */
> +       u32 phys;       /* Physical address of transfer buffer */
> +       u32 len;        /* Length of transfer buffer, in bytes */
> +       u32 app0;
> +       u32 app1;
> +       u32 app2;
> +       u32 app3;
> +       u32 app4;
> +};

Hmmm.  All this stuff looks *really* familiar.  It looks like it was
lifted directly from ll_temac.c.  Instead of cloning you should factor
out the common bits for the Xilinx dma interface out into a common
file.  In fact, ll_temac.c and this dma driver can probably share
common bd management routines.  Both drivers would gain the benefit of
having a single code base.

> +
> +struct xlldma_bd {
> +       struct cdmac_bd   dst;  /* LL recv. */
> +       struct cdmac_bd   src;  /* LL xmit */
> +} __attribute__((aligned(32)));
> +
> +struct xlldma_desc {
> +       struct xlldma_bd *bd;           /* HW buffer descriptors */
> +       struct list_head  node;         /* Position in tx_list */
> +       struct list_head  tx_list;      /* Fragments of a DMA request */
> +       struct dma_async_tx_descriptor async_tx;
> +};
> +
> +struct xlldma_chan;
> +#define XLLDMA_MAX_CHANS_PER_DEVICE 8
> +
> +struct xlldma_device {
> +       struct device       *dev;
> +       struct dma_device   common;
> +       struct kmem_cache   *desc_cache;
> +       struct xlldma_chan  *chan[XLLDMA_MAX_CHANS_PER_DEVICE];
> +};
> +
> +struct xlldma_chan {
> +#ifdef CONFIG_PPC_DCR_NATIVE
> +       dcr_host_t reg_dcrs;
> +#else
> +       struct resource reg;            /* Resource for register */
> +       u32 __iomem *reg_base;
> +#endif
> +       dma_cookie_t        completed_cookie; /* Most recent cookie completed */
> +       dma_cookie_t        stop_after_cookie;/* Last cookie released to HW */
> +       spinlock_t          desc_lock;        /* Descriptor operation lock */
> +       struct list_head    desc_queue;       /* Descriptors queue */
> +       struct list_head    pending_free;     /* Waiting for ACK */
> +       struct dma_chan     common;           /* DMA common channel */
> +       struct dma_pool     *bd_pool;         /* HW descriptors pool */
> +       struct xlldma_device *xdev;           /* Channel device */
> +       struct xlldma_bd    *bd_tail;         /* Tail of the HW desc chain */
> +       dma_addr_t          bd_tail_phys;     /* Physical addr of bd_tail */
> +       int                 rx_irq;           /* Channel receive IRQ */
> +       int                 tx_irq;           /* Channel transmit IRQ */
> +       int                 id;               /* Raw id of this channel */
> +       struct tasklet_struct tasklet;
> +};
> +
> +#define to_xlldma_chan(chan) container_of(chan, struct xlldma_chan, common)
> +#define to_xlldma_desc(lh) list_entry(lh, struct xlldma_desc, node)
> +#define tx_to_xlldma_desc(tx) container_of(tx, struct xlldma_desc, async_tx)
> +
> +/* Bus driver interface */
> +int xlldma_device_setup(struct device *dev);
> +int xlldma_device_teardown(struct device *dev);
> +void xlldma_chan_remove(struct xlldma_chan *xchan);
> +void xlldma_cleanup_tasklet(unsigned long data);
> +irqreturn_t xlldma_chan_rx_interrupt(int irq, void *data);
> +irqreturn_t xlldma_chan_tx_interrupt(int irq, void *data);
> +
> +#endif /* __DMA_XLLDMA_H */
> diff -uprN a/drivers/dma/xlldma.c b/drivers/dma/xlldma.c
> --- a/drivers/dma/xlldma.c      1969-12-31 18:00:00.000000000 -0600
> +++ b/drivers/dma/xlldma.c      2010-03-17 11:11:16.000000000 -0500
> @@ -0,0 +1,1099 @@
> +/*
> + * Xilinx Multi-Port Memory Controller (MPMC) DMA Engine support
> + *
> + * Copyright (C) 2010 Digital Design Corporation. All rights reserved.
> + *
> + * Based on fsldma code that is:
> + * Copyright (C) 2007 Freescale Semiconductor, Inc. All rights reserved.
> + *
> + * Authors:
> + *   Steve Magnani <steve@...idescorp.com>, Mar 2010
> + *
> + * fsldma authors:
> + *   Zhang Wei     <wei.zhang@...escale.com>, Jul 2007
> + *   Ebony Zhu     <ebony.zhu@...escale.com>, May 2007
> + *
> + * Description:
> + *   DMA engine driver for Xilinx MPMC Soft Direct Memory Access Controller.
> + *   The driver can be attached to "loopback" MPMC SDMA ports, i.e., those
> + *   whose LL_TX signals are connected to the LL_RX signals.
> + *
> + * This is free software; you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by
> + * the Free Software Foundation; either version 2 of the License, or
> + * (at your option) any later version.
> + *
> + */
> +
> +#include <linux/init.h>
> +#include <linux/module.h>
> +#include <linux/pci.h>
> +#include <linux/interrupt.h>
> +#include <linux/dmaengine.h>
> +#include <linux/delay.h>
> +#include <linux/dma-mapping.h>
> +#include <linux/dmapool.h>
> +#include <linux/sysfs.h>
> +
> +#include "xlldma.h"
> +
> +/* Forward declarations */
> +static enum dma_status xlldma_desc_status(struct xlldma_chan *chan,
> +                                         struct xlldma_desc *desc);
> +
> +/* ---------------------------------------------------------------------
> + * Descriptor accessors
> + */
> +inline void set_desc_cnt(struct xlldma_desc *desc, u32 count)
> +{
> +       desc->bd->src.len = count;
> +       desc->bd->dst.len = count;
> +}
> +
> +inline void set_desc_src(struct xlldma_desc *desc, dma_addr_t src)
> +{
> +       desc->bd->src.phys = src;
> +}
> +
> +inline void set_desc_dest(struct xlldma_desc *desc, dma_addr_t dest)
> +{
> +       desc->bd->dst.phys = dest;
> +}
> +
> +inline void set_desc_next_bd(struct xlldma_desc *desc, dma_addr_t next_bd_phys)
> +{
> +       desc->bd->dst.next = next_bd_phys + offsetof(struct xlldma_bd, dst);
> +       desc->bd->src.next = next_bd_phys + offsetof(struct xlldma_bd, src);
> +}
> +
> +inline void set_desc_next(struct xlldma_desc *desc, struct xlldma_desc *next)
> +{
> +       set_desc_next_bd(desc, next->async_tx.phys);
> +}
> +
> +inline void set_desc_start(struct xlldma_desc *desc)
> +{
> +       desc->bd->src.app0 |= STS_CTRL_APP0_SOP;
> +       desc->bd->dst.app0 |= STS_CTRL_APP0_SOP;
> +}
> +
> +inline void set_desc_end(struct xlldma_desc *desc)
> +{
> +       desc->bd->src.app0 |= STS_CTRL_APP0_EOP;
> +       desc->bd->dst.app0 |= STS_CTRL_APP0_EOP;
> +}
> +
> +inline void set_desc_single(struct xlldma_desc *desc)
> +{
> +       desc->bd->src.app0 |= STS_CTRL_APP0_SOP | STS_CTRL_APP0_EOP;
> +       desc->bd->dst.app0 |= STS_CTRL_APP0_SOP | STS_CTRL_APP0_EOP;
> +}
> +
> +/* ---------------------------------------------------------------------
> + * Low level register access functions
> + */
> +
> +inline u32 xlldma_in32(const struct xlldma_chan *chan, int reg)
> +{
> +#ifdef CONFIG_PPC_DCR_NATIVE
> +       return dcr_read(chan->reg_dcrs, reg);
> +#else
> +       return __raw_readl(chan->reg_base + reg);
> +#endif
> +}
> +
> +inline void xlldma_out32(const struct xlldma_chan *chan, int reg,
> +                        u32 value)
> +{
> +#ifdef CONFIG_PPC_DCR_NATIVE
> +       dcr_write(chan->reg_dcrs, reg, value);
> +#else
> +       __raw_writel(value, chan->reg_base + reg);
> +#endif
> +}
> +
> +/* ---------------------------------------------------------------------
> + * Debugging
> + */
> +
> +#ifndef DEBUG
> +static inline void xlldma_log_dma_chain(struct xlldma_chan *chan) { }
> +#else

Typically the negative case is put at the bottom for things like this.

> +/**
> + * Print the status of the specified DMA chain.
> + *
> + * @chan : DMA channel
> + */
> +static void xlldma_log_dma_chain(struct xlldma_chan *chan)
> +{
> +       unsigned long flags;
> +       struct xlldma_desc *desc;
> +       struct device *dev = chan->xdev->dev;
> +
> +       local_irq_save(flags);
> +
> +       dev_info(dev, "\n");
> +       dev_info(dev,
> +                "**************** DMA Chain, channel %d *****************\n",
> +                chan->id);
> +       dev_info(dev, "\n");
> +       dev_info(dev, "Last completed cookie: %ld\n",
> +                (unsigned long)chan->completed_cookie);
> +       dev_info(dev, "HW stop-after cookie:  %ld\n",
> +                (unsigned long)chan->stop_after_cookie);
> +       dev_info(dev, "Last allocated cookie: %ld\n",
> +                (unsigned long)chan->common.cookie);
> +       dev_info(dev, "\n");
> +
> +       dev_info(dev, "    Register            TX         RX\n");
> +       dev_info(dev, "    --------         --------   --------\n");
> +#ifndef CONFIG_PPC_DCR_NATIVE
> +       dev_info(dev, "    Base Address     %08lX   %08lX\n",
> +                (unsigned long) (chan->reg_base),
> +                (unsigned long) (chan->reg_base + RX_NXTDESC_PTR));
> +#endif
> +       dev_info(dev, "    Curr descriptor  %08X   %08X\n",
> +                xlldma_in32(chan, TX_CURDESC_PTR),
> +                xlldma_in32(chan, RX_CURDESC_PTR));
> +       dev_info(dev, "    Next descriptor  %08X   %08X\n",
> +                xlldma_in32(chan, TX_NXTDESC_PTR),
> +                xlldma_in32(chan, RX_NXTDESC_PTR));
> +       dev_info(dev, "    Tail descriptor  %08X   %08X\n",
> +                xlldma_in32(chan, TX_TAILDESC_PTR),
> +                xlldma_in32(chan, RX_TAILDESC_PTR));
> +       dev_info(dev, "    Current buffer   %08X   %08X\n",
> +                xlldma_in32(chan, TX_CURBUF_ADDR),
> +                xlldma_in32(chan, RX_CURBUF_ADDR));
> +       dev_info(dev, "    Current length   %08X   %08X\n",
> +                xlldma_in32(chan, TX_CURBUF_LENGTH),
> +                xlldma_in32(chan, RX_CURBUF_LENGTH));
> +
> +       /* Dump BDs based on the desc_queue */
> +       if (!list_empty(&chan->desc_queue)) {
> +               dev_info(dev, "\n");
> +               dev_info(dev, "          Descriptor                           "
> +                             "               Status\n");
> +               dev_info(dev, " Cookie   & RX BD      TX BD      From       "
> +                             "To     XferLen   TX  RX\n");
> +               dev_info(dev, "--------  ----------  --------  --------  "
> +                             "--------  --------  --  --\n");
> +       }
> +
> +       list_for_each_entry(desc, &chan->desc_queue, node) {
> +               dev_info(dev, "%08lX   %08lX   %08lX  %08lX  %08lX  %8lu  "
> +                             "%02X  %02X\n",
> +                        (unsigned long) desc->async_tx.cookie,
> +                        (unsigned long) desc->async_tx.phys
> +                            + offsetof(struct xlldma_bd, dst),
> +                        (unsigned long) desc->async_tx.phys
> +                            + offsetof(struct xlldma_bd, src),
> +                        (unsigned long) desc->bd->src.phys,
> +                        (unsigned long) desc->bd->dst.phys,
> +                        (unsigned long) desc->bd->src.len,
> +                        desc->bd->src.app0 >> 24, desc->bd->dst.app0 >> 24);
> +       }
> +
> +       local_irq_restore(flags);
> +}
> +#endif
> +
> +/* ---------------------------------------------------------------------
> + * SDMA HW control
> + */
> +
> +static void dma_init(struct xlldma_chan *chan)
> +{
> +       u32 timeout;
> +
> +       /* Reset the channel */
> +       xlldma_out32(chan, DMA_CONTROL_REG, DMA_CONTROL_RST);
> +       timeout = 1000;         /* microseconds, nominally */
> +       while (xlldma_in32(chan, DMA_CONTROL_REG) & DMA_CONTROL_RST) {
> +               udelay(1);

Be careful with udelays.  You may be wasting processor cycles better
used for something else.  Is the udelay strictly necessary?  Could you
sleep instead?

> +               if (--timeout == 0) {
> +                       dev_err(chan->xdev->dev,
> +                               "DMA reset timeout, channel %d!!\n", chan->id);
> +                       break;
> +               }
> +       }
> +
> +       /* NOTE: No-op on modern MPMCs (>= 3.00.a, possibly earlier) */
> +       xlldma_out32(chan, DMA_CONTROL_REG, DMA_TAIL_ENABLE);
> +
> +       /* Clear Interrupt registers of both channels, as the software reset
> +        * does not clear any register values. Not doing so will cause
> +        * interrupts asserted after the software reset if there is any
> +        * interrupt left over before.
> +        */
> +       xlldma_out32(chan, RX_IRQ_REG,
> +                    IRQ_COAL | IRQ_DLY | IRQ_ERR | IRQ_DMAERR);
> +       xlldma_out32(chan, TX_IRQ_REG,
> +                    IRQ_COAL | IRQ_DLY | IRQ_ERR | IRQ_DMAERR);
> +
> +       /* Configure interrupt enables */
> +       /* On the TX side, we optionally look only for errors */
> +       /* On the RX side, we look for errors and EOPs */
> +       if (chan->tx_irq == NO_IRQ)
> +               xlldma_out32(chan, TX_CHNL_CTRL, 0);
> +       else {
> +               xlldma_out32(chan, TX_CHNL_CTRL,
> +                            CHNL_CTRL_IRQ_EN | CHNL_CTRL_IRQ_ERR_EN);
> +       }

simplify:
reg = chan->tx_irq != NO_IRQ ? CHNL_CTRL_IRQ_EN | CHNL_CTRL_IRQ_ERR_EN : 0;
xlldma_out32(chan, TX_CHNL_CTRL, reg);

It's easier to follow if the calculation is done first and there is
only one xlldma_out32() statement.

> +
> +       if (chan->rx_irq == NO_IRQ)
> +               xlldma_out32(chan, RX_CHNL_CTRL, 0);
> +       else {
> +               /* Set up for interrupt on every EOP */
> +               u32 rx_control = (1 << 16) | CHNL_CTRL_IRQ_LD_CNT;
> +               rx_control |=   CHNL_CTRL_IRQ_EN | CHNL_CTRL_IRQ_COAL_EN
> +                             | CHNL_CTRL_IRQ_ERR_EN;
> +
> +               xlldma_out32(chan, RX_CHNL_CTRL, rx_control);
> +       }

ditto

> +}
> +
> +/**
> + * dma_set_start_bd - Tell the DMA HW where to find the first descriptor
> + *
> + * @chan         : DMA channel
> + * @start_bd_phys : Physical address of first descriptor
> + */
> +static void dma_set_start_bd(struct xlldma_chan *chan, dma_addr_t start_bd_phys)
> +{
> +       dma_addr_t tx_start_phys, rx_start_phys;
> +
> +       tx_start_phys = start_bd_phys + offsetof(struct xlldma_bd, src);
> +       rx_start_phys = start_bd_phys + offsetof(struct xlldma_bd, dst);
> +
> +       dev_dbg(chan->xdev->dev, "SDMA HW starts with 0x%lx\n",
> +               (unsigned long)start_bd_phys);
> +
> +       xlldma_out32(chan, RX_CURDESC_PTR, rx_start_phys);
> +       xlldma_out32(chan, TX_CURDESC_PTR, tx_start_phys);
> +}
> +
> +/**
> + * xlldma_set_start - Find the first transfer and tell the HW where to find
> + *                   its descriptor
> + *
> + * @chan : DMA channel
> + */
> +static void xlldma_set_start(struct xlldma_chan *chan)
> +{
> +       struct list_head *entry;
> +
> +       /* Find the first un-transferred desciptor */
> +       list_for_each(entry, &chan->desc_queue) {
> +               if (xlldma_desc_status(chan, to_xlldma_desc(entry))
> +                   != DMA_SUCCESS)
> +                       break;
> +       }
> +
> +       if (entry != &chan->desc_queue) {
> +               struct xlldma_desc *desc = to_xlldma_desc(entry);
> +               dma_set_start_bd(chan, desc->async_tx.phys);
> +       } else
> +               dma_set_start_bd(chan, chan->bd_tail_phys);

This could use a comment describing what is is making a decision on.

> +}
> +
> +/**
> + * dma_set_tail        - Tell the DMA HW where the last descriptor to proccess resides
> + *               This starts data transfer if the engine was idle.
> + *
> + * @chan       : DMA channel
> + * @tail_desc  : Last transfer to perform
> + */
> +static void dma_set_tail(struct xlldma_chan *chan,
> +                        struct xlldma_desc *tail_desc)
> +{
> +       dma_addr_t tail_desc_phys;
> +       dma_addr_t rx_tail_phys, tx_tail_phys;
> +
> +       tail_desc_phys = tail_desc->async_tx.phys;
> +
> +       tx_tail_phys  = tail_desc_phys + offsetof(struct xlldma_bd, src);
> +       rx_tail_phys  = tail_desc_phys + offsetof(struct xlldma_bd, dst);
> +
> +       /* Keep track of where we've told the HW to stop, in cookie-space */
> +       chan->stop_after_cookie = tail_desc->async_tx.cookie;
> +
> +       dev_dbg(chan->xdev->dev, "SDMA HW ends with   0x%lx (cookie %lu)\n",
> +               (unsigned long)tail_desc_phys,
> +               (unsigned long)chan->stop_after_cookie);
> +
> +       /* Go! */
> +       xlldma_out32(chan, RX_TAILDESC_PTR, rx_tail_phys);
> +       xlldma_out32(chan, TX_TAILDESC_PTR, tx_tail_phys);
> +}
> +
> +static void dma_abort(struct xlldma_chan *chan)
> +{
> +       /* Disable interrupts before resetting the core */
> +       xlldma_out32(chan, TX_CHNL_CTRL, 0);
> +       xlldma_out32(chan, RX_CHNL_CTRL, 0);
> +
> +       dma_init(chan);
> +}
> +
> +/**
> + * append_bd_chain - Link the HW buffer descriptors of a new request to
> + *                the end of the chain of requests in progress (if any).
> + *
> + * The caller is assumed to hold the desc_lock.
> + *
> + * @chan     : DMA channel
> + * @req_desc : Request to link into the HW processing chain
> + */
> +static void append_bd_chain(struct xlldma_chan *chan,
> +                           struct xlldma_desc *req_desc)
> +{
> +       struct xlldma_desc *first_desc, *last_desc;
> +       struct xlldma_bd   *first_bd;
> +       dma_addr_t      first_bd_phys;
> +
> +       first_desc = to_xlldma_desc(req_desc->tx_list.next);
> +       last_desc  = to_xlldma_desc(req_desc->tx_list.prev);
> +
> +       /* Make the bd_tail the BD for the first descriptor of this request */
> +       first_bd = first_desc->bd;
> +       first_bd_phys = first_desc->async_tx.phys;
> +
> +       memcpy(chan->bd_tail, first_bd, sizeof(*chan->bd_tail));
> +       first_desc->bd = chan->bd_tail;
> +       first_desc->async_tx.phys = chan->bd_tail_phys;
> +
> +       /* The original first descriptor BD now becomes the bd_tail. */
> +       chan->bd_tail = first_bd;
> +       chan->bd_tail_phys = first_bd_phys;
> +
> +       /* Link the tail of this request to the bd_tail */
> +       set_desc_next_bd(last_desc, chan->bd_tail_phys);
> +}
> +
> +/**
> + * xlldma_tx_submit - DMAEngine API for queuing a new request.
> + *
> + * NOTE: this function does not make the new request visible to HW.
> + *
> + * @tx : Request chain to queue
> + *
> + * Return - 'cookie' we've assigned to the request
> + */
> +static dma_cookie_t xlldma_tx_submit(struct dma_async_tx_descriptor *tx)
> +{
> +       struct xlldma_chan *chan = to_xlldma_chan(tx->chan);
> +       struct xlldma_desc *req_desc = tx_to_xlldma_desc(tx);
> +       struct xlldma_desc *child;
> +       unsigned long flags;
> +       dma_cookie_t cookie;
> +
> +       spin_lock_irqsave(&chan->desc_lock, flags);
> +
> +       /*
> +        * assign cookies to all of the software descriptors
> +        * that make up this request
> +        */
> +       cookie = chan->common.cookie;
> +       list_for_each_entry(child, &req_desc->tx_list, node) {
> +               cookie++;
> +               if (cookie < 0)
> +                       cookie = DMA_MIN_COOKIE;
> +
> +               child->async_tx.cookie = cookie;
> +       }
> +       chan->common.cookie = cookie;
> +
> +       append_bd_chain(chan, req_desc);
> +       list_splice_tail_init(&req_desc->tx_list, &chan->desc_queue);
> +
> +       spin_unlock_irqrestore(&chan->desc_lock, flags);
> +
> +       return cookie;
> +}
> +
> +/**
> + * xlldma_alloc_descriptor - Allocate descriptor from channel's DMA pool.
> + * @chan : DMA channel
> + *
> + * Return - An allocated descriptor. NULL for failed.
> + */
> +static struct xlldma_desc *xlldma_alloc_descriptor(struct xlldma_chan *chan)
> +{
> +       struct xlldma_desc *xdesc;
> +       struct xlldma_device *xdev = chan->xdev;
> +
> +       xdesc = kmem_cache_zalloc(xdev->desc_cache, GFP_ATOMIC);
> +       if (unlikely(!xdesc)) {
> +               dev_err(xdev->dev, "Descriptor allocation failed\n");
> +               return NULL;
> +       }
> +
> +       INIT_LIST_HEAD(&xdesc->tx_list);
> +       dma_async_tx_descriptor_init(&xdesc->async_tx, &chan->common);
> +       xdesc->async_tx.tx_submit = xlldma_tx_submit;
> +
> +       xdesc->bd = dma_pool_alloc(chan->bd_pool, GFP_ATOMIC,
> +                                  &xdesc->async_tx.phys);
> +       if (likely(xdesc->bd)) {
> +               memset(xdesc->bd, 0, sizeof(*xdesc->bd));
> +       } else {
> +               dev_err(xdev->dev, "DMA pool exhausted\n");
> +               kmem_cache_free(xdev->desc_cache, xdesc);
> +               return NULL;
> +       }

Again, common convention is to put the failure case first and remove
the need for an else block.

> +
> +       return xdesc;
> +}
> +
> +/**
> + * xlldma_free_descriptor - Free a previously allocated descriptor
> + * @chan  : DMA channel
> + * @xdesc : Descriptor to free (may be NULL)
> + *
> + * Return - An allocated descriptor. NULL for failed.
> + */
> +static void xlldma_free_descriptor(struct xlldma_chan *chan,
> +                                  struct xlldma_desc *xdesc)
> +{
> +       if (likely(xdesc)) {
> +               dma_pool_free(chan->bd_pool, xdesc->bd, xdesc->async_tx.phys);
> +               kmem_cache_free(chan->xdev->desc_cache, xdesc);
> +       }
> +}
> +
> +/**
> + * xlldma_alloc_chan_resources - Allocate resources for DMA channel.
> + * @dchan : DMA channel
> + *
> + * This function will create the dma pool for descriptor allocation.
> + *
> + * Return - The number of descriptors allocated.
> + *       -ENOMEM on error.
> + */
> +static int xlldma_alloc_chan_resources(struct dma_chan *dchan)
> +{
> +       struct xlldma_chan *chan = to_xlldma_chan(dchan);
> +
> +       might_sleep();
> +
> +       /* Has this channel already been allocated? */
> +       if (chan->bd_pool)
> +               return 1;
> +
> +       chan->bd_pool = dma_pool_create("xlldma_engine_bd_pool",
> +                                       chan->xdev->dev,
> +                                       sizeof(struct xlldma_bd),
> +                                       __alignof__(struct xlldma_bd), 0);
> +       if (!chan->bd_pool) {
> +               dev_err(chan->xdev->dev,
> +                       "unable to allocate channel %d BD pool.\n", chan->id);
> +               return -ENOMEM;
> +       }
> +
> +       chan->bd_tail = dma_pool_alloc(chan->bd_pool, GFP_KERNEL,
> +                                      &chan->bd_tail_phys);
> +       if (likely(chan->bd_tail)) {
> +               dma_set_start_bd(chan, chan->bd_tail_phys);
> +       } else {
> +               dma_pool_destroy(chan->bd_pool);
> +               chan->bd_pool = NULL;
> +               return -ENOMEM;
> +       }
> +
> +       /* there is at least one descriptor free to be allocated */
> +       return 1;
> +}
> +
> +/**
> + * xlldma_free_desc_list - Free all descriptors in a queue
> + * @chan: DMA channel
> + * @list: the list to free
> + *
> + * LOCKING: must hold chan->desc_lock
> + */
> +static void xlldma_free_desc_list(struct xlldma_chan *chan,
> +                                 struct list_head *list)
> +{
> +       struct xlldma_desc *desc, *_desc;
> +
> +       list_for_each_entry_safe(desc, _desc, list, node) {
> +               list_del(&desc->node);
> +               xlldma_free_descriptor(chan, desc);
> +       }
> +}
> +
> +/**
> + * xlldma_free_chan_resources - Free all resources of the channel.
> + * @dchan : DMA channel
> + */
> +static void xlldma_free_chan_resources(struct dma_chan *dchan)
> +{
> +       struct xlldma_chan *chan = to_xlldma_chan(dchan);
> +       unsigned long flags;
> +
> +       dev_dbg(chan->xdev->dev, "Free all channel resources.\n");
> +       spin_lock_irqsave(&chan->desc_lock, flags);
> +       xlldma_free_desc_list(chan, &chan->desc_queue);
> +       dma_pool_free(chan->bd_pool, chan->bd_tail, chan->bd_tail_phys);
> +       spin_unlock_irqrestore(&chan->desc_lock, flags);
> +
> +       dma_pool_destroy(chan->bd_pool);
> +       chan->bd_pool = NULL;
> +}
> +
> +static struct dma_async_tx_descriptor *
> +xlldma_prep_interrupt(struct dma_chan *dchan, unsigned long flags)
> +{
> +       struct xlldma_chan *chan;
> +       struct xlldma_desc *xdesc = NULL;
> +
> +       if (!dchan)
> +               return NULL;
> +
> +       chan = to_xlldma_chan(dchan);
> +
> +       if (chan->tx_irq == NO_IRQ) {
> +               dev_dbg(chan->xdev->dev,
> +                       "Channel does not support DMA_INTERRUPT\n");
> +               goto fail;
> +       }
> +
> +       dev_dbg(chan->xdev->dev, "DMA_INTERRUPT request\n");
> +       xdesc = xlldma_alloc_descriptor(chan);
> +       if (!xdesc)
> +               goto fail;

But is the descriptor alloc failed, then what is there to free (after
the fail label)?

> +
> +       /*
> +        * Subtle point:
> +        * Interrupts are achieved by setting up a DMA operation
> +        * that will fail with an invalid source buffer address (0)
> +        */
> +       set_desc_single(xdesc);
> +
> +       xdesc->async_tx.flags  = flags; /* client is in control of this ack */
> +       xdesc->async_tx.cookie = -EBUSY;
> +
> +       list_add_tail(&xdesc->node, &xdesc->tx_list);
> +
> +       return &xdesc->async_tx;
> +
> +fail:
> +       xlldma_free_descriptor(chan, xdesc);
> +       return NULL;
> +}
> +
> +static struct dma_async_tx_descriptor *
> +xlldma_prep_memcpy(struct dma_chan *dchan, dma_addr_t dma_dest,
> +                  dma_addr_t dma_src,  size_t len, unsigned long flags)
> +{
> +       struct xlldma_chan *chan;
> +       struct xlldma_desc *xdesc = NULL;
> +
> +       if (!dchan || !len)
> +               return NULL;
> +
> +       chan = to_xlldma_chan(dchan);
> +
> +       dev_dbg(chan->xdev->dev, "Memcpy: 0x%lX to 0x%lX (%u bytes)\n",
> +               (unsigned long) dma_src, (unsigned long) dma_dest, len);
> +
> +       /* Allocate the descriptor from DMA pool */
> +       xdesc = xlldma_alloc_descriptor(chan);
> +       if (!xdesc)
> +               goto fail;

Ditto here.

> +
> +       set_desc_cnt(xdesc,  len);
> +       set_desc_src(xdesc,  dma_src);
> +       set_desc_dest(xdesc, dma_dest);
> +       set_desc_single(xdesc);
> +
> +       xdesc->async_tx.flags  = flags; /* client is in control of this ack */
> +       xdesc->async_tx.cookie = -EBUSY;
> +
> +       list_add_tail(&xdesc->node, &xdesc->tx_list);
> +
> +       return &xdesc->async_tx;
> +
> +fail:
> +       xlldma_free_descriptor(chan, xdesc);
> +       return NULL;
> +}
> +
> +static void xlldma_device_terminate_all(struct dma_chan *dchan)
> +{
> +       struct xlldma_chan *chan;
> +       unsigned long flags;
> +
> +       if (!dchan)
> +               return;
> +
> +       chan = to_xlldma_chan(dchan);
> +
> +       /* Halt the DMA engine */
> +       dma_abort(chan);
> +
> +       spin_lock_irqsave(&chan->desc_lock, flags);
> +
> +       xlldma_free_desc_list(chan, &chan->desc_queue);
> +       xlldma_free_desc_list(chan, &chan->pending_free);
> +
> +       spin_unlock_irqrestore(&chan->desc_lock, flags);
> +}
> +
> +/**
> + * xlldma_update_completed_cookie - Update the completed cookie based on
> + *                                 the HW's current status.
> + * @chan : DMA channel
> + *
> + * CONTEXT: hardirq
> + */
> +static void xlldma_update_completed_cookie(struct xlldma_chan *chan)
> +{
> +       struct xlldma_desc *last_completed = NULL;
> +       struct xlldma_desc *desc;
> +       unsigned long flags;
> +
> +       spin_lock_irqsave(&chan->desc_lock, flags);
> +
> +       list_for_each_entry(desc, &chan->desc_queue, node) {
> +               if (desc->bd->dst.app0 & STS_CTRL_APP0_CMPLT)
> +                       last_completed = desc;
> +               else
> +                       break;

+               if (!(desc->bd->dst.app0 & STS_CTRL_APP0_CMPLT))
+                       break;
+               last_completed = desc;

> +       }
> +
> +       if (last_completed)
> +               chan->completed_cookie = last_completed->async_tx.cookie;
> +
> +       spin_unlock_irqrestore(&chan->desc_lock, flags);
> +}
> +
> +
> +/**
> + * xlldma_desc_status - Check the status of a descriptor
> + * @chan: DMA channel
> + * @desc: DMA SW descriptor
> + *
> + * This function will return the status of the given descriptor
> + */
> +static enum dma_status xlldma_desc_status(struct xlldma_chan *chan,
> +                                         struct xlldma_desc *desc)
> +{
> +       return dma_async_is_complete(desc->async_tx.cookie,
> +                                    chan->completed_cookie,
> +                                    chan->common.cookie);
> +}
> +
> +/**
> + * xlldma_unmap_desc - Unmap DMA addresses used by a descriptor
> + * @chan : DMA channel
> + * @desc : Descriptor
> + */
> +static inline void xlldma_unmap_desc(struct xlldma_chan *chan,
> +                                    struct xlldma_desc *desc)
> +{
> +       enum dma_ctrl_flags flags = desc->async_tx.flags;
> +       struct device *dev = chan->xdev->dev;
> +
> +       if (!(flags & DMA_COMPL_SKIP_DEST_UNMAP)) {
> +               if (flags & DMA_COMPL_DEST_UNMAP_SINGLE)
> +                       dma_unmap_single(dev, desc->bd->dst.phys,
> +                                        desc->bd->dst.len, DMA_FROM_DEVICE);
> +               else
> +                       dma_unmap_page(dev, desc->bd->dst.phys,
> +                                      desc->bd->dst.len, DMA_FROM_DEVICE);
> +       }
> +       if (!(flags & DMA_COMPL_SKIP_SRC_UNMAP)) {
> +               if (flags & DMA_COMPL_SRC_UNMAP_SINGLE)
> +                       dma_unmap_single(dev, desc->bd->src.phys,
> +                                        desc->bd->src.len, DMA_TO_DEVICE);
> +               else
> +                       dma_unmap_page(dev, desc->bd->src.phys,
> +                                      desc->bd->src.len, DMA_TO_DEVICE);
> +       }
> +}
> +
> +/**
> + * xlldma_chan_desc_cleanup - Clean up descriptors
> + * @chan : DMA channel
> + *
> + * This function cleans up the desc_queue of DMA channel.
> + * Completions are issued for finished descriptors.
> + */
> +static void xlldma_chan_desc_cleanup(struct xlldma_chan *chan)
> +{
> +       struct xlldma_desc *desc, *_desc;
> +       struct list_head *last_complete = &chan->desc_queue;
> +       struct list_head *entry;
> +       struct list_head cleanup_list;
> +
> +       unsigned long flags;
> +
> +       INIT_LIST_HEAD(&cleanup_list);
> +       spin_lock_irqsave(&chan->desc_lock, flags);
> +
> +       /* First try to clean up old descriptors */
> +       list_for_each_entry_safe(desc, _desc, &chan->pending_free, node) {
> +               if (async_tx_test_ack(&desc->async_tx)) {
> +                       list_del(&desc->node);
> +                       xlldma_free_descriptor(chan, desc);
> +               }
> +       }
> +
> +       /* Move all newly completed entries to the cleanup_list
> +        * so we can minimize time spent with the desc_lock held
> +        */
> +       list_for_each(entry, &chan->desc_queue) {
> +               if (xlldma_desc_status(chan, to_xlldma_desc(entry))
> +                   != DMA_IN_PROGRESS)
> +                       last_complete = entry;
> +       }
> +
> +       list_cut_position(&cleanup_list, &chan->desc_queue, last_complete);
> +       spin_unlock_irqrestore(&chan->desc_lock, flags);
> +
> +       list_for_each_entry_safe(desc, _desc, &cleanup_list, node) {
> +               dma_async_tx_callback callback;
> +               void *callback_param;
> +
> +               xlldma_unmap_desc(chan, desc);
> +
> +               /* Run the descriptor callback function */
> +               callback = desc->async_tx.callback;
> +               callback_param = desc->async_tx.callback_param;
> +               if (callback) {
> +                       dev_dbg(chan->xdev->dev, "descriptor %p callback\n",
> +                               desc);
> +                       callback(callback_param);
> +               }
> +
> +               dma_run_dependencies(&desc->async_tx);
> +
> +               if (async_tx_test_ack(&desc->async_tx))
> +                       xlldma_free_descriptor(chan, desc);
> +               else {
> +                       /* Defer free until client has ACKed */
> +                       spin_lock_irqsave(&chan->desc_lock, flags);
> +                       list_move_tail(&desc->node, &chan->pending_free);
> +                       spin_unlock_irqrestore(&chan->desc_lock, flags);
> +               }
> +       }
> +}
> +
> +/**
> + * xlldma_desc_xfer_to_hw - Release descriptors to SDMA HW for processing.
> + * @chan : DMA channel
> + */
> +static void xlldma_desc_xfer_to_hw(struct xlldma_chan *chan)
> +{
> +       unsigned long flags;
> +
> +       spin_lock_irqsave(&chan->desc_lock, flags);
> +
> +       if (chan->common.cookie != chan->stop_after_cookie) {
> +               BUG_ON(list_empty(&chan->desc_queue));
> +               dma_set_tail(chan, to_xlldma_desc(chan->desc_queue.prev));
> +       }
> +
> +       spin_unlock_irqrestore(&chan->desc_lock, flags);
> +}
> +
> +/**
> + * xlldma_memcpy_issue_pending - Issue the DMA start command
> + * @dchan : DMA channel
> + */
> +static void xlldma_memcpy_issue_pending(struct dma_chan *dchan)
> +{
> +       struct xlldma_chan *chan = to_xlldma_chan(dchan);
> +       xlldma_desc_xfer_to_hw(chan);
> +}
> +
> +/**
> + * xlldma_is_complete - Determine the DMA status
> + * @dchan : DMA channel
> + * @cookie: transaction identifier to check status of
> + * @done: returns last completed cookie, can be NULL
> + * @used: returns last issued cookie, can be NULL
> + */
> +static enum dma_status xlldma_is_complete(struct dma_chan *dchan,
> +                                         dma_cookie_t cookie,
> +                                         dma_cookie_t *done,
> +                                         dma_cookie_t *used)
> +{
> +       struct xlldma_chan *chan = to_xlldma_chan(dchan);
> +       dma_cookie_t last_used;
> +       dma_cookie_t last_complete;
> +
> +       xlldma_chan_desc_cleanup(chan);
> +
> +       last_used     = dchan->cookie;
> +       last_complete = chan->completed_cookie;
> +
> +       if (done)
> +               *done = last_complete;
> +
> +       if (used)
> +               *used = last_used;
> +
> +       return dma_async_is_complete(cookie, last_complete, last_used);
> +}
> +
> +/*--------------------------------------------------------------------------
> + * Interrupt Handling
> + */
> +
> +irqreturn_t xlldma_chan_rx_interrupt(int irq, void *data)
> +{
> +       struct xlldma_chan *chan = (struct xlldma_chan *)data;
> +       u32 irq_causes;
> +       u32 error_causes = 0;
> +
> +       irq_causes = xlldma_in32(chan, RX_IRQ_REG);
> +
> +       /* Clear the events */
> +       xlldma_out32(chan, RX_IRQ_REG, irq_causes);
> +
> +       irq_causes &= ~(IRQ_COAL | IRQ_ERR);
> +       if (!irq_causes)
> +               return IRQ_NONE;
> +
> +       /* DMA Error */
> +       if (irq_causes & IRQ_ERR) {
> +               error_causes = xlldma_in32(chan, RX_CHNL_STS);
> +               error_causes &=  (CHNL_STS_BSYWR     | CHNL_STS_CURPERR
> +                                 | CHNL_STS_NXTPERR | CHNL_STS_ADDRERR
> +                                 | CHNL_STS_CMPERR  | CHNL_STS_TAILERR);
> +               dev_emerg(chan->xdev->dev,
> +                         "Channel %d RX DMA Error, causes = 0x%X\n",
> +                         chan->id, error_causes);
> +               xlldma_log_dma_chain(chan);
> +
> +               dma_abort(chan);
> +               panic("DMA Error!\n");
> +       }
> +
> +       dev_dbg(chan->xdev->dev, "RX irq: DMA EOP\n");
> +       xlldma_update_completed_cookie(chan);
> +       xlldma_desc_xfer_to_hw(chan);
> +
> +       dev_dbg(chan->xdev->dev, "RX irq: Exit\n");
> +       tasklet_schedule(&chan->tasklet);
> +       return IRQ_HANDLED;
> +}
> +EXPORT_SYMBOL(xlldma_chan_rx_interrupt);
> +
> +irqreturn_t xlldma_chan_tx_interrupt(int irq, void *data)
> +{
> +       struct xlldma_chan *chan = (struct xlldma_chan *)data;
> +       struct xlldma_bd *bd;
> +       u32 irq_causes, error_causes;
> +       u32 desc_phys;
> +
> +       irq_causes = xlldma_in32(chan, TX_IRQ_REG);
> +       dev_dbg(chan->xdev->dev, "TX irq: channel %d, causes = 0x%X\n",
> +               chan->id, irq_causes);
> +       /* Clear the events */
> +       xlldma_out32(chan, TX_IRQ_REG, irq_causes);
> +
> +       if (!(irq_causes & IRQ_ERR))
> +               return IRQ_NONE;
> +
> +       /* DMA Error */
> +       error_causes = xlldma_in32(chan, TX_CHNL_STS);
> +       error_causes &=  (CHNL_STS_BSYWR     | CHNL_STS_CURPERR
> +                         | CHNL_STS_NXTPERR | CHNL_STS_ADDRERR
> +                         | CHNL_STS_CMPERR  | CHNL_STS_TAILERR);
> +
> +       if (!(error_causes & CHNL_STS_ADDRERR)) {
> +               dev_emerg(chan->xdev->dev,
> +                         "Channel %d TX DMA Error, causes = 0x%X\n",
> +                         chan->id, error_causes);
> +               xlldma_log_dma_chain(chan);
> +
> +               dma_abort(chan);
> +               panic("DMA Error!\n");
> +       }
> +
> +       /* DMA_INTERRUPT request complete */
> +       /* This is the only 'expected' DMA error */
> +       dev_dbg(chan->xdev->dev, "irq: DMA_INTERRUPT\n");
> +
> +       /* Set the complete flag by hand since the HW won't do it,
> +        * and we need it to update the completed cookie
> +        */
> +       desc_phys =   xlldma_in32(chan, TX_CURDESC_PTR)
> +                   - offsetof(struct xlldma_bd, src);
> +
> +       bd = (struct xlldma_bd *) bus_to_virt(desc_phys);
> +       bd->dst.app0 = STS_CTRL_APP0_CMPLT;
> +
> +       dma_abort(chan);
> +       xlldma_update_completed_cookie(chan);
> +       xlldma_set_start(chan);
> +       xlldma_desc_xfer_to_hw(chan);
> +
> +       dev_dbg(chan->xdev->dev, "TX irq: Exit\n");
> +       tasklet_schedule(&chan->tasklet);
> +       return IRQ_HANDLED;
> +}
> +EXPORT_SYMBOL(xlldma_chan_tx_interrupt);
> +
> +void xlldma_cleanup_tasklet(unsigned long data)
> +{
> +       struct xlldma_chan *chan = (struct xlldma_chan *)data;
> +       xlldma_chan_desc_cleanup(chan);
> +}
> +EXPORT_SYMBOL(xlldma_cleanup_tasklet);
> +
> +/* ---------------------------------------------------------------------
> + * Initialization / Termination / Enumeration
> + */
> +
> +void xlldma_chan_remove(struct xlldma_chan *xchan)
> +{
> +       if (xchan->tx_irq != NO_IRQ)
> +               free_irq(xchan->tx_irq, xchan);
> +       if (xchan->rx_irq != NO_IRQ)
> +               free_irq(xchan->rx_irq, xchan);
> +       list_del(&xchan->common.device_node);
> +#ifdef CONFIG_PPC_DCR_NATIVE
> +       dcr_unmap(xchan->reg_dcrs, xchan->dcr_c);
> +#else
> +       iounmap(xchan->reg_base);
> +#endif
> +       kfree(xchan);
> +}
> +EXPORT_SYMBOL(xlldma_chan_remove);
> +
> +/*---------------------------------------------------------------------------
> + * sysfs interface
> + */
> +
> +#ifdef DEBUG
> +/**
> + * Print the DMA chain in response to a userland request
> + * @dev   :  Device being accessed
> + * @buf   :  Data being written from userspace
> + *          The first character controls which channel is printed
> + *          ('0' - '7')
> + * @count :  Number of bytes of data being written
> + *
> + * @return  count   (accept all data)
> + */
> +static ssize_t do_xlldma_log_dma_chain(struct device *dev,
> +                               struct device_attribute *attr,
> +                               const char *buf, size_t count)
> +{
> +       struct xlldma_device *xdev = dev_get_drvdata(dev);
> +
> +       if (count && (*buf >= '0')) {
> +               int idx = *buf - '0';
> +               if ((idx < XLLDMA_MAX_CHANS_PER_DEVICE) && xdev->chan[idx])
> +                       xlldma_log_dma_chain(xdev->chan[idx]);
> +               else
> +                       dev_info(xdev->dev, "Channel %c not active\n", *buf);
> +       }
> +
> +       return count;
> +}
> +
> +static struct device_attribute xlldma_attrs[] = {
> +       __ATTR(log_dma_chain, S_IWUSR, NULL, do_xlldma_log_dma_chain),
> +       __ATTR_NULL
> +};
> +#endif

This should probably be using debugfs instead.

> +
> +/*----------------------------------------------------------------------------
> + * Generic probe/remove
> + */
> +
> +/**
> + * xlldma_device_setup: Complete initialization and registration of a xlldma
> + *                     device. Assumes bus-specific code has enumerated
> + *                     the channels for the device.
> + * @dev   :  Device being initialized
> + */
> +int xlldma_device_setup(struct device *dev)
> +{
> +       struct xlldma_device *xdev = dev_get_drvdata(dev);
> +       int i, rc;
> +
> +       dma_cap_set(DMA_MEMCPY,    xdev->common.cap_mask);
> +       xdev->common.device_alloc_chan_resources = xlldma_alloc_chan_resources;
> +       xdev->common.device_free_chan_resources  = xlldma_free_chan_resources;
> +       xdev->common.device_prep_dma_interrupt   = xlldma_prep_interrupt;
> +       xdev->common.device_prep_dma_memcpy      = xlldma_prep_memcpy;
> +       xdev->common.device_is_tx_complete       = xlldma_is_complete;
> +       xdev->common.device_issue_pending        = xlldma_memcpy_issue_pending;
> +       xdev->common.device_terminate_all        = xlldma_device_terminate_all;
> +
> +       for (i = 0; i < XLLDMA_MAX_CHANS_PER_DEVICE; i++) {
> +               if (xdev->chan[i] && (xdev->chan[i]->tx_irq != NO_IRQ)) {
> +                       dma_cap_set(DMA_INTERRUPT, xdev->common.cap_mask);
> +                       break;
> +               }
> +       }
> +
> +       xdev->desc_cache = KMEM_CACHE(xlldma_desc, 0);
> +       if (!xdev->desc_cache) {
> +               printk(KERN_INFO "Could not allocate descriptor cache.\n");
> +               rc = -ENOMEM;
> +               goto fail;
> +       }
> +
> +       for (i = 0; i < XLLDMA_MAX_CHANS_PER_DEVICE; i++) {
> +               if (xdev->chan[i])
> +                       dma_init(xdev->chan[i]);        /* HW initialization */
> +       }
> +
> +       rc = dma_async_device_register(&xdev->common);
> +       if (rc)
> +               goto fail;
> +
> +#ifdef DEBUG
> +       for (i = 0; attr_name(xlldma_attrs[i]); i++) {
> +               rc = device_create_file(xdev->dev, &xlldma_attrs[i]);
> +               if (rc)
> +                       goto fail;
> +       }
> +#endif

Ugly #ifdefs.  Put the debug setup code into a separate function that
becomes an empty static inline when configured out, and takes care of
its own cleanup when it fails.

> +       return 0;
> +
> +fail:
> +#ifdef DEBUG
> +       for (i = 0; attr_name(xlldma_attrs[i]); i++)
> +               device_remove_file(xdev->dev, &xlldma_attrs[i]);
> +#endif
> +       if (xdev->desc_cache)
> +               kmem_cache_destroy(xdev->desc_cache);
> +
> +       return rc;
> +}
> +EXPORT_SYMBOL(xlldma_device_setup);

Most of these symbols really shouldn't be exported.  They are driver
internal.  You should build all the .c files into a single .ko to
avoid the need of exporting.  (although and I mentioned earlier, the
bus binding code really should just be merged into this file).

Also, your irq request code probably belongs in xlldma_device_setup()
instead of the bus binding code.

> +
> +int xlldma_device_teardown(struct device *dev)
> +{
> +       struct xlldma_device *xdev = dev_get_drvdata(dev);
> +       unsigned int i;
> +
> +#ifdef DEBUG
> +       for (i = 0; attr_name(xlldma_attrs[i]); i++)
> +               device_remove_file(xdev->dev, &xlldma_attrs[i]);
> +#endif

ditto.

> +       kmem_cache_destroy(xdev->desc_cache);
> +       dma_async_device_unregister(&xdev->common);
> +
> +       for (i = 0; i < XLLDMA_MAX_CHANS_PER_DEVICE; i++)
> +               if (xdev->chan[i])
> +                       xlldma_chan_remove(xdev->chan[i]);
> +
> +       dev_set_drvdata(dev, NULL);
> +       kfree(xdev);
> +
> +       return 0;
> +}
> +EXPORT_SYMBOL(xlldma_device_teardown);
> +
> +MODULE_AUTHOR("Digital Design Corporation");
> +MODULE_DESCRIPTION("Xilinx LocalLink DMA driver");
> +MODULE_LICENSE("GPL");

g.

-- 
Grant Likely, B.Sc., P.Eng.
Secret Lab Technologies Ltd.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ