lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite for Android: free password hash cracker in your pocket
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAPDyKFpm5Vf=vzVFyhrmjZ61FE-JbMm=+xhF=6AMX96cKF5BXg@mail.gmail.com>
Date:	Mon, 20 Jul 2015 13:19:22 +0200
From:	Ulf Hansson <ulf.hansson@...aro.org>
To:	Robert Jarzmik <robert.jarzmik@...e.fr>
Cc:	linux-mmc <linux-mmc@...r.kernel.org>,
	"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
	Daniel Mack <zonque@...il.com>
Subject: Re: [PATCH] mmc: host: pxamci: switch over to dmaengine use

On 6 June 2015 at 23:15, Robert Jarzmik <robert.jarzmik@...e.fr> wrote:
> From: Daniel Mack <zonque@...il.com>
>
> Switch over pxamci to dmaengine. This prepares the devicetree full
> support of pxamci.
>
> This was successfully tested on a PXA3xx board, as well as PXA27x.
>
> Signed-off-by: Daniel Mack <zonque@...il.com>
> [adapted to pxa-dma]
> Signed-off-by: Robert Jarzmik <robert.jarzmik@...e.fr>

Thanks, applied!

Kind regards
Uffe

> ---
>  drivers/mmc/host/pxamci.c | 200 ++++++++++++++++++++++++++--------------------
>  1 file changed, 114 insertions(+), 86 deletions(-)
>
> diff --git a/drivers/mmc/host/pxamci.c b/drivers/mmc/host/pxamci.c
> index 1b6d0bf..1420f29 100644
> --- a/drivers/mmc/host/pxamci.c
> +++ b/drivers/mmc/host/pxamci.c
> @@ -22,7 +22,9 @@
>  #include <linux/platform_device.h>
>  #include <linux/delay.h>
>  #include <linux/interrupt.h>
> +#include <linux/dmaengine.h>
>  #include <linux/dma-mapping.h>
> +#include <linux/dma/pxa-dma.h>
>  #include <linux/clk.h>
>  #include <linux/err.h>
>  #include <linux/mmc/host.h>
> @@ -37,7 +39,6 @@
>  #include <asm/sizes.h>
>
>  #include <mach/hardware.h>
> -#include <mach/dma.h>
>  #include <linux/platform_data/mmc-pxamci.h>
>
>  #include "pxamci.h"
> @@ -58,7 +59,6 @@ struct pxamci_host {
>         struct clk              *clk;
>         unsigned long           clkrate;
>         int                     irq;
> -       int                     dma;
>         unsigned int            clkrt;
>         unsigned int            cmdat;
>         unsigned int            imask;
> @@ -69,8 +69,10 @@ struct pxamci_host {
>         struct mmc_command      *cmd;
>         struct mmc_data         *data;
>
> +       struct dma_chan         *dma_chan_rx;
> +       struct dma_chan         *dma_chan_tx;
> +       dma_cookie_t            dma_cookie;
>         dma_addr_t              sg_dma;
> -       struct pxa_dma_desc     *sg_cpu;
>         unsigned int            dma_len;
>
>         unsigned int            dma_dir;
> @@ -173,14 +175,18 @@ static void pxamci_disable_irq(struct pxamci_host *host, unsigned int mask)
>         spin_unlock_irqrestore(&host->lock, flags);
>  }
>
> +static void pxamci_dma_irq(void *param);
> +
>  static void pxamci_setup_data(struct pxamci_host *host, struct mmc_data *data)
>  {
> +       struct dma_async_tx_descriptor *tx;
> +       enum dma_data_direction direction;
> +       struct dma_slave_config config;
> +       struct dma_chan *chan;
>         unsigned int nob = data->blocks;
>         unsigned long long clks;
>         unsigned int timeout;
> -       bool dalgn = 0;
> -       u32 dcmd;
> -       int i;
> +       int ret;
>
>         host->data = data;
>
> @@ -195,54 +201,48 @@ static void pxamci_setup_data(struct pxamci_host *host, struct mmc_data *data)
>         timeout = (unsigned int)clks + (data->timeout_clks << host->clkrt);
>         writel((timeout + 255) / 256, host->base + MMC_RDTO);
>
> +       memset(&config, 0, sizeof(config));
> +       config.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
> +       config.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
> +       config.src_addr = host->res->start + MMC_RXFIFO;
> +       config.dst_addr = host->res->start + MMC_TXFIFO;
> +       config.src_maxburst = 32;
> +       config.dst_maxburst = 32;
> +
>         if (data->flags & MMC_DATA_READ) {
>                 host->dma_dir = DMA_FROM_DEVICE;
> -               dcmd = DCMD_INCTRGADDR | DCMD_FLOWSRC;
> -               DRCMR(host->dma_drcmrtx) = 0;
> -               DRCMR(host->dma_drcmrrx) = host->dma | DRCMR_MAPVLD;
> +               direction = DMA_DEV_TO_MEM;
> +               chan = host->dma_chan_rx;
>         } else {
>                 host->dma_dir = DMA_TO_DEVICE;
> -               dcmd = DCMD_INCSRCADDR | DCMD_FLOWTRG;
> -               DRCMR(host->dma_drcmrrx) = 0;
> -               DRCMR(host->dma_drcmrtx) = host->dma | DRCMR_MAPVLD;
> +               direction = DMA_MEM_TO_DEV;
> +               chan = host->dma_chan_tx;
>         }
>
> -       dcmd |= DCMD_BURST32 | DCMD_WIDTH1;
> +       config.direction = direction;
> +
> +       ret = dmaengine_slave_config(chan, &config);
> +       if (ret < 0) {
> +               dev_err(mmc_dev(host->mmc), "dma slave config failed\n");
> +               return;
> +       }
>
> -       host->dma_len = dma_map_sg(mmc_dev(host->mmc), data->sg, data->sg_len,
> +       host->dma_len = dma_map_sg(chan->device->dev, data->sg, data->sg_len,
>                                    host->dma_dir);
>
> -       for (i = 0; i < host->dma_len; i++) {
> -               unsigned int length = sg_dma_len(&data->sg[i]);
> -               host->sg_cpu[i].dcmd = dcmd | length;
> -               if (length & 31 && !(data->flags & MMC_DATA_READ))
> -                       host->sg_cpu[i].dcmd |= DCMD_ENDIRQEN;
> -               /* Not aligned to 8-byte boundary? */
> -               if (sg_dma_address(&data->sg[i]) & 0x7)
> -                       dalgn = 1;
> -               if (data->flags & MMC_DATA_READ) {
> -                       host->sg_cpu[i].dsadr = host->res->start + MMC_RXFIFO;
> -                       host->sg_cpu[i].dtadr = sg_dma_address(&data->sg[i]);
> -               } else {
> -                       host->sg_cpu[i].dsadr = sg_dma_address(&data->sg[i]);
> -                       host->sg_cpu[i].dtadr = host->res->start + MMC_TXFIFO;
> -               }
> -               host->sg_cpu[i].ddadr = host->sg_dma + (i + 1) *
> -                                       sizeof(struct pxa_dma_desc);
> +       tx = dmaengine_prep_slave_sg(chan, data->sg, host->dma_len, direction,
> +                                    DMA_PREP_INTERRUPT);
> +       if (!tx) {
> +               dev_err(mmc_dev(host->mmc), "prep_slave_sg() failed\n");
> +               return;
>         }
> -       host->sg_cpu[host->dma_len - 1].ddadr = DDADR_STOP;
> -       wmb();
>
> -       /*
> -        * The PXA27x DMA controller encounters overhead when working with
> -        * unaligned (to 8-byte boundaries) data, so switch on byte alignment
> -        * mode only if we have unaligned data.
> -        */
> -       if (dalgn)
> -               DALGN |= (1 << host->dma);
> -       else
> -               DALGN &= ~(1 << host->dma);
> -       DDADR(host->dma) = host->sg_dma;
> +       if (!(data->flags & MMC_DATA_READ)) {
> +               tx->callback = pxamci_dma_irq;
> +               tx->callback_param = host;
> +       }
> +
> +       host->dma_cookie = dmaengine_submit(tx);
>
>         /*
>          * workaround for erratum #91:
> @@ -251,7 +251,7 @@ static void pxamci_setup_data(struct pxamci_host *host, struct mmc_data *data)
>          * before starting DMA.
>          */
>         if (!cpu_is_pxa27x() || data->flags & MMC_DATA_READ)
> -               DCSR(host->dma) = DCSR_RUN;
> +               dma_async_issue_pending(chan);
>  }
>
>  static void pxamci_start_cmd(struct pxamci_host *host, struct mmc_command *cmd, unsigned int cmdat)
> @@ -343,7 +343,7 @@ static int pxamci_cmd_done(struct pxamci_host *host, unsigned int stat)
>                  * enable DMA late
>                  */
>                 if (cpu_is_pxa27x() && host->data->flags & MMC_DATA_WRITE)
> -                       DCSR(host->dma) = DCSR_RUN;
> +                       dma_async_issue_pending(host->dma_chan_tx);
>         } else {
>                 pxamci_finish_request(host, host->mrq);
>         }
> @@ -354,13 +354,17 @@ static int pxamci_cmd_done(struct pxamci_host *host, unsigned int stat)
>  static int pxamci_data_done(struct pxamci_host *host, unsigned int stat)
>  {
>         struct mmc_data *data = host->data;
> +       struct dma_chan *chan;
>
>         if (!data)
>                 return 0;
>
> -       DCSR(host->dma) = 0;
> -       dma_unmap_sg(mmc_dev(host->mmc), data->sg, data->sg_len,
> -                    host->dma_dir);
> +       if (data->flags & MMC_DATA_READ)
> +               chan = host->dma_chan_rx;
> +       else
> +               chan = host->dma_chan_tx;
> +       dma_unmap_sg(chan->device->dev,
> +                    data->sg, data->sg_len, host->dma_dir);
>
>         if (stat & STAT_READ_TIME_OUT)
>                 data->error = -ETIMEDOUT;
> @@ -552,20 +556,37 @@ static const struct mmc_host_ops pxamci_ops = {
>         .enable_sdio_irq        = pxamci_enable_sdio_irq,
>  };
>
> -static void pxamci_dma_irq(int dma, void *devid)
> +static void pxamci_dma_irq(void *param)
>  {
> -       struct pxamci_host *host = devid;
> -       int dcsr = DCSR(dma);
> -       DCSR(dma) = dcsr & ~DCSR_STOPIRQEN;
> +       struct pxamci_host *host = param;
> +       struct dma_tx_state state;
> +       enum dma_status status;
> +       struct dma_chan *chan;
> +       unsigned long flags;
> +
> +       spin_lock_irqsave(&host->lock, flags);
> +
> +       if (!host->data)
> +               goto out_unlock;
>
> -       if (dcsr & DCSR_ENDINTR) {
> +       if (host->data->flags & MMC_DATA_READ)
> +               chan = host->dma_chan_rx;
> +       else
> +               chan = host->dma_chan_tx;
> +
> +       status = dmaengine_tx_status(chan, host->dma_cookie, &state);
> +
> +       if (likely(status == DMA_COMPLETE)) {
>                 writel(BUF_PART_FULL, host->base + MMC_PRTBUF);
>         } else {
> -               pr_err("%s: DMA error on channel %d (DCSR=%#x)\n",
> -                      mmc_hostname(host->mmc), dma, dcsr);
> +               pr_err("%s: DMA error on %s channel\n", mmc_hostname(host->mmc),
> +                       host->data->flags & MMC_DATA_READ ? "rx" : "tx");
>                 host->data->error = -EIO;
>                 pxamci_data_done(host, 0);
>         }
> +
> +out_unlock:
> +       spin_unlock_irqrestore(&host->lock, flags);
>  }
>
>  static irqreturn_t pxamci_detect_irq(int irq, void *devid)
> @@ -625,7 +646,9 @@ static int pxamci_probe(struct platform_device *pdev)
>         struct mmc_host *mmc;
>         struct pxamci_host *host = NULL;
>         struct resource *r, *dmarx, *dmatx;
> +       struct pxad_param param_rx, param_tx;
>         int ret, irq, gpio_cd = -1, gpio_ro = -1, gpio_power = -1;
> +       dma_cap_mask_t mask;
>
>         ret = pxamci_of_init(pdev);
>         if (ret)
> @@ -671,7 +694,6 @@ static int pxamci_probe(struct platform_device *pdev)
>
>         host = mmc_priv(mmc);
>         host->mmc = mmc;
> -       host->dma = -1;
>         host->pdata = pdev->dev.platform_data;
>         host->clkrt = CLKRT_OFF;
>
> @@ -702,12 +724,6 @@ static int pxamci_probe(struct platform_device *pdev)
>                                      MMC_CAP_SD_HIGHSPEED;
>         }
>
> -       host->sg_cpu = dma_alloc_coherent(&pdev->dev, PAGE_SIZE, &host->sg_dma, GFP_KERNEL);
> -       if (!host->sg_cpu) {
> -               ret = -ENOMEM;
> -               goto out;
> -       }
> -
>         spin_lock_init(&host->lock);
>         host->res = r;
>         host->irq = irq;
> @@ -728,32 +744,45 @@ static int pxamci_probe(struct platform_device *pdev)
>         writel(64, host->base + MMC_RESTO);
>         writel(host->imask, host->base + MMC_I_MASK);
>
> -       host->dma = pxa_request_dma(DRIVER_NAME, DMA_PRIO_LOW,
> -                                   pxamci_dma_irq, host);
> -       if (host->dma < 0) {
> -               ret = -EBUSY;
> -               goto out;
> -       }
> -
>         ret = request_irq(host->irq, pxamci_irq, 0, DRIVER_NAME, host);
>         if (ret)
>                 goto out;
>
>         platform_set_drvdata(pdev, mmc);
>
> -       dmarx = platform_get_resource(pdev, IORESOURCE_DMA, 0);
> -       if (!dmarx) {
> -               ret = -ENXIO;
> +       if (!pdev->dev.of_node) {
> +               dmarx = platform_get_resource(pdev, IORESOURCE_DMA, 0);
> +               dmatx = platform_get_resource(pdev, IORESOURCE_DMA, 1);
> +               if (!dmarx || !dmatx) {
> +                       ret = -ENXIO;
> +                       goto out;
> +               }
> +               param_rx.prio = PXAD_PRIO_LOWEST;
> +               param_rx.drcmr = dmarx->start;
> +               param_tx.prio = PXAD_PRIO_LOWEST;
> +               param_tx.drcmr = dmatx->start;
> +       }
> +
> +       dma_cap_zero(mask);
> +       dma_cap_set(DMA_SLAVE, mask);
> +
> +       host->dma_chan_rx =
> +               dma_request_slave_channel_compat(mask, pxad_filter_fn,
> +                                                &param_rx, &pdev->dev, "rx");
> +       if (host->dma_chan_rx == NULL) {
> +               dev_err(&pdev->dev, "unable to request rx dma channel\n");
> +               ret = -ENODEV;
>                 goto out;
>         }
> -       host->dma_drcmrrx = dmarx->start;
>
> -       dmatx = platform_get_resource(pdev, IORESOURCE_DMA, 1);
> -       if (!dmatx) {
> -               ret = -ENXIO;
> +       host->dma_chan_tx =
> +               dma_request_slave_channel_compat(mask, pxad_filter_fn,
> +                                                &param_tx,  &pdev->dev, "tx");
> +       if (host->dma_chan_tx == NULL) {
> +               dev_err(&pdev->dev, "unable to request tx dma channel\n");
> +               ret = -ENODEV;
>                 goto out;
>         }
> -       host->dma_drcmrtx = dmatx->start;
>
>         if (host->pdata) {
>                 gpio_cd = host->pdata->gpio_card_detect;
> @@ -814,12 +843,12 @@ err_gpio_ro:
>         gpio_free(gpio_power);
>   out:
>         if (host) {
> -               if (host->dma >= 0)
> -                       pxa_free_dma(host->dma);
> +               if (host->dma_chan_rx)
> +                       dma_release_channel(host->dma_chan_rx);
> +               if (host->dma_chan_tx)
> +                       dma_release_channel(host->dma_chan_tx);
>                 if (host->base)
>                         iounmap(host->base);
> -               if (host->sg_cpu)
> -                       dma_free_coherent(&pdev->dev, PAGE_SIZE, host->sg_cpu, host->sg_dma);
>                 if (host->clk)
>                         clk_put(host->clk);
>         }
> @@ -863,13 +892,12 @@ static int pxamci_remove(struct platform_device *pdev)
>                        END_CMD_RES|PRG_DONE|DATA_TRAN_DONE,
>                        host->base + MMC_I_MASK);
>
> -               DRCMR(host->dma_drcmrrx) = 0;
> -               DRCMR(host->dma_drcmrtx) = 0;
> -
>                 free_irq(host->irq, host);
> -               pxa_free_dma(host->dma);
> +               dmaengine_terminate_all(host->dma_chan_rx);
> +               dmaengine_terminate_all(host->dma_chan_tx);
> +               dma_release_channel(host->dma_chan_rx);
> +               dma_release_channel(host->dma_chan_tx);
>                 iounmap(host->base);
> -               dma_free_coherent(&pdev->dev, PAGE_SIZE, host->sg_cpu, host->sg_dma);
>
>                 clk_put(host->clk);
>
> --
> 2.1.4
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ