[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAPDyKFrMgt9snP2NLbQ6EQ5X7gjQLA+e+TEfqgjzLYTuH+G1OA@mail.gmail.com>
Date: Wed, 10 Feb 2016 20:01:33 +0100
From: Ulf Hansson <ulf.hansson@...aro.org>
To: Matt Redfearn <matt.redfearn@...tec.com>
Cc: linux-mmc <linux-mmc@...r.kernel.org>,
David Daney <david.daney@...ium.com>,
linux-mips@...ux-mips.org,
"linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
"devicetree@...r.kernel.org" <devicetree@...r.kernel.org>,
Zubair.Kakakhel@...tec.com,
Aleksey Makarov <aleksey.makarov@...iumnetworks.com>,
Chandrakala Chavva <cchavva@...iumnetworks.com>,
Aleksey Makarov <aleksey.makarov@...iga.com>,
Leonid Rosenboim <lrosenboim@...iumnetworks.com>,
Peter Swain <pswain@...ium.com>,
Aaron Williams <aaron.williams@...ium.com>
Subject: Re: [PATCH v5] mmc: OCTEON: Add host driver for OCTEON MMC controller
On 10 February 2016 at 18:36, Matt Redfearn <matt.redfearn@...tec.com> wrote:
> From: Aleksey Makarov <aleksey.makarov@...iumnetworks.com>
>
> The OCTEON MMC controller is currently found on cn61XX and cnf71XX
> devices. Device parameters are configured from device tree data.
>
> eMMC, MMC and SD devices are supported.
>
> Tested-by: Aaro Koskinen <aaro.koskinen@....fi>
> Signed-off-by: Chandrakala Chavva <cchavva@...iumnetworks.com>
> Signed-off-by: David Daney <david.daney@...ium.com>
> Signed-off-by: Aleksey Makarov <aleksey.makarov@...iga.com>
> Signed-off-by: Leonid Rosenboim <lrosenboim@...iumnetworks.com>
> Signed-off-by: Peter Swain <pswain@...ium.com>
> Signed-off-by: Aaron Williams <aaron.williams@...ium.com>
> Signed-off-by: Matt Redfearn <matt.redfearn@...tec.com>
> ---
> v5:
> Incoroprate comments from review
> http://patchwork.linux-mips.org/patch/9558/
> - Use standard <bus-width> property instead of <cavium,bus-max-width>.
> - Use standard <max-frequency> property instead of <spi-max-frequency>.
> - Add octeon_mmc_of_parse_legacy function to deal with the above
> properties, since many devices have shipped with those properties
> embedded in firmware.
> - Allow the <vmmc-supply> binding in addition to the legacy
> <gpios-power>.
> - Remove the secondary driver for each slot.
> - Use core gpio cd/wp handling
Seems like you decided to ignore most comments realted to the DT
bindings from the earlier version.
Although, let's discuss this one more time.
Therefore I recomend you to split this patch. DT documentation should
be a separate patch preceeding the actual mmc driver patch.
The DT patch needs to be acked by the DT maintainers.
Until we somewhat agreed on the DT parts, I am going to defer the
in-depth review of the driver code as I have limited bandwidth.
Does that make sense to you?
Kind regards
Uffe
>
> Tested on Rhino labs UTM8, Cavium CN7130.
>
> For reference, the binding in the shipped devices is:
> mmc: mmc@...0000002000 {
> compatible = "cavium,octeon-6130-mmc";
> reg = <0x11800 0x00002000 0x0 0x100>,
> <0x11800 0x00000168 0x0 0x20>;
> #address-cells = <1>;
> #size-cells = <0>;
> /* EMM irq, DMA irq */
> interrupts = <1 19>, <0 63>;
>
> /* The board only has a single MMC slot */
> mmc-slot@2 {
> compatible = "cavium,octeon-6130-mmc-slot";
> reg = <2>;
> voltage-ranges = <3300 3300>;
> spi-max-frequency = <26000000>;
> /* Power on GPIO 8, active high */
> /* power-gpios = <&gpio 8 0>; */
> power-gpios = <&gpio 8 1>;
>
> /* spi-max-frequency = <52000000>; */
> /* bus width can be 1, 4 or 8 */
> cavium,bus-max-width = <8>;
> };
> mmc-slot@0 {
> compatible = "cavium,octeon-6130-mmc-slot";
> reg = <0>;
> voltage-ranges = <3300 3300>;
> spi-max-frequency = <26000000>;
> /* non-removable; */
> bus-width = <8>;
> /* bus width can be 1, 4 or 8 */
> cavium,bus-max-width = <8>;
> };
> };
>
> v3:
> https://lkml.kernel.org/g/<1425567033-31236-1-git-send-email-aleksey.makarov@...iga.com>
>
> Changes in v4:
> - The sparse error discovered by Aaro Koskinen has been fixed
> - Other sparse warnings have been silenced
>
> Changes in v3:
> - Rebased to v4.0-rc2
> - Use gpiod_*() functions instead of legacy gpio
> - Cosmetic changes
>
> Changes in v2: All the fixes suggested by Mark Rutland were implemented:
> - Device tree parsing has been fixed
> - Device tree docs have been fixed
> - Comment about errata workaroud has been added
> ---
> .../devicetree/bindings/mmc/octeon-mmc.txt | 80 ++
> drivers/mmc/host/Kconfig | 10 +
> drivers/mmc/host/Makefile | 1 +
> drivers/mmc/host/octeon_mmc.c | 1409 ++++++++++++++++++++
> 4 files changed, 1500 insertions(+)
> create mode 100644 Documentation/devicetree/bindings/mmc/octeon-mmc.txt
> create mode 100644 drivers/mmc/host/octeon_mmc.c
>
> diff --git a/Documentation/devicetree/bindings/mmc/octeon-mmc.txt b/Documentation/devicetree/bindings/mmc/octeon-mmc.txt
> new file mode 100644
> index 000000000000..a1b20753172f
> --- /dev/null
> +++ b/Documentation/devicetree/bindings/mmc/octeon-mmc.txt
> @@ -0,0 +1,80 @@
> +* OCTEON SD/MMC Host Controller
> +
> +This controller is present on some members of the Cavium OCTEON SoC
> +family, provide an interface for eMMC, MMC and SD devices. There is a
> +single controller that may have several "slots" connected. These
> +slots appear as children of the main controller node.
> +The DMA engine is an integral part of the controller block.
> +
> +1) MMC node
> +
> +Required properties:
> +- compatible : Should be "cavium,octeon-6130-mmc" or "cavium,octeon-7890-mmc"
> +- reg : Two entries:
> + 1) The base address of the MMC controller register bank.
> + 2) The base address of the MMC DMA engine register bank.
> +- interrupts :
> + For "cavium,octeon-6130-mmc": two entries:
> + 1) The MMC controller interrupt line.
> + 2) The MMC DMA engine interrupt line.
> + For "cavium,octeon-7890-mmc": nine entries:
> + 1) The next block transfer of a multiblock transfer has completed (BUF_DONE)
> + 2) Operation completed successfully (CMD_DONE).
> + 3) DMA transfer completed successfully (DMA_DONE).
> + 4) Operation encountered an error (CMD_ERR).
> + 5) DMA transfer encountered an error (DMA_ERR).
> + 6) Switch operation completed successfully (SWITCH_DONE).
> + 7) Switch operation encountered an error (SWITCH_ERR).
> + 8) Internal DMA engine request completion interrupt (DONE).
> + 9) Internal DMA FIFO underflow (FIFO).
> +- #address-cells : Must be <1>
> +- #size-cells : Must be <0>
> +
> +The node contains child nodes for each slot that the platform uses.
> +
> +Example:
> +mmc@...0000002000 {
> + compatible = "cavium,octeon-6130-mmc";
> + reg = <0x11800 0x00002000 0x0 0x100>,
> + <0x11800 0x00000168 0x0 0x20>;
> + #address-cells = <1>;
> + #size-cells = <0>;
> + /* EMM irq, DMA irq */
> + interrupts = <1 19>, <0 63>;
> +
> + [ child node definitions...]
> +};
> +
> +
> +2) Slot nodes
> +Properties in mmc.txt apply to each slot node that the platform uses.
> +
> +Required properties:
> +- reg : The slot number.
> +
> +Optional properties:
> +- cavium,cmd-clk-skew : the amount of delay (in pS) past the clock edge
> + to sample the command pin.
> +- cavium,dat-clk-skew : the amount of delay (in pS) past the clock edge
> + to sample the data pin.
> +
> +Example:
> + mmc@...0000002000 {
> + compatible = "cavium,octeon-6130-mmc";
> + reg = <0x11800 0x00002000 0x0 0x100>,
> + <0x11800 0x00000168 0x0 0x20>;
> + #address-cells = <1>;
> + #size-cells = <0>;
> + /* EMM irq, DMA irq */
> + interrupts = <1 19>, <0 63>;
> +
> + /* The board only has a single MMC slot */
> + mmc-slot@0 {
> + reg = <0>;
> + max-frequency = <20000000>;
> + bus-width = <8>;
> + vmmc-supply = <®_vmmc3>;
> + cd-gpios = <&gpio 9 0>;
> + wp-gpios = <&gpio 10 0>;
> + };
> + };
> diff --git a/drivers/mmc/host/Kconfig b/drivers/mmc/host/Kconfig
> index 1526b8a10b09..8a6c6543e48e 100644
> --- a/drivers/mmc/host/Kconfig
> +++ b/drivers/mmc/host/Kconfig
> @@ -331,6 +331,16 @@ config MMC_SDHCI_IPROC
>
> If unsure, say N.
>
> +config MMC_OCTEON
> + tristate "Cavium OCTEON Multimedia Card Interface support"
> + depends on CAVIUM_OCTEON_SOC
> + help
> + This selects Cavium OCTEON Multimedia card Interface.
> + If you have an OCTEON board with a Multimedia Card slot,
> + say Y or M here.
> +
> + If unsure, say N.
> +
> config MMC_MOXART
> tristate "MOXART SD/MMC Host Controller support"
> depends on ARCH_MOXART && MMC
> diff --git a/drivers/mmc/host/Makefile b/drivers/mmc/host/Makefile
> index 3595f83e89dd..e7cacf13a034 100644
> --- a/drivers/mmc/host/Makefile
> +++ b/drivers/mmc/host/Makefile
> @@ -21,6 +21,7 @@ obj-$(CONFIG_MMC_SDHCI_SPEAR) += sdhci-spear.o
> obj-$(CONFIG_MMC_WBSD) += wbsd.o
> obj-$(CONFIG_MMC_AU1X) += au1xmmc.o
> obj-$(CONFIG_MMC_MTK) += mtk-sd.o
> +obj-$(CONFIG_MMC_OCTEON) += octeon_mmc.o
> obj-$(CONFIG_MMC_OMAP) += omap.o
> obj-$(CONFIG_MMC_OMAP_HS) += omap_hsmmc.o
> obj-$(CONFIG_MMC_ATMELMCI) += atmel-mci.o
> diff --git a/drivers/mmc/host/octeon_mmc.c b/drivers/mmc/host/octeon_mmc.c
> new file mode 100644
> index 000000000000..8e41bffe4d30
> --- /dev/null
> +++ b/drivers/mmc/host/octeon_mmc.c
> @@ -0,0 +1,1409 @@
> +/*
> + * Driver for MMC and SSD cards for Cavium OCTEON SOCs.
> + *
> + * This file is subject to the terms and conditions of the GNU General Public
> + * License. See the file "COPYING" in the main directory of this archive
> + * for more details.
> + *
> + * Copyright (C) 2012-2015 Cavium Inc.
> + */
> +#include <linux/platform_device.h>
> +#include <linux/of_platform.h>
> +#include <linux/scatterlist.h>
> +#include <linux/interrupt.h>
> +#include <linux/blkdev.h>
> +#include <linux/device.h>
> +#include <linux/module.h>
> +#include <linux/delay.h>
> +#include <linux/init.h>
> +#include <linux/clk.h>
> +#include <linux/err.h>
> +#include <linux/io.h>
> +#include <linux/of.h>
> +
> +#include <linux/mmc/card.h>
> +#include <linux/mmc/host.h>
> +#include <linux/mmc/mmc.h>
> +#include <linux/mmc/sd.h>
> +#include <linux/mmc/slot-gpio.h>
> +#include <net/irda/parameters.h>
> +#include <linux/gpio/consumer.h>
> +#include <linux/regulator/consumer.h>
> +
> +#include <asm/byteorder.h>
> +#include <asm/octeon/octeon.h>
> +#include <asm/octeon/cvmx-mio-defs.h>
> +
> +#define DRV_NAME "octeon_mmc"
> +
> +#define OCTEON_MAX_MMC 4
> +
> +#define OCT_MIO_NDF_DMA_CFG 0x00
> +#define OCT_MIO_EMM_DMA_ADR 0x08
> +
> +#define OCT_MIO_EMM_CFG 0x00
> +#define OCT_MIO_EMM_SWITCH 0x48
> +#define OCT_MIO_EMM_DMA 0x50
> +#define OCT_MIO_EMM_CMD 0x58
> +#define OCT_MIO_EMM_RSP_STS 0x60
> +#define OCT_MIO_EMM_RSP_LO 0x68
> +#define OCT_MIO_EMM_RSP_HI 0x70
> +#define OCT_MIO_EMM_INT 0x78
> +#define OCT_MIO_EMM_INT_EN 0x80
> +#define OCT_MIO_EMM_WDOG 0x88
> +#define OCT_MIO_EMM_SAMPLE 0x90
> +#define OCT_MIO_EMM_STS_MASK 0x98
> +#define OCT_MIO_EMM_RCA 0xa0
> +#define OCT_MIO_EMM_BUF_IDX 0xe0
> +#define OCT_MIO_EMM_BUF_DAT 0xe8
> +
> +#define CVMX_MIO_BOOT_CTL CVMX_ADD_IO_SEG(0x00011800000000D0ull)
> +
> +struct octeon_mmc_host {
> + u64 base;
> + u64 ndf_base;
> + u64 emm_cfg;
> + u64 n_minus_one; /* OCTEON II workaround location */
> + int last_slot;
> +
> + struct semaphore mmc_serializer;
> + struct mmc_request *current_req;
> + unsigned int linear_buf_size;
> + void *linear_buf;
> + struct sg_mapping_iter smi;
> + int sg_idx;
> + bool dma_active;
> +
> + struct platform_device *pdev;
> + struct gpio_desc *global_pwr_gpiod;
> + bool dma_err_pending;
> + bool need_bootbus_lock;
> + bool big_dma_addr;
> + bool need_irq_handler_lock;
> + spinlock_t irq_handler_lock;
> +
> + struct octeon_mmc_slot *slot[OCTEON_MAX_MMC];
> +};
> +
> +struct octeon_mmc_slot {
> + struct mmc_host *mmc; /* slot-level mmc_core object */
> + struct octeon_mmc_host *host; /* common hw for all 4 slots */
> +
> + unsigned int clock;
> + unsigned int sclock;
> +
> + u64 cached_switch;
> + u64 cached_rca;
> +
> + unsigned int cmd_cnt; /* sample delay */
> + unsigned int dat_cnt; /* sample delay */
> +
> + int bus_id;
> +
> + /* Legacy property - in future mmc->supply.vmmc should be used */
> + struct gpio_desc *pwr_gpiod;
> +};
> +
> +static int bb_size = 1 << 18;
> +module_param(bb_size, int, S_IRUGO);
> +MODULE_PARM_DESC(bb_size,
> + "Size of DMA linearizing buffer (max transfer size).");
> +
> +static int ddr = 2;
> +module_param(ddr, int, S_IRUGO);
> +MODULE_PARM_DESC(ddr,
> + "enable DoubleDataRate clocking: 0=no, 1=always, 2=at spi-max-frequency/2");
> +
> +#if 1
> +#define octeon_mmc_dbg pr_debug
> +#else
> +static inline void octeon_mmc_dbg(const char *s, ...) { }
> +#endif
> +
> +static void octeon_mmc_acquire_bus(struct octeon_mmc_host *host)
> +{
> + if (host->need_bootbus_lock) {
> + down(&octeon_bootbus_sem);
> + /* On cn70XX switch the mmc unit onto the bus. */
> + if (OCTEON_IS_MODEL(OCTEON_CN70XX))
> + cvmx_write_csr(CVMX_MIO_BOOT_CTL, 0);
> + } else {
> + down(&host->mmc_serializer);
> + }
> +}
> +
> +static void octeon_mmc_release_bus(struct octeon_mmc_host *host)
> +{
> + if (host->need_bootbus_lock)
> + up(&octeon_bootbus_sem);
> + else
> + up(&host->mmc_serializer);
> +}
> +
> +struct octeon_mmc_cr_type {
> + u8 ctype;
> + u8 rtype;
> +};
> +
> +/*
> + * The OCTEON MMC host hardware assumes that all commands have fixed
> + * command and response types. These are correct if MMC devices are
> + * being used. However, non-MMC devices like SD use command and
> + * response types that are unexpected by the host hardware.
> + *
> + * The command and response types can be overridden by supplying an
> + * XOR value that is applied to the type. We calculate the XOR value
> + * from the values in this table and the flags passed from the MMC
> + * core.
> + */
> +static struct octeon_mmc_cr_type octeon_mmc_cr_types[] = {
> + {0, 0}, /* CMD0 */
> + {0, 3}, /* CMD1 */
> + {0, 2}, /* CMD2 */
> + {0, 1}, /* CMD3 */
> + {0, 0}, /* CMD4 */
> + {0, 1}, /* CMD5 */
> + {0, 1}, /* CMD6 */
> + {0, 1}, /* CMD7 */
> + {1, 1}, /* CMD8 */
> + {0, 2}, /* CMD9 */
> + {0, 2}, /* CMD10 */
> + {1, 1}, /* CMD11 */
> + {0, 1}, /* CMD12 */
> + {0, 1}, /* CMD13 */
> + {1, 1}, /* CMD14 */
> + {0, 0}, /* CMD15 */
> + {0, 1}, /* CMD16 */
> + {1, 1}, /* CMD17 */
> + {1, 1}, /* CMD18 */
> + {3, 1}, /* CMD19 */
> + {2, 1}, /* CMD20 */
> + {0, 0}, /* CMD21 */
> + {0, 0}, /* CMD22 */
> + {0, 1}, /* CMD23 */
> + {2, 1}, /* CMD24 */
> + {2, 1}, /* CMD25 */
> + {2, 1}, /* CMD26 */
> + {2, 1}, /* CMD27 */
> + {0, 1}, /* CMD28 */
> + {0, 1}, /* CMD29 */
> + {1, 1}, /* CMD30 */
> + {1, 1}, /* CMD31 */
> + {0, 0}, /* CMD32 */
> + {0, 0}, /* CMD33 */
> + {0, 0}, /* CMD34 */
> + {0, 1}, /* CMD35 */
> + {0, 1}, /* CMD36 */
> + {0, 0}, /* CMD37 */
> + {0, 1}, /* CMD38 */
> + {0, 4}, /* CMD39 */
> + {0, 5}, /* CMD40 */
> + {0, 0}, /* CMD41 */
> + {2, 1}, /* CMD42 */
> + {0, 0}, /* CMD43 */
> + {0, 0}, /* CMD44 */
> + {0, 0}, /* CMD45 */
> + {0, 0}, /* CMD46 */
> + {0, 0}, /* CMD47 */
> + {0, 0}, /* CMD48 */
> + {0, 0}, /* CMD49 */
> + {0, 0}, /* CMD50 */
> + {0, 0}, /* CMD51 */
> + {0, 0}, /* CMD52 */
> + {0, 0}, /* CMD53 */
> + {0, 0}, /* CMD54 */
> + {0, 1}, /* CMD55 */
> + {0xff, 0xff}, /* CMD56 */
> + {0, 0}, /* CMD57 */
> + {0, 0}, /* CMD58 */
> + {0, 0}, /* CMD59 */
> + {0, 0}, /* CMD60 */
> + {0, 0}, /* CMD61 */
> + {0, 0}, /* CMD62 */
> + {0, 0} /* CMD63 */
> +};
> +
> +struct octeon_mmc_cr_mods {
> + u8 ctype_xor;
> + u8 rtype_xor;
> +};
> +
> +/*
> + * The functions below are used for the EMMC-17978 workaround.
> + *
> + * Due to an imperfection in the design of the MMC bus hardware,
> + * the 2nd to last cache block of a DMA read must be locked into the L2 Cache.
> + * Otherwise, data corruption may occur.
> + */
> +
> +static inline void *phys_to_ptr(u64 address)
> +{
> + return (void *)(address | (1ull<<63)); /* XKPHYS */
> +}
> +
> +/**
> + * Lock a single line into L2. The line is zeroed before locking
> + * to make sure no dram accesses are made.
> + *
> + * @addr Physical address to lock
> + */
> +static void l2c_lock_line(u64 addr)
> +{
> + char *addr_ptr = phys_to_ptr(addr);
> +
> + asm volatile (
> + "cache 31, %[line]" /* Unlock the line */
> + :: [line] "m" (*addr_ptr));
> +}
> +
> +/**
> + * Locks a memory region in the L2 cache
> + *
> + * @start - start address to begin locking
> + * @len - length in bytes to lock
> + */
> +static void l2c_lock_mem_region(u64 start, u64 len)
> +{
> + u64 end;
> +
> + /* Round start/end to cache line boundaries */
> + end = ALIGN(start + len - 1, CVMX_CACHE_LINE_SIZE);
> + start = ALIGN(start, CVMX_CACHE_LINE_SIZE);
> +
> + while (start <= end) {
> + l2c_lock_line(start);
> + start += CVMX_CACHE_LINE_SIZE;
> + }
> + asm volatile("sync");
> +}
> +
> +/**
> + * Unlock a single line in the L2 cache.
> + *
> + * @addr Physical address to unlock
> + *
> + * Return Zero on success
> + */
> +static void l2c_unlock_line(u64 addr)
> +{
> + char *addr_ptr = phys_to_ptr(addr);
> +
> + asm volatile (
> + "cache 23, %[line]" /* Unlock the line */
> + :: [line] "m" (*addr_ptr));
> +}
> +
> +/**
> + * Unlock a memory region in the L2 cache
> + *
> + * @start - start address to unlock
> + * @len - length to unlock in bytes
> + */
> +static void l2c_unlock_mem_region(u64 start, u64 len)
> +{
> + u64 end;
> +
> + /* Round start/end to cache line boundaries */
> + end = ALIGN(start + len - 1, CVMX_CACHE_LINE_SIZE);
> + start = ALIGN(start, CVMX_CACHE_LINE_SIZE);
> +
> + while (start <= end) {
> + l2c_unlock_line(start);
> + start += CVMX_CACHE_LINE_SIZE;
> + }
> +}
> +
> +static struct octeon_mmc_cr_mods octeon_mmc_get_cr_mods(struct mmc_command *cmd)
> +{
> + struct octeon_mmc_cr_type *cr;
> + u8 desired_ctype, hardware_ctype;
> + u8 desired_rtype, hardware_rtype;
> + struct octeon_mmc_cr_mods r;
> +
> + desired_ctype = desired_rtype = 0;
> +
> + cr = octeon_mmc_cr_types + (cmd->opcode & 0x3f);
> + hardware_ctype = cr->ctype;
> + hardware_rtype = cr->rtype;
> + if (cmd->opcode == 56) { /* CMD56 GEN_CMD */
> + hardware_ctype = (cmd->arg & 1) ? 1 : 2;
> + }
> +
> + switch (mmc_cmd_type(cmd)) {
> + case MMC_CMD_ADTC:
> + desired_ctype = (cmd->data->flags & MMC_DATA_WRITE) ? 2 : 1;
> + break;
> + case MMC_CMD_AC:
> + case MMC_CMD_BC:
> + case MMC_CMD_BCR:
> + desired_ctype = 0;
> + break;
> + }
> +
> + switch (mmc_resp_type(cmd)) {
> + case MMC_RSP_NONE:
> + desired_rtype = 0;
> + break;
> + case MMC_RSP_R1:/* MMC_RSP_R5, MMC_RSP_R6, MMC_RSP_R7 */
> + case MMC_RSP_R1B:
> + desired_rtype = 1;
> + break;
> + case MMC_RSP_R2:
> + desired_rtype = 2;
> + break;
> + case MMC_RSP_R3: /* MMC_RSP_R4 */
> + desired_rtype = 3;
> + break;
> + }
> + r.ctype_xor = desired_ctype ^ hardware_ctype;
> + r.rtype_xor = desired_rtype ^ hardware_rtype;
> + return r;
> +}
> +
> +static bool octeon_mmc_switch_val_changed(struct octeon_mmc_slot *slot,
> + u64 new_val)
> +{
> + /* Match BUS_ID, HS_TIMING, BUS_WIDTH, POWER_CLASS, CLK_HI, CLK_LO */
> + u64 m = 0x3001070fffffffffull;
> +
> + return (slot->cached_switch & m) != (new_val & m);
> +}
> +
> +static unsigned int octeon_mmc_timeout_to_wdog(struct octeon_mmc_slot *slot,
> + unsigned int ns)
> +{
> + u64 bt = (u64)slot->clock * (u64)ns;
> +
> + return (unsigned int)(bt / 1000000000);
> +}
> +
> +static irqreturn_t octeon_mmc_interrupt(int irq, void *dev_id)
> +{
> + struct octeon_mmc_host *host = dev_id;
> + union cvmx_mio_emm_int emm_int;
> + struct mmc_request *req;
> + bool host_done;
> + union cvmx_mio_emm_rsp_sts rsp_sts;
> + unsigned long flags = 0;
> +
> + if (host->need_irq_handler_lock)
> + spin_lock_irqsave(&host->irq_handler_lock, flags);
> + else
> + __acquire(&host->irq_handler_lock);
> + emm_int.u64 = cvmx_read_csr(host->base + OCT_MIO_EMM_INT);
> + req = host->current_req;
> + cvmx_write_csr(host->base + OCT_MIO_EMM_INT, emm_int.u64);
> +
> + octeon_mmc_dbg("Got interrupt: EMM_INT = 0x%llx\n", emm_int.u64);
> +
> + if (!req)
> + goto out;
> +
> + rsp_sts.u64 = cvmx_read_csr(host->base + OCT_MIO_EMM_RSP_STS);
> + octeon_mmc_dbg("octeon_mmc_interrupt MIO_EMM_RSP_STS 0x%llx\n",
> + rsp_sts.u64);
> +
> + if (host->dma_err_pending) {
> + host->current_req = NULL;
> + host->dma_err_pending = false;
> + req->done(req);
> + host_done = true;
> + goto no_req_done;
> + }
> +
> + if (!host->dma_active && emm_int.s.buf_done && req->data) {
> + unsigned int type = (rsp_sts.u64 >> 7) & 3;
> +
> + if (type == 1) {
> + /* Read */
> + int dbuf = rsp_sts.s.dbuf;
> + struct sg_mapping_iter *smi = &host->smi;
> + unsigned int data_len =
> + req->data->blksz * req->data->blocks;
> + unsigned int bytes_xfered;
> + u64 dat = 0;
> + int shift = -1;
> +
> + /* Auto inc from offset zero */
> + cvmx_write_csr(host->base + OCT_MIO_EMM_BUF_IDX,
> + (u64)(0x10000 | (dbuf << 6)));
> +
> + for (bytes_xfered = 0; bytes_xfered < data_len;) {
> + if (smi->consumed >= smi->length) {
> + if (!sg_miter_next(smi))
> + break;
> + smi->consumed = 0;
> + }
> + if (shift < 0) {
> + dat = cvmx_read_csr(host->base +
> + OCT_MIO_EMM_BUF_DAT);
> + shift = 56;
> + }
> +
> + while (smi->consumed < smi->length &&
> + shift >= 0) {
> + ((u8 *)(smi->addr))[smi->consumed] =
> + (dat >> shift) & 0xff;
> + bytes_xfered++;
> + smi->consumed++;
> + shift -= 8;
> + }
> + }
> + sg_miter_stop(smi);
> + req->data->bytes_xfered = bytes_xfered;
> + req->data->error = 0;
> + } else if (type == 2) {
> + /* write */
> + req->data->bytes_xfered = req->data->blksz *
> + req->data->blocks;
> + req->data->error = 0;
> + }
> + }
> + host_done = emm_int.s.cmd_done || emm_int.s.dma_done ||
> + emm_int.s.cmd_err || emm_int.s.dma_err;
> + if (host_done && req->done) {
> + if (rsp_sts.s.rsp_bad_sts ||
> + rsp_sts.s.rsp_crc_err ||
> + rsp_sts.s.rsp_timeout ||
> + rsp_sts.s.blk_crc_err ||
> + rsp_sts.s.blk_timeout ||
> + rsp_sts.s.dbuf_err) {
> + req->cmd->error = -EILSEQ;
> + } else {
> + req->cmd->error = 0;
> + }
> +
> + if (host->dma_active && req->data) {
> + req->data->error = 0;
> + req->data->bytes_xfered = req->data->blocks *
> + req->data->blksz;
> + if (!(req->data->flags & MMC_DATA_WRITE) &&
> + req->data->sg_len > 1) {
> + size_t r = sg_copy_from_buffer(req->data->sg,
> + req->data->sg_len, host->linear_buf,
> + req->data->bytes_xfered);
> + WARN_ON(r != req->data->bytes_xfered);
> + }
> + }
> + if (rsp_sts.s.rsp_val) {
> + u64 rsp_hi;
> + u64 rsp_lo = cvmx_read_csr(
> + host->base + OCT_MIO_EMM_RSP_LO);
> +
> + switch (rsp_sts.s.rsp_type) {
> + case 1:
> + case 3:
> + req->cmd->resp[0] = (rsp_lo >> 8) & 0xffffffff;
> + req->cmd->resp[1] = 0;
> + req->cmd->resp[2] = 0;
> + req->cmd->resp[3] = 0;
> + break;
> + case 2:
> + req->cmd->resp[3] = rsp_lo & 0xffffffff;
> + req->cmd->resp[2] = (rsp_lo >> 32) & 0xffffffff;
> + rsp_hi = cvmx_read_csr(host->base +
> + OCT_MIO_EMM_RSP_HI);
> + req->cmd->resp[1] = rsp_hi & 0xffffffff;
> + req->cmd->resp[0] = (rsp_hi >> 32) & 0xffffffff;
> + break;
> + default:
> + octeon_mmc_dbg("octeon_mmc_interrupt unhandled rsp_val %d\n",
> + rsp_sts.s.rsp_type);
> + break;
> + }
> + octeon_mmc_dbg("octeon_mmc_interrupt resp %08x %08x %08x %08x\n",
> + req->cmd->resp[0], req->cmd->resp[1],
> + req->cmd->resp[2], req->cmd->resp[3]);
> + }
> + if (emm_int.s.dma_err && rsp_sts.s.dma_pend) {
> + /* Try to clean up failed DMA */
> + union cvmx_mio_emm_dma emm_dma;
> +
> + emm_dma.u64 =
> + cvmx_read_csr(host->base + OCT_MIO_EMM_DMA);
> + emm_dma.s.dma_val = 1;
> + emm_dma.s.dat_null = 1;
> + emm_dma.s.bus_id = rsp_sts.s.bus_id;
> + cvmx_write_csr(host->base + OCT_MIO_EMM_DMA,
> + emm_dma.u64);
> + host->dma_err_pending = true;
> + host_done = false;
> + goto no_req_done;
> + }
> +
> + host->current_req = NULL;
> + req->done(req);
> + }
> +no_req_done:
> + if (host->n_minus_one) {
> + l2c_unlock_mem_region(host->n_minus_one, 512);
> + host->n_minus_one = 0;
> + }
> + if (host_done)
> + octeon_mmc_release_bus(host);
> +out:
> + if (host->need_irq_handler_lock)
> + spin_unlock_irqrestore(&host->irq_handler_lock, flags);
> + else
> + __release(&host->irq_handler_lock);
> + return IRQ_RETVAL(emm_int.u64 != 0);
> +}
> +
> +static void octeon_mmc_switch_to(struct octeon_mmc_slot *slot)
> +{
> + struct octeon_mmc_host *host = slot->host;
> + struct octeon_mmc_slot *old_slot;
> + union cvmx_mio_emm_switch sw;
> + union cvmx_mio_emm_sample samp;
> +
> + if (slot->bus_id == host->last_slot)
> + goto out;
> +
> + if (host->last_slot >= 0 && host->slot[host->last_slot]) {
> + old_slot = host->slot[host->last_slot];
> + old_slot->cached_switch =
> + cvmx_read_csr(host->base + OCT_MIO_EMM_SWITCH);
> + old_slot->cached_rca =
> + cvmx_read_csr(host->base + OCT_MIO_EMM_RCA);
> + }
> + cvmx_write_csr(host->base + OCT_MIO_EMM_RCA, slot->cached_rca);
> + sw.u64 = slot->cached_switch;
> + sw.s.bus_id = 0;
> + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, sw.u64);
> + sw.s.bus_id = slot->bus_id;
> + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, sw.u64);
> +
> + samp.u64 = 0;
> + samp.s.cmd_cnt = slot->cmd_cnt;
> + samp.s.dat_cnt = slot->dat_cnt;
> + cvmx_write_csr(host->base + OCT_MIO_EMM_SAMPLE, samp.u64);
> +out:
> + host->last_slot = slot->bus_id;
> +}
> +
> +static void octeon_mmc_dma_request(struct mmc_host *mmc,
> + struct mmc_request *mrq)
> +{
> + struct octeon_mmc_slot *slot;
> + struct octeon_mmc_host *host;
> + struct mmc_command *cmd;
> + struct mmc_data *data;
> + union cvmx_mio_emm_int emm_int;
> + union cvmx_mio_emm_dma emm_dma;
> + union cvmx_mio_ndf_dma_cfg dma_cfg;
> +
> + cmd = mrq->cmd;
> + if (mrq->data == NULL || mrq->data->sg == NULL || !mrq->data->sg_len ||
> + mrq->stop == NULL || mrq->stop->opcode != MMC_STOP_TRANSMISSION) {
> + dev_err(&mmc->card->dev,
> + "Error: octeon_mmc_dma_request no data\n");
> + cmd->error = -EINVAL;
> + if (mrq->done)
> + mrq->done(mrq);
> + return;
> + }
> +
> + slot = mmc_priv(mmc);
> + host = slot->host;
> +
> + /* Only a single user of the bootbus at a time. */
> + octeon_mmc_acquire_bus(host);
> +
> + octeon_mmc_switch_to(slot);
> +
> + data = mrq->data;
> +
> + if (data->timeout_ns) {
> + cvmx_write_csr(host->base + OCT_MIO_EMM_WDOG,
> + octeon_mmc_timeout_to_wdog(slot, data->timeout_ns));
> + octeon_mmc_dbg("OCT_MIO_EMM_WDOG %llu\n",
> + cvmx_read_csr(host->base + OCT_MIO_EMM_WDOG));
> + }
> +
> + WARN_ON(host->current_req);
> + host->current_req = mrq;
> +
> + host->sg_idx = 0;
> +
> + WARN_ON(data->blksz * data->blocks > host->linear_buf_size);
> +
> + if ((data->flags & MMC_DATA_WRITE) && data->sg_len > 1) {
> + size_t r = sg_copy_to_buffer(data->sg, data->sg_len,
> + host->linear_buf, data->blksz * data->blocks);
> + WARN_ON(data->blksz * data->blocks != r);
> + }
> +
> + dma_cfg.u64 = 0;
> + dma_cfg.s.en = 1;
> + dma_cfg.s.rw = (data->flags & MMC_DATA_WRITE) ? 1 : 0;
> +#ifdef __LITTLE_ENDIAN
> + dma_cfg.s.endian = 1;
> +#endif
> + dma_cfg.s.size = ((data->blksz * data->blocks) / 8) - 1;
> + if (!host->big_dma_addr) {
> + if (data->sg_len > 1)
> + dma_cfg.s.adr = virt_to_phys(host->linear_buf);
> + else
> + dma_cfg.s.adr = sg_phys(data->sg);
> + }
> + cvmx_write_csr(host->ndf_base + OCT_MIO_NDF_DMA_CFG, dma_cfg.u64);
> + octeon_mmc_dbg("MIO_NDF_DMA_CFG: %016llx\n",
> + (unsigned long long)dma_cfg.u64);
> + if (host->big_dma_addr) {
> + u64 addr;
> +
> + if (data->sg_len > 1)
> + addr = virt_to_phys(host->linear_buf);
> + else
> + addr = sg_phys(data->sg);
> + cvmx_write_csr(host->ndf_base + OCT_MIO_EMM_DMA_ADR, addr);
> + octeon_mmc_dbg("MIO_EMM_DMA_ADR: %016llx\n",
> + (unsigned long long)addr);
> + }
> +
> + emm_dma.u64 = 0;
> + emm_dma.s.bus_id = slot->bus_id;
> + emm_dma.s.dma_val = 1;
> + emm_dma.s.sector = mmc_card_blockaddr(mmc->card) ? 1 : 0;
> + emm_dma.s.rw = (data->flags & MMC_DATA_WRITE) ? 1 : 0;
> + if (mmc_card_mmc(mmc->card) ||
> + (mmc_card_sd(mmc->card) &&
> + (mmc->card->scr.cmds & SD_SCR_CMD23_SUPPORT)))
> + emm_dma.s.multi = 1;
> + emm_dma.s.block_cnt = data->blocks;
> + emm_dma.s.card_addr = cmd->arg;
> +
> + emm_int.u64 = 0;
> + emm_int.s.dma_done = 1;
> + emm_int.s.cmd_err = 1;
> + emm_int.s.dma_err = 1;
> + /* Clear the bit. */
> + cvmx_write_csr(host->base + OCT_MIO_EMM_INT, emm_int.u64);
> + cvmx_write_csr(host->base + OCT_MIO_EMM_INT_EN, emm_int.u64);
> + host->dma_active = true;
> +
> + if ((OCTEON_IS_MODEL(OCTEON_CN6XXX) ||
> + OCTEON_IS_MODEL(OCTEON_CNF7XXX)) &&
> + cmd->opcode == MMC_WRITE_MULTIPLE_BLOCK &&
> + (data->blksz * data->blocks) > 1024) {
> + host->n_minus_one = dma_cfg.s.adr +
> + (data->blksz * data->blocks) - 1024;
> + l2c_lock_mem_region(host->n_minus_one, 512);
> + }
> +
> + if (mmc->card && mmc_card_sd(mmc->card))
> + cvmx_write_csr(host->base + OCT_MIO_EMM_STS_MASK,
> + 0x00b00000ull);
> + else
> + cvmx_write_csr(host->base + OCT_MIO_EMM_STS_MASK,
> + 0xe4f90080ull);
> + cvmx_write_csr(host->base + OCT_MIO_EMM_DMA, emm_dma.u64);
> + octeon_mmc_dbg("MIO_EMM_DMA: %llx\n", emm_dma.u64);
> +}
> +
> +static void octeon_mmc_request(struct mmc_host *mmc, struct mmc_request *mrq)
> +{
> + struct octeon_mmc_slot *slot;
> + struct octeon_mmc_host *host;
> + struct mmc_command *cmd;
> + union cvmx_mio_emm_int emm_int;
> + union cvmx_mio_emm_cmd emm_cmd;
> + struct octeon_mmc_cr_mods mods;
> +
> + cmd = mrq->cmd;
> +
> + if (cmd->opcode == MMC_READ_MULTIPLE_BLOCK ||
> + cmd->opcode == MMC_WRITE_MULTIPLE_BLOCK) {
> + octeon_mmc_dma_request(mmc, mrq);
> + return;
> + }
> +
> + mods = octeon_mmc_get_cr_mods(cmd);
> +
> + slot = mmc_priv(mmc);
> + host = slot->host;
> +
> + /* Only a single user of the bootbus at a time. */
> + octeon_mmc_acquire_bus(host);
> +
> + octeon_mmc_switch_to(slot);
> +
> + WARN_ON(host->current_req);
> + host->current_req = mrq;
> +
> + emm_int.u64 = 0;
> + emm_int.s.cmd_done = 1;
> + emm_int.s.cmd_err = 1;
> + if (cmd->data) {
> + octeon_mmc_dbg("command has data\n");
> + if (cmd->data->flags & MMC_DATA_READ) {
> + sg_miter_start(&host->smi, mrq->data->sg,
> + mrq->data->sg_len,
> + SG_MITER_ATOMIC | SG_MITER_TO_SG);
> + } else {
> + struct sg_mapping_iter *smi = &host->smi;
> + unsigned int data_len =
> + mrq->data->blksz * mrq->data->blocks;
> + unsigned int bytes_xfered;
> + u64 dat = 0;
> + int shift = 56;
> + /*
> + * Copy data to the xmit buffer before
> + * issuing the command
> + */
> + sg_miter_start(smi, mrq->data->sg,
> + mrq->data->sg_len, SG_MITER_FROM_SG);
> + /* Auto inc from offset zero, dbuf zero */
> + cvmx_write_csr(host->base + OCT_MIO_EMM_BUF_IDX,
> + 0x10000ull);
> +
> + for (bytes_xfered = 0; bytes_xfered < data_len;) {
> + if (smi->consumed >= smi->length) {
> + if (!sg_miter_next(smi))
> + break;
> + smi->consumed = 0;
> + }
> +
> + while (smi->consumed < smi->length &&
> + shift >= 0) {
> +
> + dat |= (u64)(((u8 *)(smi->addr))
> + [smi->consumed]) << shift;
> + bytes_xfered++;
> + smi->consumed++;
> + shift -= 8;
> + }
> + if (shift < 0) {
> + cvmx_write_csr(host->base +
> + OCT_MIO_EMM_BUF_DAT, dat);
> + shift = 56;
> + dat = 0;
> + }
> + }
> + sg_miter_stop(smi);
> + }
> + if (cmd->data->timeout_ns) {
> + cvmx_write_csr(host->base + OCT_MIO_EMM_WDOG,
> + octeon_mmc_timeout_to_wdog(slot,
> + cmd->data->timeout_ns));
> + octeon_mmc_dbg("OCT_MIO_EMM_WDOG %llu\n",
> + cvmx_read_csr(host->base +
> + OCT_MIO_EMM_WDOG));
> + }
> + } else {
> + cvmx_write_csr(host->base + OCT_MIO_EMM_WDOG,
> + ((u64)slot->clock * 850ull) / 1000ull);
> + octeon_mmc_dbg("OCT_MIO_EMM_WDOG %llu\n",
> + cvmx_read_csr(host->base + OCT_MIO_EMM_WDOG));
> + }
> + /* Clear the bit. */
> + cvmx_write_csr(host->base + OCT_MIO_EMM_INT, emm_int.u64);
> + cvmx_write_csr(host->base + OCT_MIO_EMM_INT_EN, emm_int.u64);
> + host->dma_active = false;
> +
> + emm_cmd.u64 = 0;
> + emm_cmd.s.cmd_val = 1;
> + emm_cmd.s.ctype_xor = mods.ctype_xor;
> + emm_cmd.s.rtype_xor = mods.rtype_xor;
> + if (mmc_cmd_type(cmd) == MMC_CMD_ADTC)
> + emm_cmd.s.offset = 64 -
> + ((cmd->data->blksz * cmd->data->blocks) / 8);
> + emm_cmd.s.bus_id = slot->bus_id;
> + emm_cmd.s.cmd_idx = cmd->opcode;
> + emm_cmd.s.arg = cmd->arg;
> + cvmx_write_csr(host->base + OCT_MIO_EMM_STS_MASK, 0);
> + cvmx_write_csr(host->base + OCT_MIO_EMM_CMD, emm_cmd.u64);
> + octeon_mmc_dbg("MIO_EMM_CMD: %llx\n", emm_cmd.u64);
> +}
> +
> +static void octeon_mmc_reset_bus(struct octeon_mmc_slot *slot)
> +{
> + union cvmx_mio_emm_cfg emm_cfg;
> + union cvmx_mio_emm_switch emm_switch;
> + u64 wdog = 0;
> +
> + emm_cfg.u64 = cvmx_read_csr(slot->host->base + OCT_MIO_EMM_CFG);
> + emm_switch.u64 = cvmx_read_csr(slot->host->base + OCT_MIO_EMM_SWITCH);
> + wdog = cvmx_read_csr(slot->host->base + OCT_MIO_EMM_WDOG);
> +
> + emm_switch.s.switch_exe = 0;
> + emm_switch.s.switch_err0 = 0;
> + emm_switch.s.switch_err1 = 0;
> + emm_switch.s.switch_err2 = 0;
> + emm_switch.s.bus_id = 0;
> + cvmx_write_csr(slot->host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64);
> + emm_switch.s.bus_id = slot->bus_id;
> + cvmx_write_csr(slot->host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64);
> +
> + slot->cached_switch = emm_switch.u64;
> +
> + msleep(20);
> +
> + cvmx_write_csr(slot->host->base + OCT_MIO_EMM_WDOG, wdog);
> +}
> +
> +static void octeon_mmc_set_ios(struct mmc_host *mmc, struct mmc_ios *ios)
> +{
> + struct octeon_mmc_slot *slot;
> + struct octeon_mmc_host *host;
> + int bus_width;
> + int clock;
> + bool ddr_clock;
> + int hs_timing;
> + int power_class = 10;
> + int clk_period;
> + int timeout = 2000;
> + union cvmx_mio_emm_switch emm_switch;
> + union cvmx_mio_emm_rsp_sts emm_sts;
> +
> + slot = mmc_priv(mmc);
> + host = slot->host;
> +
> + /* Only a single user of the bootbus at a time. */
> + octeon_mmc_acquire_bus(host);
> +
> + octeon_mmc_switch_to(slot);
> +
> + octeon_mmc_dbg("Calling set_ios: slot: clk = 0x%x, bus_width = %d\n",
> + slot->clock, (mmc->caps & MMC_CAP_8_BIT_DATA) ? 8 :
> + (mmc->caps & MMC_CAP_4_BIT_DATA) ? 4 : 1);
> + octeon_mmc_dbg("Calling set_ios: ios: clk = 0x%x, vdd = %u, bus_width = %u, power_mode = %u, timing = %u\n",
> + ios->clock, ios->vdd, ios->bus_width, ios->power_mode,
> + ios->timing);
> + octeon_mmc_dbg("Calling set_ios: mmc: caps = 0x%x, bus_width = %d\n",
> + mmc->caps, mmc->ios.bus_width);
> +
> + /*
> + * Reset the chip on each power off
> + */
> + if (ios->power_mode == MMC_POWER_OFF) {
> + octeon_mmc_reset_bus(slot);
> + if (!IS_ERR(mmc->supply.vmmc))
> + regulator_disable(mmc->supply.vmmc);
> + else /* Legacy power GPIO */
> + gpiod_set_value_cansleep(slot->pwr_gpiod, 0);
> + } else {
> + if (!IS_ERR(mmc->supply.vmmc))
> + regulator_enable(mmc->supply.vmmc);
> + else /* Legacy power GPIO */
> + gpiod_set_value_cansleep(slot->pwr_gpiod, 1);
> + }
> +
> + switch (ios->bus_width) {
> + case MMC_BUS_WIDTH_8:
> + bus_width = 2;
> + break;
> + case MMC_BUS_WIDTH_4:
> + bus_width = 1;
> + break;
> + case MMC_BUS_WIDTH_1:
> + bus_width = 0;
> + break;
> + default:
> + octeon_mmc_dbg("unknown bus width %d\n", ios->bus_width);
> + bus_width = 0;
> + break;
> + }
> +
> + hs_timing = (ios->timing == MMC_TIMING_MMC_HS);
> + ddr_clock = (bus_width && ios->timing >= MMC_TIMING_UHS_DDR50);
> +
> + if (ddr_clock)
> + bus_width |= 4;
> +
> + if (ios->clock) {
> + slot->clock = ios->clock;
> +
> + clock = slot->clock;
> +
> + if (clock > 52000000)
> + clock = 52000000;
> +
> + clk_period = (octeon_get_io_clock_rate() + clock - 1) /
> + (2 * clock);
> +
> + /* until clock-renengotiate-on-CRC is in */
> + if (ddr_clock && ddr > 1)
> + clk_period *= 2;
> +
> + emm_switch.u64 = 0;
> + emm_switch.s.hs_timing = hs_timing;
> + emm_switch.s.bus_width = bus_width;
> + emm_switch.s.power_class = power_class;
> + emm_switch.s.clk_hi = clk_period;
> + emm_switch.s.clk_lo = clk_period;
> +
> + if (!octeon_mmc_switch_val_changed(slot, emm_switch.u64)) {
> + octeon_mmc_dbg("No change from 0x%llx mio_emm_switch, returning.\n",
> + emm_switch.u64);
> + goto out;
> + }
> +
> + octeon_mmc_dbg("Writing 0x%llx to mio_emm_wdog\n",
> + ((u64)clock * 850ull) / 1000ull);
> + cvmx_write_csr(host->base + OCT_MIO_EMM_WDOG,
> + ((u64)clock * 850ull) / 1000ull);
> + octeon_mmc_dbg("Writing 0x%llx to mio_emm_switch\n",
> + emm_switch.u64);
> +
> + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64);
> + emm_switch.s.bus_id = slot->bus_id;
> + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64);
> + slot->cached_switch = emm_switch.u64;
> +
> + do {
> + emm_sts.u64 =
> + cvmx_read_csr(host->base + OCT_MIO_EMM_RSP_STS);
> + if (!emm_sts.s.switch_val)
> + break;
> + udelay(100);
> + } while (timeout-- > 0);
> +
> + if (timeout <= 0) {
> + octeon_mmc_dbg("switch command timed out, status=0x%llx\n",
> + emm_sts.u64);
> + goto out;
> + }
> + }
> +out:
> + octeon_mmc_release_bus(host);
> +}
> +
> +static const struct mmc_host_ops octeon_mmc_ops = {
> + .request = octeon_mmc_request,
> + .set_ios = octeon_mmc_set_ios,
> + .get_ro = mmc_gpio_get_ro,
> + .get_cd = mmc_gpio_get_cd,
> +};
> +
> +static void octeon_mmc_set_clock(struct octeon_mmc_slot *slot,
> + unsigned int clock)
> +{
> + struct mmc_host *mmc = slot->mmc;
> +
> + clock = min(clock, mmc->f_max);
> + clock = max(clock, mmc->f_min);
> + slot->clock = clock;
> +}
> +
> +static int octeon_mmc_initlowlevel(struct octeon_mmc_slot *slot)
> +{
> + union cvmx_mio_emm_switch emm_switch;
> + struct octeon_mmc_host *host = slot->host;
> +
> + host->emm_cfg |= 1ull << slot->bus_id;
> + cvmx_write_csr(slot->host->base + OCT_MIO_EMM_CFG, host->emm_cfg);
> + octeon_mmc_set_clock(slot, 400000);
> +
> + /* Program initial clock speed and power */
> + emm_switch.u64 = 0;
> + emm_switch.s.power_class = 10;
> + emm_switch.s.clk_hi = (slot->sclock / slot->clock) / 2;
> + emm_switch.s.clk_lo = (slot->sclock / slot->clock) / 2;
> +
> + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64);
> + emm_switch.s.bus_id = slot->bus_id;
> + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64);
> + slot->cached_switch = emm_switch.u64;
> +
> + cvmx_write_csr(host->base + OCT_MIO_EMM_WDOG,
> + ((u64)slot->clock * 850ull) / 1000ull);
> + cvmx_write_csr(host->base + OCT_MIO_EMM_STS_MASK, 0xe4f90080ull);
> + cvmx_write_csr(host->base + OCT_MIO_EMM_RCA, 1);
> + return 0;
> +}
> +
> +static int octeon_mmc_of_copy_legacy_u32(struct device_node *node,
> + const char *legacy_name,
> + const char *new_name)
> +{
> + u32 value;
> + int ret;
> +
> + ret = of_property_read_u32(node, legacy_name, &value);
> + if (!ret) {
> + /* Found legacy - set generic property */
> + struct property *new_p;
> + u32 *new_v;
> +
> + pr_warn(FW_WARN "%s: Legacy property '%s'. Please remove\n",
> + node->full_name, legacy_name);
> +
> + new_p = kzalloc(sizeof(*new_p), GFP_KERNEL);
> + new_v = kzalloc(sizeof(u32), GFP_KERNEL);
> + if (!new_p || !new_v)
> + return -ENOMEM;
> +
> + *new_v = value;
> + new_p->name = kstrdup(new_name, GFP_KERNEL);
> + new_p->length = sizeof(u32);
> + new_p->value = new_v;
> +
> + of_update_property(node, new_p);
> + }
> + return 0;
> +}
> +
> +/*
> + * This function parses the legacy device tree that may be found in devices
> + * shipped before the driver was upstreamed. Future devices should not require
> + * it as standard bindings should be used
> + */
> +static int octeon_mmc_of_parse_legacy(struct device *dev,
> + struct device_node *node,
> + struct octeon_mmc_slot *slot)
> +{
> + int ret;
> +
> + ret = octeon_mmc_of_copy_legacy_u32(node, "cavium,bus-max-width",
> + "bus-width");
> + if (ret)
> + return ret;
> +
> + ret = octeon_mmc_of_copy_legacy_u32(node, "spi-max-frequency",
> + "max-frequency");
> + if (ret)
> + return ret;
> +
> + slot->pwr_gpiod = devm_gpiod_get_optional(dev, "power", GPIOD_OUT_LOW);
> + if (!IS_ERR(slot->pwr_gpiod)) {
> + pr_warn(FW_WARN "%s: Legacy property '%s'. Please remove\n",
> + node->full_name, "gpios-power");
> + }
> +
> + return 0;
> +}
> +
> +static int octeon_mmc_slot_probe(struct platform_device *slot_pdev,
> + struct octeon_mmc_host *host)
> +{
> + struct mmc_host *mmc;
> + struct octeon_mmc_slot *slot;
> + struct device *dev = &slot_pdev->dev;
> + struct device_node *node = slot_pdev->dev.of_node;
> + u32 id, cmd_skew, dat_skew;
> + u64 clock_period;
> + int ret;
> +
> + ret = of_property_read_u32(node, "reg", &id);
> + if (ret) {
> + dev_err(dev, "Missing or invalid reg property on %s\n",
> + of_node_full_name(node));
> + return ret;
> + }
> +
> + if (id >= OCTEON_MAX_MMC || host->slot[id]) {
> + dev_err(dev, "Invalid reg property on %s\n",
> + of_node_full_name(node));
> + return -EINVAL;
> + }
> +
> + mmc = mmc_alloc_host(sizeof(struct octeon_mmc_slot), dev);
> + if (!mmc) {
> + dev_err(dev, "alloc host failed\n");
> + return -ENOMEM;
> + }
> +
> + slot = mmc_priv(mmc);
> + slot->mmc = mmc;
> + slot->host = host;
> +
> + /* Convert legacy DT entries into things mmc_of_parse can understand */
> + ret = octeon_mmc_of_parse_legacy(dev, node, slot);
> + if (ret)
> + return ret;
> +
> + ret = mmc_of_parse(mmc);
> + if (ret) {
> + dev_err(dev, "Failed to parse DT\n");
> + return ret;
> + }
> +
> + /* Get regulators and the supported OCR mask */
> + ret = mmc_regulator_get_supply(mmc);
> + if (ret == -EPROBE_DEFER)
> + goto err;
> +
> + /* Octeon specific DT properties */
> + ret = of_property_read_u32(node, "cavium,cmd-clk-skew", &cmd_skew);
> + if (ret)
> + cmd_skew = 0;
> +
> + ret = of_property_read_u32(node, "cavium,dat-clk-skew", &dat_skew);
> + if (ret)
> + dat_skew = 0;
> +
> + /*
> + * Set up host parameters.
> + */
> + mmc->ops = &octeon_mmc_ops;
> + mmc->f_min = 400000;
> + if (!mmc->f_max) {
> + mmc->f_max = 52000000;
> + dev_info(dev, "No max-frequency for slot %u, defaulting to %u\n",
> + id, mmc->f_max);
> + }
> +
> + mmc->caps |= MMC_CAP_MMC_HIGHSPEED | MMC_CAP_SD_HIGHSPEED |
> + MMC_CAP_ERASE;
> + mmc->ocr_avail = MMC_VDD_27_28 | MMC_VDD_28_29 | MMC_VDD_29_30 |
> + MMC_VDD_30_31 | MMC_VDD_31_32 | MMC_VDD_32_33 |
> + MMC_VDD_33_34 | MMC_VDD_34_35 | MMC_VDD_35_36;
> +
> + /* post-sdk23 caps */
> + mmc->caps |=
> + ((mmc->f_max >= 12000000) * MMC_CAP_UHS_SDR12) |
> + ((mmc->f_max >= 25000000) * MMC_CAP_UHS_SDR25) |
> + ((mmc->f_max >= 50000000) * MMC_CAP_UHS_SDR50) |
> + MMC_CAP_CMD23;
> +
> + if ((!IS_ERR(mmc->supply.vmmc)) || (slot->pwr_gpiod))
> + mmc->caps |= MMC_CAP_POWER_OFF_CARD;
> +
> + /* "1.8v" capability is actually 1.8-or-3.3v */
> + if (ddr)
> + mmc->caps |= MMC_CAP_UHS_DDR50 | MMC_CAP_1_8V_DDR;
> +
> + mmc->max_segs = 64;
> + mmc->max_seg_size = host->linear_buf_size;
> + mmc->max_req_size = host->linear_buf_size;
> + mmc->max_blk_size = 512;
> + mmc->max_blk_count = mmc->max_req_size / 512;
> +
> + slot->clock = mmc->f_min;
> + slot->sclock = octeon_get_io_clock_rate();
> +
> + clock_period = 1000000000000ull / slot->sclock; /* period in pS */
> + slot->cmd_cnt = (cmd_skew + clock_period / 2) / clock_period;
> + slot->dat_cnt = (dat_skew + clock_period / 2) / clock_period;
> +
> + slot->bus_id = id;
> + slot->cached_rca = 1;
> +
> + /* Only a single user of the bootbus at a time. */
> + octeon_mmc_acquire_bus(host);
> + host->slot[id] = slot;
> +
> + octeon_mmc_switch_to(slot);
> + /* Initialize MMC Block. */
> + octeon_mmc_initlowlevel(slot);
> +
> + octeon_mmc_release_bus(host);
> +
> + ret = mmc_add_host(mmc);
> + if (ret) {
> + dev_err(dev, "mmc_add_host() returned %d\n", ret);
> + goto err;
> + }
> +
> + return 0;
> +
> +err:
> + slot->host->slot[id] = NULL;
> +
> + gpiod_set_value_cansleep(slot->pwr_gpiod, 0);
> +
> + mmc_free_host(slot->mmc);
> + return ret;
> +}
> +
> +static int octeon_mmc_slot_remove(struct octeon_mmc_slot *slot)
> +{
> + mmc_remove_host(slot->mmc);
> +
> + slot->host->slot[slot->bus_id] = NULL;
> +
> + gpiod_set_value_cansleep(slot->pwr_gpiod, 0);
> +
> + mmc_free_host(slot->mmc);
> +
> + return 0;
> +}
> +
> +static int octeon_mmc_probe(struct platform_device *pdev)
> +{
> + struct octeon_mmc_host *host;
> + struct resource *res;
> + void __iomem *base;
> + int mmc_irq[9];
> + int i;
> + int ret = 0;
> + struct device_node *node = pdev->dev.of_node;
> + struct device_node *cn;
> + bool cn78xx_style;
> + u64 t;
> +
> + host = devm_kzalloc(&pdev->dev, sizeof(*host), GFP_KERNEL);
> + if (!host)
> + return -ENOMEM;
> +
> + spin_lock_init(&host->irq_handler_lock);
> + sema_init(&host->mmc_serializer, 1);
> +
> + cn78xx_style = of_device_is_compatible(node, "cavium,octeon-7890-mmc");
> + if (cn78xx_style) {
> + host->need_bootbus_lock = false;
> + host->big_dma_addr = true;
> + host->need_irq_handler_lock = true;
> + /*
> + * First seven are the EMM_INT bits 0..6, then two for
> + * the EMM_DMA_INT bits
> + */
> + for (i = 0; i < 9; i++) {
> + mmc_irq[i] = platform_get_irq(pdev, i);
> + if (mmc_irq[i] < 0)
> + return mmc_irq[i];
> + }
> + } else {
> + host->need_bootbus_lock = true;
> + host->big_dma_addr = false;
> + host->need_irq_handler_lock = false;
> + /* First one is EMM second NDF_DMA */
> + for (i = 0; i < 2; i++) {
> + mmc_irq[i] = platform_get_irq(pdev, i);
> + if (mmc_irq[i] < 0)
> + return mmc_irq[i];
> + }
> + }
> + host->last_slot = -1;
> +
> + if (bb_size < 512 || bb_size >= (1 << 24))
> + bb_size = 1 << 18;
> + host->linear_buf_size = bb_size;
> + host->linear_buf = devm_kzalloc(&pdev->dev, host->linear_buf_size,
> + GFP_KERNEL);
> +
> + if (!host->linear_buf) {
> + dev_err(&pdev->dev, "devm_kzalloc failed\n");
> + return -ENOMEM;
> + }
> +
> + host->pdev = pdev;
> +
> + res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
> + if (!res) {
> + dev_err(&pdev->dev, "Platform resource[0] is missing\n");
> + return -ENXIO;
> + }
> + base = devm_ioremap_resource(&pdev->dev, res);
> + if (IS_ERR(base))
> + return PTR_ERR(base);
> + host->base = (__force u64)base;
> +
> + res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
> + if (!res) {
> + dev_err(&pdev->dev, "Platform resource[1] is missing\n");
> + return -EINVAL;
> + }
> + base = devm_ioremap_resource(&pdev->dev, res);
> + if (IS_ERR(base))
> + return PTR_ERR(base);
> + host->ndf_base = (__force u64)base;
> + /*
> + * Clear out any pending interrupts that may be left over from
> + * bootloader.
> + */
> + t = cvmx_read_csr(host->base + OCT_MIO_EMM_INT);
> + cvmx_write_csr(host->base + OCT_MIO_EMM_INT, t);
> + if (cn78xx_style) {
> + /* Only CMD_DONE, DMA_DONE, CMD_ERR, DMA_ERR */
> + for (i = 1; i <= 4; i++) {
> + ret = devm_request_irq(&pdev->dev, mmc_irq[i],
> + octeon_mmc_interrupt,
> + 0, DRV_NAME, host);
> + if (ret < 0) {
> + dev_err(&pdev->dev, "Error: devm_request_irq %d\n",
> + mmc_irq[i]);
> + return ret;
> + }
> + }
> + } else {
> + ret = devm_request_irq(&pdev->dev, mmc_irq[0],
> + octeon_mmc_interrupt, 0, DRV_NAME, host);
> + if (ret < 0) {
> + dev_err(&pdev->dev, "Error: devm_request_irq %d\n",
> + mmc_irq[0]);
> + return ret;
> + }
> + }
> +
> + host->global_pwr_gpiod = devm_gpiod_get_optional(&pdev->dev, "power",
> + GPIOD_OUT_HIGH);
> + if (IS_ERR(host->global_pwr_gpiod)) {
> + dev_err(&host->pdev->dev, "Invalid POWER GPIO\n");
> + return PTR_ERR(host->global_pwr_gpiod);
> + }
> +
> + platform_set_drvdata(pdev, host);
> +
> + for_each_child_of_node(node, cn) {
> + struct platform_device *slot_pdev;
> +
> + slot_pdev = of_platform_device_create(cn, NULL, &pdev->dev);
> + ret = octeon_mmc_slot_probe(slot_pdev, host);
> + if (ret) {
> + dev_err(&host->pdev->dev, "Error populating slots\n");
> + gpiod_set_value_cansleep(host->global_pwr_gpiod, 0);
> + return ret;
> + }
> + }
> +
> + return 0;
> +}
> +
> +static int octeon_mmc_remove(struct platform_device *pdev)
> +{
> + union cvmx_mio_ndf_dma_cfg ndf_dma_cfg;
> + struct octeon_mmc_host *host = platform_get_drvdata(pdev);
> + int i;
> +
> + for (i = 0; i < OCTEON_MAX_MMC; i++) {
> + if (host->slot[i])
> + octeon_mmc_slot_remove(host->slot[i]);
> + }
> +
> + ndf_dma_cfg.u64 = cvmx_read_csr(host->ndf_base + OCT_MIO_NDF_DMA_CFG);
> + ndf_dma_cfg.s.en = 0;
> + cvmx_write_csr(host->ndf_base + OCT_MIO_NDF_DMA_CFG, ndf_dma_cfg.u64);
> +
> + gpiod_set_value_cansleep(host->global_pwr_gpiod, 0);
> +
> + return 0;
> +}
> +
> +static const struct of_device_id octeon_mmc_match[] = {
> + {
> + .compatible = "cavium,octeon-6130-mmc",
> + },
> + {
> + .compatible = "cavium,octeon-7890-mmc",
> + },
> + {},
> +};
> +MODULE_DEVICE_TABLE(of, octeon_mmc_match);
> +
> +static struct platform_driver octeon_mmc_driver = {
> + .probe = octeon_mmc_probe,
> + .remove = octeon_mmc_remove,
> + .driver = {
> + .name = DRV_NAME,
> + .of_match_table = octeon_mmc_match,
> + },
> +};
> +
> +static int __init octeon_mmc_init(void)
> +{
> + return platform_driver_register(&octeon_mmc_driver);
> +}
> +
> +static void __exit octeon_mmc_cleanup(void)
> +{
> + platform_driver_unregister(&octeon_mmc_driver);
> +}
> +
> +module_init(octeon_mmc_init);
> +module_exit(octeon_mmc_cleanup);
> +
> +MODULE_AUTHOR("Cavium Inc. <support@...ium.com>");
> +MODULE_DESCRIPTION("low-level driver for Cavium OCTEON MMC/SSD card");
> +MODULE_LICENSE("GPL");
> --
> 2.5.0
>
Powered by blists - more mailing lists