On 10 February 2016 at 18:36, Matt Redfearn <matt.redfearn@xxxxxxxxxx> wrote: > From: Aleksey Makarov <aleksey.makarov@xxxxxxxxxxxxxxxxxx> > > The OCTEON MMC controller is currently found on cn61XX and cnf71XX > devices. Device parameters are configured from device tree data. > > eMMC, MMC and SD devices are supported. > > Tested-by: Aaro Koskinen <aaro.koskinen@xxxxxx> > Signed-off-by: Chandrakala Chavva <cchavva@xxxxxxxxxxxxxxxxxx> > Signed-off-by: David Daney <david.daney@xxxxxxxxxx> > Signed-off-by: Aleksey Makarov <aleksey.makarov@xxxxxxxxxx> > Signed-off-by: Leonid Rosenboim <lrosenboim@xxxxxxxxxxxxxxxxxx> > Signed-off-by: Peter Swain <pswain@xxxxxxxxxx> > Signed-off-by: Aaron Williams <aaron.williams@xxxxxxxxxx> > Signed-off-by: Matt Redfearn <matt.redfearn@xxxxxxxxxx> > --- > v5: > Incoroprate comments from review > http://patchwork.linux-mips.org/patch/9558/ > - Use standard <bus-width> property instead of <cavium,bus-max-width>. > - Use standard <max-frequency> property instead of <spi-max-frequency>. > - Add octeon_mmc_of_parse_legacy function to deal with the above > properties, since many devices have shipped with those properties > embedded in firmware. > - Allow the <vmmc-supply> binding in addition to the legacy > <gpios-power>. > - Remove the secondary driver for each slot. > - Use core gpio cd/wp handling Seems like you decided to ignore most comments realted to the DT bindings from the earlier version. Although, let's discuss this one more time. Therefore I recomend you to split this patch. DT documentation should be a separate patch preceeding the actual mmc driver patch. The DT patch needs to be acked by the DT maintainers. Until we somewhat agreed on the DT parts, I am going to defer the in-depth review of the driver code as I have limited bandwidth. Does that make sense to you? Kind regards Uffe > > Tested on Rhino labs UTM8, Cavium CN7130. > > For reference, the binding in the shipped devices is: > mmc: mmc@1180000002000 { > compatible = "cavium,octeon-6130-mmc"; > reg = <0x11800 0x00002000 0x0 0x100>, > <0x11800 0x00000168 0x0 0x20>; > #address-cells = <1>; > #size-cells = <0>; > /* EMM irq, DMA irq */ > interrupts = <1 19>, <0 63>; > > /* The board only has a single MMC slot */ > mmc-slot@2 { > compatible = "cavium,octeon-6130-mmc-slot"; > reg = <2>; > voltage-ranges = <3300 3300>; > spi-max-frequency = <26000000>; > /* Power on GPIO 8, active high */ > /* power-gpios = <&gpio 8 0>; */ > power-gpios = <&gpio 8 1>; > > /* spi-max-frequency = <52000000>; */ > /* bus width can be 1, 4 or 8 */ > cavium,bus-max-width = <8>; > }; > mmc-slot@0 { > compatible = "cavium,octeon-6130-mmc-slot"; > reg = <0>; > voltage-ranges = <3300 3300>; > spi-max-frequency = <26000000>; > /* non-removable; */ > bus-width = <8>; > /* bus width can be 1, 4 or 8 */ > cavium,bus-max-width = <8>; > }; > }; > > v3: > https://lkml.kernel.org/g/<1425567033-31236-1-git-send-email-aleksey.makarov@xxxxxxxxxx> > > Changes in v4: > - The sparse error discovered by Aaro Koskinen has been fixed > - Other sparse warnings have been silenced > > Changes in v3: > - Rebased to v4.0-rc2 > - Use gpiod_*() functions instead of legacy gpio > - Cosmetic changes > > Changes in v2: All the fixes suggested by Mark Rutland were implemented: > - Device tree parsing has been fixed > - Device tree docs have been fixed > - Comment about errata workaroud has been added > --- > .../devicetree/bindings/mmc/octeon-mmc.txt | 80 ++ > drivers/mmc/host/Kconfig | 10 + > drivers/mmc/host/Makefile | 1 + > drivers/mmc/host/octeon_mmc.c | 1409 ++++++++++++++++++++ > 4 files changed, 1500 insertions(+) > create mode 100644 Documentation/devicetree/bindings/mmc/octeon-mmc.txt > create mode 100644 drivers/mmc/host/octeon_mmc.c > > diff --git a/Documentation/devicetree/bindings/mmc/octeon-mmc.txt b/Documentation/devicetree/bindings/mmc/octeon-mmc.txt > new file mode 100644 > index 000000000000..a1b20753172f > --- /dev/null > +++ b/Documentation/devicetree/bindings/mmc/octeon-mmc.txt > @@ -0,0 +1,80 @@ > +* OCTEON SD/MMC Host Controller > + > +This controller is present on some members of the Cavium OCTEON SoC > +family, provide an interface for eMMC, MMC and SD devices. There is a > +single controller that may have several "slots" connected. These > +slots appear as children of the main controller node. > +The DMA engine is an integral part of the controller block. > + > +1) MMC node > + > +Required properties: > +- compatible : Should be "cavium,octeon-6130-mmc" or "cavium,octeon-7890-mmc" > +- reg : Two entries: > + 1) The base address of the MMC controller register bank. > + 2) The base address of the MMC DMA engine register bank. > +- interrupts : > + For "cavium,octeon-6130-mmc": two entries: > + 1) The MMC controller interrupt line. > + 2) The MMC DMA engine interrupt line. > + For "cavium,octeon-7890-mmc": nine entries: > + 1) The next block transfer of a multiblock transfer has completed (BUF_DONE) > + 2) Operation completed successfully (CMD_DONE). > + 3) DMA transfer completed successfully (DMA_DONE). > + 4) Operation encountered an error (CMD_ERR). > + 5) DMA transfer encountered an error (DMA_ERR). > + 6) Switch operation completed successfully (SWITCH_DONE). > + 7) Switch operation encountered an error (SWITCH_ERR). > + 8) Internal DMA engine request completion interrupt (DONE). > + 9) Internal DMA FIFO underflow (FIFO). > +- #address-cells : Must be <1> > +- #size-cells : Must be <0> > + > +The node contains child nodes for each slot that the platform uses. > + > +Example: > +mmc@1180000002000 { > + compatible = "cavium,octeon-6130-mmc"; > + reg = <0x11800 0x00002000 0x0 0x100>, > + <0x11800 0x00000168 0x0 0x20>; > + #address-cells = <1>; > + #size-cells = <0>; > + /* EMM irq, DMA irq */ > + interrupts = <1 19>, <0 63>; > + > + [ child node definitions...] > +}; > + > + > +2) Slot nodes > +Properties in mmc.txt apply to each slot node that the platform uses. > + > +Required properties: > +- reg : The slot number. > + > +Optional properties: > +- cavium,cmd-clk-skew : the amount of delay (in pS) past the clock edge > + to sample the command pin. > +- cavium,dat-clk-skew : the amount of delay (in pS) past the clock edge > + to sample the data pin. > + > +Example: > + mmc@1180000002000 { > + compatible = "cavium,octeon-6130-mmc"; > + reg = <0x11800 0x00002000 0x0 0x100>, > + <0x11800 0x00000168 0x0 0x20>; > + #address-cells = <1>; > + #size-cells = <0>; > + /* EMM irq, DMA irq */ > + interrupts = <1 19>, <0 63>; > + > + /* The board only has a single MMC slot */ > + mmc-slot@0 { > + reg = <0>; > + max-frequency = <20000000>; > + bus-width = <8>; > + vmmc-supply = <®_vmmc3>; > + cd-gpios = <&gpio 9 0>; > + wp-gpios = <&gpio 10 0>; > + }; > + }; > diff --git a/drivers/mmc/host/Kconfig b/drivers/mmc/host/Kconfig > index 1526b8a10b09..8a6c6543e48e 100644 > --- a/drivers/mmc/host/Kconfig > +++ b/drivers/mmc/host/Kconfig > @@ -331,6 +331,16 @@ config MMC_SDHCI_IPROC > > If unsure, say N. > > +config MMC_OCTEON > + tristate "Cavium OCTEON Multimedia Card Interface support" > + depends on CAVIUM_OCTEON_SOC > + help > + This selects Cavium OCTEON Multimedia card Interface. > + If you have an OCTEON board with a Multimedia Card slot, > + say Y or M here. > + > + If unsure, say N. > + > config MMC_MOXART > tristate "MOXART SD/MMC Host Controller support" > depends on ARCH_MOXART && MMC > diff --git a/drivers/mmc/host/Makefile b/drivers/mmc/host/Makefile > index 3595f83e89dd..e7cacf13a034 100644 > --- a/drivers/mmc/host/Makefile > +++ b/drivers/mmc/host/Makefile > @@ -21,6 +21,7 @@ obj-$(CONFIG_MMC_SDHCI_SPEAR) += sdhci-spear.o > obj-$(CONFIG_MMC_WBSD) += wbsd.o > obj-$(CONFIG_MMC_AU1X) += au1xmmc.o > obj-$(CONFIG_MMC_MTK) += mtk-sd.o > +obj-$(CONFIG_MMC_OCTEON) += octeon_mmc.o > obj-$(CONFIG_MMC_OMAP) += omap.o > obj-$(CONFIG_MMC_OMAP_HS) += omap_hsmmc.o > obj-$(CONFIG_MMC_ATMELMCI) += atmel-mci.o > diff --git a/drivers/mmc/host/octeon_mmc.c b/drivers/mmc/host/octeon_mmc.c > new file mode 100644 > index 000000000000..8e41bffe4d30 > --- /dev/null > +++ b/drivers/mmc/host/octeon_mmc.c > @@ -0,0 +1,1409 @@ > +/* > + * Driver for MMC and SSD cards for Cavium OCTEON SOCs. > + * > + * This file is subject to the terms and conditions of the GNU General Public > + * License. See the file "COPYING" in the main directory of this archive > + * for more details. > + * > + * Copyright (C) 2012-2015 Cavium Inc. > + */ > +#include <linux/platform_device.h> > +#include <linux/of_platform.h> > +#include <linux/scatterlist.h> > +#include <linux/interrupt.h> > +#include <linux/blkdev.h> > +#include <linux/device.h> > +#include <linux/module.h> > +#include <linux/delay.h> > +#include <linux/init.h> > +#include <linux/clk.h> > +#include <linux/err.h> > +#include <linux/io.h> > +#include <linux/of.h> > + > +#include <linux/mmc/card.h> > +#include <linux/mmc/host.h> > +#include <linux/mmc/mmc.h> > +#include <linux/mmc/sd.h> > +#include <linux/mmc/slot-gpio.h> > +#include <net/irda/parameters.h> > +#include <linux/gpio/consumer.h> > +#include <linux/regulator/consumer.h> > + > +#include <asm/byteorder.h> > +#include <asm/octeon/octeon.h> > +#include <asm/octeon/cvmx-mio-defs.h> > + > +#define DRV_NAME "octeon_mmc" > + > +#define OCTEON_MAX_MMC 4 > + > +#define OCT_MIO_NDF_DMA_CFG 0x00 > +#define OCT_MIO_EMM_DMA_ADR 0x08 > + > +#define OCT_MIO_EMM_CFG 0x00 > +#define OCT_MIO_EMM_SWITCH 0x48 > +#define OCT_MIO_EMM_DMA 0x50 > +#define OCT_MIO_EMM_CMD 0x58 > +#define OCT_MIO_EMM_RSP_STS 0x60 > +#define OCT_MIO_EMM_RSP_LO 0x68 > +#define OCT_MIO_EMM_RSP_HI 0x70 > +#define OCT_MIO_EMM_INT 0x78 > +#define OCT_MIO_EMM_INT_EN 0x80 > +#define OCT_MIO_EMM_WDOG 0x88 > +#define OCT_MIO_EMM_SAMPLE 0x90 > +#define OCT_MIO_EMM_STS_MASK 0x98 > +#define OCT_MIO_EMM_RCA 0xa0 > +#define OCT_MIO_EMM_BUF_IDX 0xe0 > +#define OCT_MIO_EMM_BUF_DAT 0xe8 > + > +#define CVMX_MIO_BOOT_CTL CVMX_ADD_IO_SEG(0x00011800000000D0ull) > + > +struct octeon_mmc_host { > + u64 base; > + u64 ndf_base; > + u64 emm_cfg; > + u64 n_minus_one; /* OCTEON II workaround location */ > + int last_slot; > + > + struct semaphore mmc_serializer; > + struct mmc_request *current_req; > + unsigned int linear_buf_size; > + void *linear_buf; > + struct sg_mapping_iter smi; > + int sg_idx; > + bool dma_active; > + > + struct platform_device *pdev; > + struct gpio_desc *global_pwr_gpiod; > + bool dma_err_pending; > + bool need_bootbus_lock; > + bool big_dma_addr; > + bool need_irq_handler_lock; > + spinlock_t irq_handler_lock; > + > + struct octeon_mmc_slot *slot[OCTEON_MAX_MMC]; > +}; > + > +struct octeon_mmc_slot { > + struct mmc_host *mmc; /* slot-level mmc_core object */ > + struct octeon_mmc_host *host; /* common hw for all 4 slots */ > + > + unsigned int clock; > + unsigned int sclock; > + > + u64 cached_switch; > + u64 cached_rca; > + > + unsigned int cmd_cnt; /* sample delay */ > + unsigned int dat_cnt; /* sample delay */ > + > + int bus_id; > + > + /* Legacy property - in future mmc->supply.vmmc should be used */ > + struct gpio_desc *pwr_gpiod; > +}; > + > +static int bb_size = 1 << 18; > +module_param(bb_size, int, S_IRUGO); > +MODULE_PARM_DESC(bb_size, > + "Size of DMA linearizing buffer (max transfer size)."); > + > +static int ddr = 2; > +module_param(ddr, int, S_IRUGO); > +MODULE_PARM_DESC(ddr, > + "enable DoubleDataRate clocking: 0=no, 1=always, 2=at spi-max-frequency/2"); > + > +#if 1 > +#define octeon_mmc_dbg pr_debug > +#else > +static inline void octeon_mmc_dbg(const char *s, ...) { } > +#endif > + > +static void octeon_mmc_acquire_bus(struct octeon_mmc_host *host) > +{ > + if (host->need_bootbus_lock) { > + down(&octeon_bootbus_sem); > + /* On cn70XX switch the mmc unit onto the bus. */ > + if (OCTEON_IS_MODEL(OCTEON_CN70XX)) > + cvmx_write_csr(CVMX_MIO_BOOT_CTL, 0); > + } else { > + down(&host->mmc_serializer); > + } > +} > + > +static void octeon_mmc_release_bus(struct octeon_mmc_host *host) > +{ > + if (host->need_bootbus_lock) > + up(&octeon_bootbus_sem); > + else > + up(&host->mmc_serializer); > +} > + > +struct octeon_mmc_cr_type { > + u8 ctype; > + u8 rtype; > +}; > + > +/* > + * The OCTEON MMC host hardware assumes that all commands have fixed > + * command and response types. These are correct if MMC devices are > + * being used. However, non-MMC devices like SD use command and > + * response types that are unexpected by the host hardware. > + * > + * The command and response types can be overridden by supplying an > + * XOR value that is applied to the type. We calculate the XOR value > + * from the values in this table and the flags passed from the MMC > + * core. > + */ > +static struct octeon_mmc_cr_type octeon_mmc_cr_types[] = { > + {0, 0}, /* CMD0 */ > + {0, 3}, /* CMD1 */ > + {0, 2}, /* CMD2 */ > + {0, 1}, /* CMD3 */ > + {0, 0}, /* CMD4 */ > + {0, 1}, /* CMD5 */ > + {0, 1}, /* CMD6 */ > + {0, 1}, /* CMD7 */ > + {1, 1}, /* CMD8 */ > + {0, 2}, /* CMD9 */ > + {0, 2}, /* CMD10 */ > + {1, 1}, /* CMD11 */ > + {0, 1}, /* CMD12 */ > + {0, 1}, /* CMD13 */ > + {1, 1}, /* CMD14 */ > + {0, 0}, /* CMD15 */ > + {0, 1}, /* CMD16 */ > + {1, 1}, /* CMD17 */ > + {1, 1}, /* CMD18 */ > + {3, 1}, /* CMD19 */ > + {2, 1}, /* CMD20 */ > + {0, 0}, /* CMD21 */ > + {0, 0}, /* CMD22 */ > + {0, 1}, /* CMD23 */ > + {2, 1}, /* CMD24 */ > + {2, 1}, /* CMD25 */ > + {2, 1}, /* CMD26 */ > + {2, 1}, /* CMD27 */ > + {0, 1}, /* CMD28 */ > + {0, 1}, /* CMD29 */ > + {1, 1}, /* CMD30 */ > + {1, 1}, /* CMD31 */ > + {0, 0}, /* CMD32 */ > + {0, 0}, /* CMD33 */ > + {0, 0}, /* CMD34 */ > + {0, 1}, /* CMD35 */ > + {0, 1}, /* CMD36 */ > + {0, 0}, /* CMD37 */ > + {0, 1}, /* CMD38 */ > + {0, 4}, /* CMD39 */ > + {0, 5}, /* CMD40 */ > + {0, 0}, /* CMD41 */ > + {2, 1}, /* CMD42 */ > + {0, 0}, /* CMD43 */ > + {0, 0}, /* CMD44 */ > + {0, 0}, /* CMD45 */ > + {0, 0}, /* CMD46 */ > + {0, 0}, /* CMD47 */ > + {0, 0}, /* CMD48 */ > + {0, 0}, /* CMD49 */ > + {0, 0}, /* CMD50 */ > + {0, 0}, /* CMD51 */ > + {0, 0}, /* CMD52 */ > + {0, 0}, /* CMD53 */ > + {0, 0}, /* CMD54 */ > + {0, 1}, /* CMD55 */ > + {0xff, 0xff}, /* CMD56 */ > + {0, 0}, /* CMD57 */ > + {0, 0}, /* CMD58 */ > + {0, 0}, /* CMD59 */ > + {0, 0}, /* CMD60 */ > + {0, 0}, /* CMD61 */ > + {0, 0}, /* CMD62 */ > + {0, 0} /* CMD63 */ > +}; > + > +struct octeon_mmc_cr_mods { > + u8 ctype_xor; > + u8 rtype_xor; > +}; > + > +/* > + * The functions below are used for the EMMC-17978 workaround. > + * > + * Due to an imperfection in the design of the MMC bus hardware, > + * the 2nd to last cache block of a DMA read must be locked into the L2 Cache. > + * Otherwise, data corruption may occur. > + */ > + > +static inline void *phys_to_ptr(u64 address) > +{ > + return (void *)(address | (1ull<<63)); /* XKPHYS */ > +} > + > +/** > + * Lock a single line into L2. The line is zeroed before locking > + * to make sure no dram accesses are made. > + * > + * @addr Physical address to lock > + */ > +static void l2c_lock_line(u64 addr) > +{ > + char *addr_ptr = phys_to_ptr(addr); > + > + asm volatile ( > + "cache 31, %[line]" /* Unlock the line */ > + :: [line] "m" (*addr_ptr)); > +} > + > +/** > + * Locks a memory region in the L2 cache > + * > + * @start - start address to begin locking > + * @len - length in bytes to lock > + */ > +static void l2c_lock_mem_region(u64 start, u64 len) > +{ > + u64 end; > + > + /* Round start/end to cache line boundaries */ > + end = ALIGN(start + len - 1, CVMX_CACHE_LINE_SIZE); > + start = ALIGN(start, CVMX_CACHE_LINE_SIZE); > + > + while (start <= end) { > + l2c_lock_line(start); > + start += CVMX_CACHE_LINE_SIZE; > + } > + asm volatile("sync"); > +} > + > +/** > + * Unlock a single line in the L2 cache. > + * > + * @addr Physical address to unlock > + * > + * Return Zero on success > + */ > +static void l2c_unlock_line(u64 addr) > +{ > + char *addr_ptr = phys_to_ptr(addr); > + > + asm volatile ( > + "cache 23, %[line]" /* Unlock the line */ > + :: [line] "m" (*addr_ptr)); > +} > + > +/** > + * Unlock a memory region in the L2 cache > + * > + * @start - start address to unlock > + * @len - length to unlock in bytes > + */ > +static void l2c_unlock_mem_region(u64 start, u64 len) > +{ > + u64 end; > + > + /* Round start/end to cache line boundaries */ > + end = ALIGN(start + len - 1, CVMX_CACHE_LINE_SIZE); > + start = ALIGN(start, CVMX_CACHE_LINE_SIZE); > + > + while (start <= end) { > + l2c_unlock_line(start); > + start += CVMX_CACHE_LINE_SIZE; > + } > +} > + > +static struct octeon_mmc_cr_mods octeon_mmc_get_cr_mods(struct mmc_command *cmd) > +{ > + struct octeon_mmc_cr_type *cr; > + u8 desired_ctype, hardware_ctype; > + u8 desired_rtype, hardware_rtype; > + struct octeon_mmc_cr_mods r; > + > + desired_ctype = desired_rtype = 0; > + > + cr = octeon_mmc_cr_types + (cmd->opcode & 0x3f); > + hardware_ctype = cr->ctype; > + hardware_rtype = cr->rtype; > + if (cmd->opcode == 56) { /* CMD56 GEN_CMD */ > + hardware_ctype = (cmd->arg & 1) ? 1 : 2; > + } > + > + switch (mmc_cmd_type(cmd)) { > + case MMC_CMD_ADTC: > + desired_ctype = (cmd->data->flags & MMC_DATA_WRITE) ? 2 : 1; > + break; > + case MMC_CMD_AC: > + case MMC_CMD_BC: > + case MMC_CMD_BCR: > + desired_ctype = 0; > + break; > + } > + > + switch (mmc_resp_type(cmd)) { > + case MMC_RSP_NONE: > + desired_rtype = 0; > + break; > + case MMC_RSP_R1:/* MMC_RSP_R5, MMC_RSP_R6, MMC_RSP_R7 */ > + case MMC_RSP_R1B: > + desired_rtype = 1; > + break; > + case MMC_RSP_R2: > + desired_rtype = 2; > + break; > + case MMC_RSP_R3: /* MMC_RSP_R4 */ > + desired_rtype = 3; > + break; > + } > + r.ctype_xor = desired_ctype ^ hardware_ctype; > + r.rtype_xor = desired_rtype ^ hardware_rtype; > + return r; > +} > + > +static bool octeon_mmc_switch_val_changed(struct octeon_mmc_slot *slot, > + u64 new_val) > +{ > + /* Match BUS_ID, HS_TIMING, BUS_WIDTH, POWER_CLASS, CLK_HI, CLK_LO */ > + u64 m = 0x3001070fffffffffull; > + > + return (slot->cached_switch & m) != (new_val & m); > +} > + > +static unsigned int octeon_mmc_timeout_to_wdog(struct octeon_mmc_slot *slot, > + unsigned int ns) > +{ > + u64 bt = (u64)slot->clock * (u64)ns; > + > + return (unsigned int)(bt / 1000000000); > +} > + > +static irqreturn_t octeon_mmc_interrupt(int irq, void *dev_id) > +{ > + struct octeon_mmc_host *host = dev_id; > + union cvmx_mio_emm_int emm_int; > + struct mmc_request *req; > + bool host_done; > + union cvmx_mio_emm_rsp_sts rsp_sts; > + unsigned long flags = 0; > + > + if (host->need_irq_handler_lock) > + spin_lock_irqsave(&host->irq_handler_lock, flags); > + else > + __acquire(&host->irq_handler_lock); > + emm_int.u64 = cvmx_read_csr(host->base + OCT_MIO_EMM_INT); > + req = host->current_req; > + cvmx_write_csr(host->base + OCT_MIO_EMM_INT, emm_int.u64); > + > + octeon_mmc_dbg("Got interrupt: EMM_INT = 0x%llx\n", emm_int.u64); > + > + if (!req) > + goto out; > + > + rsp_sts.u64 = cvmx_read_csr(host->base + OCT_MIO_EMM_RSP_STS); > + octeon_mmc_dbg("octeon_mmc_interrupt MIO_EMM_RSP_STS 0x%llx\n", > + rsp_sts.u64); > + > + if (host->dma_err_pending) { > + host->current_req = NULL; > + host->dma_err_pending = false; > + req->done(req); > + host_done = true; > + goto no_req_done; > + } > + > + if (!host->dma_active && emm_int.s.buf_done && req->data) { > + unsigned int type = (rsp_sts.u64 >> 7) & 3; > + > + if (type == 1) { > + /* Read */ > + int dbuf = rsp_sts.s.dbuf; > + struct sg_mapping_iter *smi = &host->smi; > + unsigned int data_len = > + req->data->blksz * req->data->blocks; > + unsigned int bytes_xfered; > + u64 dat = 0; > + int shift = -1; > + > + /* Auto inc from offset zero */ > + cvmx_write_csr(host->base + OCT_MIO_EMM_BUF_IDX, > + (u64)(0x10000 | (dbuf << 6))); > + > + for (bytes_xfered = 0; bytes_xfered < data_len;) { > + if (smi->consumed >= smi->length) { > + if (!sg_miter_next(smi)) > + break; > + smi->consumed = 0; > + } > + if (shift < 0) { > + dat = cvmx_read_csr(host->base + > + OCT_MIO_EMM_BUF_DAT); > + shift = 56; > + } > + > + while (smi->consumed < smi->length && > + shift >= 0) { > + ((u8 *)(smi->addr))[smi->consumed] = > + (dat >> shift) & 0xff; > + bytes_xfered++; > + smi->consumed++; > + shift -= 8; > + } > + } > + sg_miter_stop(smi); > + req->data->bytes_xfered = bytes_xfered; > + req->data->error = 0; > + } else if (type == 2) { > + /* write */ > + req->data->bytes_xfered = req->data->blksz * > + req->data->blocks; > + req->data->error = 0; > + } > + } > + host_done = emm_int.s.cmd_done || emm_int.s.dma_done || > + emm_int.s.cmd_err || emm_int.s.dma_err; > + if (host_done && req->done) { > + if (rsp_sts.s.rsp_bad_sts || > + rsp_sts.s.rsp_crc_err || > + rsp_sts.s.rsp_timeout || > + rsp_sts.s.blk_crc_err || > + rsp_sts.s.blk_timeout || > + rsp_sts.s.dbuf_err) { > + req->cmd->error = -EILSEQ; > + } else { > + req->cmd->error = 0; > + } > + > + if (host->dma_active && req->data) { > + req->data->error = 0; > + req->data->bytes_xfered = req->data->blocks * > + req->data->blksz; > + if (!(req->data->flags & MMC_DATA_WRITE) && > + req->data->sg_len > 1) { > + size_t r = sg_copy_from_buffer(req->data->sg, > + req->data->sg_len, host->linear_buf, > + req->data->bytes_xfered); > + WARN_ON(r != req->data->bytes_xfered); > + } > + } > + if (rsp_sts.s.rsp_val) { > + u64 rsp_hi; > + u64 rsp_lo = cvmx_read_csr( > + host->base + OCT_MIO_EMM_RSP_LO); > + > + switch (rsp_sts.s.rsp_type) { > + case 1: > + case 3: > + req->cmd->resp[0] = (rsp_lo >> 8) & 0xffffffff; > + req->cmd->resp[1] = 0; > + req->cmd->resp[2] = 0; > + req->cmd->resp[3] = 0; > + break; > + case 2: > + req->cmd->resp[3] = rsp_lo & 0xffffffff; > + req->cmd->resp[2] = (rsp_lo >> 32) & 0xffffffff; > + rsp_hi = cvmx_read_csr(host->base + > + OCT_MIO_EMM_RSP_HI); > + req->cmd->resp[1] = rsp_hi & 0xffffffff; > + req->cmd->resp[0] = (rsp_hi >> 32) & 0xffffffff; > + break; > + default: > + octeon_mmc_dbg("octeon_mmc_interrupt unhandled rsp_val %d\n", > + rsp_sts.s.rsp_type); > + break; > + } > + octeon_mmc_dbg("octeon_mmc_interrupt resp %08x %08x %08x %08x\n", > + req->cmd->resp[0], req->cmd->resp[1], > + req->cmd->resp[2], req->cmd->resp[3]); > + } > + if (emm_int.s.dma_err && rsp_sts.s.dma_pend) { > + /* Try to clean up failed DMA */ > + union cvmx_mio_emm_dma emm_dma; > + > + emm_dma.u64 = > + cvmx_read_csr(host->base + OCT_MIO_EMM_DMA); > + emm_dma.s.dma_val = 1; > + emm_dma.s.dat_null = 1; > + emm_dma.s.bus_id = rsp_sts.s.bus_id; > + cvmx_write_csr(host->base + OCT_MIO_EMM_DMA, > + emm_dma.u64); > + host->dma_err_pending = true; > + host_done = false; > + goto no_req_done; > + } > + > + host->current_req = NULL; > + req->done(req); > + } > +no_req_done: > + if (host->n_minus_one) { > + l2c_unlock_mem_region(host->n_minus_one, 512); > + host->n_minus_one = 0; > + } > + if (host_done) > + octeon_mmc_release_bus(host); > +out: > + if (host->need_irq_handler_lock) > + spin_unlock_irqrestore(&host->irq_handler_lock, flags); > + else > + __release(&host->irq_handler_lock); > + return IRQ_RETVAL(emm_int.u64 != 0); > +} > + > +static void octeon_mmc_switch_to(struct octeon_mmc_slot *slot) > +{ > + struct octeon_mmc_host *host = slot->host; > + struct octeon_mmc_slot *old_slot; > + union cvmx_mio_emm_switch sw; > + union cvmx_mio_emm_sample samp; > + > + if (slot->bus_id == host->last_slot) > + goto out; > + > + if (host->last_slot >= 0 && host->slot[host->last_slot]) { > + old_slot = host->slot[host->last_slot]; > + old_slot->cached_switch = > + cvmx_read_csr(host->base + OCT_MIO_EMM_SWITCH); > + old_slot->cached_rca = > + cvmx_read_csr(host->base + OCT_MIO_EMM_RCA); > + } > + cvmx_write_csr(host->base + OCT_MIO_EMM_RCA, slot->cached_rca); > + sw.u64 = slot->cached_switch; > + sw.s.bus_id = 0; > + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, sw.u64); > + sw.s.bus_id = slot->bus_id; > + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, sw.u64); > + > + samp.u64 = 0; > + samp.s.cmd_cnt = slot->cmd_cnt; > + samp.s.dat_cnt = slot->dat_cnt; > + cvmx_write_csr(host->base + OCT_MIO_EMM_SAMPLE, samp.u64); > +out: > + host->last_slot = slot->bus_id; > +} > + > +static void octeon_mmc_dma_request(struct mmc_host *mmc, > + struct mmc_request *mrq) > +{ > + struct octeon_mmc_slot *slot; > + struct octeon_mmc_host *host; > + struct mmc_command *cmd; > + struct mmc_data *data; > + union cvmx_mio_emm_int emm_int; > + union cvmx_mio_emm_dma emm_dma; > + union cvmx_mio_ndf_dma_cfg dma_cfg; > + > + cmd = mrq->cmd; > + if (mrq->data == NULL || mrq->data->sg == NULL || !mrq->data->sg_len || > + mrq->stop == NULL || mrq->stop->opcode != MMC_STOP_TRANSMISSION) { > + dev_err(&mmc->card->dev, > + "Error: octeon_mmc_dma_request no data\n"); > + cmd->error = -EINVAL; > + if (mrq->done) > + mrq->done(mrq); > + return; > + } > + > + slot = mmc_priv(mmc); > + host = slot->host; > + > + /* Only a single user of the bootbus at a time. */ > + octeon_mmc_acquire_bus(host); > + > + octeon_mmc_switch_to(slot); > + > + data = mrq->data; > + > + if (data->timeout_ns) { > + cvmx_write_csr(host->base + OCT_MIO_EMM_WDOG, > + octeon_mmc_timeout_to_wdog(slot, data->timeout_ns)); > + octeon_mmc_dbg("OCT_MIO_EMM_WDOG %llu\n", > + cvmx_read_csr(host->base + OCT_MIO_EMM_WDOG)); > + } > + > + WARN_ON(host->current_req); > + host->current_req = mrq; > + > + host->sg_idx = 0; > + > + WARN_ON(data->blksz * data->blocks > host->linear_buf_size); > + > + if ((data->flags & MMC_DATA_WRITE) && data->sg_len > 1) { > + size_t r = sg_copy_to_buffer(data->sg, data->sg_len, > + host->linear_buf, data->blksz * data->blocks); > + WARN_ON(data->blksz * data->blocks != r); > + } > + > + dma_cfg.u64 = 0; > + dma_cfg.s.en = 1; > + dma_cfg.s.rw = (data->flags & MMC_DATA_WRITE) ? 1 : 0; > +#ifdef __LITTLE_ENDIAN > + dma_cfg.s.endian = 1; > +#endif > + dma_cfg.s.size = ((data->blksz * data->blocks) / 8) - 1; > + if (!host->big_dma_addr) { > + if (data->sg_len > 1) > + dma_cfg.s.adr = virt_to_phys(host->linear_buf); > + else > + dma_cfg.s.adr = sg_phys(data->sg); > + } > + cvmx_write_csr(host->ndf_base + OCT_MIO_NDF_DMA_CFG, dma_cfg.u64); > + octeon_mmc_dbg("MIO_NDF_DMA_CFG: %016llx\n", > + (unsigned long long)dma_cfg.u64); > + if (host->big_dma_addr) { > + u64 addr; > + > + if (data->sg_len > 1) > + addr = virt_to_phys(host->linear_buf); > + else > + addr = sg_phys(data->sg); > + cvmx_write_csr(host->ndf_base + OCT_MIO_EMM_DMA_ADR, addr); > + octeon_mmc_dbg("MIO_EMM_DMA_ADR: %016llx\n", > + (unsigned long long)addr); > + } > + > + emm_dma.u64 = 0; > + emm_dma.s.bus_id = slot->bus_id; > + emm_dma.s.dma_val = 1; > + emm_dma.s.sector = mmc_card_blockaddr(mmc->card) ? 1 : 0; > + emm_dma.s.rw = (data->flags & MMC_DATA_WRITE) ? 1 : 0; > + if (mmc_card_mmc(mmc->card) || > + (mmc_card_sd(mmc->card) && > + (mmc->card->scr.cmds & SD_SCR_CMD23_SUPPORT))) > + emm_dma.s.multi = 1; > + emm_dma.s.block_cnt = data->blocks; > + emm_dma.s.card_addr = cmd->arg; > + > + emm_int.u64 = 0; > + emm_int.s.dma_done = 1; > + emm_int.s.cmd_err = 1; > + emm_int.s.dma_err = 1; > + /* Clear the bit. */ > + cvmx_write_csr(host->base + OCT_MIO_EMM_INT, emm_int.u64); > + cvmx_write_csr(host->base + OCT_MIO_EMM_INT_EN, emm_int.u64); > + host->dma_active = true; > + > + if ((OCTEON_IS_MODEL(OCTEON_CN6XXX) || > + OCTEON_IS_MODEL(OCTEON_CNF7XXX)) && > + cmd->opcode == MMC_WRITE_MULTIPLE_BLOCK && > + (data->blksz * data->blocks) > 1024) { > + host->n_minus_one = dma_cfg.s.adr + > + (data->blksz * data->blocks) - 1024; > + l2c_lock_mem_region(host->n_minus_one, 512); > + } > + > + if (mmc->card && mmc_card_sd(mmc->card)) > + cvmx_write_csr(host->base + OCT_MIO_EMM_STS_MASK, > + 0x00b00000ull); > + else > + cvmx_write_csr(host->base + OCT_MIO_EMM_STS_MASK, > + 0xe4f90080ull); > + cvmx_write_csr(host->base + OCT_MIO_EMM_DMA, emm_dma.u64); > + octeon_mmc_dbg("MIO_EMM_DMA: %llx\n", emm_dma.u64); > +} > + > +static void octeon_mmc_request(struct mmc_host *mmc, struct mmc_request *mrq) > +{ > + struct octeon_mmc_slot *slot; > + struct octeon_mmc_host *host; > + struct mmc_command *cmd; > + union cvmx_mio_emm_int emm_int; > + union cvmx_mio_emm_cmd emm_cmd; > + struct octeon_mmc_cr_mods mods; > + > + cmd = mrq->cmd; > + > + if (cmd->opcode == MMC_READ_MULTIPLE_BLOCK || > + cmd->opcode == MMC_WRITE_MULTIPLE_BLOCK) { > + octeon_mmc_dma_request(mmc, mrq); > + return; > + } > + > + mods = octeon_mmc_get_cr_mods(cmd); > + > + slot = mmc_priv(mmc); > + host = slot->host; > + > + /* Only a single user of the bootbus at a time. */ > + octeon_mmc_acquire_bus(host); > + > + octeon_mmc_switch_to(slot); > + > + WARN_ON(host->current_req); > + host->current_req = mrq; > + > + emm_int.u64 = 0; > + emm_int.s.cmd_done = 1; > + emm_int.s.cmd_err = 1; > + if (cmd->data) { > + octeon_mmc_dbg("command has data\n"); > + if (cmd->data->flags & MMC_DATA_READ) { > + sg_miter_start(&host->smi, mrq->data->sg, > + mrq->data->sg_len, > + SG_MITER_ATOMIC | SG_MITER_TO_SG); > + } else { > + struct sg_mapping_iter *smi = &host->smi; > + unsigned int data_len = > + mrq->data->blksz * mrq->data->blocks; > + unsigned int bytes_xfered; > + u64 dat = 0; > + int shift = 56; > + /* > + * Copy data to the xmit buffer before > + * issuing the command > + */ > + sg_miter_start(smi, mrq->data->sg, > + mrq->data->sg_len, SG_MITER_FROM_SG); > + /* Auto inc from offset zero, dbuf zero */ > + cvmx_write_csr(host->base + OCT_MIO_EMM_BUF_IDX, > + 0x10000ull); > + > + for (bytes_xfered = 0; bytes_xfered < data_len;) { > + if (smi->consumed >= smi->length) { > + if (!sg_miter_next(smi)) > + break; > + smi->consumed = 0; > + } > + > + while (smi->consumed < smi->length && > + shift >= 0) { > + > + dat |= (u64)(((u8 *)(smi->addr)) > + [smi->consumed]) << shift; > + bytes_xfered++; > + smi->consumed++; > + shift -= 8; > + } > + if (shift < 0) { > + cvmx_write_csr(host->base + > + OCT_MIO_EMM_BUF_DAT, dat); > + shift = 56; > + dat = 0; > + } > + } > + sg_miter_stop(smi); > + } > + if (cmd->data->timeout_ns) { > + cvmx_write_csr(host->base + OCT_MIO_EMM_WDOG, > + octeon_mmc_timeout_to_wdog(slot, > + cmd->data->timeout_ns)); > + octeon_mmc_dbg("OCT_MIO_EMM_WDOG %llu\n", > + cvmx_read_csr(host->base + > + OCT_MIO_EMM_WDOG)); > + } > + } else { > + cvmx_write_csr(host->base + OCT_MIO_EMM_WDOG, > + ((u64)slot->clock * 850ull) / 1000ull); > + octeon_mmc_dbg("OCT_MIO_EMM_WDOG %llu\n", > + cvmx_read_csr(host->base + OCT_MIO_EMM_WDOG)); > + } > + /* Clear the bit. */ > + cvmx_write_csr(host->base + OCT_MIO_EMM_INT, emm_int.u64); > + cvmx_write_csr(host->base + OCT_MIO_EMM_INT_EN, emm_int.u64); > + host->dma_active = false; > + > + emm_cmd.u64 = 0; > + emm_cmd.s.cmd_val = 1; > + emm_cmd.s.ctype_xor = mods.ctype_xor; > + emm_cmd.s.rtype_xor = mods.rtype_xor; > + if (mmc_cmd_type(cmd) == MMC_CMD_ADTC) > + emm_cmd.s.offset = 64 - > + ((cmd->data->blksz * cmd->data->blocks) / 8); > + emm_cmd.s.bus_id = slot->bus_id; > + emm_cmd.s.cmd_idx = cmd->opcode; > + emm_cmd.s.arg = cmd->arg; > + cvmx_write_csr(host->base + OCT_MIO_EMM_STS_MASK, 0); > + cvmx_write_csr(host->base + OCT_MIO_EMM_CMD, emm_cmd.u64); > + octeon_mmc_dbg("MIO_EMM_CMD: %llx\n", emm_cmd.u64); > +} > + > +static void octeon_mmc_reset_bus(struct octeon_mmc_slot *slot) > +{ > + union cvmx_mio_emm_cfg emm_cfg; > + union cvmx_mio_emm_switch emm_switch; > + u64 wdog = 0; > + > + emm_cfg.u64 = cvmx_read_csr(slot->host->base + OCT_MIO_EMM_CFG); > + emm_switch.u64 = cvmx_read_csr(slot->host->base + OCT_MIO_EMM_SWITCH); > + wdog = cvmx_read_csr(slot->host->base + OCT_MIO_EMM_WDOG); > + > + emm_switch.s.switch_exe = 0; > + emm_switch.s.switch_err0 = 0; > + emm_switch.s.switch_err1 = 0; > + emm_switch.s.switch_err2 = 0; > + emm_switch.s.bus_id = 0; > + cvmx_write_csr(slot->host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64); > + emm_switch.s.bus_id = slot->bus_id; > + cvmx_write_csr(slot->host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64); > + > + slot->cached_switch = emm_switch.u64; > + > + msleep(20); > + > + cvmx_write_csr(slot->host->base + OCT_MIO_EMM_WDOG, wdog); > +} > + > +static void octeon_mmc_set_ios(struct mmc_host *mmc, struct mmc_ios *ios) > +{ > + struct octeon_mmc_slot *slot; > + struct octeon_mmc_host *host; > + int bus_width; > + int clock; > + bool ddr_clock; > + int hs_timing; > + int power_class = 10; > + int clk_period; > + int timeout = 2000; > + union cvmx_mio_emm_switch emm_switch; > + union cvmx_mio_emm_rsp_sts emm_sts; > + > + slot = mmc_priv(mmc); > + host = slot->host; > + > + /* Only a single user of the bootbus at a time. */ > + octeon_mmc_acquire_bus(host); > + > + octeon_mmc_switch_to(slot); > + > + octeon_mmc_dbg("Calling set_ios: slot: clk = 0x%x, bus_width = %d\n", > + slot->clock, (mmc->caps & MMC_CAP_8_BIT_DATA) ? 8 : > + (mmc->caps & MMC_CAP_4_BIT_DATA) ? 4 : 1); > + octeon_mmc_dbg("Calling set_ios: ios: clk = 0x%x, vdd = %u, bus_width = %u, power_mode = %u, timing = %u\n", > + ios->clock, ios->vdd, ios->bus_width, ios->power_mode, > + ios->timing); > + octeon_mmc_dbg("Calling set_ios: mmc: caps = 0x%x, bus_width = %d\n", > + mmc->caps, mmc->ios.bus_width); > + > + /* > + * Reset the chip on each power off > + */ > + if (ios->power_mode == MMC_POWER_OFF) { > + octeon_mmc_reset_bus(slot); > + if (!IS_ERR(mmc->supply.vmmc)) > + regulator_disable(mmc->supply.vmmc); > + else /* Legacy power GPIO */ > + gpiod_set_value_cansleep(slot->pwr_gpiod, 0); > + } else { > + if (!IS_ERR(mmc->supply.vmmc)) > + regulator_enable(mmc->supply.vmmc); > + else /* Legacy power GPIO */ > + gpiod_set_value_cansleep(slot->pwr_gpiod, 1); > + } > + > + switch (ios->bus_width) { > + case MMC_BUS_WIDTH_8: > + bus_width = 2; > + break; > + case MMC_BUS_WIDTH_4: > + bus_width = 1; > + break; > + case MMC_BUS_WIDTH_1: > + bus_width = 0; > + break; > + default: > + octeon_mmc_dbg("unknown bus width %d\n", ios->bus_width); > + bus_width = 0; > + break; > + } > + > + hs_timing = (ios->timing == MMC_TIMING_MMC_HS); > + ddr_clock = (bus_width && ios->timing >= MMC_TIMING_UHS_DDR50); > + > + if (ddr_clock) > + bus_width |= 4; > + > + if (ios->clock) { > + slot->clock = ios->clock; > + > + clock = slot->clock; > + > + if (clock > 52000000) > + clock = 52000000; > + > + clk_period = (octeon_get_io_clock_rate() + clock - 1) / > + (2 * clock); > + > + /* until clock-renengotiate-on-CRC is in */ > + if (ddr_clock && ddr > 1) > + clk_period *= 2; > + > + emm_switch.u64 = 0; > + emm_switch.s.hs_timing = hs_timing; > + emm_switch.s.bus_width = bus_width; > + emm_switch.s.power_class = power_class; > + emm_switch.s.clk_hi = clk_period; > + emm_switch.s.clk_lo = clk_period; > + > + if (!octeon_mmc_switch_val_changed(slot, emm_switch.u64)) { > + octeon_mmc_dbg("No change from 0x%llx mio_emm_switch, returning.\n", > + emm_switch.u64); > + goto out; > + } > + > + octeon_mmc_dbg("Writing 0x%llx to mio_emm_wdog\n", > + ((u64)clock * 850ull) / 1000ull); > + cvmx_write_csr(host->base + OCT_MIO_EMM_WDOG, > + ((u64)clock * 850ull) / 1000ull); > + octeon_mmc_dbg("Writing 0x%llx to mio_emm_switch\n", > + emm_switch.u64); > + > + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64); > + emm_switch.s.bus_id = slot->bus_id; > + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64); > + slot->cached_switch = emm_switch.u64; > + > + do { > + emm_sts.u64 = > + cvmx_read_csr(host->base + OCT_MIO_EMM_RSP_STS); > + if (!emm_sts.s.switch_val) > + break; > + udelay(100); > + } while (timeout-- > 0); > + > + if (timeout <= 0) { > + octeon_mmc_dbg("switch command timed out, status=0x%llx\n", > + emm_sts.u64); > + goto out; > + } > + } > +out: > + octeon_mmc_release_bus(host); > +} > + > +static const struct mmc_host_ops octeon_mmc_ops = { > + .request = octeon_mmc_request, > + .set_ios = octeon_mmc_set_ios, > + .get_ro = mmc_gpio_get_ro, > + .get_cd = mmc_gpio_get_cd, > +}; > + > +static void octeon_mmc_set_clock(struct octeon_mmc_slot *slot, > + unsigned int clock) > +{ > + struct mmc_host *mmc = slot->mmc; > + > + clock = min(clock, mmc->f_max); > + clock = max(clock, mmc->f_min); > + slot->clock = clock; > +} > + > +static int octeon_mmc_initlowlevel(struct octeon_mmc_slot *slot) > +{ > + union cvmx_mio_emm_switch emm_switch; > + struct octeon_mmc_host *host = slot->host; > + > + host->emm_cfg |= 1ull << slot->bus_id; > + cvmx_write_csr(slot->host->base + OCT_MIO_EMM_CFG, host->emm_cfg); > + octeon_mmc_set_clock(slot, 400000); > + > + /* Program initial clock speed and power */ > + emm_switch.u64 = 0; > + emm_switch.s.power_class = 10; > + emm_switch.s.clk_hi = (slot->sclock / slot->clock) / 2; > + emm_switch.s.clk_lo = (slot->sclock / slot->clock) / 2; > + > + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64); > + emm_switch.s.bus_id = slot->bus_id; > + cvmx_write_csr(host->base + OCT_MIO_EMM_SWITCH, emm_switch.u64); > + slot->cached_switch = emm_switch.u64; > + > + cvmx_write_csr(host->base + OCT_MIO_EMM_WDOG, > + ((u64)slot->clock * 850ull) / 1000ull); > + cvmx_write_csr(host->base + OCT_MIO_EMM_STS_MASK, 0xe4f90080ull); > + cvmx_write_csr(host->base + OCT_MIO_EMM_RCA, 1); > + return 0; > +} > + > +static int octeon_mmc_of_copy_legacy_u32(struct device_node *node, > + const char *legacy_name, > + const char *new_name) > +{ > + u32 value; > + int ret; > + > + ret = of_property_read_u32(node, legacy_name, &value); > + if (!ret) { > + /* Found legacy - set generic property */ > + struct property *new_p; > + u32 *new_v; > + > + pr_warn(FW_WARN "%s: Legacy property '%s'. Please remove\n", > + node->full_name, legacy_name); > + > + new_p = kzalloc(sizeof(*new_p), GFP_KERNEL); > + new_v = kzalloc(sizeof(u32), GFP_KERNEL); > + if (!new_p || !new_v) > + return -ENOMEM; > + > + *new_v = value; > + new_p->name = kstrdup(new_name, GFP_KERNEL); > + new_p->length = sizeof(u32); > + new_p->value = new_v; > + > + of_update_property(node, new_p); > + } > + return 0; > +} > + > +/* > + * This function parses the legacy device tree that may be found in devices > + * shipped before the driver was upstreamed. Future devices should not require > + * it as standard bindings should be used > + */ > +static int octeon_mmc_of_parse_legacy(struct device *dev, > + struct device_node *node, > + struct octeon_mmc_slot *slot) > +{ > + int ret; > + > + ret = octeon_mmc_of_copy_legacy_u32(node, "cavium,bus-max-width", > + "bus-width"); > + if (ret) > + return ret; > + > + ret = octeon_mmc_of_copy_legacy_u32(node, "spi-max-frequency", > + "max-frequency"); > + if (ret) > + return ret; > + > + slot->pwr_gpiod = devm_gpiod_get_optional(dev, "power", GPIOD_OUT_LOW); > + if (!IS_ERR(slot->pwr_gpiod)) { > + pr_warn(FW_WARN "%s: Legacy property '%s'. Please remove\n", > + node->full_name, "gpios-power"); > + } > + > + return 0; > +} > + > +static int octeon_mmc_slot_probe(struct platform_device *slot_pdev, > + struct octeon_mmc_host *host) > +{ > + struct mmc_host *mmc; > + struct octeon_mmc_slot *slot; > + struct device *dev = &slot_pdev->dev; > + struct device_node *node = slot_pdev->dev.of_node; > + u32 id, cmd_skew, dat_skew; > + u64 clock_period; > + int ret; > + > + ret = of_property_read_u32(node, "reg", &id); > + if (ret) { > + dev_err(dev, "Missing or invalid reg property on %s\n", > + of_node_full_name(node)); > + return ret; > + } > + > + if (id >= OCTEON_MAX_MMC || host->slot[id]) { > + dev_err(dev, "Invalid reg property on %s\n", > + of_node_full_name(node)); > + return -EINVAL; > + } > + > + mmc = mmc_alloc_host(sizeof(struct octeon_mmc_slot), dev); > + if (!mmc) { > + dev_err(dev, "alloc host failed\n"); > + return -ENOMEM; > + } > + > + slot = mmc_priv(mmc); > + slot->mmc = mmc; > + slot->host = host; > + > + /* Convert legacy DT entries into things mmc_of_parse can understand */ > + ret = octeon_mmc_of_parse_legacy(dev, node, slot); > + if (ret) > + return ret; > + > + ret = mmc_of_parse(mmc); > + if (ret) { > + dev_err(dev, "Failed to parse DT\n"); > + return ret; > + } > + > + /* Get regulators and the supported OCR mask */ > + ret = mmc_regulator_get_supply(mmc); > + if (ret == -EPROBE_DEFER) > + goto err; > + > + /* Octeon specific DT properties */ > + ret = of_property_read_u32(node, "cavium,cmd-clk-skew", &cmd_skew); > + if (ret) > + cmd_skew = 0; > + > + ret = of_property_read_u32(node, "cavium,dat-clk-skew", &dat_skew); > + if (ret) > + dat_skew = 0; > + > + /* > + * Set up host parameters. > + */ > + mmc->ops = &octeon_mmc_ops; > + mmc->f_min = 400000; > + if (!mmc->f_max) { > + mmc->f_max = 52000000; > + dev_info(dev, "No max-frequency for slot %u, defaulting to %u\n", > + id, mmc->f_max); > + } > + > + mmc->caps |= MMC_CAP_MMC_HIGHSPEED | MMC_CAP_SD_HIGHSPEED | > + MMC_CAP_ERASE; > + mmc->ocr_avail = MMC_VDD_27_28 | MMC_VDD_28_29 | MMC_VDD_29_30 | > + MMC_VDD_30_31 | MMC_VDD_31_32 | MMC_VDD_32_33 | > + MMC_VDD_33_34 | MMC_VDD_34_35 | MMC_VDD_35_36; > + > + /* post-sdk23 caps */ > + mmc->caps |= > + ((mmc->f_max >= 12000000) * MMC_CAP_UHS_SDR12) | > + ((mmc->f_max >= 25000000) * MMC_CAP_UHS_SDR25) | > + ((mmc->f_max >= 50000000) * MMC_CAP_UHS_SDR50) | > + MMC_CAP_CMD23; > + > + if ((!IS_ERR(mmc->supply.vmmc)) || (slot->pwr_gpiod)) > + mmc->caps |= MMC_CAP_POWER_OFF_CARD; > + > + /* "1.8v" capability is actually 1.8-or-3.3v */ > + if (ddr) > + mmc->caps |= MMC_CAP_UHS_DDR50 | MMC_CAP_1_8V_DDR; > + > + mmc->max_segs = 64; > + mmc->max_seg_size = host->linear_buf_size; > + mmc->max_req_size = host->linear_buf_size; > + mmc->max_blk_size = 512; > + mmc->max_blk_count = mmc->max_req_size / 512; > + > + slot->clock = mmc->f_min; > + slot->sclock = octeon_get_io_clock_rate(); > + > + clock_period = 1000000000000ull / slot->sclock; /* period in pS */ > + slot->cmd_cnt = (cmd_skew + clock_period / 2) / clock_period; > + slot->dat_cnt = (dat_skew + clock_period / 2) / clock_period; > + > + slot->bus_id = id; > + slot->cached_rca = 1; > + > + /* Only a single user of the bootbus at a time. */ > + octeon_mmc_acquire_bus(host); > + host->slot[id] = slot; > + > + octeon_mmc_switch_to(slot); > + /* Initialize MMC Block. */ > + octeon_mmc_initlowlevel(slot); > + > + octeon_mmc_release_bus(host); > + > + ret = mmc_add_host(mmc); > + if (ret) { > + dev_err(dev, "mmc_add_host() returned %d\n", ret); > + goto err; > + } > + > + return 0; > + > +err: > + slot->host->slot[id] = NULL; > + > + gpiod_set_value_cansleep(slot->pwr_gpiod, 0); > + > + mmc_free_host(slot->mmc); > + return ret; > +} > + > +static int octeon_mmc_slot_remove(struct octeon_mmc_slot *slot) > +{ > + mmc_remove_host(slot->mmc); > + > + slot->host->slot[slot->bus_id] = NULL; > + > + gpiod_set_value_cansleep(slot->pwr_gpiod, 0); > + > + mmc_free_host(slot->mmc); > + > + return 0; > +} > + > +static int octeon_mmc_probe(struct platform_device *pdev) > +{ > + struct octeon_mmc_host *host; > + struct resource *res; > + void __iomem *base; > + int mmc_irq[9]; > + int i; > + int ret = 0; > + struct device_node *node = pdev->dev.of_node; > + struct device_node *cn; > + bool cn78xx_style; > + u64 t; > + > + host = devm_kzalloc(&pdev->dev, sizeof(*host), GFP_KERNEL); > + if (!host) > + return -ENOMEM; > + > + spin_lock_init(&host->irq_handler_lock); > + sema_init(&host->mmc_serializer, 1); > + > + cn78xx_style = of_device_is_compatible(node, "cavium,octeon-7890-mmc"); > + if (cn78xx_style) { > + host->need_bootbus_lock = false; > + host->big_dma_addr = true; > + host->need_irq_handler_lock = true; > + /* > + * First seven are the EMM_INT bits 0..6, then two for > + * the EMM_DMA_INT bits > + */ > + for (i = 0; i < 9; i++) { > + mmc_irq[i] = platform_get_irq(pdev, i); > + if (mmc_irq[i] < 0) > + return mmc_irq[i]; > + } > + } else { > + host->need_bootbus_lock = true; > + host->big_dma_addr = false; > + host->need_irq_handler_lock = false; > + /* First one is EMM second NDF_DMA */ > + for (i = 0; i < 2; i++) { > + mmc_irq[i] = platform_get_irq(pdev, i); > + if (mmc_irq[i] < 0) > + return mmc_irq[i]; > + } > + } > + host->last_slot = -1; > + > + if (bb_size < 512 || bb_size >= (1 << 24)) > + bb_size = 1 << 18; > + host->linear_buf_size = bb_size; > + host->linear_buf = devm_kzalloc(&pdev->dev, host->linear_buf_size, > + GFP_KERNEL); > + > + if (!host->linear_buf) { > + dev_err(&pdev->dev, "devm_kzalloc failed\n"); > + return -ENOMEM; > + } > + > + host->pdev = pdev; > + > + res = platform_get_resource(pdev, IORESOURCE_MEM, 0); > + if (!res) { > + dev_err(&pdev->dev, "Platform resource[0] is missing\n"); > + return -ENXIO; > + } > + base = devm_ioremap_resource(&pdev->dev, res); > + if (IS_ERR(base)) > + return PTR_ERR(base); > + host->base = (__force u64)base; > + > + res = platform_get_resource(pdev, IORESOURCE_MEM, 1); > + if (!res) { > + dev_err(&pdev->dev, "Platform resource[1] is missing\n"); > + return -EINVAL; > + } > + base = devm_ioremap_resource(&pdev->dev, res); > + if (IS_ERR(base)) > + return PTR_ERR(base); > + host->ndf_base = (__force u64)base; > + /* > + * Clear out any pending interrupts that may be left over from > + * bootloader. > + */ > + t = cvmx_read_csr(host->base + OCT_MIO_EMM_INT); > + cvmx_write_csr(host->base + OCT_MIO_EMM_INT, t); > + if (cn78xx_style) { > + /* Only CMD_DONE, DMA_DONE, CMD_ERR, DMA_ERR */ > + for (i = 1; i <= 4; i++) { > + ret = devm_request_irq(&pdev->dev, mmc_irq[i], > + octeon_mmc_interrupt, > + 0, DRV_NAME, host); > + if (ret < 0) { > + dev_err(&pdev->dev, "Error: devm_request_irq %d\n", > + mmc_irq[i]); > + return ret; > + } > + } > + } else { > + ret = devm_request_irq(&pdev->dev, mmc_irq[0], > + octeon_mmc_interrupt, 0, DRV_NAME, host); > + if (ret < 0) { > + dev_err(&pdev->dev, "Error: devm_request_irq %d\n", > + mmc_irq[0]); > + return ret; > + } > + } > + > + host->global_pwr_gpiod = devm_gpiod_get_optional(&pdev->dev, "power", > + GPIOD_OUT_HIGH); > + if (IS_ERR(host->global_pwr_gpiod)) { > + dev_err(&host->pdev->dev, "Invalid POWER GPIO\n"); > + return PTR_ERR(host->global_pwr_gpiod); > + } > + > + platform_set_drvdata(pdev, host); > + > + for_each_child_of_node(node, cn) { > + struct platform_device *slot_pdev; > + > + slot_pdev = of_platform_device_create(cn, NULL, &pdev->dev); > + ret = octeon_mmc_slot_probe(slot_pdev, host); > + if (ret) { > + dev_err(&host->pdev->dev, "Error populating slots\n"); > + gpiod_set_value_cansleep(host->global_pwr_gpiod, 0); > + return ret; > + } > + } > + > + return 0; > +} > + > +static int octeon_mmc_remove(struct platform_device *pdev) > +{ > + union cvmx_mio_ndf_dma_cfg ndf_dma_cfg; > + struct octeon_mmc_host *host = platform_get_drvdata(pdev); > + int i; > + > + for (i = 0; i < OCTEON_MAX_MMC; i++) { > + if (host->slot[i]) > + octeon_mmc_slot_remove(host->slot[i]); > + } > + > + ndf_dma_cfg.u64 = cvmx_read_csr(host->ndf_base + OCT_MIO_NDF_DMA_CFG); > + ndf_dma_cfg.s.en = 0; > + cvmx_write_csr(host->ndf_base + OCT_MIO_NDF_DMA_CFG, ndf_dma_cfg.u64); > + > + gpiod_set_value_cansleep(host->global_pwr_gpiod, 0); > + > + return 0; > +} > + > +static const struct of_device_id octeon_mmc_match[] = { > + { > + .compatible = "cavium,octeon-6130-mmc", > + }, > + { > + .compatible = "cavium,octeon-7890-mmc", > + }, > + {}, > +}; > +MODULE_DEVICE_TABLE(of, octeon_mmc_match); > + > +static struct platform_driver octeon_mmc_driver = { > + .probe = octeon_mmc_probe, > + .remove = octeon_mmc_remove, > + .driver = { > + .name = DRV_NAME, > + .of_match_table = octeon_mmc_match, > + }, > +}; > + > +static int __init octeon_mmc_init(void) > +{ > + return platform_driver_register(&octeon_mmc_driver); > +} > + > +static void __exit octeon_mmc_cleanup(void) > +{ > + platform_driver_unregister(&octeon_mmc_driver); > +} > + > +module_init(octeon_mmc_init); > +module_exit(octeon_mmc_cleanup); > + > +MODULE_AUTHOR("Cavium Inc. <support@xxxxxxxxxx>"); > +MODULE_DESCRIPTION("low-level driver for Cavium OCTEON MMC/SSD card"); > +MODULE_LICENSE("GPL"); > -- > 2.5.0 >