Add a driver for MediaTek I3C master IP. Signed-off-by: Qii Wang <qii.wang@xxxxxxxxxxxx> --- drivers/i3c/master/Kconfig | 10 + drivers/i3c/master/Makefile | 1 + drivers/i3c/master/i3c-master-mtk.c | 1239 +++++++++++++++++++++++++++++++++++ 3 files changed, 1250 insertions(+) create mode 100644 drivers/i3c/master/i3c-master-mtk.c diff --git a/drivers/i3c/master/Kconfig b/drivers/i3c/master/Kconfig index 26c6b58..acc00d9 100644 --- a/drivers/i3c/master/Kconfig +++ b/drivers/i3c/master/Kconfig @@ -20,3 +20,13 @@ config DW_I3C_MASTER This driver can also be built as a module. If so, the module will be called dw-i3c-master. + +config MTK_I3C_MASTER + tristate "MediaTek I3C master driver" + depends on I3C + depends on HAS_IOMEM + depends on !(ALPHA || PARISC) + help + This selects the MediaTek(R) I3C master controller driver. + If you want to use MediaTek(R) I3C interface, say Y here. + If unsure, say N or M. diff --git a/drivers/i3c/master/Makefile b/drivers/i3c/master/Makefile index fc53939..fe7ccf5 100644 --- a/drivers/i3c/master/Makefile +++ b/drivers/i3c/master/Makefile @@ -1,2 +1,3 @@ obj-$(CONFIG_CDNS_I3C_MASTER) += i3c-master-cdns.o obj-$(CONFIG_DW_I3C_MASTER) += dw-i3c-master.o +obj-$(CONFIG_MTK_I3C_MASTER) += i3c-master-mtk.o diff --git a/drivers/i3c/master/i3c-master-mtk.c b/drivers/i3c/master/i3c-master-mtk.c new file mode 100644 index 0000000..3981149 --- /dev/null +++ b/drivers/i3c/master/i3c-master-mtk.c @@ -0,0 +1,1239 @@ +// SPDX-License-Identifier: GPL-2.0 +/* + * Copyright (C) 2019 MediaTek Design Systems Inc. + * + * Author: Qii Wang <qii.wang@xxxxxxxxxxxx> + */ + +#include <linux/clk.h> +#include <linux/dma-mapping.h> +#include <linux/i3c/master.h> +#include <linux/interrupt.h> +#include <linux/iopoll.h> +#include <linux/platform_device.h> +#include <linux/slab.h> + +#define DRV_NAME "i3c-master-mtk" + +#define SLAVE_ADDR 0x04 +#define INTR_MASK 0x08 +#define INTR_STAT 0x0c +#define INTR_TRANSAC_COMP BIT(0) +#define INTR_ACKERR GENMASK(2, 1) +#define INTR_ARB_LOST BIT(3) +#define INTR_RS_MULTI BIT(4) +#define INTR_MAS_ERR BIT(8) +#define INTR_ALL (INTR_MAS_ERR | INTR_ARB_LOST |\ + INTR_ACKERR | INTR_TRANSAC_COMP) + +#define CONTROL 0x10 +#define CONTROL_WRAPPER BIT(0) +#define CONTROL_RS BIT(1) +#define CONTROL_DMA_EN BIT(2) +#define CONTROL_CLK_EXT_EN BIT(3) +#define CONTROL_DIR_CHANGE BIT(4) +#define CONTROL_ACKERR_DET_EN BIT(5) +#define CONTROL_LEN_CHANGE BIT(6) +#define CONTROL_DMAACK_EN BIT(8) +#define CONTROL_ASYNC_MODE BIT(9) + +#define TRANSFER_LEN 0x14 +#define TRANSAC_LEN 0x18 +#define TRANSAC_LEN_WRRD 0x0002 +#define TRANS_ONE_LEN 0x0001 + +#define DELAY_LEN 0x1c +#define DELAY_LEN_DEFAULT 0x000a + +#define TIMING 0x20 +#define TIMING_VALUE(sample_cnt, step_cnt) ({ \ + typeof(sample_cnt) sample_cnt_ = (sample_cnt); \ + typeof(step_cnt) step_cnt_ = (step_cnt); \ + (((sample_cnt_) << 8) | (step_cnt_)); \ +}) + +#define START 0x24 +#define START_EN BIT(0) +#define START_MUL_TRIG BIT(14) +#define START_MUL_CNFG BIT(15) + +#define EXT_CONF 0x28 +#define EXT_CONF_DEFAULT 0x0a1f + +#define LTIMING 0x2c +#define LTIMING_VALUE(sample_cnt, step_cnt) ({ \ + typeof(sample_cnt) sample_cnt_ = (sample_cnt); \ + typeof(step_cnt) step_cnt_ = (step_cnt); \ + (((sample_cnt_) << 6) | (step_cnt_) | \ + ((sample_cnt_) << 12) | ((step_cnt_) << 9)); \ +}) + +#define HS 0x30 +#define HS_CLR_VALUE 0x0000 +#define HS_DEFAULT_VALUE 0x0083 +#define HS_VALUE(sample_cnt, step_cnt) ({ \ + typeof(sample_cnt) sample_cnt_ = (sample_cnt); \ + typeof(step_cnt) step_cnt_ = (step_cnt); \ + (HS_DEFAULT_VALUE | \ + ((sample_cnt_) << 12) | ((step_cnt_) << 8)); \ +}) + +#define IO_CONFIG 0x34 +#define IO_CONFIG_PUSH_PULL 0x0000 + +#define FIFO_ADDR_CLR 0x38 +#define FIFO_CLR 0x0003 + +#define MCU_INTR 0x40 +#define MCU_INTR_EN BIT(0) + +#define TRANSFER_LEN_AUX 0x44 +#define CLOCK_DIV 0x48 +#define CLOCK_DIV_DEFAULT ((INTER_CLK_DIV - 1) << 8 |\ + (INTER_CLK_DIV - 1)) + +#define SOFTRESET 0x50 +#define SOFT_RST BIT(0) + +#define TRAFFIC 0x54 +#define TRAFFIC_DAA_EN BIT(4) +#define TRAFFIC_TBIT BIT(7) +#define TRAFFIC_HEAD_ONLY BIT(9) +#define TRAFFIC_SKIP_SLV_ADDR BIT(10) +#define TRAFFIC_HANDOFF BIT(14) + +#define DEF_DA 0x68 +#define DEF_DAA_SLV_PARITY BIT(8) + +#define SHAPE 0x6c +#define SHAPE_T_STALL BIT(1) +#define SHAPE_T_PARITY BIT(2) + +#define HFIFO_DATA 0x70 +#define NINTH_BIT_IGNORE 0 +#define NINTH_BIT_ACK 1 +#define NINTH_BIT_NACK 2 +#define NINTH_BIT_ODD_PAR 3 +#define INST_WITH_HS BIT(10) +#define UNLOCK_HFIFO BIT(15) +#define HFIFO_DATA_08 0x8208 +#define HFIFO_DATA_7E (UNLOCK_HFIFO |\ + (I3C_BROADCAST_ADDR << 1) |\ + (NINTH_BIT_ACK << 8)) +#define HFIFO_HEAD (UNLOCK_HFIFO | INST_WITH_HS |\ + (NINTH_BIT_ODD_PAR << 8)) + +#define FIFO_STAT 0xf4 +#define FIFO_DAA 0x0080 + +#define DMA_INT_FLAG 0x0 +#define DMA_EN 0x08 +#define DMA_RST 0x0c +#define DMA_CON 0x18 +#define DMA_TX_MEM_ADDR 0x1c +#define DMA_RX_MEM_ADDR 0x20 +#define DMA_TX_LEN 0x24 +#define DMA_RX_LEN 0x28 +#define DMA_TX_4G_MODE 0x54 +#define DMA_RX_4G_MODE 0x58 + +#define DMA_EN_START BIT(0) +#define DMA_RST_HARD BIT(1) +#define DMA_4G_MODE BIT(0) +#define DMA_CLR_FLAG 0x0000 +#define DMA_CON_TX 0x0000 +#define DMA_CON_RX 0x0001 + +#define INTER_CLK_DIV 5 +#define MAX_SAMPLE_CNT 8 +#define MAX_STEP_CNT 64 +#define MAX_HS_STEP_CNT 8 +#define MAX_I3C_DEVS 16 + +#define MTK_I3C_BCR 0x01 +#define MTK_I3C_DCR 0x02 +#define MTK_I3C_PID 0x03040506 + +enum mtk_trans_op { + MASTER_WR = 1, + MASTER_RD, + MASTER_WRRD, + /* I3C private op */ + MASTER_CCC_BROADCAST, + MASTER_DAA, +}; + +enum mtk_trans_mode { + I2C_TRANSFER = 1, + I3C_TRANSFER, + I3C_CCC, +}; + +struct daa_addr_ary { + u8 addr; + bool used; +}; + +struct daa_anchor { + struct daa_addr_ary daa_addr[MAX_I3C_DEVS]; + int idx; +}; + +struct mtk_i3c_cmd { + enum mtk_trans_op op; + u8 ccc_id; + u16 addr; /* device addr */ + + u16 tx_len; + const void *tx_buf; + u16 rx_len; + void *rx_buf; +}; + +struct mtk_i3c_xfer { + struct list_head node; + struct completion complete; /* i3c transfer stop */ + + int ret; + u16 irq_stat; /* interrupt status */ + enum mtk_trans_mode mode; + bool ignore_restart_irq; + bool auto_restart; + unsigned int ncmds; + unsigned int trans_num; + u8 *msg_buf_w; + u8 *msg_buf_r; + dma_addr_t rpaddr; + dma_addr_t wpaddr; + struct mtk_i3c_cmd cmds[0]; +}; + +struct mtk_i3c_master { + struct device *dev; + struct i3c_master_controller mas_ctrler; + + /* set in i3c probe */ + void __iomem *regs; /* i3c base addr */ + void __iomem *dma_regs; /* apdma base address*/ + struct clk *clk_main; /* main clock for i3c bus */ + struct clk *clk_dma; /* DMA clock for i3c via DMA */ + struct clk *clk_arb; /* Arbitrator clock for i3c */ + struct daa_anchor daa_anchor; + + u16 timing_reg[2]; + u16 ltiming_reg[2]; + u16 high_speed_reg[2]; + + struct { + struct list_head list; + struct mtk_i3c_xfer *cur; + spinlock_t lock; /* Lock for stats update */ + } xferqueue; +}; + +static u16 mtk_i3c_readw(struct mtk_i3c_master *master, u16 offset) +{ + return readw(master->regs + offset); +} + +static void mtk_i3c_writew(struct mtk_i3c_master *master, u16 val, u16 offset) +{ + writew(val, master->regs + offset); +} + +/** + * Calculate i3c port speed + * + * Hardware design: + * i3c_bus_freq = parent_clk / (clock_div * 2 * sample_cnt * step_cnt) + * clock_div: fixed in hardware, but may be various in different SoCs + * + * The calculation want to pick the highest bus frequency that is still + * less than or equal to master->speed_hz. The calculation try to get + * sample_cnt and step_cn + */ +static int mtk_i3c_calculate_speed(struct mtk_i3c_master *master, + unsigned int clk_src, + unsigned int target_speed, + unsigned int *timing_step_cnt, + unsigned int *timing_sample_cnt) +{ + unsigned int sample_cnt, step_cnt, max_step_cnt, opt_div; + unsigned int best_mul, cnt_mul, base_step_cnt; + unsigned int base_sample_cnt = MAX_SAMPLE_CNT; + + if (target_speed > I3C_BUS_I2C_FM_PLUS_SCL_RATE) + max_step_cnt = MAX_HS_STEP_CNT; + else + max_step_cnt = MAX_STEP_CNT; + + base_step_cnt = max_step_cnt; + /* Find the best combination */ + opt_div = DIV_ROUND_UP(clk_src >> 1, target_speed); + best_mul = MAX_SAMPLE_CNT * max_step_cnt; + + /* Search for the best pair (sample_cnt, step_cnt) with + * 1 < sample_cnt < MAX_SAMPLE_CNT + * 1 < step_cnt < max_step_cnt + * sample_cnt * step_cnt >= opt_div + * optimizing for sample_cnt * step_cnt being minimal + */ + for (sample_cnt = 2; sample_cnt <= MAX_SAMPLE_CNT; sample_cnt++) { + step_cnt = DIV_ROUND_UP(opt_div, sample_cnt); + cnt_mul = step_cnt * sample_cnt; + if (step_cnt > max_step_cnt || step_cnt < 2) + continue; + + if (cnt_mul < best_mul) { + best_mul = cnt_mul; + base_sample_cnt = sample_cnt; + base_step_cnt = step_cnt; + if (best_mul == opt_div) + break; + } + } + + sample_cnt = base_sample_cnt; + step_cnt = base_step_cnt; + + if ((clk_src / (2 * sample_cnt * step_cnt)) > target_speed) { + dev_err(master->dev, "Unsupport speed (%uhz)\n", target_speed); + return -EINVAL; + } + + *timing_step_cnt = step_cnt - 1; + *timing_sample_cnt = sample_cnt - 1; + + return 0; +} + +static int mtk_i3c_set_speed(struct mtk_i3c_master *master, + unsigned int clk_src) +{ + struct i3c_bus *bus = i3c_master_get_bus(&master->mas_ctrler); + unsigned int step_cnt, sample_cnt; + int ret; + + ret = mtk_i3c_calculate_speed(master, clk_src, bus->scl_rate.i2c, + &step_cnt, &sample_cnt); + if (ret < 0) + return ret; + + master->timing_reg[0] = TIMING_VALUE(sample_cnt, step_cnt); + /* Disable the high speed transaction */ + master->high_speed_reg[0] = HS_CLR_VALUE; + master->ltiming_reg[0] = LTIMING_VALUE(sample_cnt, step_cnt); + + ret = mtk_i3c_calculate_speed(master, clk_src, bus->scl_rate.i3c, + &step_cnt, &sample_cnt); + if (ret < 0) + return ret; + + master->timing_reg[1] = TIMING_VALUE(sample_cnt, step_cnt); + master->high_speed_reg[1] = HS_VALUE(sample_cnt, step_cnt); + master->ltiming_reg[1] = LTIMING_VALUE(sample_cnt, step_cnt); + + return 0; +} + +static inline u32 mtk_i3c_set_4g_mode(dma_addr_t addr) +{ + return (addr & BIT_ULL(32)) ? DMA_4G_MODE : DMA_CLR_FLAG; +} + +static inline struct mtk_i3c_master * +to_mtk_i3c_master(struct i3c_master_controller *master) +{ + return container_of(master, struct mtk_i3c_master, mas_ctrler); +} + +static u8 mtk_i3c_master_get_free_addr(struct mtk_i3c_master *master) +{ + int slot = 0; + + for (slot = 0; slot < MAX_I3C_DEVS; slot++) { + if (master->daa_anchor.daa_addr[slot].used) + continue; + + master->daa_anchor.idx = slot; + master->daa_anchor.daa_addr[slot].used = true; + return master->daa_anchor.daa_addr[slot].addr; + } + + return 0; +} + +static int mtk_i3c_master_clock_enable(struct mtk_i3c_master *master) +{ + int ret; + + ret = clk_prepare_enable(master->clk_dma); + if (ret) + return ret; + + ret = clk_prepare_enable(master->clk_main); + if (ret) + goto err_main; + + ret = clk_prepare_enable(master->clk_arb); + if (ret) + goto err_arb; + + return 0; + +err_arb: + clk_disable_unprepare(master->clk_main); +err_main: + clk_disable_unprepare(master->clk_dma); + + return ret; +} + +static void mtk_i3c_master_clock_disable(struct mtk_i3c_master *master) +{ + clk_disable_unprepare(master->clk_arb); + clk_disable_unprepare(master->clk_main); + clk_disable_unprepare(master->clk_dma); +} + +static inline void mtk_i3c_master_init_hw(struct mtk_i3c_master *master) +{ + struct mtk_i3c_xfer *mtk_xfer = master->xferqueue.cur; + + if (!mtk_xfer->trans_num) + mtk_i3c_writew(master, SOFT_RST, SOFTRESET); + + mtk_i3c_writew(master, IO_CONFIG_PUSH_PULL, IO_CONFIG); + mtk_i3c_writew(master, CLOCK_DIV_DEFAULT, CLOCK_DIV); + + mtk_i3c_writew(master, DELAY_LEN_DEFAULT, DELAY_LEN); + mtk_i3c_writew(master, EXT_CONF_DEFAULT, EXT_CONF); + mtk_i3c_writew(master, FIFO_CLR, FIFO_ADDR_CLR); + + /* DMA hard reset */ + writel(DMA_RST_HARD, master->dma_regs + DMA_RST); + writel(DMA_CLR_FLAG, master->dma_regs + DMA_RST); +} + +static int mtk_i3c_master_apdma_tx(struct mtk_i3c_master *master) +{ + struct mtk_i3c_xfer *xfer = master->xferqueue.cur; + struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num]; + u32 reg_4g_mode; + + xfer->msg_buf_w = kzalloc(cmd->tx_len, GFP_KERNEL); + if (!xfer->msg_buf_w) + return -ENOMEM; + + memcpy(xfer->msg_buf_w, cmd->tx_buf, cmd->tx_len); + + writel(DMA_CLR_FLAG, master->dma_regs + DMA_INT_FLAG); + writel(DMA_CON_TX, master->dma_regs + DMA_CON); + + xfer->wpaddr = dma_map_single(master->dev, xfer->msg_buf_w, + cmd->tx_len, DMA_TO_DEVICE); + if (dma_mapping_error(master->dev, xfer->wpaddr)) { + kfree(xfer->msg_buf_w); + return -ENOMEM; + } + + reg_4g_mode = mtk_i3c_set_4g_mode(xfer->wpaddr); + writel(reg_4g_mode, master->dma_regs + DMA_TX_4G_MODE); + writel((u32)xfer->wpaddr, master->dma_regs + DMA_TX_MEM_ADDR); + writel(cmd->tx_len, master->dma_regs + DMA_TX_LEN); + + return 0; +} + +static int mtk_i3c_master_apdma_rx(struct mtk_i3c_master *master) +{ + struct mtk_i3c_xfer *xfer = master->xferqueue.cur; + struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num]; + u32 reg_4g_mode; + + xfer->msg_buf_r = kzalloc(cmd->rx_len, GFP_KERNEL); + if (!xfer->msg_buf_r) + return -ENOMEM; + + writel(DMA_CLR_FLAG, master->dma_regs + DMA_INT_FLAG); + writel(DMA_CON_RX, master->dma_regs + DMA_CON); + + xfer->rpaddr = dma_map_single(master->dev, xfer->msg_buf_r, + cmd->rx_len, DMA_FROM_DEVICE); + if (dma_mapping_error(master->dev, xfer->rpaddr)) { + kfree(xfer->msg_buf_r); + return -ENOMEM; + } + + reg_4g_mode = mtk_i3c_set_4g_mode(xfer->rpaddr); + writel(reg_4g_mode, master->dma_regs + DMA_RX_4G_MODE); + writel((u32)xfer->rpaddr, master->dma_regs + DMA_RX_MEM_ADDR); + writel(cmd->rx_len, master->dma_regs + DMA_RX_LEN); + + return 0; +} + +static void mtk_i3c_master_apdma_end(struct mtk_i3c_master *master) +{ + struct mtk_i3c_xfer *xfer = master->xferqueue.cur; + struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num]; + u16 int_reg = 0; + + if (xfer->auto_restart) + int_reg = INTR_RS_MULTI; + + int_reg |= INTR_ALL; + + /* Clear interrupt mask */ + mtk_i3c_writew(master, ~int_reg, INTR_MASK); + + if (cmd->op == MASTER_WR) { + dma_unmap_single(master->dev, xfer->wpaddr, + cmd->tx_len, DMA_TO_DEVICE); + kfree(xfer->msg_buf_w); + + } else if (cmd->op == MASTER_RD) { + dma_unmap_single(master->dev, xfer->rpaddr, + cmd->rx_len, DMA_FROM_DEVICE); + memcpy(cmd->rx_buf, xfer->msg_buf_r, cmd->rx_len); + kfree(xfer->msg_buf_r); + } else if (cmd->op == MASTER_WRRD) { + dma_unmap_single(master->dev, xfer->wpaddr, cmd->tx_len, + DMA_TO_DEVICE); + dma_unmap_single(master->dev, xfer->rpaddr, cmd->rx_len, + DMA_FROM_DEVICE); + memcpy(cmd->rx_buf, xfer->msg_buf_r, cmd->rx_len); + kfree(xfer->msg_buf_w); + kfree(xfer->msg_buf_r); + } +} + +static int mtk_i3c_master_start_transfer(struct mtk_i3c_master *master) +{ + struct mtk_i3c_xfer *xfer = master->xferqueue.cur; + struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num]; + u16 addr_reg, control_reg, start_reg, tmp_rg_val; + u16 int_reg = 0, data_size; + int ret = 0; + u8 *ptr; + + xfer->irq_stat = 0; + + mtk_i3c_master_init_hw(master); + + control_reg = CONTROL_ACKERR_DET_EN | CONTROL_CLK_EXT_EN | CONTROL_RS; + if (cmd->op != MASTER_DAA) + control_reg |= CONTROL_DMA_EN | CONTROL_DMAACK_EN | + CONTROL_ASYNC_MODE; + + if (cmd->op == MASTER_WRRD) + control_reg |= CONTROL_DIR_CHANGE; + + mtk_i3c_writew(master, control_reg, CONTROL); + + addr_reg = cmd->addr << 1; + if (cmd->op == MASTER_RD || cmd->op == MASTER_DAA) + addr_reg |= 0x1; + + mtk_i3c_writew(master, addr_reg, SLAVE_ADDR); + + if (xfer->auto_restart) + int_reg = INTR_RS_MULTI; + + int_reg |= INTR_ALL; + /* Clear interrupt status */ + mtk_i3c_writew(master, int_reg, INTR_STAT); + /* Enable interrupt */ + mtk_i3c_writew(master, int_reg, INTR_MASK); + + /* Set transfer and transaction len */ + switch (cmd->op) { + case MASTER_WRRD: + mtk_i3c_writew(master, cmd->tx_len, TRANSFER_LEN); + mtk_i3c_writew(master, cmd->rx_len, TRANSFER_LEN_AUX); + mtk_i3c_writew(master, TRANSAC_LEN_WRRD, TRANSAC_LEN); + break; + case MASTER_WR: + mtk_i3c_writew(master, cmd->tx_len, TRANSFER_LEN); + mtk_i3c_writew(master, xfer->ncmds, TRANSAC_LEN); + break; + case MASTER_RD: + mtk_i3c_writew(master, cmd->rx_len, TRANSFER_LEN); + mtk_i3c_writew(master, xfer->ncmds, TRANSAC_LEN); + break; + case MASTER_CCC_BROADCAST: + mtk_i3c_writew(master, TRANS_ONE_LEN, TRANSFER_LEN); + mtk_i3c_writew(master, TRANS_ONE_LEN, TRANSAC_LEN); + break; + case MASTER_DAA: + mtk_i3c_writew(master, cmd->rx_len + cmd->tx_len, TRANSFER_LEN); + mtk_i3c_writew(master, MAX_I3C_DEVS, TRANSAC_LEN); + break; + } + + mtk_i3c_writew(master, master->timing_reg[1], TIMING); + mtk_i3c_writew(master, master->ltiming_reg[1], LTIMING); + mtk_i3c_writew(master, master->high_speed_reg[1], HS); + + if (xfer->mode == I2C_TRANSFER) { + mtk_i3c_writew(master, master->timing_reg[0], TIMING); + mtk_i3c_writew(master, master->ltiming_reg[0], LTIMING); + mtk_i3c_writew(master, master->high_speed_reg[0], HS); + mtk_i3c_writew(master, 0, TRAFFIC); + mtk_i3c_writew(master, 0, SHAPE); + } else if ((xfer->mode == I3C_TRANSFER) && (!xfer->trans_num)) { + mtk_i3c_writew(master, HFIFO_DATA_7E, HFIFO_DATA); + mtk_i3c_writew(master, TRAFFIC_HANDOFF | TRAFFIC_TBIT, TRAFFIC); + mtk_i3c_writew(master, SHAPE_T_STALL | SHAPE_T_PARITY, SHAPE); + } else if (xfer->mode == I3C_CCC) { + tmp_rg_val = TRAFFIC_HANDOFF | TRAFFIC_TBIT; + mtk_i3c_writew(master, HFIFO_DATA_7E, HFIFO_DATA); + mtk_i3c_writew(master, HFIFO_HEAD | cmd->ccc_id, HFIFO_DATA); + + if (cmd->op == MASTER_DAA) { + addr_reg = mtk_i3c_master_get_free_addr(master); + addr_reg |= DEF_DAA_SLV_PARITY; + tmp_rg_val |= TRAFFIC_DAA_EN; + mtk_i3c_writew(master, addr_reg, DEF_DA); + } else if (cmd->op == MASTER_CCC_BROADCAST) { + tmp_rg_val |= TRAFFIC_HEAD_ONLY | TRAFFIC_SKIP_SLV_ADDR; + data_size = cmd->tx_len; + ptr = (u8 *)cmd->tx_buf; + while (data_size--) + mtk_i3c_writew(master, HFIFO_HEAD | *ptr++, + HFIFO_DATA); + } + + mtk_i3c_writew(master, tmp_rg_val, TRAFFIC); + mtk_i3c_writew(master, SHAPE_T_STALL, SHAPE); + } + + /* Prepare buffer data to start transfer */ + if (cmd->op == MASTER_RD) { + ret = mtk_i3c_master_apdma_rx(master); + } else if (cmd->op == MASTER_WR) { + ret = mtk_i3c_master_apdma_tx(master); + } else if (cmd->op == MASTER_WRRD) { + ret = mtk_i3c_master_apdma_rx(master); + ret |= mtk_i3c_master_apdma_tx(master); + } + + if (ret) + return ret; + + if (cmd->op != MASTER_DAA) + writel(DMA_EN_START, master->dma_regs + DMA_EN); + + start_reg = START_EN; + if (xfer->auto_restart) { + start_reg |= START_MUL_TRIG; + if ((xfer->ncmds - xfer->trans_num) >= 2 || + cmd->op == MASTER_DAA) + start_reg |= START_MUL_CNFG; + } + + mtk_i3c_writew(master, MCU_INTR_EN, MCU_INTR); + mtk_i3c_writew(master, start_reg, START); + + return 0; +} + +static struct mtk_i3c_xfer * +mtk_i3c_master_alloc_xfer(struct mtk_i3c_master *master, unsigned int ncmds) +{ + struct mtk_i3c_xfer *xfer; + + xfer = kzalloc(struct_size(xfer, cmds, ncmds), GFP_KERNEL); + if (!xfer) + return NULL; + + INIT_LIST_HEAD(&xfer->node); + xfer->ncmds = ncmds; + xfer->ret = -ETIMEDOUT; + + return xfer; +} + +static void mtk_i3c_master_free_xfer(struct mtk_i3c_xfer *xfer) +{ + kfree(xfer); +} + +static void mtk_i3c_master_start_xfer_locked(struct mtk_i3c_master *master) +{ + struct mtk_i3c_xfer *xfer = master->xferqueue.cur; + unsigned int ret = 0; + + if (!xfer) + return; + + xfer->trans_num = 0; + ret = mtk_i3c_master_start_transfer(master); + xfer->ret = ret; +} + +static void mtk_i3c_master_end_xfer_locked(struct mtk_i3c_master *master) +{ + struct mtk_i3c_xfer *xfer = master->xferqueue.cur; + struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num]; + int slot = 0; + + if (!xfer) + return; + + if (xfer->irq_stat & (INTR_ACKERR | INTR_MAS_ERR)) { + if (cmd->op == MASTER_DAA) { + slot = master->daa_anchor.idx; + master->daa_anchor.daa_addr[slot].used = false; + } + + dev_err(master->dev, "Addr: %x, device ACK error\n", cmd->addr); + xfer->ret = -ENXIO; + } + + mtk_i3c_master_apdma_end(master); + + complete(&xfer->complete); + xfer = list_first_entry_or_null(&master->xferqueue.list, + struct mtk_i3c_xfer, node); + if (xfer) + list_del_init(&xfer->node); + + master->xferqueue.cur = xfer; + mtk_i3c_master_start_xfer_locked(master); +} + +static irqreturn_t mtk_i3c_master_irq(int irqno, void *dev_id) +{ + struct mtk_i3c_master *master = dev_id; + struct mtk_i3c_xfer *xfer = master->xferqueue.cur; + struct mtk_i3c_cmd *cmd = &xfer->cmds[xfer->trans_num]; + u16 restart_flag = 0, intr_stat, fifo_stat; + unsigned long flags; + int ret = 0; + u8 addr; + + if (cmd->op == MASTER_DAA) { + fifo_stat = mtk_i3c_readw(master, FIFO_STAT); + intr_stat = mtk_i3c_readw(master, INTR_STAT); + + if (!(fifo_stat & FIFO_DAA) && !(intr_stat & INTR_TRANSAC_COMP)) + return IRQ_NONE; + + mtk_i3c_writew(master, FIFO_CLR, FIFO_ADDR_CLR); + } + + if (xfer->auto_restart) + restart_flag = INTR_RS_MULTI; + + intr_stat = mtk_i3c_readw(master, INTR_STAT); + mtk_i3c_writew(master, intr_stat, INTR_STAT); + + spin_lock_irqsave(&master->xferqueue.lock, flags); + /* + * when occurs ack error, i3c controller generate two interrupts + * first is the ack error interrupt, then the complete interrupt + * i3c->irq_stat need keep the two interrupt value. + */ + xfer->irq_stat |= intr_stat; + + if (xfer->irq_stat & INTR_TRANSAC_COMP) { + mtk_i3c_master_end_xfer_locked(master); + goto exit_irq; + } + + if (xfer->irq_stat & restart_flag) { + xfer->irq_stat = 0; + + if (xfer->ignore_restart_irq) { + xfer->ignore_restart_irq = false; + + mtk_i3c_writew(master, START_MUL_CNFG | START_MUL_TRIG | + START_EN, START); + } else if (cmd->op == MASTER_DAA) { + addr = mtk_i3c_master_get_free_addr(master); + mtk_i3c_writew(master, addr, DEF_DA); + mtk_i3c_writew(master, START_MUL_CNFG | START_MUL_TRIG | + START_EN, START); + } else { + mtk_i3c_master_apdma_end(master); + + xfer->trans_num++; + ret = mtk_i3c_master_start_transfer(master); + xfer->ret = ret; + } + } + +exit_irq: + spin_unlock_irqrestore(&master->xferqueue.lock, flags); + return IRQ_HANDLED; +} + +static void mtk_i3c_master_queue_xfer(struct mtk_i3c_master *master, + struct mtk_i3c_xfer *xfer) +{ + unsigned long flags; + + init_completion(&xfer->complete); + spin_lock_irqsave(&master->xferqueue.lock, flags); + if (master->xferqueue.cur) { + list_add_tail(&xfer->node, &master->xferqueue.list); + } else { + master->xferqueue.cur = xfer; + mtk_i3c_master_start_xfer_locked(master); + } + spin_unlock_irqrestore(&master->xferqueue.lock, flags); +} + +static void mtk_i3c_master_unqueue_xfer(struct mtk_i3c_master *master, + struct mtk_i3c_xfer *xfer) +{ + struct mtk_i3c_cmd *ccmd = &xfer->cmds[xfer->trans_num]; + unsigned long flags; + + dev_err(master->dev, "addr: %x, transfer timeout\n", ccmd->addr); + xfer->ret = -ETIMEDOUT; + + spin_lock_irqsave(&master->xferqueue.lock, flags); + mtk_i3c_master_apdma_end(master); + if (master->xferqueue.cur == xfer) + master->xferqueue.cur = NULL; + else + list_del_init(&xfer->node); + spin_unlock_irqrestore(&master->xferqueue.lock, flags); +} + +static bool mtk_i3c_master_supports_ccc_cmd(struct i3c_master_controller *m, + const struct i3c_ccc_cmd *cmd) +{ + if (cmd->ndests > 1) + return false; + + switch (cmd->id) { + case I3C_CCC_ENEC(true): + case I3C_CCC_ENEC(false): + case I3C_CCC_DISEC(true): + case I3C_CCC_DISEC(false): + case I3C_CCC_ENTAS(0, true): + case I3C_CCC_ENTAS(0, false): + case I3C_CCC_RSTDAA(true): + case I3C_CCC_RSTDAA(false): + /* FALL THROUGH */ + case I3C_CCC_ENTDAA: + case I3C_CCC_SETMWL(true): + case I3C_CCC_SETMWL(false): + case I3C_CCC_SETMRL(true): + case I3C_CCC_SETMRL(false): + /* FALL THROUGH */ + case I3C_CCC_DEFSLVS: + case I3C_CCC_ENTHDR(0): + case I3C_CCC_SETDASA: + case I3C_CCC_SETNEWDA: + case I3C_CCC_GETMWL: + case I3C_CCC_GETMRL: + case I3C_CCC_GETPID: + case I3C_CCC_GETBCR: + case I3C_CCC_GETDCR: + case I3C_CCC_GETSTATUS: + case I3C_CCC_GETACCMST: + case I3C_CCC_GETMXDS: + case I3C_CCC_GETHDRCAP: + return true; + default: + break; + } + + return false; +} + +static int mtk_i3c_master_send_ccc_cmd(struct i3c_master_controller *m, + struct i3c_ccc_cmd *cmd) +{ + struct mtk_i3c_master *master = to_mtk_i3c_master(m); + struct mtk_i3c_xfer *mtk_xfer; + struct mtk_i3c_cmd *ccmd; + int ret = 0; + + mtk_xfer = mtk_i3c_master_alloc_xfer(master, 1); + if (!mtk_xfer) + return -ENOMEM; + + mtk_xfer->mode = I3C_CCC; + mtk_xfer->ignore_restart_irq = false; + mtk_xfer->auto_restart = false; + + ccmd = mtk_xfer->cmds; + ccmd->addr = cmd->dests[0].addr; + ccmd->ccc_id = cmd->id; + + if (cmd->rnw) { + ccmd->op = MASTER_RD; + ccmd->rx_len = cmd->dests[0].payload.len; + ccmd->rx_buf = cmd->dests[0].payload.data; + } else { + ccmd->op = MASTER_WR; + ccmd->tx_len = cmd->dests[0].payload.len; + ccmd->tx_buf = cmd->dests[0].payload.data; + } + + if (ccmd->ccc_id < I3C_CCC_DIRECT) { + ccmd->op = MASTER_CCC_BROADCAST; + if (ccmd->ccc_id == I3C_CCC_ENTDAA) { + ccmd->op = MASTER_DAA; + ccmd->rx_len = 8; + ccmd->tx_len = 1; + mtk_xfer->ignore_restart_irq = true; + mtk_xfer->auto_restart = true; + } + } + + mtk_i3c_master_queue_xfer(master, mtk_xfer); + if (!wait_for_completion_timeout(&mtk_xfer->complete, + msecs_to_jiffies(2000))) + mtk_i3c_master_unqueue_xfer(master, mtk_xfer); + + ret = mtk_xfer->ret; + + mtk_i3c_master_free_xfer(mtk_xfer); + if (ret == -ENXIO) + return I3C_ERROR_M2; + + return ret; +} + +static int mtk_i3c_master_priv_xfers(struct i3c_dev_desc *dev, + struct i3c_priv_xfer *xfers, + int nxfers) +{ + struct i3c_master_controller *m = i3c_dev_get_master(dev); + struct mtk_i3c_master *master = to_mtk_i3c_master(m); + struct mtk_i3c_xfer *mtk_xfer; + int i, ret = 0; + + if (!nxfers) + return 0; + + mtk_xfer = mtk_i3c_master_alloc_xfer(master, nxfers); + if (!mtk_xfer) + return -ENOMEM; + + mtk_xfer->mode = I3C_TRANSFER; + + if (nxfers == 2 && !xfers[0].rnw && xfers[1].rnw) + mtk_xfer->auto_restart = false; + else + mtk_xfer->auto_restart = true; + + if (mtk_xfer->auto_restart && nxfers >= 2) + mtk_xfer->ignore_restart_irq = true; + else + mtk_xfer->ignore_restart_irq = false; + + for (i = 0; i < nxfers; i++) { + struct mtk_i3c_cmd *ccmd = &mtk_xfer->cmds[i]; + + ccmd->addr = dev->info.dyn_addr; + ccmd->ccc_id = 0; + + if (!mtk_xfer->auto_restart) { + /* combined two messages into one transaction */ + ccmd->op = MASTER_WRRD; + ccmd->tx_len = xfers[i].len; + ccmd->rx_len = xfers[i + 1].len; + ccmd->tx_buf = xfers->data.out; + ccmd->rx_buf = xfers[i + 1].data.in; + break; + } + + if (xfers[i].rnw) { + ccmd->op = MASTER_RD; + ccmd->rx_len = xfers->len; + ccmd->rx_buf = xfers->data.in; + } else { + ccmd->op = MASTER_WR; + ccmd->tx_len = xfers->len; + ccmd->tx_buf = xfers->data.out; + } + } + + mtk_i3c_master_queue_xfer(master, mtk_xfer); + if (!wait_for_completion_timeout(&mtk_xfer->complete, + msecs_to_jiffies(2000))) + mtk_i3c_master_unqueue_xfer(master, mtk_xfer); + + ret = mtk_xfer->ret; + + mtk_i3c_master_free_xfer(mtk_xfer); + return ret; +} + +static int mtk_i3c_master_i2c_xfers(struct i2c_dev_desc *dev, + const struct i2c_msg *xfers, + int nxfers) +{ + struct i3c_master_controller *m = i2c_dev_get_master(dev); + struct mtk_i3c_master *master = to_mtk_i3c_master(m); + struct mtk_i3c_xfer *mtk_xfer; + int i, ret = 0; + + mtk_xfer = mtk_i3c_master_alloc_xfer(master, nxfers); + if (!mtk_xfer) + return -ENOMEM; + + mtk_xfer->mode = I2C_TRANSFER; + + if (!(xfers[0].flags & I2C_M_RD) && (xfers[1].flags & I2C_M_RD) && + xfers[0].addr == xfers[1].addr && nxfers == 2) + mtk_xfer->auto_restart = false; + else + mtk_xfer->auto_restart = true; + + mtk_xfer->ignore_restart_irq = false; + + for (i = 0; i < nxfers; i++) { + struct mtk_i3c_cmd *ccmd = &mtk_xfer->cmds[i]; + + ccmd->addr = xfers->addr; + ccmd->ccc_id = 0; + + if (!mtk_xfer->auto_restart) { + /* combined two messages into one transaction */ + ccmd->op = MASTER_WRRD; + ccmd->tx_len = xfers[i].len; + ccmd->rx_len = xfers[i + 1].len; + ccmd->tx_buf = xfers[i].buf; + ccmd->rx_buf = xfers[i + 1].buf; + break; + } + + if (xfers[i].flags & I2C_M_RD) { + ccmd->op = MASTER_RD; + ccmd->rx_len = xfers[i].len; + ccmd->rx_buf = xfers[i].buf; + } else { + ccmd->op = MASTER_WR; + ccmd->tx_len = xfers[i].len; + ccmd->tx_buf = xfers[i].buf; + } + } + + mtk_i3c_master_queue_xfer(master, mtk_xfer); + if (!wait_for_completion_timeout(&mtk_xfer->complete, + msecs_to_jiffies(2000))) + mtk_i3c_master_unqueue_xfer(master, mtk_xfer); + + ret = mtk_xfer->ret; + + mtk_i3c_master_free_xfer(mtk_xfer); + return ret; +} + +static u32 mtk_i3c_master_i2c_funcs(struct i3c_master_controller *m) +{ + return I2C_FUNC_SMBUS_EMUL | I2C_FUNC_I2C; +} + +static int mtk_i3c_master_do_daa(struct i3c_master_controller *m) +{ + struct mtk_i3c_master *master = to_mtk_i3c_master(m); + int ret = 0, slot; + u8 last_addr = 0; + + for (slot = 0; slot < MAX_I3C_DEVS; slot++) { + if (master->daa_anchor.daa_addr[slot].used) + continue; + + ret = i3c_master_get_free_addr(m, last_addr + 1); + if (ret < 0) + return -ENOSPC; + + last_addr = ret; + master->daa_anchor.daa_addr[slot].addr = last_addr; + } + + ret = i3c_master_entdaa_locked(&master->mas_ctrler); + if (ret && ret != I3C_ERROR_M2) + return ret; + + /* + * Clear all retaining registers filled during DAA. We already + * have the addressed assigned to them in the addrs array. + */ + for (slot = 1; slot < MAX_I3C_DEVS; slot++) { + if (master->daa_anchor.daa_addr[slot].used) { + last_addr = master->daa_anchor.daa_addr[slot].addr; + i3c_master_add_i3c_dev_locked(m, last_addr); + } + } + + /* + * Clear slots that ended up not being used. Can be caused by I3C + * device creation failure or when the I3C device was already known + * by the system but with a different address (in this case the device + * already has a slot and does not need a new one). + */ + i3c_master_defslvs_locked(&master->mas_ctrler); + + /* Unmask Hot-Join and Mastership request interrupts. */ + i3c_master_enec_locked(m, I3C_BROADCAST_ADDR, I3C_CCC_EVENT_HJ | + I3C_CCC_EVENT_MR); + + return 0; +} + +static int mtk_i3c_master_bus_init(struct i3c_master_controller *m) +{ + struct mtk_i3c_master *master = to_mtk_i3c_master(m); + struct i3c_device_info info = { }; + unsigned long source_clk_rate; + int ret; + + source_clk_rate = clk_get_rate(master->clk_main) / INTER_CLK_DIV; + + ret = mtk_i3c_set_speed(master, source_clk_rate); + if (ret) { + dev_err(master->dev, "Failed to set the bus speed.\n"); + return -EINVAL; + } + + /* Get an address for the master. */ + ret = i3c_master_get_free_addr(m, 0); + if (ret < 0) + return ret; + + master->daa_anchor.daa_addr[0].addr = (u8)ret; + master->daa_anchor.daa_addr[0].used = true; + master->daa_anchor.idx = 0; + + /* + * In I3C protocol host controller is also with device role, + * so the driver should provide dcr, bcr, and pid info + * of host controller itself + */ + memset(&info, 0, sizeof(info)); + info.dyn_addr = ret; + info.dcr = MTK_I3C_DCR; + info.bcr = MTK_I3C_BCR; + info.pid = MTK_I3C_PID; + + if (info.bcr & I3C_BCR_HDR_CAP) + info.hdr_cap = I3C_CCC_HDR_MODE(I3C_HDR_DDR); + + return i3c_master_set_info(&master->mas_ctrler, &info); +} + +static const struct i3c_master_controller_ops mtk_i3c_master_ops = { + .bus_init = mtk_i3c_master_bus_init, + .do_daa = mtk_i3c_master_do_daa, + .supports_ccc_cmd = mtk_i3c_master_supports_ccc_cmd, + .send_ccc_cmd = mtk_i3c_master_send_ccc_cmd, + .priv_xfers = mtk_i3c_master_priv_xfers, + .i2c_xfers = mtk_i3c_master_i2c_xfers, + .i2c_funcs = mtk_i3c_master_i2c_funcs, +}; + +static int mtk_i3c_master_probe(struct platform_device *pdev) +{ + struct device *dev = &pdev->dev; + struct mtk_i3c_master *master; + struct resource *res; + int ret, irqnr; + + master = devm_kzalloc(dev, sizeof(*master), GFP_KERNEL); + if (!master) + return -ENOMEM; + + res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "main"); + master->regs = devm_ioremap_resource(dev, res); + if (IS_ERR(master->regs)) + return PTR_ERR(master->regs); + + res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "dma"); + master->dma_regs = devm_ioremap_resource(dev, res); + if (IS_ERR(master->dma_regs)) + return PTR_ERR(master->dma_regs); + + irqnr = platform_get_irq(pdev, 0); + if (irqnr < 0) + return irqnr; + + ret = devm_request_irq(dev, irqnr, mtk_i3c_master_irq, + IRQF_TRIGGER_NONE, DRV_NAME, master); + if (ret < 0) { + dev_err(dev, "Request I3C IRQ %d fail\n", irqnr); + return ret; + } + + spin_lock_init(&master->xferqueue.lock); + INIT_LIST_HEAD(&master->xferqueue.list); + + if (dma_set_mask(dev, DMA_BIT_MASK(33))) { + dev_err(dev, "dma_set_mask return error.\n"); + return -EINVAL; + } + + master->clk_main = devm_clk_get(dev, "main"); + if (IS_ERR(master->clk_main)) { + dev_err(dev, "cannot get main clock\n"); + return PTR_ERR(master->clk_main); + } + master->clk_dma = devm_clk_get(dev, "dma"); + if (IS_ERR(master->clk_dma)) { + dev_err(dev, "cannot get dma clock\n"); + return PTR_ERR(master->clk_dma); + } + + master->clk_arb = devm_clk_get_optional(dev, "arb"); + if (IS_ERR(master->clk_arb)) + return PTR_ERR(master->clk_arb); + + ret = mtk_i3c_master_clock_enable(master); + if (ret) { + dev_err(dev, "clock enable failed!\n"); + return ret; + } + + master->dev = dev; + platform_set_drvdata(pdev, master); + + ret = i3c_master_register(&master->mas_ctrler, dev, + &mtk_i3c_master_ops, false); + if (ret) { + dev_err(dev, "Failed to add i3c bus to i3c core\n"); + mtk_i3c_master_clock_disable(master); + return ret; + } + + return 0; +} + +static int mtk_i3c_master_remove(struct platform_device *pdev) +{ + struct mtk_i3c_master *master = platform_get_drvdata(pdev); + + i3c_master_unregister(&master->mas_ctrler); + mtk_i3c_master_clock_disable(master); + + return 0; +} + +static const struct of_device_id mtk_i3c_master_of_ids[] = { + { .compatible = "mediatek,i3c-master" }, + { /* sentinel */ }, +}; +MODULE_DEVICE_TABLE(of, mtk_i3c_master_of_ids); + +static struct platform_driver mtk_i3c_master_driver = { + .probe = mtk_i3c_master_probe, + .remove = mtk_i3c_master_remove, + .driver = { + .name = DRV_NAME, + .of_match_table = mtk_i3c_master_of_ids, + }, +}; +module_platform_driver(mtk_i3c_master_driver); + +MODULE_AUTHOR("Qii Wang <qii.wang@xxxxxxxxxxxx>"); +MODULE_DESCRIPTION("MediaTek I3C master driver"); +MODULE_LICENSE("GPL v2"); +MODULE_ALIAS("platform:mtk-i3c-master"); -- 1.7.9.5