The sh_sci driver supports dma with pio mode as fallback. When the DMA Rx time out happens, it switches to pio mode and the timer function has the below dma callbacks rx_timer_fn() of the sh-sci.c: dmaengine_pause(); /* [1] */ ... dmaengine_tx_status(); /* [2] */ ... dmaengine_terminate_all(); /* [3] */ Update [1] to re-enable the interrupt by clearing the DMARS. RZ/G2L SoC use the same signal for both interrupt and DMA transfer requests. The signal works as a DMA transfer request signal by setting DMARS, and subsequent interrupt requests to the interrupt controller are masked. Update [2] to calculate residue, so that sh_sci driver can work on leftover data from DMA operation during pio mode. Update [3] to invalidate hw descriptors for reuse. Based on similar work done for rcar_dmac for supporting scif dma. Signed-off-by: Biju Das <biju.das.jz@xxxxxxxxxxxxxx> --- drivers/dma/sh/rz-dmac.c | 193 ++++++++++++++++++++++++++++++++++++++- 1 file changed, 192 insertions(+), 1 deletion(-) diff --git a/drivers/dma/sh/rz-dmac.c b/drivers/dma/sh/rz-dmac.c index 65a27c5a7bce..3ef4dda51c8e 100644 --- a/drivers/dma/sh/rz-dmac.c +++ b/drivers/dma/sh/rz-dmac.c @@ -109,10 +109,12 @@ struct rz_dmac { * Registers */ +#define CRTB 0x0020 #define CHSTAT 0x0024 #define CHCTRL 0x0028 #define CHCFG 0x002c #define NXLA 0x0038 +#define CRLA 0x003c #define DCTRL 0x0000 @@ -533,11 +535,15 @@ rz_dmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl, static int rz_dmac_terminate_all(struct dma_chan *chan) { struct rz_dmac_chan *channel = to_rz_dmac_chan(chan); + struct rz_lmdesc *lmdesc = channel->lmdesc.base; unsigned long flags; LIST_HEAD(head); rz_dmac_disable_hw(channel); spin_lock_irqsave(&channel->vc.lock, flags); + for (; lmdesc < channel->lmdesc.base + DMAC_NR_LMDESC; lmdesc++) + lmdesc->header = 0; + list_splice_tail_init(&channel->ld_active, &channel->ld_free); list_splice_tail_init(&channel->ld_queue, &channel->ld_free); vchan_get_all_descriptors(&channel->vc, &head); @@ -647,6 +653,190 @@ static void rz_dmac_device_synchronize(struct dma_chan *chan) rz_dmac_set_dmars_register(dmac, channel->index, 0); } +static struct rz_lmdesc * +rz_dmac_get_next_lmdesc(struct rz_lmdesc *base, struct rz_lmdesc *lmdesc) +{ + struct rz_lmdesc *next = lmdesc++; + + if (next >= base + DMAC_NR_LMDESC) + next = base; + + return next; +} + +static unsigned int +rz_dmac_calculate_residue_bytes_in_vd(struct rz_dmac_chan *channel) +{ + struct rz_lmdesc *lmdesc = channel->lmdesc.head; + struct dma_chan *chan = &channel->vc.chan; + struct rz_dmac *dmac = to_rz_dmac(chan->device); + unsigned int residue = 0, i = 0; + unsigned int crla; + + crla = rz_dmac_ch_readl(channel, CRLA, 1); + while (!(lmdesc->nxla == crla)) { + lmdesc = rz_dmac_get_next_lmdesc(channel->lmdesc.base, lmdesc); + if (i++ > DMAC_NR_LMDESC) + return 0; + } + + /* Get current processing lmdesc in hardware */ + lmdesc = rz_dmac_get_next_lmdesc(channel->lmdesc.base, lmdesc); + /* Calculate residue from next lmdesc to end of virtual desc*/ + while (lmdesc->chcfg & CHCFG_DEM) { + lmdesc = rz_dmac_get_next_lmdesc(channel->lmdesc.base, lmdesc); + residue += lmdesc->tb; + } + + dev_dbg(dmac->dev, "%s: Getting residue is %d\n", __func__, residue); + + return residue; +} + +static unsigned int rz_dmac_calculate_total_bytes_in_vd(struct rz_dmac_desc *desc) +{ + unsigned int i, size = 0; + struct scatterlist *sg; + + for_each_sg(desc->sg, sg, desc->sgcount, i) + size += sg_dma_len(sg); + + return size; +} + +static unsigned int rz_dmac_chan_get_residue(struct rz_dmac_chan *channel, + dma_cookie_t cookie) +{ + struct rz_dmac_desc *current_desc, *desc; + enum dma_status status; + unsigned int residue; + unsigned int crla; + unsigned int crtb; + unsigned int i; + + /* Get current processing virtual descriptor */ + current_desc = list_first_entry(&channel->ld_active, + struct rz_dmac_desc, node); + if (!current_desc) + return 0; + + /* + * If the cookie corresponds to a descriptor that has been completed + * there is no residue. The same check has already been performed by the + * caller but without holding the channel lock, so the descriptor could + * now be complete. + */ + status = dma_cookie_status(&channel->vc.chan, cookie, NULL); + if (status == DMA_COMPLETE) + return 0; + + /* + * If the cookie doesn't correspond to the currently processing virtual + * descriptor then the descriptor hasn't been processed yet, and the + * residue is equal to the full descriptor size. + * Also, a client driver is possible to call this function before + * rz_dmac_irq_handler_thread() runs. In this case, the running + * descriptor will be the next descriptor, and the done list will + * appear. So, if the argument cookie matches the done list's cookie, + * we can assume the residue is zero. + */ + if (cookie != current_desc->vd.tx.cookie) { + list_for_each_entry(desc, &channel->ld_free, node) { + if (cookie == desc->vd.tx.cookie) + return 0; + } + + list_for_each_entry(desc, &channel->ld_queue, node) { + if (cookie == desc->vd.tx.cookie) + return rz_dmac_calculate_total_bytes_in_vd(desc); + } + + list_for_each_entry(desc, &channel->ld_active, node) { + if (cookie == desc->vd.tx.cookie) + return rz_dmac_calculate_total_bytes_in_vd(desc); + } + + /* + * No descriptor found for the cookie, there's thus no residue. + * This shouldn't happen if the calling driver passes a correct + * cookie value. + */ + WARN_ONCE(1, "No descriptor for cookie!"); + return 0; + } + + /* + * We need to read two registers. + * Make sure the hardware does not move to next lmdesc while reading + * the current lmdesc. + * Trying it 3 times should be enough: Initial read, retry, retry + * for the paranoid. + */ + for (i = 0; i < 3; i++) { + crla = rz_dmac_ch_readl(channel, CRLA, 1); + crtb = rz_dmac_ch_readl(channel, CRTB, 1); + /* Still the same? */ + if (crla == rz_dmac_ch_readl(channel, CRLA, 1)) + break; + } + + WARN_ONCE(i >= 3, "residue might be not continuous!"); + + /* + * Calculate number of byte transferred in processing virtual descriptor + * One virtual descriptor can have many lmdesc + */ + residue = crtb; + residue += rz_dmac_calculate_residue_bytes_in_vd(channel); + + return residue; +} + +static enum dma_status rz_dmac_tx_status(struct dma_chan *chan, + dma_cookie_t cookie, + struct dma_tx_state *txstate) +{ + struct rz_dmac_chan *channel = to_rz_dmac_chan(chan); + enum dma_status status; + unsigned int residue; + unsigned long flags; + + status = dma_cookie_status(chan, cookie, txstate); + if (status == DMA_COMPLETE || !txstate) + return status; + + spin_lock_irqsave(&channel->vc.lock, flags); + residue = rz_dmac_chan_get_residue(channel, cookie); + spin_unlock_irqrestore(&channel->vc.lock, flags); + + /* if there's no residue, the cookie is complete */ + if (!residue) + return DMA_COMPLETE; + + dma_set_residue(txstate, residue); + + return status; +} + +static int rz_dmac_device_pause(struct dma_chan *chan) +{ + struct rz_dmac_chan *channel = to_rz_dmac_chan(chan); + struct rz_dmac *dmac = to_rz_dmac(chan->device); + unsigned int i; + u32 chstat; + + for (i = 0; i < 1024; i++) { + chstat = rz_dmac_ch_readl(channel, CHSTAT, 1); + if (!(chstat & CHSTAT_EN)) + break; + udelay(1); + } + + rz_dmac_set_dmars_register(dmac, channel->index, 0); + + return 0; +} + /* * ----------------------------------------------------------------------------- * IRQ handling @@ -929,13 +1119,14 @@ static int rz_dmac_probe(struct platform_device *pdev) engine->device_alloc_chan_resources = rz_dmac_alloc_chan_resources; engine->device_free_chan_resources = rz_dmac_free_chan_resources; - engine->device_tx_status = dma_cookie_status; + engine->device_tx_status = rz_dmac_tx_status; engine->device_prep_slave_sg = rz_dmac_prep_slave_sg; engine->device_prep_dma_memcpy = rz_dmac_prep_dma_memcpy; engine->device_config = rz_dmac_config; engine->device_terminate_all = rz_dmac_terminate_all; engine->device_issue_pending = rz_dmac_issue_pending; engine->device_synchronize = rz_dmac_device_synchronize; + engine->device_pause = rz_dmac_device_pause; engine->copy_align = DMAENGINE_ALIGN_1_BYTE; dma_set_max_seg_size(engine->dev, U32_MAX); -- 2.43.0