On 17:05 Tue 12 Mar , Dave Stevenson wrote: > Hi Andrea > > On Tue, 12 Mar 2024 at 09:13, Andrea della Porta <andrea.porta@xxxxxxxx> wrote: > > > > This finally moves all platform specific stuff into a separate structure, > > which is initialized on the OF compatible during probing. Since the DMA > > control block is different on the BCM2711 platform, we introduce a common > > control block to reserve the necessary space and adequate methods for > > access. > > > > Signed-off-by: Stefan Wahren <stefan.wahren@xxxxxxxx> > > Signed-off-by: Andrea della Porta <andrea.porta@xxxxxxxx> > > --- > > drivers/dma/bcm2835-dma.c | 336 +++++++++++++++++++++++++++++--------- > > 1 file changed, 260 insertions(+), 76 deletions(-) > > > > diff --git a/drivers/dma/bcm2835-dma.c b/drivers/dma/bcm2835-dma.c > > index 88ae5d05402e..b015eae29b08 100644 > > --- a/drivers/dma/bcm2835-dma.c > > +++ b/drivers/dma/bcm2835-dma.c > > @@ -48,6 +48,11 @@ struct bcm2835_dmadev { > > struct dma_device ddev; > > void __iomem *base; > > dma_addr_t zero_page; > > + const struct bcm2835_dma_cfg *cfg; > > +}; > > + > > +struct bcm_dma_cb { > > + uint32_t rsvd[8]; > > }; > > > > struct bcm2835_dma_cb { > > @@ -61,7 +66,7 @@ struct bcm2835_dma_cb { > > }; > > > > struct bcm2835_cb_entry { > > - struct bcm2835_dma_cb *cb; > > + struct bcm_dma_cb *cb; > > dma_addr_t paddr; > > }; > > > > @@ -82,6 +87,44 @@ struct bcm2835_chan { > > bool is_lite_channel; > > }; > > > > +struct bcm2835_dma_cfg { > > + dma_addr_t addr_offset; > > + u32 cs_reg; > > + u32 cb_reg; > > + u32 next_reg; > > + u32 ti_reg; > > + > > + u32 wait_mask; > > + u32 reset_mask; > > + u32 int_mask; > > + u32 active_mask; > > + u32 abort_mask; > > + u32 s_dreq_mask; > > + u32 d_dreq_mask; > > + > > + u32 (*cb_get_length)(void *data); > > + dma_addr_t (*cb_get_addr)(void *data, enum dma_transfer_direction); > > + > > + void (*cb_init)(void *data, struct bcm2835_chan *c, > > + enum dma_transfer_direction, u32 src, u32 dst, > > + bool zero_page); > > + void (*cb_set_src)(void *data, enum dma_transfer_direction, u32 src); > > + void (*cb_set_dst)(void *data, enum dma_transfer_direction, u32 dst); > > + void (*cb_set_next)(void *data, u32 next); > > + void (*cb_set_length)(void *data, u32 length); > > + void (*cb_append_extra)(void *data, > > + struct bcm2835_chan *c, > > + enum dma_transfer_direction direction, > > + bool cyclic, bool final, unsigned long flags); > > + > > + dma_addr_t (*to_cb_addr)(dma_addr_t addr); > > + > > + void (*chan_plat_init)(struct bcm2835_chan *c); > > + dma_addr_t (*read_addr)(struct bcm2835_chan *c, > > + enum dma_transfer_direction); > > + u32 (*cs_flags)(struct bcm2835_chan *c); > > +}; > > + > > struct bcm2835_desc { > > struct bcm2835_chan *c; > > struct virt_dma_desc vd; > > @@ -215,6 +258,13 @@ static inline struct bcm2835_dmadev *to_bcm2835_dma_dev(struct dma_device *d) > > return container_of(d, struct bcm2835_dmadev, ddev); > > } > > > > +static inline const struct bcm2835_dma_cfg *to_bcm2835_cfg(struct dma_device *d) > > +{ > > + struct bcm2835_dmadev *od = container_of(d, struct bcm2835_dmadev, ddev); > > + > > + return od->cfg; > > +} > > + > > static inline struct bcm2835_chan *to_bcm2835_dma_chan(struct dma_chan *c) > > { > > return container_of(c, struct bcm2835_chan, vc.chan); > > @@ -292,6 +342,109 @@ static inline bool need_dst_incr(enum dma_transfer_direction direction) > > return false; > > } > > > > +static inline u32 bcm2835_dma_cb_get_length(void *data) > > +{ > > + struct bcm2835_dma_cb *cb = data; > > + > > + return cb->length; > > +} > > + > > +static inline dma_addr_t > > +bcm2835_dma_cb_get_addr(void *data, enum dma_transfer_direction direction) > > +{ > > + struct bcm2835_dma_cb *cb = data; > > + > > + if (direction == DMA_DEV_TO_MEM) > > + return cb->dst; > > + > > + return cb->src; > > +} > > + > > +static inline void > > +bcm2835_dma_cb_init(void *data, struct bcm2835_chan *c, > > + enum dma_transfer_direction direction, u32 src, u32 dst, > > + bool zero_page) > > +{ > > + struct bcm2835_dma_cb *cb = data; > > + > > + cb->info = bcm2835_dma_prepare_cb_info(c, direction, zero_page); > > + cb->src = src; > > + cb->dst = dst; > > + cb->stride = 0; > > + cb->next = 0; > > +} > > + > > +static inline void > > +bcm2835_dma_cb_set_src(void *data, enum dma_transfer_direction direction, > > + u32 src) > > +{ > > + struct bcm2835_dma_cb *cb = data; > > + > > + cb->src = src; > > +} > > + > > +static inline void > > +bcm2835_dma_cb_set_dst(void *data, enum dma_transfer_direction direction, > > + u32 dst) > > +{ > > + struct bcm2835_dma_cb *cb = data; > > + > > + cb->dst = dst; > > +} > > + > > +static inline void bcm2835_dma_cb_set_next(void *data, u32 next) > > +{ > > + struct bcm2835_dma_cb *cb = data; > > + > > + cb->next = next; > > +} > > + > > +static inline void bcm2835_dma_cb_set_length(void *data, u32 length) > > +{ > > + struct bcm2835_dma_cb *cb = data; > > + > > + cb->length = length; > > +} > > + > > +static inline void > > +bcm2835_dma_cb_append_extra(void *data, struct bcm2835_chan *c, > > + enum dma_transfer_direction direction, > > + bool cyclic, bool final, unsigned long flags) > > +{ > > + struct bcm2835_dma_cb *cb = data; > > + > > + cb->info |= bcm2835_dma_prepare_cb_extra(c, direction, cyclic, final, > > + flags); > > +} > > + > > +static inline dma_addr_t bcm2835_dma_to_cb_addr(dma_addr_t addr) > > +{ > > + return addr; > > +} > > + > > +static void bcm2835_dma_chan_plat_init(struct bcm2835_chan *c) > > +{ > > + /* check in DEBUG register if this is a LITE channel */ > > + if (readl(c->chan_base + BCM2835_DMA_DEBUG) & BCM2835_DMA_DEBUG_LITE) > > + c->is_lite_channel = true; > > +} > > + > > +static dma_addr_t bcm2835_dma_read_addr(struct bcm2835_chan *c, > > + enum dma_transfer_direction direction) > > +{ > > + if (direction == DMA_MEM_TO_DEV) > > + return readl(c->chan_base + BCM2835_DMA_SOURCE_AD); > > + else if (direction == DMA_DEV_TO_MEM) > > + return readl(c->chan_base + BCM2835_DMA_DEST_AD); > > + > > + return 0; > > +} > > + > > +static u32 bcm2835_dma_cs_flags(struct bcm2835_chan *c) > > +{ > > + return BCM2835_DMA_CS_FLAGS(c->dreq); > > +} > > + > > static void bcm2835_dma_free_cb_chain(struct bcm2835_desc *desc) > > { > > size_t i; > > @@ -309,16 +462,19 @@ static void bcm2835_dma_desc_free(struct virt_dma_desc *vd) > > } > > > > static bool bcm2835_dma_create_cb_set_length(struct dma_chan *chan, > > - struct bcm2835_dma_cb *control_block, > > + void *data, > > size_t len, > > size_t period_len, > > size_t *total_len) > > { > > + const struct bcm2835_dma_cfg *cfg = to_bcm2835_cfg(chan->device); > > struct bcm2835_chan *c = to_bcm2835_dma_chan(chan); > > size_t max_len = bcm2835_dma_max_frame_length(c); > > > > /* set the length taking lite-channel limitations into account */ > > - control_block->length = min_t(u32, len, max_len); > > + u32 length = min_t(u32, len, max_len); > > + > > + cfg->cb_set_length(data, length); > > > > /* finished if we have no period_length */ > > if (!period_len) > > @@ -333,14 +489,14 @@ static bool bcm2835_dma_create_cb_set_length(struct dma_chan *chan, > > */ > > > > /* have we filled in period_length yet? */ > > - if (*total_len + control_block->length < period_len) { > > + if (*total_len + length < period_len) { > > /* update number of bytes in this period so far */ > > - *total_len += control_block->length; > > + *total_len += length; > > return false; > > } > > > > /* calculate the length that remains to reach period_length */ > > - control_block->length = period_len - *total_len; > > + cfg->cb_set_length(data, period_len - *total_len); > > > > /* reset total_length for next period */ > > *total_len = 0; > > @@ -388,15 +544,14 @@ static struct bcm2835_desc *bcm2835_dma_create_cb_chain( > > size_t buf_len, size_t period_len, > > gfp_t gfp, unsigned long flags) > > { > > + const struct bcm2835_dma_cfg *cfg = to_bcm2835_cfg(chan->device); > > struct bcm2835_dmadev *od = to_bcm2835_dma_dev(chan->device); > > struct bcm2835_chan *c = to_bcm2835_dma_chan(chan); > > size_t len = buf_len, total_len; > > size_t frame; > > struct bcm2835_desc *d; > > struct bcm2835_cb_entry *cb_entry; > > - struct bcm2835_dma_cb *control_block; > > - u32 extrainfo = bcm2835_dma_prepare_cb_extra(c, direction, cyclic, > > - false, flags); > > + struct bcm_dma_cb *control_block; > > bool zero_page = false; > > > > if (!frames) > > @@ -432,12 +587,7 @@ static struct bcm2835_desc *bcm2835_dma_create_cb_chain( > > > > /* fill in the control block */ > > control_block = cb_entry->cb; > > - control_block->info = bcm2835_dma_prepare_cb_info(c, direction, > > - zero_page); > > - control_block->src = src; > > - control_block->dst = dst; > > - control_block->stride = 0; > > - control_block->next = 0; > > + cfg->cb_init(control_block, c, src, dst, direction, zero_page); > > Can I ask how you've been testing these patches? Sure, most of these issues were easily spotted by mem to mem transactions, using dmatest. > > This line was one of the bugs that I found during my work. The > prototype for cb_init is > + void (*cb_init)(void *data, struct bcm2835_chan *c, > + enum dma_transfer_direction, u32 src, u32 dst, > + bool zero_page); > So this call has direction in the wrong place, which leads to quite > comical failures. Exactly, that was one of the funniest. This is indeed resolved in patch number 14. The others being related to address shifting in the 40 bit case, basically I've just replaced << 8 with << 32 on srci and dsti control block fields. Many thanks, Andrea > > Thanks > Dave >