On Fri, 26 May 2023 17:07:40 +0100 Jonathan Cameron <Jonathan.Cameron@xxxxxxxxxx> wrote: > On Thu, 25 May 2023 15:31:34 +0100 > Catalin Marinas <catalin.marinas@xxxxxxx> wrote: > > > On Thu, May 25, 2023 at 01:31:38PM +0100, Jonathan Cameron wrote: > > > On Wed, 24 May 2023 18:18:49 +0100 > > > Catalin Marinas <catalin.marinas@xxxxxxx> wrote: > > > > Another version of the series reducing the kmalloc() minimum alignment > > > > on arm64 to 8 (from 128). Other architectures can easily opt in by > > > > defining ARCH_KMALLOC_MINALIGN as 8 and selecting > > > > DMA_BOUNCE_UNALIGNED_KMALLOC. > > > > > > > > The first 10 patches decouple ARCH_KMALLOC_MINALIGN from > > > > ARCH_DMA_MINALIGN and, for arm64, limit the kmalloc() caches to those > > > > aligned to the run-time probed cache_line_size(). On arm64 we gain the > > > > kmalloc-{64,192} caches. > > > > > > > > The subsequent patches (11 to 15) further reduce the kmalloc() caches to > > > > kmalloc-{8,16,32,96} if the default swiotlb is present by bouncing small > > > > buffers in the DMA API. > > > > > > I think IIO_DMA_MINALIGN needs to switch to ARCH_DMA_MINALIGN as well. > > > > > > It's used to force static alignement of buffers with larger structures, > > > to make them suitable for non coherent DMA, similar to your other cases. > > > > Ah, I forgot that you introduced that macro. However, at a quick grep, I > > don't think this forced alignment always works as intended (irrespective > > of this series). Let's take an example: > > > > struct ltc2496_driverdata { > > /* this must be the first member */ > > struct ltc2497core_driverdata common_ddata; > > struct spi_device *spi; > > > > /* > > * DMA (thus cache coherency maintenance) may require the > > * transfer buffers to live in their own cache lines. > > */ > > unsigned char rxbuf[3] __aligned(IIO_DMA_MINALIGN); > > unsigned char txbuf[3]; > > }; > > > > The rxbuf is aligned to IIO_DMA_MINALIGN, the structure and its size as > > well but txbuf is at an offset of 3 bytes from the aligned > > IIO_DMA_MINALIGN. So basically any cache maintenance on rxbuf would > > corrupt txbuf. > > That was intentional (though possibly wrong if I've misunderstood > the underlying issue). > > For SPI controllers at least my understanding was that it is safe to > assume that they won't trample on themselves. The driver doesn't > touch the buffers when DMA is in flight - to do so would indeed result > in corruption. > > So whilst we could end up with the SPI master writing stale data back > to txbuf after the transfer it will never matter (as the value is unchanged). > Any flushes in the other direction will end up flushing both rxbuf and > txbuf anyway which is also harmless. Adding missing detail. As the driver never writes txbuf whilst any DMA is going on, the second cache evict (to flush out any lines that have crept back into cache after the flush - and write backs - pre DMA) will find a clean line and will drop it without writing back - thus no corruption. > > > You need rxbuf to be the only resident of a > > cache line, therefore the next member needs such alignment as well. > > > > With this series and SWIOTLB enabled, however, if you try to transfer 3 > > bytes, they will be bounced, so the missing alignment won't matter much. > > > > Only on arm64? If the above is wrong, it might cause trouble on > some other architectures. > > As a side note spi_write_then_read() goes through a bounce buffer dance > to avoid using dma unsafe buffers. Superficially that looks to me like it might > now end up with an undersized buffer and hence up bouncing which rather > defeats the point of it. It uses SMP_CACHE_BYTES for the size. > > Jonathan > >