Re: [PATCH v4 13/15] iommu/dma: Force bouncing if the size is not cacheline-aligned

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 19/05/2023 3:02 pm, Catalin Marinas wrote:
On Fri, May 19, 2023 at 01:29:38PM +0100, Robin Murphy wrote:
On 2023-05-18 18:34, Catalin Marinas wrote:
diff --git a/drivers/iommu/dma-iommu.c b/drivers/iommu/dma-iommu.c
index 7a9f0b0bddbd..ab1c1681c06e 100644
--- a/drivers/iommu/dma-iommu.c
+++ b/drivers/iommu/dma-iommu.c
@@ -956,7 +956,7 @@ static void iommu_dma_sync_sg_for_cpu(struct device *dev,
   	struct scatterlist *sg;
   	int i;
-	if (dev_use_swiotlb(dev))
+	if (dev_use_swiotlb(dev) || sg_is_dma_bounced(sgl))
   		for_each_sg(sgl, sg, nelems, i)
   			iommu_dma_sync_single_for_cpu(dev, sg_dma_address(sg),
   						      sg->length, dir);
@@ -972,7 +972,7 @@ static void iommu_dma_sync_sg_for_device(struct device *dev,
   	struct scatterlist *sg;
   	int i;
-	if (dev_use_swiotlb(dev))
+	if (dev_use_swiotlb(dev) || sg_is_dma_bounced(sgl))
   		for_each_sg(sgl, sg, nelems, i)
   			iommu_dma_sync_single_for_device(dev,
   							 sg_dma_address(sg),
@@ -998,7 +998,8 @@ static dma_addr_t iommu_dma_map_page(struct device *dev, struct page *page,
   	 * If both the physical buffer start address and size are
   	 * page aligned, we don't need to use a bounce page.
   	 */
-	if (dev_use_swiotlb(dev) && iova_offset(iovad, phys | size)) {
+	if ((dev_use_swiotlb(dev) && iova_offset(iovad, phys | size)) ||
+	    dma_kmalloc_needs_bounce(dev, size, dir)) {
   		void *padding_start;
   		size_t padding_size, aligned_size;
@@ -1210,7 +1211,21 @@ static int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg,
   			goto out;
   	}
-	if (dev_use_swiotlb(dev))
+	/*
+	 * If kmalloc() buffers are not DMA-safe for this device and
+	 * direction, check the individual lengths in the sg list. If one of
+	 * the buffers is deemed unsafe, follow the iommu_dma_map_sg_swiotlb()
+	 * path for potential bouncing.
+	 */
+	if (!dma_kmalloc_safe(dev, dir)) {
+		for_each_sg(sg, s, nents, i)
+			if (!dma_kmalloc_size_aligned(s->length)) {

Just to remind myself, we're not checking s->offset on the grounds that if
anyone wants to DMA into an unaligned part of a larger allocation that
remains at their own risk, is that right?

Right. That's the case currently as well and those users that were
relying on ARCH_KMALLOC_MINALIGN for this have either been migrated to
ARCH_DMA_MINALIGN in this series or the logic rewritten (as in the
crypto code).

OK, I did manage to summon a vague memory of this being discussed before, which at least stopped me asking "Should we be checking..." - perhaps a comment on dma_kmalloc_safe() to help remember that reasoning might not go amiss?

Do we care about the (probably theoretical) case where someone might build a
scatterlist for multiple small allocations such that ones which happen to be
adjacent might get combined into a single segment of apparently "safe"
length but still at "unsafe" alignment?

I'd say that's theoretical only. One could write such code but normally
you'd go for an array rather than relying on the randomness of the
kmalloc pointers to figure out adjacent objects. It also only works if
the individual struct size is exactly one of the kmalloc cache sizes, so
not generic enough.

FWIW I was imagining something like sg_alloc_table_from_pages() but at a smaller scale, queueing up some list/array of, say, 32-byte buffers into a scatterlist to submit as a single DMA job. I'm not aware that such a thing exists though, and I'm inclined to agree that it probably is sufficiently unrealistic to be concerned about. As usual I just want to feel comfortable that we've explored all the possibilities :)

+				sg_dma_mark_bounced(sg);

I'd prefer to have iommu_dma_map_sg_swiotlb() mark the segments, since
that's in charge of the actual bouncing. Then we can fold the alignment
check into dev_use_swiotlb() (with the dev_is_untrusted() condition taking
priority), and sync/unmap can simply rely on sg_is_dma_bounced() alone.

With this patch we only set the SG_DMA_BOUNCED on the first element of
the sglist. Do you want to set this flag only on individual elements
being bounced? It makes some sense in principle but the
iommu_dma_unmap_sg() path would need to scan the list again to decide
whether to go the swiotlb path.

If we keep the SG_DMA_BOUNCED flag only on the first element, I can
change it to your suggestion, assuming I understood it.

Indeed that should be fine - sync_sg/unmap_sg always have to be given the same arguments which were passed to map_sg (and note that in the normal case, the DMA address/length will often end up concatenated entirely into the first element), so while we still have the two distinct flows internally, I don't think there's any issue with only tagging the head of the list to steer between them. Of course if it then works out to be trivial enough to tag *all* the segments for good measure, there should be no harm in that either - at the moment the flag is destined to have more of a "this might be bounced, so needs checking" meaning than "this definitely is bounced" either way.

Cheers,
Robin.




[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux