Make all cache invalidation conditional on sg_has_page(). Signed-off-by: Christoph Hellwig <hch@xxxxxx> --- arch/xtensa/include/asm/dma-mapping.h | 17 ++++++++++------- 1 file changed, 10 insertions(+), 7 deletions(-) diff --git a/arch/xtensa/include/asm/dma-mapping.h b/arch/xtensa/include/asm/dma-mapping.h index 1f5f6dc..262a1d1 100644 --- a/arch/xtensa/include/asm/dma-mapping.h +++ b/arch/xtensa/include/asm/dma-mapping.h @@ -61,10 +61,9 @@ dma_map_sg(struct device *dev, struct scatterlist *sglist, int nents, BUG_ON(direction == DMA_NONE); for_each_sg(sglist, sg, nents, i) { - BUG_ON(!sg_page(sg)); - sg->dma_address = sg_phys(sg); - consistent_sync(sg_virt(sg), sg->length, direction); + if (sg_has_page(sg)) + consistent_sync(sg_virt(sg), sg->length, direction); } return nents; @@ -131,8 +130,10 @@ dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sglist, int nelems, int i; struct scatterlist *sg; - for_each_sg(sglist, sg, nelems, i) - consistent_sync(sg_virt(sg), sg->length, dir); + for_each_sg(sglist, sg, nelems, i) { + if (sg_has_page(sg)) + consistent_sync(sg_virt(sg), sg->length, dir); + } } static inline void @@ -142,8 +143,10 @@ dma_sync_sg_for_device(struct device *dev, struct scatterlist *sglist, int i; struct scatterlist *sg; - for_each_sg(sglist, sg, nelems, i) - consistent_sync(sg_virt(sg), sg->length, dir); + for_each_sg(sglist, sg, nelems, i) { + if (sg_has_page(sg)) + consistent_sync(sg_virt(sg), sg->length, dir); + } } static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr) -- 1.9.1