On Mon, May 30, 2016 at 01:54:06PM +0200, Krzysztof Kozlowski wrote: > The dma-mapping core and the implementations do not change the > DMA attributes passed by pointer. Thus the pointer can point to const > data. However the attributes do not have to be a bitfield. Instead > unsigned long will do fine: > > 1. This is just simpler. Both in terms of reading the code and setting > attributes. Instead of initializing local attributes on the stack and > passing pointer to it to dma_set_attr(), just set the bits. > > 2. It brings safeness and checking for const correctness because the > attributes are passed by value. .. why not go the next step a do an enum? Perhaps that should be mentioned as part of the description? Thanks. > > Please have in mind that this is RFC, not finished yet. Only ARM and > ARM64 are fixed (and not everywhere). > However other API users also have to be converted which is quite > intrusive. I would rather avoid it until the overall approach is > accepted. > > Signed-off-by: Krzysztof Kozlowski <k.kozlowski@xxxxxxxxxxx> > --- > Documentation/DMA-API.txt | 2 +- > Documentation/DMA-attributes.txt | 2 +- > arch/arm/include/asm/dma-mapping.h | 13 ++-- > arch/arm/include/asm/xen/page-coherent.h | 16 ++--- > arch/arm/mm/dma-mapping.c | 82 +++++++++++------------ > arch/arm/xen/mm.c | 4 +- > arch/arm64/mm/dma-mapping.c | 57 ++++++++-------- > drivers/gpu/drm/exynos/exynos_drm_fbdev.c | 2 +- > drivers/gpu/drm/exynos/exynos_drm_g2d.c | 1 - > drivers/gpu/drm/exynos/exynos_drm_gem.c | 20 +++--- > drivers/gpu/drm/exynos/exynos_drm_gem.h | 2 +- > drivers/iommu/dma-iommu.c | 6 +- > drivers/xen/swiotlb-xen.c | 14 ++-- > include/linux/dma-attrs.h | 71 -------------------- > include/linux/dma-iommu.h | 6 +- > include/linux/dma-mapping.h | 105 +++++++++++++++++------------- > include/linux/swiotlb.h | 10 +-- > include/xen/swiotlb-xen.h | 12 ++-- > lib/dma-noop.c | 9 +-- > lib/swiotlb.c | 13 ++-- > 20 files changed, 195 insertions(+), 252 deletions(-) > delete mode 100644 include/linux/dma-attrs.h > > diff --git a/Documentation/DMA-API.txt b/Documentation/DMA-API.txt > index 45ef3f279c3b..0b55cb7c5aaa 100644 > --- a/Documentation/DMA-API.txt > +++ b/Documentation/DMA-API.txt > @@ -391,7 +391,7 @@ without the _attrs suffixes, except that they pass an optional > struct dma_attrs*. > > struct dma_attrs encapsulates a set of "DMA attributes". For the > -definition of struct dma_attrs see linux/dma-attrs.h. > +definition of struct dma_attrs see linux/dma-mapping.h. > > The interpretation of DMA attributes is architecture-specific, and > each attribute should be documented in Documentation/DMA-attributes.txt. > diff --git a/Documentation/DMA-attributes.txt b/Documentation/DMA-attributes.txt > index e8cf9cf873b3..2d455a5cf671 100644 > --- a/Documentation/DMA-attributes.txt > +++ b/Documentation/DMA-attributes.txt > @@ -2,7 +2,7 @@ > ============== > > This document describes the semantics of the DMA attributes that are > -defined in linux/dma-attrs.h. > +defined in linux/dma-mapping.h. > > DMA_ATTR_WRITE_BARRIER > ---------------------- > diff --git a/arch/arm/include/asm/dma-mapping.h b/arch/arm/include/asm/dma-mapping.h > index a83570f10124..d009f7911ffc 100644 > --- a/arch/arm/include/asm/dma-mapping.h > +++ b/arch/arm/include/asm/dma-mapping.h > @@ -5,7 +5,6 @@ > > #include <linux/mm_types.h> > #include <linux/scatterlist.h> > -#include <linux/dma-attrs.h> > #include <linux/dma-debug.h> > > #include <asm/memory.h> > @@ -174,7 +173,7 @@ static inline void dma_mark_clean(void *addr, size_t size) { } > * to be the device-viewed address. > */ > extern void *arm_dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, > - gfp_t gfp, struct dma_attrs *attrs); > + gfp_t gfp, unsigned long attrs); > > /** > * arm_dma_free - free memory allocated by arm_dma_alloc > @@ -191,7 +190,7 @@ extern void *arm_dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, > * during and after this call executing are illegal. > */ > extern void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, > - dma_addr_t handle, struct dma_attrs *attrs); > + dma_addr_t handle, unsigned long attrs); > > /** > * arm_dma_mmap - map a coherent DMA allocation into user space > @@ -208,7 +207,7 @@ extern void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, > */ > extern int arm_dma_mmap(struct device *dev, struct vm_area_struct *vma, > void *cpu_addr, dma_addr_t dma_addr, size_t size, > - struct dma_attrs *attrs); > + unsigned long attrs); > > /* > * This can be called during early boot to increase the size of the atomic > @@ -262,16 +261,16 @@ extern void dmabounce_unregister_dev(struct device *); > * The scatter list versions of the above methods. > */ > extern int arm_dma_map_sg(struct device *, struct scatterlist *, int, > - enum dma_data_direction, struct dma_attrs *attrs); > + enum dma_data_direction, unsigned long attrs); > extern void arm_dma_unmap_sg(struct device *, struct scatterlist *, int, > - enum dma_data_direction, struct dma_attrs *attrs); > + enum dma_data_direction, unsigned long attrs); > extern void arm_dma_sync_sg_for_cpu(struct device *, struct scatterlist *, int, > enum dma_data_direction); > extern void arm_dma_sync_sg_for_device(struct device *, struct scatterlist *, int, > enum dma_data_direction); > extern int arm_dma_get_sgtable(struct device *dev, struct sg_table *sgt, > void *cpu_addr, dma_addr_t dma_addr, size_t size, > - struct dma_attrs *attrs); > + unsigned long attrs); > > #endif /* __KERNEL__ */ > #endif > diff --git a/arch/arm/include/asm/xen/page-coherent.h b/arch/arm/include/asm/xen/page-coherent.h > index 9408a994cc91..95ce6ac3a971 100644 > --- a/arch/arm/include/asm/xen/page-coherent.h > +++ b/arch/arm/include/asm/xen/page-coherent.h > @@ -2,15 +2,14 @@ > #define _ASM_ARM_XEN_PAGE_COHERENT_H > > #include <asm/page.h> > -#include <linux/dma-attrs.h> > #include <linux/dma-mapping.h> > > void __xen_dma_map_page(struct device *hwdev, struct page *page, > dma_addr_t dev_addr, unsigned long offset, size_t size, > - enum dma_data_direction dir, struct dma_attrs *attrs); > + enum dma_data_direction dir, unsigned long attrs); > void __xen_dma_unmap_page(struct device *hwdev, dma_addr_t handle, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > void __xen_dma_sync_single_for_cpu(struct device *hwdev, > dma_addr_t handle, size_t size, enum dma_data_direction dir); > > @@ -18,22 +17,20 @@ void __xen_dma_sync_single_for_device(struct device *hwdev, > dma_addr_t handle, size_t size, enum dma_data_direction dir); > > static inline void *xen_alloc_coherent_pages(struct device *hwdev, size_t size, > - dma_addr_t *dma_handle, gfp_t flags, > - struct dma_attrs *attrs) > + dma_addr_t *dma_handle, gfp_t flags, unsigned long attrs) > { > return __generic_dma_ops(hwdev)->alloc(hwdev, size, dma_handle, flags, attrs); > } > > static inline void xen_free_coherent_pages(struct device *hwdev, size_t size, > - void *cpu_addr, dma_addr_t dma_handle, > - struct dma_attrs *attrs) > + void *cpu_addr, dma_addr_t dma_handle, unsigned long attrs) > { > __generic_dma_ops(hwdev)->free(hwdev, size, cpu_addr, dma_handle, attrs); > } > > static inline void xen_dma_map_page(struct device *hwdev, struct page *page, > dma_addr_t dev_addr, unsigned long offset, size_t size, > - enum dma_data_direction dir, struct dma_attrs *attrs) > + enum dma_data_direction dir, unsigned long attrs) > { > unsigned long page_pfn = page_to_xen_pfn(page); > unsigned long dev_pfn = XEN_PFN_DOWN(dev_addr); > @@ -58,8 +55,7 @@ static inline void xen_dma_map_page(struct device *hwdev, struct page *page, > } > > static inline void xen_dma_unmap_page(struct device *hwdev, dma_addr_t handle, > - size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + size_t size, enum dma_data_direction dir, unsigned long attrs) > { > unsigned long pfn = PFN_DOWN(handle); > /* > diff --git a/arch/arm/mm/dma-mapping.c b/arch/arm/mm/dma-mapping.c > index ff7ed5697d3e..fe31fbfd926d 100644 > --- a/arch/arm/mm/dma-mapping.c > +++ b/arch/arm/mm/dma-mapping.c > @@ -124,7 +124,7 @@ static void __dma_page_dev_to_cpu(struct page *, unsigned long, > */ > static dma_addr_t arm_dma_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) > __dma_page_cpu_to_dev(page, offset, size, dir); > @@ -133,7 +133,7 @@ static dma_addr_t arm_dma_map_page(struct device *dev, struct page *page, > > static dma_addr_t arm_coherent_dma_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > return pfn_to_dma(dev, page_to_pfn(page)) + offset; > } > @@ -153,8 +153,7 @@ static dma_addr_t arm_coherent_dma_map_page(struct device *dev, struct page *pag > * whatever the device wrote there. > */ > static void arm_dma_unmap_page(struct device *dev, dma_addr_t handle, > - size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + size_t size, enum dma_data_direction dir, unsigned long attrs) > { > if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) > __dma_page_dev_to_cpu(pfn_to_page(dma_to_pfn(dev, handle)), > @@ -194,12 +193,12 @@ struct dma_map_ops arm_dma_ops = { > EXPORT_SYMBOL(arm_dma_ops); > > static void *arm_coherent_dma_alloc(struct device *dev, size_t size, > - dma_addr_t *handle, gfp_t gfp, struct dma_attrs *attrs); > + dma_addr_t *handle, gfp_t gfp, unsigned long attrs); > static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr, > - dma_addr_t handle, struct dma_attrs *attrs); > + dma_addr_t handle, unsigned long attrs); > static int arm_coherent_dma_mmap(struct device *dev, struct vm_area_struct *vma, > void *cpu_addr, dma_addr_t dma_addr, size_t size, > - struct dma_attrs *attrs); > + unsigned long attrs); > > struct dma_map_ops arm_coherent_dma_ops = { > .alloc = arm_coherent_dma_alloc, > @@ -621,7 +620,7 @@ static void __free_from_contiguous(struct device *dev, struct page *page, > dma_release_from_contiguous(dev, page, size >> PAGE_SHIFT); > } > > -static inline pgprot_t __get_dma_pgprot(struct dma_attrs *attrs, pgprot_t prot) > +static inline pgprot_t __get_dma_pgprot(unsigned long attrs, pgprot_t prot) > { > prot = dma_get_attr(DMA_ATTR_WRITE_COMBINE, attrs) ? > pgprot_writecombine(prot) : > @@ -732,7 +731,7 @@ static struct arm_dma_allocator remap_allocator = { > > static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, > gfp_t gfp, pgprot_t prot, bool is_coherent, > - struct dma_attrs *attrs, const void *caller) > + unsigned long attrs, const void *caller) > { > u64 mask = get_coherent_dma_mask(dev); > struct page *page = NULL; > @@ -814,7 +813,7 @@ static void *__dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, > * virtual and bus address for that space. > */ > void *arm_dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, > - gfp_t gfp, struct dma_attrs *attrs) > + gfp_t gfp, unsigned long attrs) > { > pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL); > > @@ -823,7 +822,7 @@ void *arm_dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, > } > > static void *arm_coherent_dma_alloc(struct device *dev, size_t size, > - dma_addr_t *handle, gfp_t gfp, struct dma_attrs *attrs) > + dma_addr_t *handle, gfp_t gfp, unsigned long attrs) > { > return __dma_alloc(dev, size, handle, gfp, PAGE_KERNEL, true, > attrs, __builtin_return_address(0)); > @@ -831,7 +830,7 @@ static void *arm_coherent_dma_alloc(struct device *dev, size_t size, > > static int __arm_dma_mmap(struct device *dev, struct vm_area_struct *vma, > void *cpu_addr, dma_addr_t dma_addr, size_t size, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > int ret = -ENXIO; > #ifdef CONFIG_MMU > @@ -859,14 +858,14 @@ static int __arm_dma_mmap(struct device *dev, struct vm_area_struct *vma, > */ > static int arm_coherent_dma_mmap(struct device *dev, struct vm_area_struct *vma, > void *cpu_addr, dma_addr_t dma_addr, size_t size, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > return __arm_dma_mmap(dev, vma, cpu_addr, dma_addr, size, attrs); > } > > int arm_dma_mmap(struct device *dev, struct vm_area_struct *vma, > void *cpu_addr, dma_addr_t dma_addr, size_t size, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > #ifdef CONFIG_MMU > vma->vm_page_prot = __get_dma_pgprot(attrs, vma->vm_page_prot); > @@ -878,7 +877,7 @@ int arm_dma_mmap(struct device *dev, struct vm_area_struct *vma, > * Free a buffer as defined by the above mapping. > */ > static void __arm_dma_free(struct device *dev, size_t size, void *cpu_addr, > - dma_addr_t handle, struct dma_attrs *attrs, > + dma_addr_t handle, unsigned long attrs, > bool is_coherent) > { > struct page *page = pfn_to_page(dma_to_pfn(dev, handle)); > @@ -900,20 +899,20 @@ static void __arm_dma_free(struct device *dev, size_t size, void *cpu_addr, > } > > void arm_dma_free(struct device *dev, size_t size, void *cpu_addr, > - dma_addr_t handle, struct dma_attrs *attrs) > + dma_addr_t handle, unsigned long attrs) > { > __arm_dma_free(dev, size, cpu_addr, handle, attrs, false); > } > > static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr, > - dma_addr_t handle, struct dma_attrs *attrs) > + dma_addr_t handle, unsigned long attrs) > { > __arm_dma_free(dev, size, cpu_addr, handle, attrs, true); > } > > int arm_dma_get_sgtable(struct device *dev, struct sg_table *sgt, > void *cpu_addr, dma_addr_t handle, size_t size, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct page *page = pfn_to_page(dma_to_pfn(dev, handle)); > int ret; > @@ -1046,7 +1045,7 @@ static void __dma_page_dev_to_cpu(struct page *page, unsigned long off, > * here. > */ > int arm_dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, > - enum dma_data_direction dir, struct dma_attrs *attrs) > + enum dma_data_direction dir, unsigned long attrs) > { > struct dma_map_ops *ops = get_dma_ops(dev); > struct scatterlist *s; > @@ -1080,7 +1079,7 @@ int arm_dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, > * rules concerning calls here are the same as for dma_unmap_single(). > */ > void arm_dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, > - enum dma_data_direction dir, struct dma_attrs *attrs) > + enum dma_data_direction dir, unsigned long attrs) > { > struct dma_map_ops *ops = get_dma_ops(dev); > struct scatterlist *s; > @@ -1253,7 +1252,7 @@ static inline void __free_iova(struct dma_iommu_mapping *mapping, > static const int iommu_order_array[] = { 9, 8, 4, 0 }; > > static struct page **__iommu_alloc_buffer(struct device *dev, size_t size, > - gfp_t gfp, struct dma_attrs *attrs) > + gfp_t gfp, unsigned long attrs) > { > struct page **pages; > int count = size >> PAGE_SHIFT; > @@ -1342,7 +1341,7 @@ error: > } > > static int __iommu_free_buffer(struct device *dev, struct page **pages, > - size_t size, struct dma_attrs *attrs) > + size_t size, unsigned long attrs) > { > int count = size >> PAGE_SHIFT; > int i; > @@ -1439,7 +1438,7 @@ static struct page **__atomic_get_pages(void *addr) > return (struct page **)page; > } > > -static struct page **__iommu_get_pages(void *cpu_addr, struct dma_attrs *attrs) > +static struct page **__iommu_get_pages(void *cpu_addr, unsigned long attrs) > { > struct vm_struct *area; > > @@ -1484,7 +1483,7 @@ static void __iommu_free_atomic(struct device *dev, void *cpu_addr, > } > > static void *arm_iommu_alloc_attrs(struct device *dev, size_t size, > - dma_addr_t *handle, gfp_t gfp, struct dma_attrs *attrs) > + dma_addr_t *handle, gfp_t gfp, unsigned long attrs) > { > pgprot_t prot = __get_dma_pgprot(attrs, PAGE_KERNEL); > struct page **pages; > @@ -1532,7 +1531,7 @@ err_buffer: > > static int arm_iommu_mmap_attrs(struct device *dev, struct vm_area_struct *vma, > void *cpu_addr, dma_addr_t dma_addr, size_t size, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > unsigned long uaddr = vma->vm_start; > unsigned long usize = vma->vm_end - vma->vm_start; > @@ -1568,7 +1567,7 @@ static int arm_iommu_mmap_attrs(struct device *dev, struct vm_area_struct *vma, > * Must not be called with IRQs disabled. > */ > void arm_iommu_free_attrs(struct device *dev, size_t size, void *cpu_addr, > - dma_addr_t handle, struct dma_attrs *attrs) > + dma_addr_t handle, unsigned long attrs) > { > struct page **pages; > size = PAGE_ALIGN(size); > @@ -1595,7 +1594,7 @@ void arm_iommu_free_attrs(struct device *dev, size_t size, void *cpu_addr, > > static int arm_iommu_get_sgtable(struct device *dev, struct sg_table *sgt, > void *cpu_addr, dma_addr_t dma_addr, > - size_t size, struct dma_attrs *attrs) > + size_t size, unsigned long attrs) > { > unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT; > struct page **pages = __iommu_get_pages(cpu_addr, attrs); > @@ -1633,7 +1632,7 @@ static int __dma_direction_to_prot(enum dma_data_direction dir) > */ > static int __map_sg_chunk(struct device *dev, struct scatterlist *sg, > size_t size, dma_addr_t *handle, > - enum dma_data_direction dir, struct dma_attrs *attrs, > + enum dma_data_direction dir, unsigned long attrs, > bool is_coherent) > { > struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); > @@ -1676,7 +1675,7 @@ fail: > } > > static int __iommu_map_sg(struct device *dev, struct scatterlist *sg, int nents, > - enum dma_data_direction dir, struct dma_attrs *attrs, > + enum dma_data_direction dir, unsigned long attrs, > bool is_coherent) > { > struct scatterlist *s = sg, *dma = sg, *start = sg; > @@ -1734,7 +1733,7 @@ bad_mapping: > * obtained via sg_dma_{address,length}. > */ > int arm_coherent_iommu_map_sg(struct device *dev, struct scatterlist *sg, > - int nents, enum dma_data_direction dir, struct dma_attrs *attrs) > + int nents, enum dma_data_direction dir, unsigned long attrs) > { > return __iommu_map_sg(dev, sg, nents, dir, attrs, true); > } > @@ -1752,14 +1751,14 @@ int arm_coherent_iommu_map_sg(struct device *dev, struct scatterlist *sg, > * sg_dma_{address,length}. > */ > int arm_iommu_map_sg(struct device *dev, struct scatterlist *sg, > - int nents, enum dma_data_direction dir, struct dma_attrs *attrs) > + int nents, enum dma_data_direction dir, unsigned long attrs) > { > return __iommu_map_sg(dev, sg, nents, dir, attrs, false); > } > > static void __iommu_unmap_sg(struct device *dev, struct scatterlist *sg, > - int nents, enum dma_data_direction dir, struct dma_attrs *attrs, > - bool is_coherent) > + int nents, enum dma_data_direction dir, > + unsigned long attrs, bool is_coherent) > { > struct scatterlist *s; > int i; > @@ -1786,7 +1785,8 @@ static void __iommu_unmap_sg(struct device *dev, struct scatterlist *sg, > * rules concerning calls here are the same as for dma_unmap_single(). > */ > void arm_coherent_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, > - int nents, enum dma_data_direction dir, struct dma_attrs *attrs) > + int nents, enum dma_data_direction dir, > + unsigned long attrs) > { > __iommu_unmap_sg(dev, sg, nents, dir, attrs, true); > } > @@ -1802,7 +1802,8 @@ void arm_coherent_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, > * rules concerning calls here are the same as for dma_unmap_single(). > */ > void arm_iommu_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, > - enum dma_data_direction dir, struct dma_attrs *attrs) > + enum dma_data_direction dir, > + unsigned long attrs) > { > __iommu_unmap_sg(dev, sg, nents, dir, attrs, false); > } > @@ -1855,7 +1856,7 @@ void arm_iommu_sync_sg_for_device(struct device *dev, struct scatterlist *sg, > */ > static dma_addr_t arm_coherent_iommu_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); > dma_addr_t dma_addr; > @@ -1889,7 +1890,7 @@ fail: > */ > static dma_addr_t arm_iommu_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) > __dma_page_cpu_to_dev(page, offset, size, dir); > @@ -1907,8 +1908,7 @@ static dma_addr_t arm_iommu_map_page(struct device *dev, struct page *page, > * Coherent IOMMU aware version of arm_dma_unmap_page() > */ > static void arm_coherent_iommu_unmap_page(struct device *dev, dma_addr_t handle, > - size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + size_t size, enum dma_data_direction dir, unsigned long attrs) > { > struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); > dma_addr_t iova = handle & PAGE_MASK; > @@ -1932,8 +1932,7 @@ static void arm_coherent_iommu_unmap_page(struct device *dev, dma_addr_t handle, > * IOMMU aware version of arm_dma_unmap_page() > */ > static void arm_iommu_unmap_page(struct device *dev, dma_addr_t handle, > - size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + size_t size, enum dma_data_direction dir, unsigned long attrs) > { > struct dma_iommu_mapping *mapping = to_dma_iommu_mapping(dev); > dma_addr_t iova = handle & PAGE_MASK; > @@ -1944,6 +1943,7 @@ static void arm_iommu_unmap_page(struct device *dev, dma_addr_t handle, > if (!iova) > return; > > + // FIXME: replace get with simple check > if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) > __dma_page_dev_to_cpu(page, offset, size, dir); > > diff --git a/arch/arm/xen/mm.c b/arch/arm/xen/mm.c > index c5f9a9e3d1f3..fc67ed236a10 100644 > --- a/arch/arm/xen/mm.c > +++ b/arch/arm/xen/mm.c > @@ -98,7 +98,7 @@ static void __xen_dma_page_cpu_to_dev(struct device *hwdev, dma_addr_t handle, > > void __xen_dma_map_page(struct device *hwdev, struct page *page, > dma_addr_t dev_addr, unsigned long offset, size_t size, > - enum dma_data_direction dir, struct dma_attrs *attrs) > + enum dma_data_direction dir, unsigned long attrs) > { > if (is_device_dma_coherent(hwdev)) > return; > @@ -110,7 +110,7 @@ void __xen_dma_map_page(struct device *hwdev, struct page *page, > > void __xen_dma_unmap_page(struct device *hwdev, dma_addr_t handle, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > > { > if (is_device_dma_coherent(hwdev)) > diff --git a/arch/arm64/mm/dma-mapping.c b/arch/arm64/mm/dma-mapping.c > index c566ec83719f..a7686028dfeb 100644 > --- a/arch/arm64/mm/dma-mapping.c > +++ b/arch/arm64/mm/dma-mapping.c > @@ -29,7 +29,7 @@ > > #include <asm/cacheflush.h> > > -static pgprot_t __get_dma_pgprot(struct dma_attrs *attrs, pgprot_t prot, > +static pgprot_t __get_dma_pgprot(unsigned long attrs, pgprot_t prot, > bool coherent) > { > if (!coherent || dma_get_attr(DMA_ATTR_WRITE_COMBINE, attrs)) > @@ -88,7 +88,7 @@ static int __free_from_pool(void *start, size_t size) > > static void *__dma_alloc_coherent(struct device *dev, size_t size, > dma_addr_t *dma_handle, gfp_t flags, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > if (dev == NULL) { > WARN_ONCE(1, "Use an actual device structure for DMA allocation\n"); > @@ -118,7 +118,7 @@ static void *__dma_alloc_coherent(struct device *dev, size_t size, > > static void __dma_free_coherent(struct device *dev, size_t size, > void *vaddr, dma_addr_t dma_handle, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > bool freed; > phys_addr_t paddr = dma_to_phys(dev, dma_handle); > @@ -137,7 +137,7 @@ static void __dma_free_coherent(struct device *dev, size_t size, > > static void *__dma_alloc(struct device *dev, size_t size, > dma_addr_t *dma_handle, gfp_t flags, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct page *page; > void *ptr, *coherent_ptr; > @@ -185,7 +185,7 @@ no_mem: > > static void __dma_free(struct device *dev, size_t size, > void *vaddr, dma_addr_t dma_handle, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > void *swiotlb_addr = phys_to_virt(dma_to_phys(dev, dma_handle)); > > @@ -202,7 +202,7 @@ static void __dma_free(struct device *dev, size_t size, > static dma_addr_t __swiotlb_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > dma_addr_t dev_addr; > > @@ -216,7 +216,7 @@ static dma_addr_t __swiotlb_map_page(struct device *dev, struct page *page, > > static void __swiotlb_unmap_page(struct device *dev, dma_addr_t dev_addr, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > if (!is_device_dma_coherent(dev)) > __dma_unmap_area(phys_to_virt(dma_to_phys(dev, dev_addr)), size, dir); > @@ -225,7 +225,7 @@ static void __swiotlb_unmap_page(struct device *dev, dma_addr_t dev_addr, > > static int __swiotlb_map_sg_attrs(struct device *dev, struct scatterlist *sgl, > int nelems, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct scatterlist *sg; > int i, ret; > @@ -242,7 +242,7 @@ static int __swiotlb_map_sg_attrs(struct device *dev, struct scatterlist *sgl, > static void __swiotlb_unmap_sg_attrs(struct device *dev, > struct scatterlist *sgl, int nelems, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct scatterlist *sg; > int i; > @@ -303,7 +303,7 @@ static void __swiotlb_sync_sg_for_device(struct device *dev, > static int __swiotlb_mmap(struct device *dev, > struct vm_area_struct *vma, > void *cpu_addr, dma_addr_t dma_addr, size_t size, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > int ret = -ENXIO; > unsigned long nr_vma_pages = (vma->vm_end - vma->vm_start) >> > @@ -330,7 +330,7 @@ static int __swiotlb_mmap(struct device *dev, > > static int __swiotlb_get_sgtable(struct device *dev, struct sg_table *sgt, > void *cpu_addr, dma_addr_t handle, size_t size, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > int ret = sg_alloc_table(sgt, 1, GFP_KERNEL); > > @@ -425,21 +425,21 @@ out: > > static void *__dummy_alloc(struct device *dev, size_t size, > dma_addr_t *dma_handle, gfp_t flags, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > return NULL; > } > > static void __dummy_free(struct device *dev, size_t size, > void *vaddr, dma_addr_t dma_handle, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > } > > static int __dummy_mmap(struct device *dev, > struct vm_area_struct *vma, > void *cpu_addr, dma_addr_t dma_addr, size_t size, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > return -ENXIO; > } > @@ -447,20 +447,20 @@ static int __dummy_mmap(struct device *dev, > static dma_addr_t __dummy_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > return DMA_ERROR_CODE; > } > > static void __dummy_unmap_page(struct device *dev, dma_addr_t dev_addr, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > } > > static int __dummy_map_sg(struct device *dev, struct scatterlist *sgl, > int nelems, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > return 0; > } > @@ -468,7 +468,7 @@ static int __dummy_map_sg(struct device *dev, struct scatterlist *sgl, > static void __dummy_unmap_sg(struct device *dev, > struct scatterlist *sgl, int nelems, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > } > > @@ -540,7 +540,7 @@ static void flush_page(struct device *dev, const void *virt, phys_addr_t phys) > > static void *__iommu_alloc_attrs(struct device *dev, size_t size, > dma_addr_t *handle, gfp_t gfp, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > bool coherent = is_device_dma_coherent(dev); > int ioprot = dma_direction_to_prot(DMA_BIDIRECTIONAL, coherent); > @@ -600,7 +600,8 @@ static void *__iommu_alloc_attrs(struct device *dev, size_t size, > } > > static void __iommu_free_attrs(struct device *dev, size_t size, void *cpu_addr, > - dma_addr_t handle, struct dma_attrs *attrs) > + dma_addr_t handle, > + unsigned long attrs) > { > size_t iosize = size; > > @@ -616,7 +617,7 @@ static void __iommu_free_attrs(struct device *dev, size_t size, void *cpu_addr, > * Hence how dodgy the below logic looks... > */ > if (__in_atomic_pool(cpu_addr, size)) { > - iommu_dma_unmap_page(dev, handle, iosize, 0, NULL); > + iommu_dma_unmap_page(dev, handle, iosize, 0, 0); > __free_from_pool(cpu_addr, size); > } else if (is_vmalloc_addr(cpu_addr)){ > struct vm_struct *area = find_vm_area(cpu_addr); > @@ -626,14 +627,14 @@ static void __iommu_free_attrs(struct device *dev, size_t size, void *cpu_addr, > iommu_dma_free(dev, area->pages, iosize, &handle); > dma_common_free_remap(cpu_addr, size, VM_USERMAP); > } else { > - iommu_dma_unmap_page(dev, handle, iosize, 0, NULL); > + iommu_dma_unmap_page(dev, handle, iosize, 0, 0); > __free_pages(virt_to_page(cpu_addr), get_order(size)); > } > } > > static int __iommu_mmap_attrs(struct device *dev, struct vm_area_struct *vma, > void *cpu_addr, dma_addr_t dma_addr, size_t size, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct vm_struct *area; > int ret; > @@ -653,7 +654,7 @@ static int __iommu_mmap_attrs(struct device *dev, struct vm_area_struct *vma, > > static int __iommu_get_sgtable(struct device *dev, struct sg_table *sgt, > void *cpu_addr, dma_addr_t dma_addr, > - size_t size, struct dma_attrs *attrs) > + size_t size, unsigned long attrs) > { > unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT; > struct vm_struct *area = find_vm_area(cpu_addr); > @@ -694,7 +695,7 @@ static void __iommu_sync_single_for_device(struct device *dev, > static dma_addr_t __iommu_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > bool coherent = is_device_dma_coherent(dev); > int prot = dma_direction_to_prot(dir, coherent); > @@ -709,7 +710,7 @@ static dma_addr_t __iommu_map_page(struct device *dev, struct page *page, > > static void __iommu_unmap_page(struct device *dev, dma_addr_t dev_addr, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) > __iommu_sync_single_for_cpu(dev, dev_addr, size, dir); > @@ -747,7 +748,7 @@ static void __iommu_sync_sg_for_device(struct device *dev, > > static int __iommu_map_sg_attrs(struct device *dev, struct scatterlist *sgl, > int nelems, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > bool coherent = is_device_dma_coherent(dev); > > @@ -761,7 +762,7 @@ static int __iommu_map_sg_attrs(struct device *dev, struct scatterlist *sgl, > static void __iommu_unmap_sg_attrs(struct device *dev, > struct scatterlist *sgl, int nelems, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > if (!dma_get_attr(DMA_ATTR_SKIP_CPU_SYNC, attrs)) > __iommu_sync_sg_for_cpu(dev, sgl, nelems, dir); > diff --git a/drivers/gpu/drm/exynos/exynos_drm_fbdev.c b/drivers/gpu/drm/exynos/exynos_drm_fbdev.c > index 67dcd6831291..dd091175fc2d 100644 > --- a/drivers/gpu/drm/exynos/exynos_drm_fbdev.c > +++ b/drivers/gpu/drm/exynos/exynos_drm_fbdev.c > @@ -52,7 +52,7 @@ static int exynos_drm_fb_mmap(struct fb_info *info, > > ret = dma_mmap_attrs(to_dma_dev(helper->dev), vma, exynos_gem->cookie, > exynos_gem->dma_addr, exynos_gem->size, > - &exynos_gem->dma_attrs); > + exynos_gem->dma_attrs); > if (ret < 0) { > DRM_ERROR("failed to mmap.\n"); > return ret; > diff --git a/drivers/gpu/drm/exynos/exynos_drm_g2d.c b/drivers/gpu/drm/exynos/exynos_drm_g2d.c > index 493552368295..f65e6b7ef93b 100644 > --- a/drivers/gpu/drm/exynos/exynos_drm_g2d.c > +++ b/drivers/gpu/drm/exynos/exynos_drm_g2d.c > @@ -17,7 +17,6 @@ > #include <linux/slab.h> > #include <linux/workqueue.h> > #include <linux/dma-mapping.h> > -#include <linux/dma-attrs.h> > #include <linux/of.h> > > #include <drm/drmP.h> > diff --git a/drivers/gpu/drm/exynos/exynos_drm_gem.c b/drivers/gpu/drm/exynos/exynos_drm_gem.c > index cdf9f1af4347..f2ae72ba7d5a 100644 > --- a/drivers/gpu/drm/exynos/exynos_drm_gem.c > +++ b/drivers/gpu/drm/exynos/exynos_drm_gem.c > @@ -24,7 +24,7 @@ > static int exynos_drm_alloc_buf(struct exynos_drm_gem *exynos_gem) > { > struct drm_device *dev = exynos_gem->base.dev; > - enum dma_attr attr; > + unsigned long attr; > unsigned int nr_pages; > struct sg_table sgt; > int ret = -ENOMEM; > @@ -34,7 +34,7 @@ static int exynos_drm_alloc_buf(struct exynos_drm_gem *exynos_gem) > return 0; > } > > - init_dma_attrs(&exynos_gem->dma_attrs); > + exynos_gem->dma_attrs = 0; > > /* > * if EXYNOS_BO_CONTIG, fully physically contiguous memory > @@ -42,7 +42,7 @@ static int exynos_drm_alloc_buf(struct exynos_drm_gem *exynos_gem) > * as possible. > */ > if (!(exynos_gem->flags & EXYNOS_BO_NONCONTIG)) > - dma_set_attr(DMA_ATTR_FORCE_CONTIGUOUS, &exynos_gem->dma_attrs); > + exynos_gem->dma_attrs |= DMA_ATTR_FORCE_CONTIGUOUS; > > /* > * if EXYNOS_BO_WC or EXYNOS_BO_NONCACHABLE, writecombine mapping > @@ -54,8 +54,8 @@ static int exynos_drm_alloc_buf(struct exynos_drm_gem *exynos_gem) > else > attr = DMA_ATTR_NON_CONSISTENT; > > - dma_set_attr(attr, &exynos_gem->dma_attrs); > - dma_set_attr(DMA_ATTR_NO_KERNEL_MAPPING, &exynos_gem->dma_attrs); > + exynos_gem->dma_attrs |= attr; > + exynos_gem->dma_attrs |= DMA_ATTR_NO_KERNEL_MAPPING; > > nr_pages = exynos_gem->size >> PAGE_SHIFT; > > @@ -67,7 +67,7 @@ static int exynos_drm_alloc_buf(struct exynos_drm_gem *exynos_gem) > > exynos_gem->cookie = dma_alloc_attrs(to_dma_dev(dev), exynos_gem->size, > &exynos_gem->dma_addr, GFP_KERNEL, > - &exynos_gem->dma_attrs); > + exynos_gem->dma_attrs); > if (!exynos_gem->cookie) { > DRM_ERROR("failed to allocate buffer.\n"); > goto err_free; > @@ -75,7 +75,7 @@ static int exynos_drm_alloc_buf(struct exynos_drm_gem *exynos_gem) > > ret = dma_get_sgtable_attrs(to_dma_dev(dev), &sgt, exynos_gem->cookie, > exynos_gem->dma_addr, exynos_gem->size, > - &exynos_gem->dma_attrs); > + exynos_gem->dma_attrs); > if (ret < 0) { > DRM_ERROR("failed to get sgtable.\n"); > goto err_dma_free; > @@ -99,7 +99,7 @@ err_sgt_free: > sg_free_table(&sgt); > err_dma_free: > dma_free_attrs(to_dma_dev(dev), exynos_gem->size, exynos_gem->cookie, > - exynos_gem->dma_addr, &exynos_gem->dma_attrs); > + exynos_gem->dma_addr, exynos_gem->dma_attrs); > err_free: > drm_free_large(exynos_gem->pages); > > @@ -120,7 +120,7 @@ static void exynos_drm_free_buf(struct exynos_drm_gem *exynos_gem) > > dma_free_attrs(to_dma_dev(dev), exynos_gem->size, exynos_gem->cookie, > (dma_addr_t)exynos_gem->dma_addr, > - &exynos_gem->dma_attrs); > + exynos_gem->dma_attrs); > > drm_free_large(exynos_gem->pages); > } > @@ -346,7 +346,7 @@ static int exynos_drm_gem_mmap_buffer(struct exynos_drm_gem *exynos_gem, > > ret = dma_mmap_attrs(to_dma_dev(drm_dev), vma, exynos_gem->cookie, > exynos_gem->dma_addr, exynos_gem->size, > - &exynos_gem->dma_attrs); > + exynos_gem->dma_attrs); > if (ret < 0) { > DRM_ERROR("failed to mmap.\n"); > return ret; > diff --git a/drivers/gpu/drm/exynos/exynos_drm_gem.h b/drivers/gpu/drm/exynos/exynos_drm_gem.h > index 78100742281d..df7c543d6558 100644 > --- a/drivers/gpu/drm/exynos/exynos_drm_gem.h > +++ b/drivers/gpu/drm/exynos/exynos_drm_gem.h > @@ -50,7 +50,7 @@ struct exynos_drm_gem { > void *cookie; > void __iomem *kvaddr; > dma_addr_t dma_addr; > - struct dma_attrs dma_attrs; > + unsigned long dma_attrs; > struct page **pages; > struct sg_table *sgt; > }; > diff --git a/drivers/iommu/dma-iommu.c b/drivers/iommu/dma-iommu.c > index ea5a9ebf0f78..6c1bda504fb1 100644 > --- a/drivers/iommu/dma-iommu.c > +++ b/drivers/iommu/dma-iommu.c > @@ -286,7 +286,7 @@ void iommu_dma_free(struct device *dev, struct page **pages, size_t size, > * or NULL on failure. > */ > struct page **iommu_dma_alloc(struct device *dev, size_t size, gfp_t gfp, > - struct dma_attrs *attrs, int prot, dma_addr_t *handle, > + unsigned long attrs, int prot, dma_addr_t *handle, > void (*flush_page)(struct device *, const void *, phys_addr_t)) > { > struct iommu_domain *domain = iommu_get_domain_for_dev(dev); > @@ -400,7 +400,7 @@ dma_addr_t iommu_dma_map_page(struct device *dev, struct page *page, > } > > void iommu_dma_unmap_page(struct device *dev, dma_addr_t handle, size_t size, > - enum dma_data_direction dir, struct dma_attrs *attrs) > + enum dma_data_direction dir, unsigned long attrs) > { > __iommu_dma_unmap(iommu_get_domain_for_dev(dev), handle); > } > @@ -560,7 +560,7 @@ out_restore_sg: > } > > void iommu_dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, > - enum dma_data_direction dir, struct dma_attrs *attrs) > + enum dma_data_direction dir, unsigned long attrs) > { > /* > * The scatterlist segments are mapped into a single > diff --git a/drivers/xen/swiotlb-xen.c b/drivers/xen/swiotlb-xen.c > index 7399782c0998..87e6035c9e81 100644 > --- a/drivers/xen/swiotlb-xen.c > +++ b/drivers/xen/swiotlb-xen.c > @@ -294,7 +294,7 @@ error: > void * > xen_swiotlb_alloc_coherent(struct device *hwdev, size_t size, > dma_addr_t *dma_handle, gfp_t flags, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > void *ret; > int order = get_order(size); > @@ -346,7 +346,7 @@ EXPORT_SYMBOL_GPL(xen_swiotlb_alloc_coherent); > > void > xen_swiotlb_free_coherent(struct device *hwdev, size_t size, void *vaddr, > - dma_addr_t dev_addr, struct dma_attrs *attrs) > + dma_addr_t dev_addr, unsigned long attrs) > { > int order = get_order(size); > phys_addr_t phys; > @@ -378,7 +378,7 @@ EXPORT_SYMBOL_GPL(xen_swiotlb_free_coherent); > dma_addr_t xen_swiotlb_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > phys_addr_t map, phys = page_to_phys(page) + offset; > dma_addr_t dev_addr = xen_phys_to_bus(phys); > @@ -434,7 +434,7 @@ EXPORT_SYMBOL_GPL(xen_swiotlb_map_page); > */ > static void xen_unmap_single(struct device *hwdev, dma_addr_t dev_addr, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > phys_addr_t paddr = xen_bus_to_phys(dev_addr); > > @@ -462,7 +462,7 @@ static void xen_unmap_single(struct device *hwdev, dma_addr_t dev_addr, > > void xen_swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > xen_unmap_single(hwdev, dev_addr, size, dir, attrs); > } > @@ -538,7 +538,7 @@ EXPORT_SYMBOL_GPL(xen_swiotlb_sync_single_for_device); > int > xen_swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, > int nelems, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct scatterlist *sg; > int i; > @@ -599,7 +599,7 @@ EXPORT_SYMBOL_GPL(xen_swiotlb_map_sg_attrs); > void > xen_swiotlb_unmap_sg_attrs(struct device *hwdev, struct scatterlist *sgl, > int nelems, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct scatterlist *sg; > int i; > diff --git a/include/linux/dma-attrs.h b/include/linux/dma-attrs.h > deleted file mode 100644 > index 5246239a4953..000000000000 > --- a/include/linux/dma-attrs.h > +++ /dev/null > @@ -1,71 +0,0 @@ > -#ifndef _DMA_ATTR_H > -#define _DMA_ATTR_H > - > -#include <linux/bitmap.h> > -#include <linux/bitops.h> > -#include <linux/bug.h> > - > -/** > - * an enum dma_attr represents an attribute associated with a DMA > - * mapping. The semantics of each attribute should be defined in > - * Documentation/DMA-attributes.txt. > - */ > -enum dma_attr { > - DMA_ATTR_WRITE_BARRIER, > - DMA_ATTR_WEAK_ORDERING, > - DMA_ATTR_WRITE_COMBINE, > - DMA_ATTR_NON_CONSISTENT, > - DMA_ATTR_NO_KERNEL_MAPPING, > - DMA_ATTR_SKIP_CPU_SYNC, > - DMA_ATTR_FORCE_CONTIGUOUS, > - DMA_ATTR_ALLOC_SINGLE_PAGES, > - DMA_ATTR_MAX, > -}; > - > -#define __DMA_ATTRS_LONGS BITS_TO_LONGS(DMA_ATTR_MAX) > - > -/** > - * struct dma_attrs - an opaque container for DMA attributes > - * @flags - bitmask representing a collection of enum dma_attr > - */ > -struct dma_attrs { > - unsigned long flags[__DMA_ATTRS_LONGS]; > -}; > - > -#define DEFINE_DMA_ATTRS(x) \ > - struct dma_attrs x = { \ > - .flags = { [0 ... __DMA_ATTRS_LONGS-1] = 0 }, \ > - } > - > -static inline void init_dma_attrs(struct dma_attrs *attrs) > -{ > - bitmap_zero(attrs->flags, __DMA_ATTRS_LONGS); > -} > - > -/** > - * dma_set_attr - set a specific attribute > - * @attr: attribute to set > - * @attrs: struct dma_attrs (may be NULL) > - */ > -static inline void dma_set_attr(enum dma_attr attr, struct dma_attrs *attrs) > -{ > - if (attrs == NULL) > - return; > - BUG_ON(attr >= DMA_ATTR_MAX); > - __set_bit(attr, attrs->flags); > -} > - > -/** > - * dma_get_attr - check for a specific attribute > - * @attr: attribute to set > - * @attrs: struct dma_attrs (may be NULL) > - */ > -static inline int dma_get_attr(enum dma_attr attr, struct dma_attrs *attrs) > -{ > - if (attrs == NULL) > - return 0; > - BUG_ON(attr >= DMA_ATTR_MAX); > - return test_bit(attr, attrs->flags); > -} > - > -#endif /* _DMA_ATTR_H */ > diff --git a/include/linux/dma-iommu.h b/include/linux/dma-iommu.h > index 8443bbb5c071..81c5c8d167ad 100644 > --- a/include/linux/dma-iommu.h > +++ b/include/linux/dma-iommu.h > @@ -39,7 +39,7 @@ int dma_direction_to_prot(enum dma_data_direction dir, bool coherent); > * the arch code to take care of attributes and cache maintenance > */ > struct page **iommu_dma_alloc(struct device *dev, size_t size, gfp_t gfp, > - struct dma_attrs *attrs, int prot, dma_addr_t *handle, > + unsigned long attrs, int prot, dma_addr_t *handle, > void (*flush_page)(struct device *, const void *, phys_addr_t)); > void iommu_dma_free(struct device *dev, struct page **pages, size_t size, > dma_addr_t *handle); > @@ -56,9 +56,9 @@ int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg, > * directly as DMA mapping callbacks for simplicity > */ > void iommu_dma_unmap_page(struct device *dev, dma_addr_t handle, size_t size, > - enum dma_data_direction dir, struct dma_attrs *attrs); > + enum dma_data_direction dir, unsigned long attrs); > void iommu_dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, > - enum dma_data_direction dir, struct dma_attrs *attrs); > + enum dma_data_direction dir, unsigned long attrs); > int iommu_dma_supported(struct device *dev, u64 mask); > int iommu_dma_mapping_error(struct device *dev, dma_addr_t dma_addr); > > diff --git a/include/linux/dma-mapping.h b/include/linux/dma-mapping.h > index 71c1b215ef66..19e581d5f8b4 100644 > --- a/include/linux/dma-mapping.h > +++ b/include/linux/dma-mapping.h > @@ -5,13 +5,25 @@ > #include <linux/string.h> > #include <linux/device.h> > #include <linux/err.h> > -#include <linux/dma-attrs.h> > #include <linux/dma-debug.h> > #include <linux/dma-direction.h> > #include <linux/scatterlist.h> > #include <linux/kmemcheck.h> > #include <linux/bug.h> > > +/** > + * List of possible attributes associated with a DMA mapping. The semantics > + * of each attribute should be defined in Documentation/DMA-attributes.txt. > + */ > +#define DMA_ATTR_WRITE_BARRIER BIT(1) > +#define DMA_ATTR_WEAK_ORDERING BIT(2) > +#define DMA_ATTR_WRITE_COMBINE BIT(3) > +#define DMA_ATTR_NON_CONSISTENT BIT(4) > +#define DMA_ATTR_NO_KERNEL_MAPPING BIT(5) > +#define DMA_ATTR_SKIP_CPU_SYNC BIT(6) > +#define DMA_ATTR_FORCE_CONTIGUOUS BIT(7) > +#define DMA_ATTR_ALLOC_SINGLE_PAGES BIT(8) > + > /* > * A dma_addr_t can hold any valid DMA or bus address for the platform. > * It can be given to a device to use as a DMA source or target. A CPU cannot > @@ -21,34 +33,35 @@ > struct dma_map_ops { > void* (*alloc)(struct device *dev, size_t size, > dma_addr_t *dma_handle, gfp_t gfp, > - struct dma_attrs *attrs); > + unsigned long attrs); > void (*free)(struct device *dev, size_t size, > void *vaddr, dma_addr_t dma_handle, > - struct dma_attrs *attrs); > + unsigned long attrs); > int (*mmap)(struct device *, struct vm_area_struct *, > - void *, dma_addr_t, size_t, struct dma_attrs *attrs); > + void *, dma_addr_t, size_t, > + unsigned long attrs); > > int (*get_sgtable)(struct device *dev, struct sg_table *sgt, void *, > - dma_addr_t, size_t, struct dma_attrs *attrs); > + dma_addr_t, size_t, unsigned long attrs); > > dma_addr_t (*map_page)(struct device *dev, struct page *page, > unsigned long offset, size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > void (*unmap_page)(struct device *dev, dma_addr_t dma_handle, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > /* > * map_sg returns 0 on error and a value > 0 on success. > * It should never return a value < 0. > */ > int (*map_sg)(struct device *dev, struct scatterlist *sg, > int nents, enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > void (*unmap_sg)(struct device *dev, > struct scatterlist *sg, int nents, > enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > void (*sync_single_for_cpu)(struct device *dev, > dma_addr_t dma_handle, size_t size, > enum dma_data_direction dir); > @@ -88,6 +101,16 @@ static inline int is_device_dma_capable(struct device *dev) > return dev->dma_mask != NULL && *dev->dma_mask != DMA_MASK_NONE; > } > > +/** > + * dma_get_attr - check for a specific attribute > + * @attr: attribute to look for > + * @attrs: attributes to check within > + */ > +static inline bool dma_get_attr(unsigned long attr, unsigned long attrs) > +{ > + return !!(attr & attrs); > +} > + > #ifdef CONFIG_HAVE_GENERIC_DMA_COHERENT > /* > * These three functions are only for dma allocator. > @@ -123,7 +146,7 @@ static inline struct dma_map_ops *get_dma_ops(struct device *dev) > static inline dma_addr_t dma_map_single_attrs(struct device *dev, void *ptr, > size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct dma_map_ops *ops = get_dma_ops(dev); > dma_addr_t addr; > @@ -142,7 +165,7 @@ static inline dma_addr_t dma_map_single_attrs(struct device *dev, void *ptr, > static inline void dma_unmap_single_attrs(struct device *dev, dma_addr_t addr, > size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct dma_map_ops *ops = get_dma_ops(dev); > > @@ -158,7 +181,7 @@ static inline void dma_unmap_single_attrs(struct device *dev, dma_addr_t addr, > */ > static inline int dma_map_sg_attrs(struct device *dev, struct scatterlist *sg, > int nents, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct dma_map_ops *ops = get_dma_ops(dev); > int i, ents; > @@ -176,7 +199,7 @@ static inline int dma_map_sg_attrs(struct device *dev, struct scatterlist *sg, > > static inline void dma_unmap_sg_attrs(struct device *dev, struct scatterlist *sg, > int nents, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct dma_map_ops *ops = get_dma_ops(dev); > > @@ -195,7 +218,7 @@ static inline dma_addr_t dma_map_page(struct device *dev, struct page *page, > > kmemcheck_mark_initialized(page_address(page) + offset, size); > BUG_ON(!valid_dma_direction(dir)); > - addr = ops->map_page(dev, page, offset, size, dir, NULL); > + addr = ops->map_page(dev, page, offset, size, dir, 0); > debug_dma_map_page(dev, page, offset, size, dir, addr, false); > > return addr; > @@ -208,7 +231,7 @@ static inline void dma_unmap_page(struct device *dev, dma_addr_t addr, > > BUG_ON(!valid_dma_direction(dir)); > if (ops->unmap_page) > - ops->unmap_page(dev, addr, size, dir, NULL); > + ops->unmap_page(dev, addr, size, dir, 0); > debug_dma_unmap_page(dev, addr, size, dir, false); > } > > @@ -289,10 +312,10 @@ dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, > > } > > -#define dma_map_single(d, a, s, r) dma_map_single_attrs(d, a, s, r, NULL) > -#define dma_unmap_single(d, a, s, r) dma_unmap_single_attrs(d, a, s, r, NULL) > -#define dma_map_sg(d, s, n, r) dma_map_sg_attrs(d, s, n, r, NULL) > -#define dma_unmap_sg(d, s, n, r) dma_unmap_sg_attrs(d, s, n, r, NULL) > +#define dma_map_single(d, a, s, r) dma_map_single_attrs(d, a, s, r, 0) > +#define dma_unmap_single(d, a, s, r) dma_unmap_single_attrs(d, a, s, r, 0) > +#define dma_map_sg(d, s, n, r) dma_map_sg_attrs(d, s, n, r, 0) > +#define dma_unmap_sg(d, s, n, r) dma_unmap_sg_attrs(d, s, n, r, 0) > > extern int dma_common_mmap(struct device *dev, struct vm_area_struct *vma, > void *cpu_addr, dma_addr_t dma_addr, size_t size); > @@ -321,7 +344,7 @@ void dma_common_free_remap(void *cpu_addr, size_t size, unsigned long vm_flags); > */ > static inline int > dma_mmap_attrs(struct device *dev, struct vm_area_struct *vma, void *cpu_addr, > - dma_addr_t dma_addr, size_t size, struct dma_attrs *attrs) > + dma_addr_t dma_addr, size_t size, unsigned long attrs) > { > struct dma_map_ops *ops = get_dma_ops(dev); > BUG_ON(!ops); > @@ -330,7 +353,7 @@ dma_mmap_attrs(struct device *dev, struct vm_area_struct *vma, void *cpu_addr, > return dma_common_mmap(dev, vma, cpu_addr, dma_addr, size); > } > > -#define dma_mmap_coherent(d, v, c, h, s) dma_mmap_attrs(d, v, c, h, s, NULL) > +#define dma_mmap_coherent(d, v, c, h, s) dma_mmap_attrs(d, v, c, h, s, 0) > > int > dma_common_get_sgtable(struct device *dev, struct sg_table *sgt, > @@ -338,7 +361,8 @@ dma_common_get_sgtable(struct device *dev, struct sg_table *sgt, > > static inline int > dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt, void *cpu_addr, > - dma_addr_t dma_addr, size_t size, struct dma_attrs *attrs) > + dma_addr_t dma_addr, size_t size, > + unsigned long attrs) > { > struct dma_map_ops *ops = get_dma_ops(dev); > BUG_ON(!ops); > @@ -348,7 +372,7 @@ dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt, void *cpu_addr, > return dma_common_get_sgtable(dev, sgt, cpu_addr, dma_addr, size); > } > > -#define dma_get_sgtable(d, t, v, h, s) dma_get_sgtable_attrs(d, t, v, h, s, NULL) > +#define dma_get_sgtable(d, t, v, h, s) dma_get_sgtable_attrs(d, t, v, h, s, 0) > > #ifndef arch_dma_alloc_attrs > #define arch_dma_alloc_attrs(dev, flag) (true) > @@ -356,7 +380,7 @@ dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt, void *cpu_addr, > > static inline void *dma_alloc_attrs(struct device *dev, size_t size, > dma_addr_t *dma_handle, gfp_t flag, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct dma_map_ops *ops = get_dma_ops(dev); > void *cpu_addr; > @@ -378,7 +402,7 @@ static inline void *dma_alloc_attrs(struct device *dev, size_t size, > > static inline void dma_free_attrs(struct device *dev, size_t size, > void *cpu_addr, dma_addr_t dma_handle, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > struct dma_map_ops *ops = get_dma_ops(dev); > > @@ -398,31 +422,27 @@ static inline void dma_free_attrs(struct device *dev, size_t size, > static inline void *dma_alloc_coherent(struct device *dev, size_t size, > dma_addr_t *dma_handle, gfp_t flag) > { > - return dma_alloc_attrs(dev, size, dma_handle, flag, NULL); > + return dma_alloc_attrs(dev, size, dma_handle, flag, 0); > } > > static inline void dma_free_coherent(struct device *dev, size_t size, > void *cpu_addr, dma_addr_t dma_handle) > { > - return dma_free_attrs(dev, size, cpu_addr, dma_handle, NULL); > + return dma_free_attrs(dev, size, cpu_addr, dma_handle, 0); > } > > static inline void *dma_alloc_noncoherent(struct device *dev, size_t size, > dma_addr_t *dma_handle, gfp_t gfp) > { > - DEFINE_DMA_ATTRS(attrs); > - > - dma_set_attr(DMA_ATTR_NON_CONSISTENT, &attrs); > - return dma_alloc_attrs(dev, size, dma_handle, gfp, &attrs); > + return dma_alloc_attrs(dev, size, dma_handle, gfp, > + DMA_ATTR_NON_CONSISTENT); > } > > static inline void dma_free_noncoherent(struct device *dev, size_t size, > void *cpu_addr, dma_addr_t dma_handle) > { > - DEFINE_DMA_ATTRS(attrs); > - > - dma_set_attr(DMA_ATTR_NON_CONSISTENT, &attrs); > - dma_free_attrs(dev, size, cpu_addr, dma_handle, &attrs); > + dma_free_attrs(dev, size, cpu_addr, dma_handle, > + DMA_ATTR_NON_CONSISTENT); > } > > static inline int dma_mapping_error(struct device *dev, dma_addr_t dma_addr) > @@ -646,9 +666,8 @@ static inline void dmam_release_declared_memory(struct device *dev) > static inline void *dma_alloc_wc(struct device *dev, size_t size, > dma_addr_t *dma_addr, gfp_t gfp) > { > - DEFINE_DMA_ATTRS(attrs); > - dma_set_attr(DMA_ATTR_WRITE_COMBINE, &attrs); > - return dma_alloc_attrs(dev, size, dma_addr, gfp, &attrs); > + return dma_alloc_attrs(dev, size, dma_addr, gfp, > + DMA_ATTR_WRITE_COMBINE); > } > #ifndef dma_alloc_writecombine > #define dma_alloc_writecombine dma_alloc_wc > @@ -657,9 +676,8 @@ static inline void *dma_alloc_wc(struct device *dev, size_t size, > static inline void dma_free_wc(struct device *dev, size_t size, > void *cpu_addr, dma_addr_t dma_addr) > { > - DEFINE_DMA_ATTRS(attrs); > - dma_set_attr(DMA_ATTR_WRITE_COMBINE, &attrs); > - return dma_free_attrs(dev, size, cpu_addr, dma_addr, &attrs); > + return dma_free_attrs(dev, size, cpu_addr, dma_addr, > + DMA_ATTR_WRITE_COMBINE); > } > #ifndef dma_free_writecombine > #define dma_free_writecombine dma_free_wc > @@ -670,9 +688,8 @@ static inline int dma_mmap_wc(struct device *dev, > void *cpu_addr, dma_addr_t dma_addr, > size_t size) > { > - DEFINE_DMA_ATTRS(attrs); > - dma_set_attr(DMA_ATTR_WRITE_COMBINE, &attrs); > - return dma_mmap_attrs(dev, vma, cpu_addr, dma_addr, size, &attrs); > + return dma_mmap_attrs(dev, vma, cpu_addr, dma_addr, size, > + DMA_ATTR_WRITE_COMBINE); > } > #ifndef dma_mmap_writecombine > #define dma_mmap_writecombine dma_mmap_wc > diff --git a/include/linux/swiotlb.h b/include/linux/swiotlb.h > index 017fced60242..5f81f8a187f2 100644 > --- a/include/linux/swiotlb.h > +++ b/include/linux/swiotlb.h > @@ -6,7 +6,6 @@ > #include <linux/types.h> > > struct device; > -struct dma_attrs; > struct page; > struct scatterlist; > > @@ -68,10 +67,10 @@ swiotlb_free_coherent(struct device *hwdev, size_t size, > extern dma_addr_t swiotlb_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > extern void swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > > extern int > swiotlb_map_sg(struct device *hwdev, struct scatterlist *sg, int nents, > @@ -83,12 +82,13 @@ swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sg, int nents, > > extern int > swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, int nelems, > - enum dma_data_direction dir, struct dma_attrs *attrs); > + enum dma_data_direction dir, > + unsigned long attrs); > > extern void > swiotlb_unmap_sg_attrs(struct device *hwdev, struct scatterlist *sgl, > int nelems, enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > > extern void > swiotlb_sync_single_for_cpu(struct device *hwdev, dma_addr_t dev_addr, > diff --git a/include/xen/swiotlb-xen.h b/include/xen/swiotlb-xen.h > index 8b2eb93ae8ba..7c35e279d1e3 100644 > --- a/include/xen/swiotlb-xen.h > +++ b/include/xen/swiotlb-xen.h > @@ -9,30 +9,30 @@ extern int xen_swiotlb_init(int verbose, bool early); > extern void > *xen_swiotlb_alloc_coherent(struct device *hwdev, size_t size, > dma_addr_t *dma_handle, gfp_t flags, > - struct dma_attrs *attrs); > + unsigned long attrs); > > extern void > xen_swiotlb_free_coherent(struct device *hwdev, size_t size, > void *vaddr, dma_addr_t dma_handle, > - struct dma_attrs *attrs); > + unsigned long attrs); > > extern dma_addr_t xen_swiotlb_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > > extern void xen_swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > extern int > xen_swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, > int nelems, enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > > extern void > xen_swiotlb_unmap_sg_attrs(struct device *hwdev, struct scatterlist *sgl, > int nelems, enum dma_data_direction dir, > - struct dma_attrs *attrs); > + unsigned long attrs); > > extern void > xen_swiotlb_sync_single_for_cpu(struct device *hwdev, dma_addr_t dev_addr, > diff --git a/lib/dma-noop.c b/lib/dma-noop.c > index 72145646857e..3d766e78fbe2 100644 > --- a/lib/dma-noop.c > +++ b/lib/dma-noop.c > @@ -10,7 +10,7 @@ > > static void *dma_noop_alloc(struct device *dev, size_t size, > dma_addr_t *dma_handle, gfp_t gfp, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > void *ret; > > @@ -22,7 +22,7 @@ static void *dma_noop_alloc(struct device *dev, size_t size, > > static void dma_noop_free(struct device *dev, size_t size, > void *cpu_addr, dma_addr_t dma_addr, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > free_pages((unsigned long)cpu_addr, get_order(size)); > } > @@ -30,13 +30,14 @@ static void dma_noop_free(struct device *dev, size_t size, > static dma_addr_t dma_noop_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > return page_to_phys(page) + offset; > } > > static int dma_noop_map_sg(struct device *dev, struct scatterlist *sgl, int nents, > - enum dma_data_direction dir, struct dma_attrs *attrs) > + enum dma_data_direction dir, > + unsigned long attrs) > { > int i; > struct scatterlist *sg; > diff --git a/lib/swiotlb.c b/lib/swiotlb.c > index 76f29ecba8f4..22e13a0e19d7 100644 > --- a/lib/swiotlb.c > +++ b/lib/swiotlb.c > @@ -738,7 +738,7 @@ swiotlb_full(struct device *dev, size_t size, enum dma_data_direction dir, > dma_addr_t swiotlb_map_page(struct device *dev, struct page *page, > unsigned long offset, size_t size, > enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > phys_addr_t map, phys = page_to_phys(page) + offset; > dma_addr_t dev_addr = phys_to_dma(dev, phys); > @@ -807,7 +807,7 @@ static void unmap_single(struct device *hwdev, dma_addr_t dev_addr, > > void swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr, > size_t size, enum dma_data_direction dir, > - struct dma_attrs *attrs) > + unsigned long attrs) > { > unmap_single(hwdev, dev_addr, size, dir); > } > @@ -877,7 +877,7 @@ EXPORT_SYMBOL(swiotlb_sync_single_for_device); > */ > int > swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, int nelems, > - enum dma_data_direction dir, struct dma_attrs *attrs) > + enum dma_data_direction dir, unsigned long attrs) > { > struct scatterlist *sg; > int i; > @@ -914,7 +914,7 @@ int > swiotlb_map_sg(struct device *hwdev, struct scatterlist *sgl, int nelems, > enum dma_data_direction dir) > { > - return swiotlb_map_sg_attrs(hwdev, sgl, nelems, dir, NULL); > + return swiotlb_map_sg_attrs(hwdev, sgl, nelems, dir, 0); > } > EXPORT_SYMBOL(swiotlb_map_sg); > > @@ -924,7 +924,8 @@ EXPORT_SYMBOL(swiotlb_map_sg); > */ > void > swiotlb_unmap_sg_attrs(struct device *hwdev, struct scatterlist *sgl, > - int nelems, enum dma_data_direction dir, struct dma_attrs *attrs) > + int nelems, enum dma_data_direction dir, > + unsigned long attrs) > { > struct scatterlist *sg; > int i; > @@ -941,7 +942,7 @@ void > swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sgl, int nelems, > enum dma_data_direction dir) > { > - return swiotlb_unmap_sg_attrs(hwdev, sgl, nelems, dir, NULL); > + return swiotlb_unmap_sg_attrs(hwdev, sgl, nelems, dir, 0); > } > EXPORT_SYMBOL(swiotlb_unmap_sg); > > -- > 1.9.1 > > _______________________________________________ > iommu mailing list > iommu@xxxxxxxxxxxxxxxxxxxxxxxxxx > https://lists.linuxfoundation.org/mailman/listinfo/iommu -- To unsubscribe from this list: send the line "unsubscribe linux-samsung-soc" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html