On 2016?06?24? 10:13, Shunqian Zheng wrote: > From: Tomasz Figa <tfiga at chromium.org> > > The API is not suitable for subsystems consisting of multiple devices > and requires severe hacks to use it. To mitigate this, this patch > implements allocation and address space management locally by using > helpers provided by DRM framework, like other DRM drivers do, e.g. > Tegra. > > This patch should not introduce any functional changes until the driver > is made to attach subdevices into an IOMMU domain with the generic IOMMU > API, which will happen in following patch. Based heavily on GEM > implementation of Tegra DRM driver. Acked-by: Mark Yao <mark.yao at rock-chips.com> > > Signed-off-by: Tomasz Figa <tfiga at chromium.org> > Signed-off-by: Shunqian Zheng <zhengsq at rock-chips.com> > --- > drivers/gpu/drm/rockchip/rockchip_drm_drv.h | 3 + > drivers/gpu/drm/rockchip/rockchip_drm_gem.c | 221 ++++++++++++++++++++++++++-- > drivers/gpu/drm/rockchip/rockchip_drm_gem.h | 9 ++ > 3 files changed, 222 insertions(+), 11 deletions(-) > > diff --git a/drivers/gpu/drm/rockchip/rockchip_drm_drv.h b/drivers/gpu/drm/rockchip/rockchip_drm_drv.h > index ea39329..5ab1223 100644 > --- a/drivers/gpu/drm/rockchip/rockchip_drm_drv.h > +++ b/drivers/gpu/drm/rockchip/rockchip_drm_drv.h > @@ -30,6 +30,7 @@ > > struct drm_device; > struct drm_connector; > +struct iommu_domain; > > /* > * Rockchip drm private crtc funcs. > @@ -61,6 +62,8 @@ struct rockchip_drm_private { > struct drm_gem_object *fbdev_bo; > const struct rockchip_crtc_funcs *crtc_funcs[ROCKCHIP_MAX_CRTC]; > struct drm_atomic_state *state; > + struct iommu_domain *domain; > + struct drm_mm mm; > }; > > int rockchip_register_crtc_funcs(struct drm_crtc *crtc, > diff --git a/drivers/gpu/drm/rockchip/rockchip_drm_gem.c b/drivers/gpu/drm/rockchip/rockchip_drm_gem.c > index 394f92b..e7cd93d 100644 > --- a/drivers/gpu/drm/rockchip/rockchip_drm_gem.c > +++ b/drivers/gpu/drm/rockchip/rockchip_drm_gem.c > @@ -19,11 +19,135 @@ > #include <drm/rockchip_drm.h> > > #include <linux/dma-attrs.h> > +#include <linux/iommu.h> > > #include "rockchip_drm_drv.h" > #include "rockchip_drm_gem.h" > > -static int rockchip_gem_alloc_buf(struct rockchip_gem_object *rk_obj, > +static int rockchip_gem_iommu_map(struct rockchip_gem_object *rk_obj) > +{ > + struct drm_device *drm = rk_obj->base.dev; > + struct rockchip_drm_private *private = drm->dev_private; > + int prot = IOMMU_READ | IOMMU_WRITE; > + ssize_t ret; > + > + ret = drm_mm_insert_node_generic(&private->mm, &rk_obj->mm, > + rk_obj->base.size, PAGE_SIZE, > + 0, 0, 0); > + if (ret < 0) { > + DRM_ERROR("out of I/O virtual memory: %zd\n", ret); > + return ret; > + } > + > + rk_obj->dma_addr = rk_obj->mm.start; > + > + ret = iommu_map_sg(private->domain, rk_obj->dma_addr, rk_obj->sgt->sgl, > + rk_obj->sgt->nents, prot); > + if (ret < 0) { > + DRM_ERROR("failed to map buffer: %zd\n", ret); > + goto err_remove_node; > + } > + > + rk_obj->size = ret; > + > + return 0; > + > +err_remove_node: > + drm_mm_remove_node(&rk_obj->mm); > + > + return ret; > +} > + > +static int rockchip_gem_iommu_unmap(struct rockchip_gem_object *rk_obj) > +{ > + struct drm_device *drm = rk_obj->base.dev; > + struct rockchip_drm_private *private = drm->dev_private; > + > + iommu_unmap(private->domain, rk_obj->dma_addr, rk_obj->size); > + drm_mm_remove_node(&rk_obj->mm); > + > + return 0; > +} > + > +static int rockchip_gem_get_pages(struct rockchip_gem_object *rk_obj) > +{ > + struct drm_device *drm = rk_obj->base.dev; > + int ret, i; > + struct scatterlist *s; > + > + rk_obj->pages = drm_gem_get_pages(&rk_obj->base); > + if (IS_ERR(rk_obj->pages)) > + return PTR_ERR(rk_obj->pages); > + > + rk_obj->num_pages = rk_obj->base.size >> PAGE_SHIFT; > + > + rk_obj->sgt = drm_prime_pages_to_sg(rk_obj->pages, rk_obj->num_pages); > + if (IS_ERR(rk_obj->sgt)) { > + ret = PTR_ERR(rk_obj->sgt); > + goto err_put_pages; > + } > + > + /* > + * Fake up the SG table so that dma_sync_sg_for_device() can be used > + * to flush the pages associated with it. > + * > + * TODO: Replace this by drm_clflush_sg() once it can be implemented > + * without relying on symbols that are not exported. > + */ > + for_each_sg(rk_obj->sgt->sgl, s, rk_obj->sgt->nents, i) > + sg_dma_address(s) = sg_phys(s); > + > + dma_sync_sg_for_device(drm->dev, rk_obj->sgt->sgl, rk_obj->sgt->nents, > + DMA_TO_DEVICE); > + > + return 0; > + > +err_put_pages: > + drm_gem_put_pages(&rk_obj->base, rk_obj->pages, false, false); > + return ret; > +} > + > +static void rockchip_gem_put_pages(struct rockchip_gem_object *rk_obj) > +{ > + sg_free_table(rk_obj->sgt); > + kfree(rk_obj->sgt); > + drm_gem_put_pages(&rk_obj->base, rk_obj->pages, false, false); > +} > + > +static int rockchip_gem_alloc_iommu(struct rockchip_gem_object *rk_obj, > + bool alloc_kmap) > +{ > + int ret; > + > + ret = rockchip_gem_get_pages(rk_obj); > + if (ret < 0) > + return ret; > + > + ret = rockchip_gem_iommu_map(rk_obj); > + if (ret < 0) > + goto err_free; > + > + if (alloc_kmap) { > + rk_obj->kvaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP, > + pgprot_writecombine(PAGE_KERNEL)); > + if (!rk_obj->kvaddr) { > + DRM_ERROR("failed to vmap() buffer\n"); > + ret = -ENOMEM; > + goto err_unmap; > + } > + } > + > + return 0; > + > +err_unmap: > + rockchip_gem_iommu_unmap(rk_obj); > +err_free: > + rockchip_gem_put_pages(rk_obj); > + > + return ret; > +} > + > +static int rockchip_gem_alloc_dma(struct rockchip_gem_object *rk_obj, > bool alloc_kmap) > { > struct drm_gem_object *obj = &rk_obj->base; > @@ -46,32 +170,93 @@ static int rockchip_gem_alloc_buf(struct rockchip_gem_object *rk_obj, > return 0; > } > > -static void rockchip_gem_free_buf(struct rockchip_gem_object *rk_obj) > +static int rockchip_gem_alloc_buf(struct rockchip_gem_object *rk_obj, > + bool alloc_kmap) > { > struct drm_gem_object *obj = &rk_obj->base; > struct drm_device *drm = obj->dev; > + struct rockchip_drm_private *private = drm->dev_private; > > - dma_free_attrs(drm->dev, obj->size, rk_obj->kvaddr, rk_obj->dma_addr, > - &rk_obj->dma_attrs); > + if (private->domain) > + return rockchip_gem_alloc_iommu(rk_obj, alloc_kmap); > + else > + return rockchip_gem_alloc_dma(rk_obj, alloc_kmap); > } > > -static int rockchip_drm_gem_object_mmap(struct drm_gem_object *obj, > - struct vm_area_struct *vma) > +static void rockchip_gem_free_iommu(struct rockchip_gem_object *rk_obj) > +{ > + vunmap(rk_obj->kvaddr); > + rockchip_gem_iommu_unmap(rk_obj); > + rockchip_gem_put_pages(rk_obj); > +} > > +static void rockchip_gem_free_dma(struct rockchip_gem_object *rk_obj) > { > + struct drm_gem_object *obj = &rk_obj->base; > + struct drm_device *drm = obj->dev; > + > + dma_free_attrs(drm->dev, obj->size, rk_obj->kvaddr, > + rk_obj->dma_addr, &rk_obj->dma_attrs); > +} > + > +static void rockchip_gem_free_buf(struct rockchip_gem_object *rk_obj) > +{ > + if (rk_obj->pages) > + rockchip_gem_free_iommu(rk_obj); > + else > + rockchip_gem_free_dma(rk_obj); > +} > + > +static int rockchip_drm_gem_object_mmap_iommu(struct drm_gem_object *obj, > + struct vm_area_struct *vma) > +{ > + struct rockchip_gem_object *rk_obj = to_rockchip_obj(obj); > + unsigned int i, count = obj->size >> PAGE_SHIFT; > + unsigned long user_count = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; > + unsigned long uaddr = vma->vm_start; > int ret; > + > + if (user_count == 0 || user_count > count) > + return -ENXIO; > + > + for (i = 0; i < user_count; i++) { > + ret = vm_insert_page(vma, uaddr, rk_obj->pages[i]); > + if (ret) > + return ret; > + uaddr += PAGE_SIZE; > + } > + > + return 0; > +} > + > +static int rockchip_drm_gem_object_mmap_dma(struct drm_gem_object *obj, > + struct vm_area_struct *vma) > +{ > struct rockchip_gem_object *rk_obj = to_rockchip_obj(obj); > struct drm_device *drm = obj->dev; > > + return dma_mmap_attrs(drm->dev, vma, rk_obj->kvaddr, rk_obj->dma_addr, > + obj->size, &rk_obj->dma_attrs); > +} > + > +static int rockchip_drm_gem_object_mmap(struct drm_gem_object *obj, > + struct vm_area_struct *vma) > +{ > + int ret; > + struct rockchip_gem_object *rk_obj = to_rockchip_obj(obj); > + > /* > - * dma_alloc_attrs() allocated a struct page table for rk_obj, so clear > + * We allocated a struct page table for rk_obj, so clear > * VM_PFNMAP flag that was set by drm_gem_mmap_obj()/drm_gem_mmap(). > */ > vma->vm_flags &= ~VM_PFNMAP; > vma->vm_pgoff = 0; > > - ret = dma_mmap_attrs(drm->dev, vma, rk_obj->kvaddr, rk_obj->dma_addr, > - obj->size, &rk_obj->dma_attrs); > + if (rk_obj->pages) > + ret = rockchip_drm_gem_object_mmap_iommu(obj, vma); > + else > + ret = rockchip_drm_gem_object_mmap_dma(obj, vma); > + > if (ret) > drm_gem_vm_close(vma); > > @@ -121,7 +306,7 @@ struct rockchip_gem_object * > > obj = &rk_obj->base; > > - drm_gem_private_object_init(drm, obj, size); > + drm_gem_object_init(drm, obj, size); > > ret = rockchip_gem_alloc_buf(rk_obj, alloc_kmap); > if (ret) > @@ -277,6 +462,9 @@ struct sg_table *rockchip_gem_prime_get_sg_table(struct drm_gem_object *obj) > struct sg_table *sgt; > int ret; > > + if (rk_obj->pages) > + return drm_prime_pages_to_sg(rk_obj->pages, rk_obj->num_pages); > + > sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); > if (!sgt) > return ERR_PTR(-ENOMEM); > @@ -297,6 +485,10 @@ void *rockchip_gem_prime_vmap(struct drm_gem_object *obj) > { > struct rockchip_gem_object *rk_obj = to_rockchip_obj(obj); > > + if (rk_obj->pages) > + return vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP, > + pgprot_writecombine(PAGE_KERNEL)); > + > if (dma_get_attr(DMA_ATTR_NO_KERNEL_MAPPING, &rk_obj->dma_attrs)) > return NULL; > > @@ -305,5 +497,12 @@ void *rockchip_gem_prime_vmap(struct drm_gem_object *obj) > > void rockchip_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr) > { > - /* Nothing to do */ > + struct rockchip_gem_object *rk_obj = to_rockchip_obj(obj); > + > + if (rk_obj->pages) { > + vunmap(vaddr); > + return; > + } > + > + /* Nothing to do if allocated by DMA mapping API. */ > } > diff --git a/drivers/gpu/drm/rockchip/rockchip_drm_gem.h b/drivers/gpu/drm/rockchip/rockchip_drm_gem.h > index 3584b94..6aa61b2 100644 > --- a/drivers/gpu/drm/rockchip/rockchip_drm_gem.h > +++ b/drivers/gpu/drm/rockchip/rockchip_drm_gem.h > @@ -23,7 +23,16 @@ struct rockchip_gem_object { > > void *kvaddr; > dma_addr_t dma_addr; > + > + /* Used when IOMMU is disabled */ > struct dma_attrs dma_attrs; > + > + /* Used when IOMMU is enabled */ > + struct drm_mm_node mm; > + unsigned long num_pages; > + struct page **pages; > + struct sg_table *sgt; > + size_t size; > }; > > struct sg_table *rockchip_gem_prime_get_sg_table(struct drm_gem_object *obj); -- ?ark Yao