thanks, it's nice to see a2xx getting some attention upstream.. few comments inline.. On Wed, Nov 14, 2018 at 5:28 PM Jonathan Marek <jonathan@xxxxxxxx> wrote: > > A2XX has its own very simple MMU. > > Added a msm_use_mmu() function because we can't rely on iommu_present to > decide to use MMU or not. > > Signed-off-by: Jonathan Marek <jonathan@xxxxxxxx> > --- > drivers/gpu/drm/msm/Makefile | 3 +- > drivers/gpu/drm/msm/adreno/a2xx_gpu.c | 57 +++++++++-- > drivers/gpu/drm/msm/adreno/adreno_device.c | 3 + > drivers/gpu/drm/msm/adreno/adreno_gpu.c | 15 +++ > drivers/gpu/drm/msm/msm_drv.c | 11 +- > drivers/gpu/drm/msm/msm_drv.h | 8 ++ > drivers/gpu/drm/msm/msm_gem.c | 4 +- > drivers/gpu/drm/msm/msm_gem_vma.c | 23 +++++ > drivers/gpu/drm/msm/msm_gpu.c | 26 +++-- > drivers/gpu/drm/msm/msm_gpummu.c | 113 +++++++++++++++++++++ > drivers/gpu/drm/msm/msm_mmu.h | 3 + > 11 files changed, 243 insertions(+), 23 deletions(-) > create mode 100644 drivers/gpu/drm/msm/msm_gpummu.c > > diff --git a/drivers/gpu/drm/msm/Makefile b/drivers/gpu/drm/msm/Makefile > index 0170766393ea..004574bc9bd3 100644 > --- a/drivers/gpu/drm/msm/Makefile > +++ b/drivers/gpu/drm/msm/Makefile > @@ -91,7 +91,8 @@ msm-y := \ > msm_perf.o \ > msm_rd.o \ > msm_ringbuffer.o \ > - msm_submitqueue.o > + msm_submitqueue.o \ > + msm_gpummu.o > > msm-$(CONFIG_DEBUG_FS) += adreno/a5xx_debugfs.o \ > disp/dpu1/dpu_dbg.o > diff --git a/drivers/gpu/drm/msm/adreno/a2xx_gpu.c b/drivers/gpu/drm/msm/adreno/a2xx_gpu.c > index 65b2352408fa..dd669f046389 100644 > --- a/drivers/gpu/drm/msm/adreno/a2xx_gpu.c > +++ b/drivers/gpu/drm/msm/adreno/a2xx_gpu.c > @@ -2,6 +2,8 @@ > /* Copyright (c) 2018 The Linux Foundation. All rights reserved. */ > > #include "a2xx_gpu.h" > +#include "msm_gem.h" > +#include "msm_mmu.h" > > extern bool hang_debug; > > @@ -58,9 +60,12 @@ static bool a2xx_me_init(struct msm_gpu *gpu) > static int a2xx_hw_init(struct msm_gpu *gpu) > { > struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); > + dma_addr_t pt_base, tran_error; > uint32_t *ptr, len; > int i, ret; > > + msm_gpummu_params(gpu->aspace->mmu, &pt_base, &tran_error); > + > DBG("%s", gpu->name); > > gpu_write(gpu, REG_A2XX_RBBM_PM_OVERRIDE1, 0xfffffffe); > @@ -77,9 +82,34 @@ static int a2xx_hw_init(struct msm_gpu *gpu) > /* XXX kgsl uses 0x0000ffff for a20x */ > gpu_write(gpu, REG_A2XX_RBBM_CNTL, 0x00004442); > > - gpu_write(gpu, REG_A2XX_MH_MMU_CONFIG, 0); > - gpu_write(gpu, REG_A2XX_MH_MMU_MPU_BASE, 0); > + /* MPU: physical range */ > + gpu_write(gpu, REG_A2XX_MH_MMU_MPU_BASE, 0x00000000); > gpu_write(gpu, REG_A2XX_MH_MMU_MPU_END, 0xfffff000); > + > + gpu_write(gpu, REG_A2XX_MH_MMU_CONFIG, A2XX_MH_MMU_CONFIG_MMU_ENABLE | > + A2XX_MH_MMU_CONFIG_RB_W_CLNT_BEHAVIOR(BEH_TRAN_RNG) | > + A2XX_MH_MMU_CONFIG_CP_W_CLNT_BEHAVIOR(BEH_TRAN_RNG) | > + A2XX_MH_MMU_CONFIG_CP_R0_CLNT_BEHAVIOR(BEH_TRAN_RNG) | > + A2XX_MH_MMU_CONFIG_CP_R1_CLNT_BEHAVIOR(BEH_TRAN_RNG) | > + A2XX_MH_MMU_CONFIG_CP_R2_CLNT_BEHAVIOR(BEH_TRAN_RNG) | > + A2XX_MH_MMU_CONFIG_CP_R3_CLNT_BEHAVIOR(BEH_TRAN_RNG) | > + A2XX_MH_MMU_CONFIG_CP_R4_CLNT_BEHAVIOR(BEH_TRAN_RNG) | > + A2XX_MH_MMU_CONFIG_VGT_R0_CLNT_BEHAVIOR(BEH_TRAN_RNG) | > + A2XX_MH_MMU_CONFIG_VGT_R1_CLNT_BEHAVIOR(BEH_TRAN_RNG) | > + A2XX_MH_MMU_CONFIG_TC_R_CLNT_BEHAVIOR(BEH_TRAN_RNG) | > + A2XX_MH_MMU_CONFIG_PA_W_CLNT_BEHAVIOR(BEH_TRAN_RNG)); > + > + /* same as parameters in adreno_gpu */ > + gpu_write(gpu, REG_A2XX_MH_MMU_VA_RANGE, SZ_16M | > + A2XX_MH_MMU_VA_RANGE_NUM_64KB_REGIONS(0xfff)); > + > + gpu_write(gpu, REG_A2XX_MH_MMU_PT_BASE, pt_base); > + gpu_write(gpu, REG_A2XX_MH_MMU_TRAN_ERROR, tran_error); > + > + gpu_write(gpu, REG_A2XX_MH_MMU_INVALIDATE, > + A2XX_MH_MMU_INVALIDATE_INVALIDATE_ALL | > + A2XX_MH_MMU_INVALIDATE_INVALIDATE_TC); > + > gpu_write(gpu, REG_A2XX_MH_ARBITER_CONFIG, > A2XX_MH_ARBITER_CONFIG_SAME_PAGE_LIMIT(16) | > A2XX_MH_ARBITER_CONFIG_L1_ARB_ENABLE | > @@ -106,9 +136,21 @@ static int a2xx_hw_init(struct msm_gpu *gpu) > /* XXX gsl doesn't set this */ > gpu_write(gpu, REG_A2XX_RBBM_DEBUG, 0x00080000); > > - gpu_write(gpu, REG_A2XX_RBBM_INT_CNTL, 0); > - gpu_write(gpu, REG_AXXX_CP_INT_CNTL, 0x80000000); /* RB INT */ > + gpu_write(gpu, REG_A2XX_RBBM_INT_CNTL, > + A2XX_RBBM_INT_CNTL_RDERR_INT_MASK); > + gpu_write(gpu, REG_AXXX_CP_INT_CNTL, > + AXXX_CP_INT_CNTL_T0_PACKET_IN_IB_MASK | > + AXXX_CP_INT_CNTL_OPCODE_ERROR_MASK | > + AXXX_CP_INT_CNTL_PROTECTED_MODE_ERROR_MASK | > + AXXX_CP_INT_CNTL_RESERVED_BIT_ERROR_MASK | > + AXXX_CP_INT_CNTL_IB_ERROR_MASK | > + AXXX_CP_INT_CNTL_IB1_INT_MASK | > + AXXX_CP_INT_CNTL_RB_INT_MASK); > gpu_write(gpu, REG_A2XX_SQ_INT_CNTL, 0); > + gpu_write(gpu, REG_A2XX_MH_INTERRUPT_MASK, > + A2XX_MH_INTERRUPT_MASK_AXI_READ_ERROR | > + A2XX_MH_INTERRUPT_MASK_AXI_WRITE_ERROR | > + A2XX_MH_INTERRUPT_MASK_MMU_PAGE_FAULT); > > for (i = 3; i <= 5; i++) > if ((SZ_16K << i) == adreno_gpu->gmem) > @@ -205,7 +247,6 @@ static bool a2xx_idle(struct msm_gpu *gpu) > > static irqreturn_t a2xx_irq(struct msm_gpu *gpu) > { > - could you fold this whitespace change back into the patch that adds this? > uint32_t mstatus, status; > > mstatus = gpu_read(gpu, REG_A2XX_MASTER_INT_SIGNAL); > @@ -216,6 +257,7 @@ static irqreturn_t a2xx_irq(struct msm_gpu *gpu) > dev_warn(gpu->dev->dev, "MH_INT: %08X\n", status); > dev_warn(gpu->dev->dev, "MMU_PAGE_FAULT: %08X\n", > gpu_read(gpu, REG_A2XX_MH_MMU_PAGE_FAULT)); > + > gpu_write(gpu, REG_A2XX_MH_INTERRUPT_CLEAR, status); > } > > @@ -430,11 +472,6 @@ struct msm_gpu *a2xx_gpu_init(struct drm_device *dev) > if (ret) > goto fail; > > - if (!gpu->aspace) { > - /* MMU is not implemented... */ > - dev_err(dev->dev, "No memory protection without MMU\n"); > - } > - hmm, could mmu init still fail? I guess it looks like it does not need to request any additional clks/etc. And hopefully there is nothing out there with a non-functional gpummu? > return gpu; > > fail: > diff --git a/drivers/gpu/drm/msm/adreno/adreno_device.c b/drivers/gpu/drm/msm/adreno/adreno_device.c > index 6e3f78d4da5d..f651cd51449b 100644 > --- a/drivers/gpu/drm/msm/adreno/adreno_device.c > +++ b/drivers/gpu/drm/msm/adreno/adreno_device.c > @@ -307,6 +307,7 @@ static int adreno_bind(struct device *dev, struct device *master, void *data) > static struct adreno_platform_config config = {}; > const struct adreno_info *info; > struct drm_device *drm = dev_get_drvdata(master); > + struct msm_drm_private *priv = drm->dev_private; > struct msm_gpu *gpu; > int ret; > > @@ -337,6 +338,8 @@ static int adreno_bind(struct device *dev, struct device *master, void *data) > > dev_set_drvdata(dev, gpu); > > + priv->is_a2xx = config.rev.core == 2; > + > return 0; > } > > diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.c b/drivers/gpu/drm/msm/adreno/adreno_gpu.c > index e09671324b05..18736637c933 100644 > --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.c > +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.c > @@ -24,6 +24,7 @@ > #include "adreno_gpu.h" > #include "msm_gem.h" > #include "msm_mmu.h" > +#include "a2xx_gpu.h" > > int adreno_get_param(struct msm_gpu *gpu, uint32_t param, uint64_t *value) > { > @@ -290,6 +291,17 @@ void adreno_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit, > struct msm_ringbuffer *ring = submit->ring; > unsigned i; > > + if (adreno_is_a2xx(adreno_gpu)) { > + /* tlb flush for a2xx MMU. TODO: only do this when required */ > + OUT_PKT3(ring, CP_SET_PROTECTED_MODE, 1); > + OUT_RING(ring, 0); > + OUT_PKT0(ring, REG_A2XX_MH_MMU_INVALIDATE, 1); > + OUT_RING(ring, A2XX_MH_MMU_INVALIDATE_INVALIDATE_ALL | > + A2XX_MH_MMU_INVALIDATE_INVALIDATE_TC); > + OUT_PKT3(ring, CP_SET_PROTECTED_MODE, 1); > + OUT_RING(ring, 1); can we only invalidate from the ring? If so, defn keeping a flag about whether inv needed would be a good idea. Userspace tries to use a bo_cache[1] to re-use already mapped buffers, so I'd expect in steady-state this wouldn't be needed on every submit. [1] actually two of them, a separate bo cache is used for things that get vmap'd on kernel side > + } > + > for (i = 0; i < submit->nr_cmds; i++) { > switch (submit->cmd[i].type) { > case MSM_SUBMIT_CMD_IB_TARGET_BUF: > @@ -728,6 +740,9 @@ int adreno_gpu_init(struct drm_device *drm, struct platform_device *pdev, > > adreno_gpu_config.va_start = SZ_16M; > adreno_gpu_config.va_end = 0xffffffff; > + /* maximum range of a2xx mmu */ > + if (adreno_is_a2xx(adreno_gpu)) > + adreno_gpu_config.va_end = SZ_16M + 0xfff * SZ_64K; > > adreno_gpu_config.nr_rings = nr_rings; > > diff --git a/drivers/gpu/drm/msm/msm_drv.c b/drivers/gpu/drm/msm/msm_drv.c > index a10cc0298d38..bda23011494d 100644 > --- a/drivers/gpu/drm/msm/msm_drv.c > +++ b/drivers/gpu/drm/msm/msm_drv.c > @@ -25,6 +25,7 @@ > #include "msm_fence.h" > #include "msm_gpu.h" > #include "msm_kms.h" > +#include "adreno/adreno_gpu.h" > > > /* > @@ -358,6 +359,14 @@ static int get_mdp_ver(struct platform_device *pdev) > > #include <linux/of_address.h> > > +bool msm_use_mmu(struct drm_device *dev) > +{ > + struct msm_drm_private *priv = dev->dev_private; > + > + /* a2xx comes with its own MMU */ > + return priv->is_a2xx || iommu_present(&platform_bus_type); tbh, iommu_present(bus) is not even really sufficient for other devices where some platform devices have iommu and other do not but I guess this is the best we can do at this point :-( > +} > + > static int msm_init_vram(struct drm_device *dev) > { > struct msm_drm_private *priv = dev->dev_private; > @@ -396,7 +405,7 @@ static int msm_init_vram(struct drm_device *dev) > * Grab the entire CMA chunk carved out in early startup in > * mach-msm: > */ > - } else if (!iommu_present(&platform_bus_type)) { > + } else if (!msm_use_mmu(dev)) { > DRM_INFO("using %s VRAM carveout\n", vram); > size = memparse(vram, NULL); > } > diff --git a/drivers/gpu/drm/msm/msm_drv.h b/drivers/gpu/drm/msm/msm_drv.h > index 9d11f321f5a9..99a37962e92b 100644 > --- a/drivers/gpu/drm/msm/msm_drv.h > +++ b/drivers/gpu/drm/msm/msm_drv.h > @@ -179,6 +179,8 @@ struct msm_drm_private { > /* when we have more than one 'msm_gpu' these need to be an array: */ > struct msm_gpu *gpu; > struct msm_file_private *lastctx; > + /* gpu is only set on open(), but we need this info earlier */ > + bool is_a2xx; > > struct drm_fb_helper *fbdev; > > @@ -252,9 +254,15 @@ struct msm_gem_address_space * > msm_gem_address_space_create(struct device *dev, struct iommu_domain *domain, > const char *name); > > +struct msm_gem_address_space * > +msm_gem_address_space_create_a2xx(struct device *dev, struct msm_gpu *gpu, > + const char *name, uint64_t va_start, uint64_t va_end); > + > int msm_register_mmu(struct drm_device *dev, struct msm_mmu *mmu); > void msm_unregister_mmu(struct drm_device *dev, struct msm_mmu *mmu); > > +bool msm_use_mmu(struct drm_device *dev); > + > void msm_gem_submit_free(struct msm_gem_submit *submit); > int msm_ioctl_gem_submit(struct drm_device *dev, void *data, > struct drm_file *file); > diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c > index 6657453a3a58..c1ef8780f686 100644 > --- a/drivers/gpu/drm/msm/msm_gem.c > +++ b/drivers/gpu/drm/msm/msm_gem.c > @@ -912,7 +912,7 @@ static struct drm_gem_object *_msm_gem_new(struct drm_device *dev, > > size = PAGE_ALIGN(size); > > - if (!iommu_present(&platform_bus_type)) > + if (!msm_use_mmu(dev)) > use_vram = true; > else if ((flags & (MSM_BO_STOLEN | MSM_BO_SCANOUT)) && priv->vram.size) > use_vram = true; > @@ -991,7 +991,7 @@ struct drm_gem_object *msm_gem_import(struct drm_device *dev, > int ret, npages; > > /* if we don't have IOMMU, don't bother pretending we can import: */ > - if (!iommu_present(&platform_bus_type)) { > + if (!msm_use_mmu(dev)) { > dev_err(dev->dev, "cannot import without IOMMU\n"); > return ERR_PTR(-EINVAL); > } > diff --git a/drivers/gpu/drm/msm/msm_gem_vma.c b/drivers/gpu/drm/msm/msm_gem_vma.c > index ffbec224551b..8075fd699803 100644 > --- a/drivers/gpu/drm/msm/msm_gem_vma.c > +++ b/drivers/gpu/drm/msm/msm_gem_vma.c > @@ -114,3 +114,26 @@ msm_gem_address_space_create(struct device *dev, struct iommu_domain *domain, > > return aspace; > } > + > +struct msm_gem_address_space * > +msm_gem_address_space_create_a2xx(struct device *dev, struct msm_gpu *gpu, > + const char *name, uint64_t va_start, uint64_t va_end) > +{ > + struct msm_gem_address_space *aspace; > + u64 size = va_end - va_start; > + > + aspace = kzalloc(sizeof(*aspace), GFP_KERNEL); > + if (!aspace) > + return ERR_PTR(-ENOMEM); > + > + spin_lock_init(&aspace->lock); > + aspace->name = name; > + aspace->mmu = msm_gpummu_new(dev, gpu); > + > + drm_mm_init(&aspace->mm, (va_start >> PAGE_SHIFT), > + size >> PAGE_SHIFT); > + > + kref_init(&aspace->kref); > + > + return aspace; > +} > diff --git a/drivers/gpu/drm/msm/msm_gpu.c b/drivers/gpu/drm/msm/msm_gpu.c > index 11aac8337066..9af24a1a70a0 100644 > --- a/drivers/gpu/drm/msm/msm_gpu.c > +++ b/drivers/gpu/drm/msm/msm_gpu.c > @@ -19,6 +19,7 @@ > #include "msm_gem.h" > #include "msm_mmu.h" > #include "msm_fence.h" > +#include "adreno/adreno_gpu.h" > > #include <generated/utsrelease.h> > #include <linux/string_helpers.h> > @@ -800,7 +801,7 @@ static struct msm_gem_address_space * > msm_gpu_create_address_space(struct msm_gpu *gpu, struct platform_device *pdev, > uint64_t va_start, uint64_t va_end) > { > - struct iommu_domain *iommu; > + struct iommu_domain *iommu = NULL; > struct msm_gem_address_space *aspace; > int ret; > > @@ -809,20 +810,27 @@ msm_gpu_create_address_space(struct msm_gpu *gpu, struct platform_device *pdev, > * and have separate page tables per context. For now, to keep things > * simple and to get something working, just use a single address space: > */ > - iommu = iommu_domain_alloc(&platform_bus_type); > - if (!iommu) > - return NULL; > + if (!adreno_is_a2xx(to_adreno_gpu(gpu))) { > + iommu = iommu_domain_alloc(&platform_bus_type); > + if (!iommu) > + return NULL; > > - iommu->geometry.aperture_start = va_start; > - iommu->geometry.aperture_end = va_end; > + iommu->geometry.aperture_start = va_start; > + iommu->geometry.aperture_end = va_end; > > - dev_info(gpu->dev->dev, "%s: using IOMMU\n", gpu->name); > + dev_info(gpu->dev->dev, "%s: using IOMMU\n", gpu->name); > + > + aspace = msm_gem_address_space_create(&pdev->dev, iommu, "gpu"); > + } else { > + aspace = msm_gem_address_space_create_a2xx(&pdev->dev, gpu, "gpu", > + va_start, va_end); > + } > > - aspace = msm_gem_address_space_create(&pdev->dev, iommu, "gpu"); > if (IS_ERR(aspace)) { > dev_err(gpu->dev->dev, "failed to init iommu: %ld\n", > PTR_ERR(aspace)); > - iommu_domain_free(iommu); > + if (iommu) > + iommu_domain_free(iommu); minor nit, maybe just move iommu and the iommu_domain_free() into the if(!a2xx) case? > return ERR_CAST(aspace); > } > > diff --git a/drivers/gpu/drm/msm/msm_gpummu.c b/drivers/gpu/drm/msm/msm_gpummu.c > new file mode 100644 > index 000000000000..f14cf7759c7f > --- /dev/null > +++ b/drivers/gpu/drm/msm/msm_gpummu.c > @@ -0,0 +1,113 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* Copyright (c) 2018 The Linux Foundation. All rights reserved. */ > + > +#include "msm_drv.h" > +#include "msm_mmu.h" > + > +struct msm_gpummu { > + struct msm_mmu base; > + struct msm_gpu *gpu; > + dma_addr_t pt_base; > + uint32_t *table; > +}; > +#define to_msm_gpummu(x) container_of(x, struct msm_gpummu, base) > + > +#define VA_START SZ_16M > +#define VA_RANGE (0xfff * SZ_64K) > +#define MMU_PAGE_SIZE SZ_4K > +#define TABLE_SIZE (sizeof(uint32_t) * VA_RANGE / MMU_PAGE_SIZE) > + > +static int msm_gpummu_attach(struct msm_mmu *mmu, const char * const *names, > + int cnt) > +{ > + return 0; > +} > + > +static void msm_gpummu_detach(struct msm_mmu *mmu, const char * const *names, > + int cnt) > +{ > +} > + > +static void update_pt(struct msm_mmu *mmu, uint64_t iova, > + struct sg_table *sgt, unsigned len, unsigned prot) > +{ > + struct msm_gpummu *gpummu = to_msm_gpummu(mmu); > + unsigned idx = (iova - VA_START) / MMU_PAGE_SIZE; > + struct scatterlist *sg; > + unsigned i, j; > + > + for_each_sg(sgt->sgl, sg, sgt->nents, i) { > + dma_addr_t addr = sg->dma_address; > + for (j = 0; j < sg->length / MMU_PAGE_SIZE; j++, idx++) { > + gpummu->table[idx] = prot ? addr | prot : 0; > + addr += MMU_PAGE_SIZE; > + } > + } > +} > + > +static int msm_gpummu_map(struct msm_mmu *mmu, uint64_t iova, > + struct sg_table *sgt, unsigned len, int prot) > +{ > + unsigned prot_bits = 0; > + if (prot & IOMMU_WRITE) > + prot_bits |= 1; > + if (prot & IOMMU_READ) > + prot_bits |= 2; > + update_pt(mmu, iova, sgt, len, prot_bits); > + return 0; > +} > + > +static int msm_gpummu_unmap(struct msm_mmu *mmu, uint64_t iova, > + struct sg_table *sgt, unsigned len) > +{ > + update_pt(mmu, iova, sgt, len, 0); so one thing I'm a bit concerned about is tlb flush on unmap not happening until the next submit, while meanwhile we've returned the buffer's pages to the OS. If we can trigger tlb inv from CPU, then it is probably a good idea to do so for unmap. (But map is fine to defer.) If we can't tlb inv from the CPU then maybe we need to keep a list of bo's to defer freeing the pages. Otherwise this looks fine to me. BR, -R Side note: I have kinda wanted to do something similar for a while for the IOMMU case, and teach the IOMMU about deferred TLB flush.. since the common usage pattern is to map N bo's that we haven't used before in the submit ioctl, and unmap/free M buffers from retire_worker. For a2xx and older iommu's (which had to inv on map) batching the mapping and doing a single inv seems like a win. And for everyone keeping a list of "zombie" bo's which are unmapped but waiting to be free'd until tlb inv, could let us back multiple unmaps and do a single tlb inv. > + return 0; > +} > + > +static void msm_gpummu_destroy(struct msm_mmu *mmu) > +{ > + struct msm_gpummu *gpummu = to_msm_gpummu(mmu); > + > + dma_free_attrs(mmu->dev, TABLE_SIZE, gpummu->table, gpummu->pt_base, > + DMA_ATTR_FORCE_CONTIGUOUS); > + > + kfree(gpummu); > +} > + > +static const struct msm_mmu_funcs funcs = { > + .attach = msm_gpummu_attach, > + .detach = msm_gpummu_detach, > + .map = msm_gpummu_map, > + .unmap = msm_gpummu_unmap, > + .destroy = msm_gpummu_destroy, > +}; > + > +struct msm_mmu *msm_gpummu_new(struct device *dev, struct msm_gpu *gpu) > +{ > + struct msm_gpummu *gpummu; > + > + gpummu = kzalloc(sizeof(*gpummu), GFP_KERNEL); > + if (!gpummu) > + return ERR_PTR(-ENOMEM); > + > + gpummu->table = dma_alloc_attrs(dev, TABLE_SIZE + 32, &gpummu->pt_base, > + GFP_KERNEL | __GFP_ZERO, DMA_ATTR_FORCE_CONTIGUOUS); > + if (!gpummu->table) { > + kfree(gpummu); > + return ERR_PTR(-ENOMEM); > + } > + > + gpummu->gpu = gpu; > + msm_mmu_init(&gpummu->base, dev, &funcs); > + > + return &gpummu->base; > +} > + > +void msm_gpummu_params(struct msm_mmu *mmu, dma_addr_t *pt_base, > + dma_addr_t *tran_error) > +{ > + dma_addr_t base = to_msm_gpummu(mmu)->pt_base; > + > + *pt_base = base; > + *tran_error = base + TABLE_SIZE; /* 32-byte aligned */ > +} > diff --git a/drivers/gpu/drm/msm/msm_mmu.h b/drivers/gpu/drm/msm/msm_mmu.h > index aa2c5d4580c8..c5bbb49dc1be 100644 > --- a/drivers/gpu/drm/msm/msm_mmu.h > +++ b/drivers/gpu/drm/msm/msm_mmu.h > @@ -54,4 +54,7 @@ static inline void msm_mmu_set_fault_handler(struct msm_mmu *mmu, void *arg, > mmu->handler = handler; > } > > +void msm_gpummu_params(struct msm_mmu *mmu, dma_addr_t *pt_base, > + dma_addr_t *tran_error); > + > #endif /* __MSM_MMU_H__ */ > -- > 2.17.1 > _______________________________________________ dri-devel mailing list dri-devel@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/dri-devel