On Tue, Oct 25, 2022 at 1:04 PM Dmitry Baryshkov <dmitry.baryshkov@xxxxxxxxxx> wrote: > > After the msm_iommu instance is created, the IOMMU domain is completely > handled inside the msm_iommu code. Move the iommu_domain_alloc() call > into the msm_iommu_new() to simplify callers code. > > Reported-by: kernel test robot <lkp@xxxxxxxxx> > Signed-off-by: Dmitry Baryshkov <dmitry.baryshkov@xxxxxxxxxx> > --- > drivers/gpu/drm/msm/adreno/a6xx_gmu.c | 12 +++++------- > drivers/gpu/drm/msm/adreno/a6xx_gpu.c | 25 +++++++++--------------- > drivers/gpu/drm/msm/adreno/adreno_gpu.c | 25 +++++++++--------------- > drivers/gpu/drm/msm/adreno/adreno_gpu.h | 2 -- > drivers/gpu/drm/msm/disp/mdp4/mdp4_kms.c | 19 +++++++++--------- > drivers/gpu/drm/msm/msm_drv.c | 18 ++++++++--------- > drivers/gpu/drm/msm/msm_iommu.c | 20 ++++++++++++++++--- > drivers/gpu/drm/msm/msm_mmu.h | 3 ++- > 8 files changed, 60 insertions(+), 64 deletions(-) > > diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c > index e033d6a67a20..6484b97c5344 100644 > --- a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c > +++ b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c > @@ -1213,19 +1213,17 @@ static int a6xx_gmu_memory_alloc(struct a6xx_gmu *gmu, struct a6xx_gmu_bo *bo, > > static int a6xx_gmu_memory_probe(struct a6xx_gmu *gmu) > { > - struct iommu_domain *domain; > struct msm_mmu *mmu; > > - domain = iommu_domain_alloc(&platform_bus_type); > - if (!domain) > + mmu = msm_iommu_new(gmu->dev, 0); > + if (!mmu) > return -ENODEV; > + if (IS_ERR(mmu)) > + return PTR_ERR(mmu); > > - mmu = msm_iommu_new(gmu->dev, domain); > gmu->aspace = msm_gem_address_space_create(mmu, "gmu", 0x0, 0x80000000); > - if (IS_ERR(gmu->aspace)) { > - iommu_domain_free(domain); > + if (IS_ERR(gmu->aspace)) > return PTR_ERR(gmu->aspace); > - } > > return 0; > } > diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c > index fdc578016e0b..7a1b4397b842 100644 > --- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c > +++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c > @@ -1784,37 +1784,30 @@ static void a6xx_gpu_set_freq(struct msm_gpu *gpu, struct dev_pm_opp *opp, > static struct msm_gem_address_space * > a6xx_create_address_space(struct msm_gpu *gpu, struct platform_device *pdev) > { > - struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); > - struct a6xx_gpu *a6xx_gpu = to_a6xx_gpu(adreno_gpu); > - struct iommu_domain *iommu; > struct msm_mmu *mmu; > struct msm_gem_address_space *aspace; > + struct iommu_domain_geometry *geometry; > u64 start, size; > > - iommu = iommu_domain_alloc(&platform_bus_type); > - if (!iommu) > - return NULL; > - > /* > * This allows GPU to set the bus attributes required to use system > * cache on behalf of the iommu page table walker. > */ > - if (!IS_ERR_OR_NULL(a6xx_gpu->htw_llc_slice)) > - adreno_set_llc_attributes(iommu); > - > - mmu = msm_iommu_new(&pdev->dev, iommu); > - if (IS_ERR(mmu)) { > - iommu_domain_free(iommu); > + mmu = msm_iommu_new(&pdev->dev, IO_PGTABLE_QUIRK_ARM_OUTER_WBWA); I think/assume the quirk still needs to be conditional on a6xx_gpu->htw_llc_slice.. or at least I'm not sure what happens if we set it but do not have an LLCC (or allocated slice) BR, -R > + if (IS_ERR_OR_NULL(mmu)) > return ERR_CAST(mmu); > - } > + > + geometry = msm_iommu_get_geometry(mmu); > + if (IS_ERR(geometry)) > + return ERR_CAST(geometry); > > /* > * Use the aperture start or SZ_16M, whichever is greater. This will > * ensure that we align with the allocated pagetable range while still > * allowing room in the lower 32 bits for GMEM and whatnot > */ > - start = max_t(u64, SZ_16M, iommu->geometry.aperture_start); > - size = iommu->geometry.aperture_end - start + 1; > + start = max_t(u64, SZ_16M, geometry->aperture_start); > + size = geometry->aperture_end - start + 1; > > aspace = msm_gem_address_space_create(mmu, "gpu", > start & GENMASK_ULL(48, 0), size); > diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.c b/drivers/gpu/drm/msm/adreno/adreno_gpu.c > index 382fb7f9e497..5808911899c7 100644 > --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.c > +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.c > @@ -191,37 +191,30 @@ int adreno_zap_shader_load(struct msm_gpu *gpu, u32 pasid) > return zap_shader_load_mdt(gpu, adreno_gpu->info->zapfw, pasid); > } > > -void adreno_set_llc_attributes(struct iommu_domain *iommu) > -{ > - iommu_set_pgtable_quirks(iommu, IO_PGTABLE_QUIRK_ARM_OUTER_WBWA); > -} > - > struct msm_gem_address_space * > adreno_iommu_create_address_space(struct msm_gpu *gpu, > struct platform_device *pdev) > { > - struct iommu_domain *iommu; > struct msm_mmu *mmu; > struct msm_gem_address_space *aspace; > + struct iommu_domain_geometry *geometry; > u64 start, size; > > - iommu = iommu_domain_alloc(&platform_bus_type); > - if (!iommu) > - return NULL; > - > - mmu = msm_iommu_new(&pdev->dev, iommu); > - if (IS_ERR(mmu)) { > - iommu_domain_free(iommu); > + mmu = msm_iommu_new(&pdev->dev, 0); > + if (IS_ERR_OR_NULL(mmu)) > return ERR_CAST(mmu); > - } > + > + geometry = msm_iommu_get_geometry(mmu); > + if (IS_ERR(geometry)) > + return ERR_CAST(geometry); > > /* > * Use the aperture start or SZ_16M, whichever is greater. This will > * ensure that we align with the allocated pagetable range while still > * allowing room in the lower 32 bits for GMEM and whatnot > */ > - start = max_t(u64, SZ_16M, iommu->geometry.aperture_start); > - size = iommu->geometry.aperture_end - start + 1; > + start = max_t(u64, SZ_16M, geometry->aperture_start); > + size = geometry->aperture_end - start + 1; > > aspace = msm_gem_address_space_create(mmu, "gpu", > start & GENMASK_ULL(48, 0), size); > diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h b/drivers/gpu/drm/msm/adreno/adreno_gpu.h > index e7adc5c632d0..707273339969 100644 > --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h > +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h > @@ -338,8 +338,6 @@ struct msm_gem_address_space * > adreno_iommu_create_address_space(struct msm_gpu *gpu, > struct platform_device *pdev); > > -void adreno_set_llc_attributes(struct iommu_domain *iommu); > - > int adreno_read_speedbin(struct device *dev, u32 *speedbin); > > /* > diff --git a/drivers/gpu/drm/msm/disp/mdp4/mdp4_kms.c b/drivers/gpu/drm/msm/disp/mdp4/mdp4_kms.c > index 964573d26d26..9a1a0769575d 100644 > --- a/drivers/gpu/drm/msm/disp/mdp4/mdp4_kms.c > +++ b/drivers/gpu/drm/msm/disp/mdp4/mdp4_kms.c > @@ -387,7 +387,7 @@ static int mdp4_kms_init(struct drm_device *dev) > struct msm_drm_private *priv = dev->dev_private; > struct mdp4_kms *mdp4_kms; > struct msm_kms *kms = NULL; > - struct iommu_domain *iommu; > + struct msm_mmu *mmu; > struct msm_gem_address_space *aspace; > int irq, ret; > u32 major, minor; > @@ -499,10 +499,15 @@ static int mdp4_kms_init(struct drm_device *dev) > mdp4_disable(mdp4_kms); > mdelay(16); > > - iommu = iommu_domain_alloc(pdev->dev.bus); > - if (iommu) { > - struct msm_mmu *mmu = msm_iommu_new(&pdev->dev, iommu); > - > + mmu = msm_iommu_new(&pdev->dev, 0); > + if (IS_ERR(mmu)) { > + ret = PTR_ERR(mmu); > + goto fail; > + } else if (!mmu) { > + DRM_DEV_INFO(dev->dev, "no iommu, fallback to phys " > + "contig buffers for scanout\n"); > + aspace = NULL; > + } else { > aspace = msm_gem_address_space_create(mmu, > "mdp4", 0x1000, 0x100000000 - 0x1000); > > @@ -514,10 +519,6 @@ static int mdp4_kms_init(struct drm_device *dev) > } > > kms->aspace = aspace; > - } else { > - DRM_DEV_INFO(dev->dev, "no iommu, fallback to phys " > - "contig buffers for scanout\n"); > - aspace = NULL; > } > > ret = modeset_init(mdp4_kms); > diff --git a/drivers/gpu/drm/msm/msm_drv.c b/drivers/gpu/drm/msm/msm_drv.c > index 28034c21f6bc..be32b4460e94 100644 > --- a/drivers/gpu/drm/msm/msm_drv.c > +++ b/drivers/gpu/drm/msm/msm_drv.c > @@ -276,7 +276,6 @@ static int msm_drm_uninit(struct device *dev) > > struct msm_gem_address_space *msm_kms_init_aspace(struct drm_device *dev) > { > - struct iommu_domain *domain; > struct msm_gem_address_space *aspace; > struct msm_mmu *mmu; > struct device *mdp_dev = dev->dev; > @@ -292,22 +291,21 @@ struct msm_gem_address_space *msm_kms_init_aspace(struct drm_device *dev) > else > iommu_dev = mdss_dev; > > - domain = iommu_domain_alloc(iommu_dev->bus); > - if (!domain) { > + mmu = msm_iommu_new(iommu_dev, 0); > + if (IS_ERR(mmu)) > + return ERR_CAST(mmu); > + > + if (!mmu) { > drm_info(dev, "no IOMMU, fallback to phys contig buffers for scanout\n"); > return NULL; > } > > - mmu = msm_iommu_new(iommu_dev, domain); > - if (IS_ERR(mmu)) { > - iommu_domain_free(domain); > - return ERR_CAST(mmu); > - } > - > aspace = msm_gem_address_space_create(mmu, "mdp_kms", > 0x1000, 0x100000000 - 0x1000); > - if (IS_ERR(aspace)) > + if (IS_ERR(aspace)) { > + dev_err(mdp_dev, "aspace create, error %pe\n", aspace); > mmu->funcs->destroy(mmu); > + } > > return aspace; > } > diff --git a/drivers/gpu/drm/msm/msm_iommu.c b/drivers/gpu/drm/msm/msm_iommu.c > index 5577cea7c009..c2507582ecf3 100644 > --- a/drivers/gpu/drm/msm/msm_iommu.c > +++ b/drivers/gpu/drm/msm/msm_iommu.c > @@ -186,6 +186,13 @@ int msm_iommu_pagetable_params(struct msm_mmu *mmu, > return 0; > } > > +struct iommu_domain_geometry *msm_iommu_get_geometry(struct msm_mmu *mmu) > +{ > + struct msm_iommu *iommu = to_msm_iommu(mmu); > + > + return &iommu->domain->geometry; > +} > + > static const struct msm_mmu_funcs pagetable_funcs = { > .map = msm_iommu_pagetable_map, > .unmap = msm_iommu_pagetable_unmap, > @@ -367,17 +374,23 @@ static const struct msm_mmu_funcs funcs = { > .resume_translation = msm_iommu_resume_translation, > }; > > -struct msm_mmu *msm_iommu_new(struct device *dev, struct iommu_domain *domain) > +struct msm_mmu *msm_iommu_new(struct device *dev, unsigned long quirks) > { > + struct iommu_domain *domain; > struct msm_iommu *iommu; > int ret; > > + domain = iommu_domain_alloc(dev->bus); > if (!domain) > - return ERR_PTR(-ENODEV); > + return NULL; > + > + iommu_set_pgtable_quirks(domain, quirks); > > iommu = kzalloc(sizeof(*iommu), GFP_KERNEL); > - if (!iommu) > + if (!iommu) { > + iommu_domain_free(domain); > return ERR_PTR(-ENOMEM); > + } > > iommu->domain = domain; > msm_mmu_init(&iommu->base, dev, &funcs, MSM_MMU_IOMMU); > @@ -386,6 +399,7 @@ struct msm_mmu *msm_iommu_new(struct device *dev, struct iommu_domain *domain) > > ret = iommu_attach_device(iommu->domain, dev); > if (ret) { > + iommu_domain_free(domain); > kfree(iommu); > return ERR_PTR(ret); > } > diff --git a/drivers/gpu/drm/msm/msm_mmu.h b/drivers/gpu/drm/msm/msm_mmu.h > index de158e1bf765..74cd81e701ff 100644 > --- a/drivers/gpu/drm/msm/msm_mmu.h > +++ b/drivers/gpu/drm/msm/msm_mmu.h > @@ -40,7 +40,7 @@ static inline void msm_mmu_init(struct msm_mmu *mmu, struct device *dev, > mmu->type = type; > } > > -struct msm_mmu *msm_iommu_new(struct device *dev, struct iommu_domain *domain); > +struct msm_mmu *msm_iommu_new(struct device *dev, unsigned long quirks); > struct msm_mmu *msm_gpummu_new(struct device *dev, struct msm_gpu *gpu); > > static inline void msm_mmu_set_fault_handler(struct msm_mmu *mmu, void *arg, > @@ -58,5 +58,6 @@ void msm_gpummu_params(struct msm_mmu *mmu, dma_addr_t *pt_base, > > int msm_iommu_pagetable_params(struct msm_mmu *mmu, phys_addr_t *ttbr, > int *asid); > +struct iommu_domain_geometry *msm_iommu_get_geometry(struct msm_mmu *mmu); > > #endif /* __MSM_MMU_H__ */ > -- > 2.35.1 >