On Tue, Jun 13, 2017 at 02:49:47PM -0400, Rob Clark wrote: > It means we have to do a list traversal where we once had an index into > a table. But the list will normally have one or two entries. > > Signed-off-by: Rob Clark <robdclark@xxxxxxxxx> I dig the rename - makes more sense. Acked-by: Jordan Crouse <jcrouse@xxxxxxxxxxxxxx> > --- > drivers/gpu/drm/msm/msm_gem.c | 138 +++++++++++++++++++++++++++++------------- > drivers/gpu/drm/msm/msm_gem.h | 4 +- > 2 files changed, 99 insertions(+), 43 deletions(-) > > diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c > index 754432c..410368f 100644 > --- a/drivers/gpu/drm/msm/msm_gem.c > +++ b/drivers/gpu/drm/msm/msm_gem.c > @@ -283,21 +283,59 @@ uint64_t msm_gem_mmap_offset(struct drm_gem_object *obj) > return offset; > } > > +static struct msm_gem_vma *add_vma(struct drm_gem_object *obj, > + struct msm_gem_address_space *aspace) > +{ > + struct msm_gem_object *msm_obj = to_msm_bo(obj); > + struct msm_gem_vma *vma; > + > + vma = kzalloc(sizeof(*vma), GFP_KERNEL); > + if (!vma) > + return ERR_PTR(-ENOMEM); > + > + vma->aspace = aspace; > + > + list_add_tail(&vma->list, &msm_obj->vmas); > + > + return vma; > +} > + > +static struct msm_gem_vma *lookup_vma(struct drm_gem_object *obj, > + struct msm_gem_address_space *aspace) > +{ > + struct msm_gem_object *msm_obj = to_msm_bo(obj); > + struct msm_gem_vma *vma; > + > + WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex)); > + > + list_for_each_entry(vma, &msm_obj->vmas, list) { > + if (vma->aspace == aspace) > + return vma; > + } > + > + return NULL; > +} > + > +static void del_vma(struct msm_gem_vma *vma) > +{ > + if (!vma) > + return; > + > + list_del(&vma->list); > + kfree(vma); > +} > + > static void > put_iova(struct drm_gem_object *obj) > { > - struct drm_device *dev = obj->dev; > - struct msm_drm_private *priv = obj->dev->dev_private; > struct msm_gem_object *msm_obj = to_msm_bo(obj); > - int id; > + struct msm_gem_vma *vma, *tmp; > > - WARN_ON(!mutex_is_locked(&dev->struct_mutex)); > + WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex)); > > - for (id = 0; id < ARRAY_SIZE(msm_obj->domain); id++) { > - if (!priv->aspace[id]) > - continue; > - msm_gem_unmap_vma(priv->aspace[id], > - &msm_obj->domain[id], msm_obj->sgt); > + list_for_each_entry_safe(vma, tmp, &msm_obj->vmas, list) { > + msm_gem_unmap_vma(vma->aspace, vma, msm_obj->sgt); > + del_vma(vma); > } > } > > @@ -312,24 +350,37 @@ int msm_gem_get_iova_locked(struct drm_gem_object *obj, > struct msm_gem_address_space *aspace, uint64_t *iova) > { > struct msm_gem_object *msm_obj = to_msm_bo(obj); > - int id = aspace ? aspace->id : 0; > + struct msm_gem_vma *vma; > int ret = 0; > > WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex)); > > - if (!msm_obj->domain[id].iova) { > - struct msm_drm_private *priv = obj->dev->dev_private; > - struct page **pages = get_pages(obj); > + vma = lookup_vma(obj, aspace); > > - if (IS_ERR(pages)) > - return PTR_ERR(pages); > + if (!vma) { > + struct page **pages; > + > + vma = add_vma(obj, aspace); > + if (IS_ERR(vma)) > + return PTR_ERR(vma); > + > + pages = get_pages(obj); > + if (IS_ERR(pages)) { > + ret = PTR_ERR(pages); > + goto fail; > + } > > - ret = msm_gem_map_vma(priv->aspace[id], &msm_obj->domain[id], > - msm_obj->sgt, obj->size >> PAGE_SHIFT); > + ret = msm_gem_map_vma(aspace, vma, msm_obj->sgt, > + obj->size >> PAGE_SHIFT); > + if (ret) > + goto fail; > } > > - if (!ret) > - *iova = msm_obj->domain[id].iova; > + *iova = vma->iova; > + return 0; > + > +fail: > + del_vma(vma); > > return ret; > } > @@ -338,22 +389,12 @@ int msm_gem_get_iova_locked(struct drm_gem_object *obj, > int msm_gem_get_iova(struct drm_gem_object *obj, > struct msm_gem_address_space *aspace, uint64_t *iova) > { > - struct msm_gem_object *msm_obj = to_msm_bo(obj); > - int id = aspace ? aspace->id : 0; > int ret; > > - /* this is safe right now because we don't unmap until the > - * bo is deleted: > - */ > - if (msm_obj->domain[id].iova) { > - might_lock(&obj->dev->struct_mutex); > - *iova = msm_obj->domain[id].iova; > - return 0; > - } > - > mutex_lock(&obj->dev->struct_mutex); > ret = msm_gem_get_iova_locked(obj, aspace, iova); > mutex_unlock(&obj->dev->struct_mutex); > + > return ret; > } > > @@ -363,10 +404,14 @@ int msm_gem_get_iova(struct drm_gem_object *obj, > uint64_t msm_gem_iova(struct drm_gem_object *obj, > struct msm_gem_address_space *aspace) > { > - struct msm_gem_object *msm_obj = to_msm_bo(obj); > - int id = aspace ? aspace->id : 0; > - WARN_ON(!msm_obj->domain[id].iova); > - return msm_obj->domain[id].iova; > + struct msm_gem_vma *vma; > + > + mutex_lock(&obj->dev->struct_mutex); > + vma = lookup_vma(obj, aspace); > + mutex_unlock(&obj->dev->struct_mutex); > + WARN_ON(!vma); > + > + return vma ? vma->iova : 0; > } > > void msm_gem_put_iova(struct drm_gem_object *obj, > @@ -624,11 +669,10 @@ void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m) > struct msm_gem_object *msm_obj = to_msm_bo(obj); > struct reservation_object *robj = msm_obj->resv; > struct reservation_object_list *fobj; > - struct msm_drm_private *priv = obj->dev->dev_private; > struct dma_fence *fence; > + struct msm_gem_vma *vma; > uint64_t off = drm_vma_node_start(&obj->vma_node); > const char *madv; > - unsigned id; > > WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex)); > > @@ -650,8 +694,9 @@ void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m) > obj->name, kref_read(&obj->refcount), > off, msm_obj->vaddr); > > - for (id = 0; id < priv->num_aspaces; id++) > - seq_printf(m, " %08llx", msm_obj->domain[id].iova); > + /* FIXME: we need to print the address space here too */ > + list_for_each_entry(vma, &msm_obj->vmas, list) > + seq_printf(m, " %08llx", vma->iova); > > seq_printf(m, " %zu%s\n", obj->size, madv); > > @@ -788,6 +833,8 @@ static int msm_gem_new_impl(struct drm_device *dev, > } > > INIT_LIST_HEAD(&msm_obj->submit_entry); > + INIT_LIST_HEAD(&msm_obj->vmas); > + > list_add_tail(&msm_obj->mm_list, &priv->inactive_list); > > *obj = &msm_obj->base; > @@ -826,19 +873,26 @@ struct drm_gem_object *msm_gem_new(struct drm_device *dev, > goto fail; > > if (use_vram) { > - struct msm_gem_object *msm_obj = to_msm_bo(obj); > + struct msm_gem_vma *vma; > struct page **pages; > > - msm_obj->vram_node = &msm_obj->domain[0].node; > + vma = add_vma(obj, NULL); > + if (IS_ERR(vma)) { > + ret = PTR_ERR(vma); > + goto fail; > + } > + > + to_msm_bo(obj)->vram_node = &vma->node; > + > drm_gem_private_object_init(dev, obj, size); > > - msm_obj->pages = get_pages(obj); > pages = get_pages(obj); > if (IS_ERR(pages)) { > ret = PTR_ERR(pages); > goto fail; > } > - msm_obj->domain[0].iova = physaddr(obj); > + > + vma->iova = physaddr(obj); > } else { > ret = drm_gem_object_init(dev, obj, size); > if (ret) > diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h > index 4b4b352..ff468da 100644 > --- a/drivers/gpu/drm/msm/msm_gem.h > +++ b/drivers/gpu/drm/msm/msm_gem.h > @@ -39,6 +39,8 @@ struct msm_gem_address_space { > struct msm_gem_vma { > struct drm_mm_node node; > uint64_t iova; > + struct msm_gem_address_space *aspace; > + struct list_head list; /* node in msm_gem_object::vmas */ > }; > > struct msm_gem_object { > @@ -78,7 +80,7 @@ struct msm_gem_object { > struct sg_table *sgt; > void *vaddr; > > - struct msm_gem_vma domain[NUM_DOMAINS]; > + struct list_head vmas; /* list of msm_gem_vma */ > > /* normally (resv == &_resv) except for imported bo's */ > struct reservation_object *resv; > -- > 2.9.4 > -- The Qualcomm Innovation Center, Inc. is a member of Code Aurora Forum, a Linux Foundation Collaborative Project _______________________________________________ dri-devel mailing list dri-devel@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/dri-devel