Adding a pile of people since there's lots of discussoins going on around this. -Daniel On Wed, Aug 7, 2019 at 10:50 PM Daniel Vetter <daniel@xxxxxxxx> wrote: > > On Fri, Jun 14, 2019 at 08:10:20AM +0100, Chris Wilson wrote: > > Signed-off-by: Chris Wilson <chris@xxxxxxxxxxxxxxxxxx> > > Not sure this works, 2 thoughts way down ... > > > --- > > drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c | 27 +-- > > .../gpu/drm/i915/gem/i915_gem_execbuffer.c | 3 - > > drivers/gpu/drm/i915/gem/i915_gem_internal.c | 30 +-- > > .../gpu/drm/i915/gem/i915_gem_object_types.h | 11 +- > > drivers/gpu/drm/i915/gem/i915_gem_pages.c | 137 ++++++++++- > > drivers/gpu/drm/i915/gem/i915_gem_phys.c | 33 ++- > > drivers/gpu/drm/i915/gem/i915_gem_shmem.c | 30 +-- > > drivers/gpu/drm/i915/gem/i915_gem_stolen.c | 25 +- > > drivers/gpu/drm/i915/gem/i915_gem_userptr.c | 221 ++++-------------- > > .../drm/i915/gem/selftests/huge_gem_object.c | 17 +- > > .../gpu/drm/i915/gem/selftests/huge_pages.c | 45 ++-- > > drivers/gpu/drm/i915/gvt/dmabuf.c | 17 +- > > drivers/gpu/drm/i915/i915_drv.h | 9 +- > > drivers/gpu/drm/i915/i915_gem.c | 5 +- > > drivers/gpu/drm/i915/selftests/i915_gem_gtt.c | 14 +- > > 15 files changed, 300 insertions(+), 324 deletions(-) > > > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c b/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c > > index 84992d590da5..a44d6d2ef7ed 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_dmabuf.c > > @@ -225,33 +225,24 @@ struct dma_buf *i915_gem_prime_export(struct drm_device *dev, > > return drm_gem_dmabuf_export(dev, &exp_info); > > } > > > > -static int i915_gem_object_get_pages_dmabuf(struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +dmabuf_get_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > - struct sg_table *pages; > > - unsigned int sg_page_sizes; > > - > > - pages = dma_buf_map_attachment(obj->base.import_attach, > > - DMA_BIDIRECTIONAL); > > - if (IS_ERR(pages)) > > - return PTR_ERR(pages); > > - > > - sg_page_sizes = i915_sg_page_sizes(pages->sgl); > > - > > - __i915_gem_object_set_pages(obj, pages, sg_page_sizes); > > - > > - return 0; > > + return dma_buf_map_attachment(ctx->object->base.import_attach, > > + DMA_BIDIRECTIONAL); > > } > > > > -static void i915_gem_object_put_pages_dmabuf(struct drm_i915_gem_object *obj, > > - struct sg_table *pages) > > +static void dmabuf_put_pages(struct drm_i915_gem_object *obj, > > + struct sg_table *pages) > > { > > dma_buf_unmap_attachment(obj->base.import_attach, pages, > > DMA_BIDIRECTIONAL); > > } > > > > static const struct drm_i915_gem_object_ops i915_gem_object_dmabuf_ops = { > > - .get_pages = i915_gem_object_get_pages_dmabuf, > > - .put_pages = i915_gem_object_put_pages_dmabuf, > > + .get_pages = dmabuf_get_pages, > > + .put_pages = dmabuf_put_pages, > > }; > > > > struct drm_gem_object *i915_gem_prime_import(struct drm_device *dev, > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > > index 44bcb681c168..68faf1a71c97 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > > @@ -1784,9 +1784,6 @@ static noinline int eb_relocate_slow(struct i915_execbuffer *eb) > > goto out; > > } > > > > - /* A frequent cause for EAGAIN are currently unavailable client pages */ > > - flush_workqueue(eb->i915->mm.userptr_wq); > > - > > err = i915_mutex_lock_interruptible(dev); > > if (err) { > > mutex_lock(&dev->struct_mutex); > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_internal.c b/drivers/gpu/drm/i915/gem/i915_gem_internal.c > > index 0c41e04ab8fa..aa0bd5de313b 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_internal.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_internal.c > > @@ -32,11 +32,14 @@ static void internal_free_pages(struct sg_table *st) > > kfree(st); > > } > > > > -static int i915_gem_object_get_pages_internal(struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +internal_get_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > + struct drm_i915_gem_object *obj = ctx->object; > > struct drm_i915_private *i915 = to_i915(obj->base.dev); > > - struct sg_table *st; > > struct scatterlist *sg; > > + struct sg_table *st; > > unsigned int sg_page_sizes; > > unsigned int npages; > > int max_order; > > @@ -66,12 +69,12 @@ static int i915_gem_object_get_pages_internal(struct drm_i915_gem_object *obj) > > create_st: > > st = kmalloc(sizeof(*st), GFP_KERNEL); > > if (!st) > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > > > npages = obj->base.size / PAGE_SIZE; > > if (sg_alloc_table(st, npages, GFP_KERNEL)) { > > kfree(st); > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > } > > > > sg = st->sgl; > > @@ -117,27 +120,26 @@ static int i915_gem_object_get_pages_internal(struct drm_i915_gem_object *obj) > > goto err; > > } > > > > - /* Mark the pages as dontneed whilst they are still pinned. As soon > > + /* > > + * Mark the pages as dontneed whilst they are still pinned. As soon > > * as they are unpinned they are allowed to be reaped by the shrinker, > > * and the caller is expected to repopulate - the contents of this > > * object are only valid whilst active and pinned. > > */ > > obj->mm.madv = I915_MADV_DONTNEED; > > > > - __i915_gem_object_set_pages(obj, st, sg_page_sizes); > > - > > - return 0; > > + *sizes = sg_page_sizes; > > + return st; > > > > err: > > sg_set_page(sg, NULL, 0, 0); > > sg_mark_end(sg); > > internal_free_pages(st); > > - > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > } > > > > -static void i915_gem_object_put_pages_internal(struct drm_i915_gem_object *obj, > > - struct sg_table *pages) > > +static void internal_put_pages(struct drm_i915_gem_object *obj, > > + struct sg_table *pages) > > { > > i915_gem_gtt_finish_pages(obj, pages); > > internal_free_pages(pages); > > @@ -149,8 +151,8 @@ static void i915_gem_object_put_pages_internal(struct drm_i915_gem_object *obj, > > static const struct drm_i915_gem_object_ops i915_gem_object_internal_ops = { > > .flags = I915_GEM_OBJECT_HAS_STRUCT_PAGE | > > I915_GEM_OBJECT_IS_SHRINKABLE, > > - .get_pages = i915_gem_object_get_pages_internal, > > - .put_pages = i915_gem_object_put_pages_internal, > > + .get_pages = internal_get_pages, > > + .put_pages = internal_put_pages, > > }; > > > > /** > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > > index f792953b8a71..0ea404cfbc1c 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > > @@ -19,6 +19,7 @@ > > > > struct drm_i915_gem_object; > > struct intel_fronbuffer; > > +struct task_struct; > > > > /* > > * struct i915_lut_handle tracks the fast lookups from handle to vma used > > @@ -32,6 +33,11 @@ struct i915_lut_handle { > > u32 handle; > > }; > > > > +struct i915_gem_object_get_pages_context { > > + struct drm_i915_gem_object *object; > > + struct task_struct *task; > > +}; > > + > > struct drm_i915_gem_object_ops { > > unsigned int flags; > > #define I915_GEM_OBJECT_HAS_STRUCT_PAGE BIT(0) > > @@ -52,9 +58,11 @@ struct drm_i915_gem_object_ops { > > * being released or under memory pressure (where we attempt to > > * reap pages for the shrinker). > > */ > > - int (*get_pages)(struct drm_i915_gem_object *obj); > > + struct sg_table *(*get_pages)(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes); > > void (*put_pages)(struct drm_i915_gem_object *obj, > > struct sg_table *pages); > > + > > void (*truncate)(struct drm_i915_gem_object *obj); > > void (*writeback)(struct drm_i915_gem_object *obj); > > > > @@ -252,7 +260,6 @@ struct drm_i915_gem_object { > > > > struct i915_mm_struct *mm; > > struct i915_mmu_object *mmu_object; > > - struct work_struct *work; > > } userptr; > > > > unsigned long scratch; > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_pages.c b/drivers/gpu/drm/i915/gem/i915_gem_pages.c > > index 6bec301cee79..f65a983248c6 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_pages.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_pages.c > > @@ -8,16 +8,49 @@ > > #include "i915_gem_object.h" > > #include "i915_scatterlist.h" > > > > -void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj, > > - struct sg_table *pages, > > - unsigned int sg_page_sizes) > > +static DEFINE_SPINLOCK(fence_lock); > > + > > +struct get_pages_work { > > + struct dma_fence dma; /* Must be first for dma_fence_free() */ > > + struct i915_sw_fence wait; > > + struct work_struct work; > > + struct i915_gem_object_get_pages_context ctx; > > +}; > > + > > +static const char *get_pages_work_driver_name(struct dma_fence *fence) > > +{ > > + return DRIVER_NAME; > > +} > > + > > +static const char *get_pages_work_timeline_name(struct dma_fence *fence) > > +{ > > + return "allocation"; > > +} > > + > > +static void get_pages_work_release(struct dma_fence *fence) > > +{ > > + struct get_pages_work *w = container_of(fence, typeof(*w), dma); > > + > > + i915_sw_fence_fini(&w->wait); > > + > > + BUILD_BUG_ON(offsetof(typeof(*w), dma)); > > + dma_fence_free(&w->dma); > > +} > > + > > +static const struct dma_fence_ops get_pages_work_ops = { > > + .get_driver_name = get_pages_work_driver_name, > > + .get_timeline_name = get_pages_work_timeline_name, > > + .release = get_pages_work_release, > > +}; > > + > > +static void __set_pages(struct drm_i915_gem_object *obj, > > + struct sg_table *pages, > > + unsigned int sg_page_sizes) > > { > > struct drm_i915_private *i915 = to_i915(obj->base.dev); > > unsigned long supported = INTEL_INFO(i915)->page_sizes; > > int i; > > > > - lockdep_assert_held(&obj->mm.lock); > > - > > /* Make the pages coherent with the GPU (flushing any swapin). */ > > if (obj->cache_dirty) { > > obj->write_domain = 0; > > @@ -29,8 +62,6 @@ void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj, > > obj->mm.get_page.sg_pos = pages->sgl; > > obj->mm.get_page.sg_idx = 0; > > > > - obj->mm.pages = pages; > > - > > if (i915_gem_object_is_tiled(obj) && > > i915->quirks & QUIRK_PIN_SWIZZLED_PAGES) { > > GEM_BUG_ON(obj->mm.quirked); > > @@ -38,7 +69,8 @@ void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj, > > obj->mm.quirked = true; > > } > > > > - GEM_BUG_ON(!sg_page_sizes); > > + if (!sg_page_sizes) > > + sg_page_sizes = i915_sg_page_sizes(pages->sgl); > > obj->mm.page_sizes.phys = sg_page_sizes; > > > > /* > > @@ -73,18 +105,105 @@ void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj, > > > > spin_unlock_irqrestore(&i915->mm.obj_lock, flags); > > } > > +} > > > > +static void > > +get_pages_worker(struct work_struct *_work) > > +{ > > + struct get_pages_work *work = container_of(_work, typeof(*work), work); > > + struct drm_i915_gem_object *obj = work->ctx.object; > > + struct sg_table *pages; > > + unsigned int sizes = 0; > > + > > + if (!work->dma.error) { > > + pages = obj->ops->get_pages(&work->ctx, &sizes); > > + if (!IS_ERR(pages)) > > + __set_pages(obj, pages, sizes); > > + else > > + dma_fence_set_error(&work->dma, PTR_ERR(pages)); > > + } else { > > + pages = ERR_PTR(work->dma.error); > > + } > > + > > + obj->mm.pages = pages; > > complete_all(&obj->mm.completion); > > + atomic_dec(&obj->mm.pages_pin_count); > > + > > + i915_gem_object_put(obj); > > + put_task_struct(work->ctx.task); > > + > > + dma_fence_signal(&work->dma); > > + dma_fence_put(&work->dma); > > +} > > + > > +static int __i915_sw_fence_call > > +get_pages_notify(struct i915_sw_fence *fence, enum i915_sw_fence_notify state) > > +{ > > + struct get_pages_work *w = container_of(fence, typeof(*w), wait); > > + > > + switch (state) { > > + case FENCE_COMPLETE: > > + if (fence->error) > > + dma_fence_set_error(&w->dma, fence->error); > > + queue_work(system_unbound_wq, &w->work); > > + break; > > + > > + case FENCE_FREE: > > + dma_fence_put(&w->dma); > > + break; > > + } > > + > > + return NOTIFY_DONE; > > } > > > > int ____i915_gem_object_get_pages(struct drm_i915_gem_object *obj) > > { > > + struct get_pages_work *work; > > + int err; > > + > > if (unlikely(obj->mm.madv != I915_MADV_WILLNEED)) { > > DRM_DEBUG("Attempting to obtain a purgeable object\n"); > > return -EFAULT; > > } > > > > - return obj->ops->get_pages(obj); > > + /* XXX inline? */ > > + > > + work = kmalloc(sizeof(*work), GFP_KERNEL); > > + if (!work) > > + return -ENOMEM; > > + > > + dma_fence_init(&work->dma, > > + &get_pages_work_ops, > > + &fence_lock, > > + to_i915(obj->base.dev)->mm.unordered_timeline, > > + 0); > > With all the shrinkers (across the tree, not just i915) we have that just > blindly assume you can wait for a fence (at least if you manage to acquire > the reservation_lock) I expect this to deadlock. I think the rule is that > no published fence can depend on any memory allocation anywhere. > > Yes it sucks real hard. > > > + i915_sw_fence_init(&work->wait, get_pages_notify); > > + > > + work->ctx.object = i915_gem_object_get(obj); > > + > > + work->ctx.task = current; > > + get_task_struct(work->ctx.task); > > + > > + INIT_WORK(&work->work, get_pages_worker); > > + > > + i915_gem_object_lock(obj); > > Other bit is nesting reservation_object lock within obj->mm.lock. > > obj->mm.lock is a very neat&tidy answer to dealing with bo alloc vs. > shrinker madness. But looking at where dma-buf is headed we'll need to > have the reservation_object nest outside of the obj->mm.lock (because the > importer is going to acquire that for us, before calling down into > exporter code). > > Of course on the importer side we're currently nesting the other way > round, and Christian König already set CI on fire figuring that out. > > I think what could work is pulling the reservation_object out of > obj->mm.lock, and entirely bypassing it for imported dma-buf (since for > that case dealing with shrinker lolz isn't our problem, but the exporters, > so we don't need the special shrinker lock). > > 3rd part is just irrational me freaking out when we hide userptr behind a > wait_completion, that didn't work out well last time around because > lockdep fails us. But that's really just an aside. > -Daniel > > > + GEM_BUG_ON(!reservation_object_test_signaled_rcu(obj->resv, true)); > > + err = i915_sw_fence_await_reservation(&work->wait, > > + obj->resv, NULL, > > + true, I915_FENCE_TIMEOUT, > > + I915_FENCE_GFP); > > + if (err == 0) { > > + reservation_object_add_excl_fence(obj->resv, &work->dma); > > + atomic_inc(&obj->mm.pages_pin_count); > > + } else { > > + dma_fence_set_error(&work->dma, err); > > + } > > + i915_gem_object_unlock(obj); > > + > > + dma_fence_get(&work->dma); > > + i915_sw_fence_commit(&work->wait); > > + > > + return err; > > } > > > > /* Ensure that the associated pages are gathered from the backing storage > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_phys.c b/drivers/gpu/drm/i915/gem/i915_gem_phys.c > > index 2deac933cf59..6b4a5fb52055 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_phys.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_phys.c > > @@ -17,18 +17,21 @@ > > #include "i915_gem_object.h" > > #include "i915_scatterlist.h" > > > > -static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +phys_get_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > + struct drm_i915_gem_object *obj = ctx->object; > > struct address_space *mapping = obj->base.filp->f_mapping; > > struct drm_dma_handle *phys; > > - struct sg_table *st; > > struct scatterlist *sg; > > + struct sg_table *st; > > char *vaddr; > > int i; > > int err; > > > > if (WARN_ON(i915_gem_object_needs_bit17_swizzle(obj))) > > - return -EINVAL; > > + return ERR_PTR(-EINVAL); > > > > /* Always aligning to the object size, allows a single allocation > > * to handle all possible callers, and given typical object sizes, > > @@ -38,7 +41,7 @@ static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj) > > roundup_pow_of_two(obj->base.size), > > roundup_pow_of_two(obj->base.size)); > > if (!phys) > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > > > vaddr = phys->vaddr; > > for (i = 0; i < obj->base.size / PAGE_SIZE; i++) { > > @@ -83,19 +86,16 @@ static int i915_gem_object_get_pages_phys(struct drm_i915_gem_object *obj) > > > > obj->phys_handle = phys; > > > > - __i915_gem_object_set_pages(obj, st, sg->length); > > - > > - return 0; > > + *sizes = sg->length; > > + return st; > > > > err_phys: > > drm_pci_free(obj->base.dev, phys); > > - > > - return err; > > + return ERR_PTR(err); > > } > > > > static void > > -i915_gem_object_put_pages_phys(struct drm_i915_gem_object *obj, > > - struct sg_table *pages) > > +phys_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) > > { > > __i915_gem_object_release_shmem(obj, pages, false); > > > > @@ -139,8 +139,8 @@ i915_gem_object_release_phys(struct drm_i915_gem_object *obj) > > } > > > > static const struct drm_i915_gem_object_ops i915_gem_phys_ops = { > > - .get_pages = i915_gem_object_get_pages_phys, > > - .put_pages = i915_gem_object_put_pages_phys, > > + .get_pages = phys_get_pages, > > + .put_pages = phys_put_pages, > > .release = i915_gem_object_release_phys, > > }; > > > > @@ -193,15 +193,12 @@ int i915_gem_object_attach_phys(struct drm_i915_gem_object *obj, int align) > > if (!IS_ERR_OR_NULL(pages)) > > i915_gem_shmem_ops.put_pages(obj, pages); > > mutex_unlock(&obj->mm.lock); > > + > > + wait_for_completion(&obj->mm.completion); > > return 0; > > > > err_xfer: > > obj->ops = &i915_gem_shmem_ops; > > - if (!IS_ERR_OR_NULL(pages)) { > > - unsigned int sg_page_sizes = i915_sg_page_sizes(pages->sgl); > > - > > - __i915_gem_object_set_pages(obj, pages, sg_page_sizes); > > - } > > err_unlock: > > mutex_unlock(&obj->mm.lock); > > return err; > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c > > index 19d9ecdb2894..c43304e3bada 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c > > @@ -22,11 +22,13 @@ static void check_release_pagevec(struct pagevec *pvec) > > cond_resched(); > > } > > > > -static int shmem_get_pages(struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +shmem_get_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > + struct drm_i915_gem_object *obj = ctx->object; > > struct drm_i915_private *i915 = to_i915(obj->base.dev); > > const unsigned long page_count = obj->base.size / PAGE_SIZE; > > - unsigned long i; > > struct address_space *mapping; > > struct sg_table *st; > > struct scatterlist *sg; > > @@ -37,31 +39,24 @@ static int shmem_get_pages(struct drm_i915_gem_object *obj) > > unsigned int sg_page_sizes; > > struct pagevec pvec; > > gfp_t noreclaim; > > + unsigned long i; > > int ret; > > > > - /* > > - * Assert that the object is not currently in any GPU domain. As it > > - * wasn't in the GTT, there shouldn't be any way it could have been in > > - * a GPU cache > > - */ > > - GEM_BUG_ON(obj->read_domains & I915_GEM_GPU_DOMAINS); > > - GEM_BUG_ON(obj->write_domain & I915_GEM_GPU_DOMAINS); > > - > > /* > > * If there's no chance of allocating enough pages for the whole > > * object, bail early. > > */ > > if (page_count > totalram_pages()) > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > > > st = kmalloc(sizeof(*st), GFP_KERNEL); > > if (!st) > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > > > rebuild_st: > > if (sg_alloc_table(st, page_count, GFP_KERNEL)) { > > kfree(st); > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > } > > > > /* > > @@ -179,9 +174,8 @@ static int shmem_get_pages(struct drm_i915_gem_object *obj) > > if (i915_gem_object_needs_bit17_swizzle(obj)) > > i915_gem_object_do_bit_17_swizzle(obj, st); > > > > - __i915_gem_object_set_pages(obj, st, sg_page_sizes); > > - > > - return 0; > > + *sizes = sg_page_sizes; > > + return st; > > > > err_sg: > > sg_mark_end(sg); > > @@ -209,7 +203,7 @@ static int shmem_get_pages(struct drm_i915_gem_object *obj) > > if (ret == -ENOSPC) > > ret = -ENOMEM; > > > > - return ret; > > + return ERR_PTR(ret); > > } > > > > static void > > @@ -276,8 +270,6 @@ __i915_gem_object_release_shmem(struct drm_i915_gem_object *obj, > > struct sg_table *pages, > > bool needs_clflush) > > { > > - GEM_BUG_ON(obj->mm.madv == __I915_MADV_PURGED); > > - > > if (obj->mm.madv == I915_MADV_DONTNEED) > > obj->mm.dirty = false; > > > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > > index 7066044d63cf..28ea06e667cd 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > > @@ -499,22 +499,19 @@ i915_pages_create_for_stolen(struct drm_device *dev, > > return st; > > } > > > > -static int i915_gem_object_get_pages_stolen(struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +stolen_get_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > - struct sg_table *pages = > > - i915_pages_create_for_stolen(obj->base.dev, > > - obj->stolen->start, > > - obj->stolen->size); > > - if (IS_ERR(pages)) > > - return PTR_ERR(pages); > > - > > - __i915_gem_object_set_pages(obj, pages, obj->stolen->size); > > + struct drm_i915_gem_object *obj = ctx->object; > > > > - return 0; > > + return i915_pages_create_for_stolen(obj->base.dev, > > + obj->stolen->start, > > + obj->stolen->size); > > } > > > > -static void i915_gem_object_put_pages_stolen(struct drm_i915_gem_object *obj, > > - struct sg_table *pages) > > +static void > > +stolen_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) > > { > > /* Should only be called from i915_gem_object_release_stolen() */ > > sg_free_table(pages); > > @@ -536,8 +533,8 @@ i915_gem_object_release_stolen(struct drm_i915_gem_object *obj) > > } > > > > static const struct drm_i915_gem_object_ops i915_gem_object_stolen_ops = { > > - .get_pages = i915_gem_object_get_pages_stolen, > > - .put_pages = i915_gem_object_put_pages_stolen, > > + .get_pages = stolen_get_pages, > > + .put_pages = stolen_put_pages, > > .release = i915_gem_object_release_stolen, > > }; > > > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c > > index f093deaeb5c0..6748e15bf89a 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c > > @@ -378,7 +378,7 @@ __i915_mm_struct_free(struct kref *kref) > > mutex_unlock(&mm->i915->mm_lock); > > > > INIT_WORK(&mm->work, __i915_mm_struct_free__worker); > > - queue_work(mm->i915->mm.userptr_wq, &mm->work); > > + schedule_work(&mm->work); > > } > > > > static void > > @@ -393,19 +393,12 @@ i915_gem_userptr_release__mm_struct(struct drm_i915_gem_object *obj) > > obj->userptr.mm = NULL; > > } > > > > -struct get_pages_work { > > - struct work_struct work; > > - struct drm_i915_gem_object *obj; > > - struct task_struct *task; > > -}; > > - > > static struct sg_table * > > __i915_gem_userptr_alloc_pages(struct drm_i915_gem_object *obj, > > struct page **pvec, int num_pages) > > { > > unsigned int max_segment = i915_sg_segment_size(); > > struct sg_table *st; > > - unsigned int sg_page_sizes; > > int ret; > > > > st = kmalloc(sizeof(*st), GFP_KERNEL); > > @@ -435,131 +428,23 @@ __i915_gem_userptr_alloc_pages(struct drm_i915_gem_object *obj, > > return ERR_PTR(ret); > > } > > > > - sg_page_sizes = i915_sg_page_sizes(st->sgl); > > - > > - __i915_gem_object_set_pages(obj, st, sg_page_sizes); > > - > > return st; > > } > > > > -static void > > -__i915_gem_userptr_get_pages_worker(struct work_struct *_work) > > -{ > > - struct get_pages_work *work = container_of(_work, typeof(*work), work); > > - struct drm_i915_gem_object *obj = work->obj; > > - const int npages = obj->base.size >> PAGE_SHIFT; > > - struct page **pvec; > > - int pinned, ret; > > - > > - ret = -ENOMEM; > > - pinned = 0; > > - > > - pvec = kvmalloc_array(npages, sizeof(struct page *), GFP_KERNEL); > > - if (pvec != NULL) { > > - struct mm_struct *mm = obj->userptr.mm->mm; > > - unsigned int flags = 0; > > - > > - if (!i915_gem_object_is_readonly(obj)) > > - flags |= FOLL_WRITE; > > - > > - ret = -EFAULT; > > - if (mmget_not_zero(mm)) { > > - down_read(&mm->mmap_sem); > > - while (pinned < npages) { > > - ret = get_user_pages_remote > > - (work->task, mm, > > - obj->userptr.ptr + pinned * PAGE_SIZE, > > - npages - pinned, > > - flags, > > - pvec + pinned, NULL, NULL); > > - if (ret < 0) > > - break; > > - > > - pinned += ret; > > - } > > - up_read(&mm->mmap_sem); > > - mmput(mm); > > - } > > - } > > - > > - mutex_lock(&obj->mm.lock); > > - if (obj->userptr.work == &work->work) { > > - struct sg_table *pages = ERR_PTR(ret); > > - > > - if (pinned == npages) { > > - pages = __i915_gem_userptr_alloc_pages(obj, pvec, > > - npages); > > - if (!IS_ERR(pages)) { > > - pinned = 0; > > - pages = NULL; > > - } > > - } > > - > > - obj->userptr.work = ERR_CAST(pages); > > - if (IS_ERR(pages)) > > - __i915_gem_userptr_set_active(obj, false); > > - } > > - mutex_unlock(&obj->mm.lock); > > - > > - release_pages(pvec, pinned); > > - kvfree(pvec); > > - > > - i915_gem_object_put(obj); > > - put_task_struct(work->task); > > - kfree(work); > > -} > > - > > static struct sg_table * > > -__i915_gem_userptr_get_pages_schedule(struct drm_i915_gem_object *obj) > > -{ > > - struct get_pages_work *work; > > - > > - /* Spawn a worker so that we can acquire the > > - * user pages without holding our mutex. Access > > - * to the user pages requires mmap_sem, and we have > > - * a strict lock ordering of mmap_sem, struct_mutex - > > - * we already hold struct_mutex here and so cannot > > - * call gup without encountering a lock inversion. > > - * > > - * Userspace will keep on repeating the operation > > - * (thanks to EAGAIN) until either we hit the fast > > - * path or the worker completes. If the worker is > > - * cancelled or superseded, the task is still run > > - * but the results ignored. (This leads to > > - * complications that we may have a stray object > > - * refcount that we need to be wary of when > > - * checking for existing objects during creation.) > > - * If the worker encounters an error, it reports > > - * that error back to this function through > > - * obj->userptr.work = ERR_PTR. > > - */ > > - work = kmalloc(sizeof(*work), GFP_KERNEL); > > - if (work == NULL) > > - return ERR_PTR(-ENOMEM); > > - > > - obj->userptr.work = &work->work; > > - > > - work->obj = i915_gem_object_get(obj); > > - > > - work->task = current; > > - get_task_struct(work->task); > > - > > - INIT_WORK(&work->work, __i915_gem_userptr_get_pages_worker); > > - queue_work(to_i915(obj->base.dev)->mm.userptr_wq, &work->work); > > - > > - return ERR_PTR(-EAGAIN); > > -} > > - > > -static int i915_gem_userptr_get_pages(struct drm_i915_gem_object *obj) > > +userptr_get_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > + struct drm_i915_gem_object *obj = ctx->object; > > const int num_pages = obj->base.size >> PAGE_SHIFT; > > struct mm_struct *mm = obj->userptr.mm->mm; > > - struct page **pvec; > > struct sg_table *pages; > > - bool active; > > + struct page **pvec; > > int pinned; > > + int ret; > > > > - /* If userspace should engineer that these pages are replaced in > > + /* > > + * If userspace should engineer that these pages are replaced in > > * the vma between us binding this page into the GTT and completion > > * of rendering... Their loss. If they change the mapping of their > > * pages they need to create a new bo to point to the new vma. > > @@ -576,59 +461,58 @@ static int i915_gem_userptr_get_pages(struct drm_i915_gem_object *obj) > > * egregious cases from causing harm. > > */ > > > > - if (obj->userptr.work) { > > - /* active flag should still be held for the pending work */ > > - if (IS_ERR(obj->userptr.work)) > > - return PTR_ERR(obj->userptr.work); > > - else > > - return -EAGAIN; > > - } > > + pvec = kvmalloc_array(num_pages, sizeof(struct page *), > > + GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN); > > + if (!pvec) > > + return ERR_PTR(-ENOMEM); > > + > > + __i915_gem_userptr_set_active(obj, true); > > > > - pvec = NULL; > > pinned = 0; > > + ret = -EFAULT; > > + if (mmget_not_zero(mm)) { > > + unsigned int flags; > > > > - if (mm == current->mm) { > > - pvec = kvmalloc_array(num_pages, sizeof(struct page *), > > - GFP_KERNEL | > > - __GFP_NORETRY | > > - __GFP_NOWARN); > > - if (pvec) /* defer to worker if malloc fails */ > > - pinned = __get_user_pages_fast(obj->userptr.ptr, > > - num_pages, > > - !i915_gem_object_is_readonly(obj), > > - pvec); > > - } > > + flags = 0; > > + if (!i915_gem_object_is_readonly(obj)) > > + flags |= FOLL_WRITE; > > > > - active = false; > > - if (pinned < 0) { > > - pages = ERR_PTR(pinned); > > - pinned = 0; > > - } else if (pinned < num_pages) { > > - pages = __i915_gem_userptr_get_pages_schedule(obj); > > - active = pages == ERR_PTR(-EAGAIN); > > - } else { > > - pages = __i915_gem_userptr_alloc_pages(obj, pvec, num_pages); > > - active = !IS_ERR(pages); > > + down_read(&mm->mmap_sem); > > + while (pinned < num_pages) { > > + ret = get_user_pages_remote > > + (ctx->task, mm, > > + obj->userptr.ptr + pinned * PAGE_SIZE, > > + num_pages - pinned, > > + flags, > > + pvec + pinned, NULL, NULL); > > + if (ret < 0) > > + break; > > + > > + pinned += ret; > > + } > > + up_read(&mm->mmap_sem); > > + mmput(mm); > > } > > - if (active) > > - __i915_gem_userptr_set_active(obj, true); > > > > - if (IS_ERR(pages)) > > + if (ret) > > + pages = ERR_PTR(ret); > > + else > > + pages = __i915_gem_userptr_alloc_pages(obj, pvec, num_pages); > > + if (IS_ERR(pages)) { > > release_pages(pvec, pinned); > > + __i915_gem_userptr_set_active(obj, false); > > + } > > kvfree(pvec); > > > > - return PTR_ERR_OR_ZERO(pages); > > + return pages; > > } > > > > static void > > -i915_gem_userptr_put_pages(struct drm_i915_gem_object *obj, > > - struct sg_table *pages) > > +userptr_put_pages(struct drm_i915_gem_object *obj, struct sg_table *pages) > > { > > struct sgt_iter sgt_iter; > > struct page *page; > > > > - /* Cancel any inflight work and force them to restart their gup */ > > - obj->userptr.work = NULL; > > __i915_gem_userptr_set_active(obj, false); > > if (!pages) > > return; > > @@ -669,8 +553,8 @@ static const struct drm_i915_gem_object_ops i915_gem_userptr_ops = { > > .flags = I915_GEM_OBJECT_HAS_STRUCT_PAGE | > > I915_GEM_OBJECT_IS_SHRINKABLE | > > I915_GEM_OBJECT_ASYNC_CANCEL, > > - .get_pages = i915_gem_userptr_get_pages, > > - .put_pages = i915_gem_userptr_put_pages, > > + .get_pages = userptr_get_pages, > > + .put_pages = userptr_put_pages, > > .dmabuf_export = i915_gem_userptr_dmabuf_export, > > .release = i915_gem_userptr_release, > > }; > > @@ -786,22 +670,13 @@ i915_gem_userptr_ioctl(struct drm_device *dev, > > return 0; > > } > > > > -int i915_gem_init_userptr(struct drm_i915_private *dev_priv) > > +void i915_gem_init_userptr(struct drm_i915_private *dev_priv) > > { > > mutex_init(&dev_priv->mm_lock); > > hash_init(dev_priv->mm_structs); > > - > > - dev_priv->mm.userptr_wq = > > - alloc_workqueue("i915-userptr-acquire", > > - WQ_HIGHPRI | WQ_UNBOUND, > > - 0); > > - if (!dev_priv->mm.userptr_wq) > > - return -ENOMEM; > > - > > - return 0; > > } > > > > void i915_gem_cleanup_userptr(struct drm_i915_private *dev_priv) > > { > > - destroy_workqueue(dev_priv->mm.userptr_wq); > > + mutex_destroy(&dev_priv->mm_lock); > > } > > diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c b/drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c > > index 3c5d17b2b670..02e6edce715e 100644 > > --- a/drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c > > +++ b/drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c > > @@ -21,9 +21,12 @@ static void huge_free_pages(struct drm_i915_gem_object *obj, > > kfree(pages); > > } > > > > -static int huge_get_pages(struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +huge_get_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > #define GFP (GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY) > > + struct drm_i915_gem_object *obj = ctx->object; > > const unsigned long nreal = obj->scratch / PAGE_SIZE; > > const unsigned long npages = obj->base.size / PAGE_SIZE; > > struct scatterlist *sg, *src, *end; > > @@ -32,11 +35,11 @@ static int huge_get_pages(struct drm_i915_gem_object *obj) > > > > pages = kmalloc(sizeof(*pages), GFP); > > if (!pages) > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > > > if (sg_alloc_table(pages, npages, GFP)) { > > kfree(pages); > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > } > > > > sg = pages->sgl; > > @@ -64,14 +67,12 @@ static int huge_get_pages(struct drm_i915_gem_object *obj) > > if (i915_gem_gtt_prepare_pages(obj, pages)) > > goto err; > > > > - __i915_gem_object_set_pages(obj, pages, PAGE_SIZE); > > - > > - return 0; > > + *sizes = PAGE_SIZE; > > + return pages; > > > > err: > > huge_free_pages(obj, pages); > > - > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > #undef GFP > > } > > > > diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > > index b03a29366bd1..3b93a7337f27 100644 > > --- a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > > +++ b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > > @@ -51,9 +51,12 @@ static void huge_pages_free_pages(struct sg_table *st) > > kfree(st); > > } > > > > -static int get_huge_pages(struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +get_huge_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > #define GFP (GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY) > > + struct drm_i915_gem_object *obj = ctx->object; > > unsigned int page_mask = obj->mm.page_mask; > > struct sg_table *st; > > struct scatterlist *sg; > > @@ -62,11 +65,11 @@ static int get_huge_pages(struct drm_i915_gem_object *obj) > > > > st = kmalloc(sizeof(*st), GFP); > > if (!st) > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > > > if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) { > > kfree(st); > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > } > > > > rem = obj->base.size; > > @@ -114,16 +117,14 @@ static int get_huge_pages(struct drm_i915_gem_object *obj) > > obj->mm.madv = I915_MADV_DONTNEED; > > > > GEM_BUG_ON(sg_page_sizes != obj->mm.page_mask); > > - __i915_gem_object_set_pages(obj, st, sg_page_sizes); > > - > > - return 0; > > + *sizes = sg_page_sizes; > > + return st; > > > > err: > > sg_set_page(sg, NULL, 0, 0); > > sg_mark_end(sg); > > huge_pages_free_pages(st); > > - > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > } > > > > static void put_huge_pages(struct drm_i915_gem_object *obj, > > @@ -175,8 +176,11 @@ huge_pages_object(struct drm_i915_private *i915, > > return obj; > > } > > > > -static int fake_get_huge_pages(struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +fake_get_huge_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > + struct drm_i915_gem_object *obj = ctx->object; > > struct drm_i915_private *i915 = to_i915(obj->base.dev); > > const u64 max_len = rounddown_pow_of_two(UINT_MAX); > > struct sg_table *st; > > @@ -186,11 +190,11 @@ static int fake_get_huge_pages(struct drm_i915_gem_object *obj) > > > > st = kmalloc(sizeof(*st), GFP); > > if (!st) > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > > > if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) { > > kfree(st); > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > } > > > > /* Use optimal page sized chunks to fill in the sg table */ > > @@ -227,13 +231,15 @@ static int fake_get_huge_pages(struct drm_i915_gem_object *obj) > > > > obj->mm.madv = I915_MADV_DONTNEED; > > > > - __i915_gem_object_set_pages(obj, st, sg_page_sizes); > > - > > - return 0; > > + *sizes = sg_page_sizes; > > + return st; > > } > > > > -static int fake_get_huge_pages_single(struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +fake_get_huge_pages_single(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > + struct drm_i915_gem_object *obj = ctx->object; > > struct drm_i915_private *i915 = to_i915(obj->base.dev); > > struct sg_table *st; > > struct scatterlist *sg; > > @@ -241,11 +247,11 @@ static int fake_get_huge_pages_single(struct drm_i915_gem_object *obj) > > > > st = kmalloc(sizeof(*st), GFP); > > if (!st) > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > > > if (sg_alloc_table(st, 1, GFP)) { > > kfree(st); > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > } > > > > sg = st->sgl; > > @@ -261,9 +267,8 @@ static int fake_get_huge_pages_single(struct drm_i915_gem_object *obj) > > > > obj->mm.madv = I915_MADV_DONTNEED; > > > > - __i915_gem_object_set_pages(obj, st, sg->length); > > - > > - return 0; > > + *sizes = sg->length; > > + return st; > > #undef GFP > > } > > > > diff --git a/drivers/gpu/drm/i915/gvt/dmabuf.c b/drivers/gpu/drm/i915/gvt/dmabuf.c > > index 41c8ebc60c63..c2a81939f698 100644 > > --- a/drivers/gpu/drm/i915/gvt/dmabuf.c > > +++ b/drivers/gpu/drm/i915/gvt/dmabuf.c > > @@ -36,9 +36,11 @@ > > > > #define GEN8_DECODE_PTE(pte) (pte & GENMASK_ULL(63, 12)) > > > > -static int vgpu_gem_get_pages( > > - struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +vgpu_gem_get_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > + struct drm_i915_gem_object *obj = ctx->object; > > struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > struct sg_table *st; > > struct scatterlist *sg; > > @@ -49,17 +51,17 @@ static int vgpu_gem_get_pages( > > > > fb_info = (struct intel_vgpu_fb_info *)obj->gvt_info; > > if (WARN_ON(!fb_info)) > > - return -ENODEV; > > + return ERR_PTR(-ENODEV); > > > > st = kmalloc(sizeof(*st), GFP_KERNEL); > > if (unlikely(!st)) > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > > > page_num = obj->base.size >> PAGE_SHIFT; > > ret = sg_alloc_table(st, page_num, GFP_KERNEL); > > if (ret) { > > kfree(st); > > - return ret; > > + return ERR_PTR(ret); > > } > > gtt_entries = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + > > (fb_info->start >> PAGE_SHIFT); > > @@ -71,9 +73,8 @@ static int vgpu_gem_get_pages( > > sg_dma_len(sg) = PAGE_SIZE; > > } > > > > - __i915_gem_object_set_pages(obj, st, PAGE_SIZE); > > - > > - return 0; > > + *sizes = PAGE_SIZE; > > + return st; > > } > > > > static void vgpu_gem_put_pages(struct drm_i915_gem_object *obj, > > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h > > index 9a1ec487e8b1..ef9e4cb49b4a 100644 > > --- a/drivers/gpu/drm/i915/i915_drv.h > > +++ b/drivers/gpu/drm/i915/i915_drv.h > > @@ -786,13 +786,6 @@ struct i915_gem_mm { > > struct notifier_block vmap_notifier; > > struct shrinker shrinker; > > > > - /** > > - * Workqueue to fault in userptr pages, flushed by the execbuf > > - * when required but otherwise left to userspace to try again > > - * on EAGAIN. > > - */ > > - struct workqueue_struct *userptr_wq; > > - > > u64 unordered_timeline; > > > > /* the indicator for dispatch video commands on two BSD rings */ > > @@ -2514,7 +2507,7 @@ static inline bool intel_vgpu_active(struct drm_i915_private *dev_priv) > > } > > > > /* i915_gem.c */ > > -int i915_gem_init_userptr(struct drm_i915_private *dev_priv); > > +void i915_gem_init_userptr(struct drm_i915_private *dev_priv); > > void i915_gem_cleanup_userptr(struct drm_i915_private *dev_priv); > > void i915_gem_sanitize(struct drm_i915_private *i915); > > int i915_gem_init_early(struct drm_i915_private *dev_priv); > > diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c > > index 21416b87905e..1571c707ad15 100644 > > --- a/drivers/gpu/drm/i915/i915_gem.c > > +++ b/drivers/gpu/drm/i915/i915_gem.c > > @@ -1544,10 +1544,7 @@ int i915_gem_init(struct drm_i915_private *dev_priv) > > dev_priv->mm.unordered_timeline = dma_fence_context_alloc(1); > > > > i915_timelines_init(dev_priv); > > - > > - ret = i915_gem_init_userptr(dev_priv); > > - if (ret) > > - return ret; > > + i915_gem_init_userptr(dev_priv); > > > > ret = intel_uc_init_misc(dev_priv); > > if (ret) > > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > > index 664b7e3e1d7c..dae1f8634a38 100644 > > --- a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > > @@ -54,10 +54,13 @@ static void fake_free_pages(struct drm_i915_gem_object *obj, > > kfree(pages); > > } > > > > -static int fake_get_pages(struct drm_i915_gem_object *obj) > > +static struct sg_table * > > +fake_get_pages(struct i915_gem_object_get_pages_context *ctx, > > + unsigned int *sizes) > > { > > #define GFP (GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY) > > #define PFN_BIAS 0x1000 > > + struct drm_i915_gem_object *obj = ctx->object; > > struct sg_table *pages; > > struct scatterlist *sg; > > unsigned int sg_page_sizes; > > @@ -65,12 +68,12 @@ static int fake_get_pages(struct drm_i915_gem_object *obj) > > > > pages = kmalloc(sizeof(*pages), GFP); > > if (!pages) > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > > > rem = round_up(obj->base.size, BIT(31)) >> 31; > > if (sg_alloc_table(pages, rem, GFP)) { > > kfree(pages); > > - return -ENOMEM; > > + return ERR_PTR(-ENOMEM); > > } > > > > sg_page_sizes = 0; > > @@ -90,9 +93,8 @@ static int fake_get_pages(struct drm_i915_gem_object *obj) > > > > obj->mm.madv = I915_MADV_DONTNEED; > > > > - __i915_gem_object_set_pages(obj, pages, sg_page_sizes); > > - > > - return 0; > > + *sizes = sg_page_sizes; > > + return pages; > > #undef GFP > > } > > > > -- > > 2.20.1 > > > > _______________________________________________ > > Intel-gfx mailing list > > Intel-gfx@xxxxxxxxxxxxxxxxxxxxx > > https://lists.freedesktop.org/mailman/listinfo/intel-gfx > > -- > Daniel Vetter > Software Engineer, Intel Corporation > http://blog.ffwll.ch -- Daniel Vetter Software Engineer, Intel Corporation +41 (0) 79 365 57 48 - http://blog.ffwll.ch _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx