On 7/6/20 8:19 AM, Chris Wilson wrote:
This is the easy part; pulling reservation of multiple objects under an ww acquire context. With one simple rule that eviction is handled by the ww acquire context, we can carefully transition the driver over to using eviction. Instead of feeding the acquire context everywhere, we make the caller gather up all the objects they need to acquire into the context, then acquire their backing store. The major boon here is that by providing a clean set of objects (that we have not yet started to acquire any auxiliary attachments for) to the acquire context, it can handle all the EDEADLK processing for us [since it is a pure locking operation and does not need to release attachments upon revoking the locks]. As a sketch of what that would look like, to illustrate the major work remaining: static int evict(struct drm_i915_gem_object *obj, struct i915_acquire_ctx *ctx) { struct intel_memory_region *mem = obj->mm->region; struct drm_i915_gem_object *swap; // struct i915_mm_bo *swap struct i915_request *rq; int err; /* swap = mem->create_eviction_target(obj); */ swap = i915_gem_object_create_shmem(mem->i915, obj->base.size); if (IS_ERR(swap)) return PTR_ERR(swap); err = dma_resv_lock_interruptible(swap, &ctx->ctx); GEM_BUG_ON(err == -EALREADY); if (err == -EDEADLK) goto out; /* Obviously swap has to be carefully chosen so that this may succeed */ err = __i915_gem_object_get_pages_locked(swap); if (err) goto out_unlock; rq = pinned_evict_copy(ctx, obj, swap); if (IS_ERR(rq)) { err = PTR_ERR(rq); goto out_unlock; } err = i915_gem_revoke_mm(obj); if (err) goto out_request; /* Alternatively you could wait synchronously! */ mem->release_blocks(&obj->mm->blocks, rq); i915_mm_bo_put(xchg(&obj->mm, i915_mm_bo_get(swap))); dma_resv_add_exclusive_fence(obj->base.resv, &rq->fence); out_request: i915_request_put(rq); out_unlock: dma_resv_unlock(swap); out: i915_gem_object_put(swap); return err; } static int relock_all(struct i915_acquire_ctx *ctx) { struct i915_acquire_link *lnk, *lock; int err; for (lnk = ctx->locked; lnk; lnk = lnk->next) dma_resv_unlock(lnk->obj->base.resv); lock = fetch_and_zero(&ctx->locked); while ((lnk = lock)) { struct drm_i915_gem_object *obj; obj = lnk->obj; lock = lnk->next; if (ctx->locked) err = dma_resv_lock_interruptible(obj->base.resv, &ctx->ctx); else err = dma_resv_lock_slow_interruptible(obj->base.resv, &ctx->ctx); GEM_BUG_ON(err == -EALREADY); if (err == -EDEADLK) { struct i915_acquire *old; while ((old = ctx->locked)) { dma_resv_unlock(old->obj->base.resv); ctx->locked = old->next; old->next = lock; lock = old; } lnk->next = lock; lock = lnk; continue; } if (err) { lock = lnk; break; } lnk->next = ctx->locked; ctx->locked = lnk; } while ((lnk = lock)) { lock = lnk->next; i915_gem_object_put(lnk->obj); i915_acquire_free(lnk); } return err; } int i915_acquire_mm(struct i915_acquire_ctx *ctx) { struct i915_acquire_link *lnk; int n, err; restart: for (lnk = ctx->locked; lnk; lnk = lnk->next) { for (n = 0; !i915_gem_object_has_pages(lnk->obj); n++) { struct drm_i915_gem_object *evictee = NULL; mem = get_preferred_memregion_for_object(lnk->obj, n); if (!mem) return -ENXIO; while (!i915_gem_object_get_pages(lnk->obj)) { struct i915_acquire_link *this; evictee = mem->get_eviction_candidate(mem, evictee); if (!evictee) break; err = dma_resv_lock_interruptible(evictee, &ctx->ctx); if (err == -EALREADY) continue; /* XXX fragmentation? */ if (err == 0) err = evict(evictee); dma_resv_unlock(evictee);
There was a discussion on dri-devel not too long ago, where Christian mentioned there is a point holding on to the evictee(s) locks until the get_pages() succeeds to avoid racing with threads wanting to move the evictee back into vram. Perhaps something worth considering.
/Thomas _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx