On Wed, Jun 09, 2021 at 01:34:00PM +0200, Daniel Vetter wrote: > On Tue, Jun 08, 2021 at 11:36:07PM -0500, Jason Ekstrand wrote: > > When the APIs were added to manage VMs more directly from userspace, the > > questionable choice was made to allow changing out the VM on a context > > at any time. This is horribly racy and there's absolutely no reason why > > any userspace would want to do this outside of testing that exact race. > > By removing support for CONTEXT_PARAM_VM from ctx_setparam, we make it > > impossible to change out the VM after the context has been fully > > created. This lets us delete a bunch of deferred task code as well as a > > duplicated (and slightly different) copy of the code which programs the > > PPGTT registers. > > > > v2 (Jason Ekstrand): > > - Expand the commit message Also tsk, tsk for non mentioning you're dropping the __rcu here, that's a rather crucial change and without 0day complaining I'd have overlooked it. -Daniel > > > > Signed-off-by: Jason Ekstrand <jason@xxxxxxxxxxxxxx> > > Reviewed-by: Daniel Vetter <daniel.vetter@xxxxxxxx> > > Need to retract this r-b here until the issue below is fixed. > > > --- > > drivers/gpu/drm/i915/gem/i915_gem_context.c | 262 ------------------ > > .../gpu/drm/i915/gem/i915_gem_context_types.h | 2 +- > > .../drm/i915/gem/selftests/i915_gem_context.c | 119 -------- > > .../drm/i915/selftests/i915_mock_selftests.h | 1 - > > 4 files changed, 1 insertion(+), 383 deletions(-) > > > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.c b/drivers/gpu/drm/i915/gem/i915_gem_context.c > > index f74c22dc506ec..2f3d92224d2fe 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_context.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_context.c > > @@ -1633,120 +1633,6 @@ int i915_gem_vm_destroy_ioctl(struct drm_device *dev, void *data, > > return 0; > > } > > > > -struct context_barrier_task { > > - struct i915_active base; > > - void (*task)(void *data); > > - void *data; > > -}; > > - > > -static void cb_retire(struct i915_active *base) > > -{ > > - struct context_barrier_task *cb = container_of(base, typeof(*cb), base); > > - > > - if (cb->task) > > - cb->task(cb->data); > > - > > - i915_active_fini(&cb->base); > > - kfree(cb); > > -} > > - > > -I915_SELFTEST_DECLARE(static intel_engine_mask_t context_barrier_inject_fault); > > -static int context_barrier_task(struct i915_gem_context *ctx, > > - intel_engine_mask_t engines, > > - bool (*skip)(struct intel_context *ce, void *data), > > - int (*pin)(struct intel_context *ce, struct i915_gem_ww_ctx *ww, void *data), > > - int (*emit)(struct i915_request *rq, void *data), > > - void (*task)(void *data), > > - void *data) > > -{ > > - struct context_barrier_task *cb; > > - struct i915_gem_engines_iter it; > > - struct i915_gem_engines *e; > > - struct i915_gem_ww_ctx ww; > > - struct intel_context *ce; > > - int err = 0; > > - > > - GEM_BUG_ON(!task); > > - > > - cb = kmalloc(sizeof(*cb), GFP_KERNEL); > > - if (!cb) > > - return -ENOMEM; > > - > > - i915_active_init(&cb->base, NULL, cb_retire, 0); > > - err = i915_active_acquire(&cb->base); > > - if (err) { > > - kfree(cb); > > - return err; > > - } > > - > > - e = __context_engines_await(ctx, NULL); > > - if (!e) { > > - i915_active_release(&cb->base); > > - return -ENOENT; > > - } > > - > > - for_each_gem_engine(ce, e, it) { > > - struct i915_request *rq; > > - > > - if (I915_SELFTEST_ONLY(context_barrier_inject_fault & > > - ce->engine->mask)) { > > - err = -ENXIO; > > - break; > > - } > > - > > - if (!(ce->engine->mask & engines)) > > - continue; > > - > > - if (skip && skip(ce, data)) > > - continue; > > - > > - i915_gem_ww_ctx_init(&ww, true); > > -retry: > > - err = intel_context_pin_ww(ce, &ww); > > - if (err) > > - goto err; > > - > > - if (pin) > > - err = pin(ce, &ww, data); > > - if (err) > > - goto err_unpin; > > - > > - rq = i915_request_create(ce); > > - if (IS_ERR(rq)) { > > - err = PTR_ERR(rq); > > - goto err_unpin; > > - } > > - > > - err = 0; > > - if (emit) > > - err = emit(rq, data); > > - if (err == 0) > > - err = i915_active_add_request(&cb->base, rq); > > - > > - i915_request_add(rq); > > -err_unpin: > > - intel_context_unpin(ce); > > -err: > > - if (err == -EDEADLK) { > > - err = i915_gem_ww_ctx_backoff(&ww); > > - if (!err) > > - goto retry; > > - } > > - i915_gem_ww_ctx_fini(&ww); > > - > > - if (err) > > - break; > > - } > > - i915_sw_fence_complete(&e->fence); > > - > > - cb->task = err ? NULL : task; /* caller needs to unwind instead */ > > - cb->data = data; > > - > > - i915_active_release(&cb->base); > > - > > - return err; > > -} > > - > > static int get_ppgtt(struct drm_i915_file_private *file_priv, > > struct i915_gem_context *ctx, > > struct drm_i915_gem_context_param *args) > > @@ -1779,150 +1665,6 @@ static int get_ppgtt(struct drm_i915_file_private *file_priv, > > return err; > > } > > > > -static void set_ppgtt_barrier(void *data) > > -{ > > - struct i915_address_space *old = data; > > - > > - if (GRAPHICS_VER(old->i915) < 8) > > - gen6_ppgtt_unpin_all(i915_vm_to_ppgtt(old)); > > - > > - i915_vm_close(old); > > -} > > - > > -static int pin_ppgtt_update(struct intel_context *ce, struct i915_gem_ww_ctx *ww, void *data) > > -{ > > - struct i915_address_space *vm = ce->vm; > > - > > - if (!HAS_LOGICAL_RING_CONTEXTS(vm->i915)) > > - /* ppGTT is not part of the legacy context image */ > > - return gen6_ppgtt_pin(i915_vm_to_ppgtt(vm), ww); > > - > > - return 0; > > -} > > - > > -static int emit_ppgtt_update(struct i915_request *rq, void *data) > > -{ > > - struct i915_address_space *vm = rq->context->vm; > > - struct intel_engine_cs *engine = rq->engine; > > - u32 base = engine->mmio_base; > > - u32 *cs; > > - int i; > > - > > - if (i915_vm_is_4lvl(vm)) { > > - struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); > > - const dma_addr_t pd_daddr = px_dma(ppgtt->pd); > > - > > - cs = intel_ring_begin(rq, 6); > > - if (IS_ERR(cs)) > > - return PTR_ERR(cs); > > - > > - *cs++ = MI_LOAD_REGISTER_IMM(2); > > - > > - *cs++ = i915_mmio_reg_offset(GEN8_RING_PDP_UDW(base, 0)); > > - *cs++ = upper_32_bits(pd_daddr); > > - *cs++ = i915_mmio_reg_offset(GEN8_RING_PDP_LDW(base, 0)); > > - *cs++ = lower_32_bits(pd_daddr); > > - > > - *cs++ = MI_NOOP; > > - intel_ring_advance(rq, cs); > > - } else if (HAS_LOGICAL_RING_CONTEXTS(engine->i915)) { > > - struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); > > - int err; > > - > > - /* Magic required to prevent forcewake errors! */ > > - err = engine->emit_flush(rq, EMIT_INVALIDATE); > > - if (err) > > - return err; > > - > > - cs = intel_ring_begin(rq, 4 * GEN8_3LVL_PDPES + 2); > > - if (IS_ERR(cs)) > > - return PTR_ERR(cs); > > - > > - *cs++ = MI_LOAD_REGISTER_IMM(2 * GEN8_3LVL_PDPES) | MI_LRI_FORCE_POSTED; > > - for (i = GEN8_3LVL_PDPES; i--; ) { > > - const dma_addr_t pd_daddr = i915_page_dir_dma_addr(ppgtt, i); > > - > > - *cs++ = i915_mmio_reg_offset(GEN8_RING_PDP_UDW(base, i)); > > - *cs++ = upper_32_bits(pd_daddr); > > - *cs++ = i915_mmio_reg_offset(GEN8_RING_PDP_LDW(base, i)); > > - *cs++ = lower_32_bits(pd_daddr); > > - } > > - *cs++ = MI_NOOP; > > - intel_ring_advance(rq, cs); > > - } > > - > > - return 0; > > -} > > - > > -static bool skip_ppgtt_update(struct intel_context *ce, void *data) > > -{ > > - if (HAS_LOGICAL_RING_CONTEXTS(ce->engine->i915)) > > - return !ce->state; > > - else > > - return !atomic_read(&ce->pin_count); > > -} > > - > > -static int set_ppgtt(struct drm_i915_file_private *file_priv, > > - struct i915_gem_context *ctx, > > - struct drm_i915_gem_context_param *args) > > -{ > > - struct i915_address_space *vm, *old; > > - int err; > > - > > - if (args->size) > > - return -EINVAL; > > - > > - if (!rcu_access_pointer(ctx->vm)) > > - return -ENODEV; > > - > > - if (upper_32_bits(args->value)) > > - return -ENOENT; > > - > > - vm = i915_gem_vm_lookup(file_priv, args->value); > > - if (!vm) > > - return -ENOENT; > > - > > - err = mutex_lock_interruptible(&ctx->mutex); > > - if (err) > > - goto out; > > - > > - if (i915_gem_context_is_closed(ctx)) { > > - err = -ENOENT; > > - goto unlock; > > - } > > - > > - if (vm == rcu_access_pointer(ctx->vm)) > > - goto unlock; > > - > > - old = __set_ppgtt(ctx, vm); > > - > > - /* Teardown the existing obj:vma cache, it will have to be rebuilt. */ > > - lut_close(ctx); > > - > > - /* > > - * We need to flush any requests using the current ppgtt before > > - * we release it as the requests do not hold a reference themselves, > > - * only indirectly through the context. > > - */ > > - err = context_barrier_task(ctx, ALL_ENGINES, > > - skip_ppgtt_update, > > - pin_ppgtt_update, > > - emit_ppgtt_update, > > - set_ppgtt_barrier, > > - old); > > - if (err) { > > - i915_vm_close(__set_ppgtt(ctx, old)); > > - i915_vm_close(old); > > - lut_close(ctx); /* force a rebuild of the old obj:vma cache */ > > - } > > - > > -unlock: > > - mutex_unlock(&ctx->mutex); > > -out: > > - i915_vm_put(vm); > > - return err; > > -} > > - > > int > > i915_gem_user_to_context_sseu(struct intel_gt *gt, > > const struct drm_i915_gem_context_param_sseu *user, > > @@ -2458,10 +2200,6 @@ static int ctx_setparam(struct drm_i915_file_private *fpriv, > > ret = set_sseu(ctx, args); > > break; > > > > - case I915_CONTEXT_PARAM_VM: > > - ret = set_ppgtt(fpriv, ctx, args); > > - break; > > - > > case I915_CONTEXT_PARAM_ENGINES: > > ret = set_engines(ctx, args); > > break; > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context_types.h b/drivers/gpu/drm/i915/gem/i915_gem_context_types.h > > index 94c03a97cb77c..540ad16204a97 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_context_types.h > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_context_types.h > > @@ -262,7 +262,7 @@ struct i915_gem_context { > > * In other modes, this is a NULL pointer with the expectation that > > * the caller uses the shared global GTT. > > */ > > - struct i915_address_space __rcu *vm; > > + struct i915_address_space *vm; > > Ok, you fixed this wrong. We can't just drop the __rcu here because in > various places we're probably relying on rcu_read_lock to give us a > temporary reference. Until that is sorted, the __rcu here needs to stay. > > That also takes of the 0day issue the kernel reported. > > To fixe the __rcu mismatches in i915_gem_context you probably need to > sprinkle some rcu_assign_pointer around. > > With that addressed again > > Reviewed-by: Daniel Vetter <daniel.vetter@xxxxxxxx> > > > > > /** > > * @pid: process id of creator > > diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c > > index dbcfa28a9d91b..92544a174cc9a 100644 > > --- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c > > +++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c > > @@ -1875,125 +1875,6 @@ static int igt_vm_isolation(void *arg) > > return err; > > } > > > > -static bool skip_unused_engines(struct intel_context *ce, void *data) > > -{ > > - return !ce->state; > > -} > > - > > -static void mock_barrier_task(void *data) > > -{ > > - unsigned int *counter = data; > > - > > - ++*counter; > > -} > > - > > -static int mock_context_barrier(void *arg) > > -{ > > -#undef pr_fmt > > -#define pr_fmt(x) "context_barrier_task():" # x > > - struct drm_i915_private *i915 = arg; > > - struct i915_gem_context *ctx; > > - struct i915_request *rq; > > - unsigned int counter; > > - int err; > > - > > - /* > > - * The context barrier provides us with a callback after it emits > > - * a request; useful for retiring old state after loading new. > > - */ > > - > > - ctx = mock_context(i915, "mock"); > > - if (!ctx) > > - return -ENOMEM; > > - > > - counter = 0; > > - err = context_barrier_task(ctx, 0, NULL, NULL, NULL, > > - mock_barrier_task, &counter); > > - if (err) { > > - pr_err("Failed at line %d, err=%d\n", __LINE__, err); > > - goto out; > > - } > > - if (counter == 0) { > > - pr_err("Did not retire immediately with 0 engines\n"); > > - err = -EINVAL; > > - goto out; > > - } > > - > > - counter = 0; > > - err = context_barrier_task(ctx, ALL_ENGINES, skip_unused_engines, > > - NULL, NULL, mock_barrier_task, &counter); > > - if (err) { > > - pr_err("Failed at line %d, err=%d\n", __LINE__, err); > > - goto out; > > - } > > - if (counter == 0) { > > - pr_err("Did not retire immediately for all unused engines\n"); > > - err = -EINVAL; > > - goto out; > > - } > > - > > - rq = igt_request_alloc(ctx, i915->gt.engine[RCS0]); > > - if (IS_ERR(rq)) { > > - pr_err("Request allocation failed!\n"); > > - goto out; > > - } > > - i915_request_add(rq); > > - > > - counter = 0; > > - context_barrier_inject_fault = BIT(RCS0); > > - err = context_barrier_task(ctx, ALL_ENGINES, NULL, NULL, NULL, > > - mock_barrier_task, &counter); > > - context_barrier_inject_fault = 0; > > - if (err == -ENXIO) > > - err = 0; > > - else > > - pr_err("Did not hit fault injection!\n"); > > - if (counter != 0) { > > - pr_err("Invoked callback on error!\n"); > > - err = -EIO; > > - } > > - if (err) > > - goto out; > > - > > - counter = 0; > > - err = context_barrier_task(ctx, ALL_ENGINES, skip_unused_engines, > > - NULL, NULL, mock_barrier_task, &counter); > > - if (err) { > > - pr_err("Failed at line %d, err=%d\n", __LINE__, err); > > - goto out; > > - } > > - mock_device_flush(i915); > > - if (counter == 0) { > > - pr_err("Did not retire on each active engines\n"); > > - err = -EINVAL; > > - goto out; > > - } > > - > > -out: > > - mock_context_close(ctx); > > - return err; > > -#undef pr_fmt > > -#define pr_fmt(x) x > > -} > > - > > -int i915_gem_context_mock_selftests(void) > > -{ > > - static const struct i915_subtest tests[] = { > > - SUBTEST(mock_context_barrier), > > - }; > > - struct drm_i915_private *i915; > > - int err; > > - > > - i915 = mock_gem_device(); > > - if (!i915) > > - return -ENOMEM; > > - > > - err = i915_subtests(tests, i915); > > - > > - mock_destroy_device(i915); > > - return err; > > -} > > - > > int i915_gem_context_live_selftests(struct drm_i915_private *i915) > > { > > static const struct i915_subtest tests[] = { > > diff --git a/drivers/gpu/drm/i915/selftests/i915_mock_selftests.h b/drivers/gpu/drm/i915/selftests/i915_mock_selftests.h > > index 34e5caf380933..0c22e0fc9059c 100644 > > --- a/drivers/gpu/drm/i915/selftests/i915_mock_selftests.h > > +++ b/drivers/gpu/drm/i915/selftests/i915_mock_selftests.h > > @@ -32,5 +32,4 @@ selftest(vma, i915_vma_mock_selftests) > > selftest(evict, i915_gem_evict_mock_selftests) > > selftest(gtt, i915_gem_gtt_mock_selftests) > > selftest(hugepages, i915_gem_huge_page_mock_selftests) > > -selftest(contexts, i915_gem_context_mock_selftests) > > selftest(memory_region, intel_memory_region_mock_selftests) > > -- > > 2.31.1 > > > > _______________________________________________ > > Intel-gfx mailing list > > Intel-gfx@xxxxxxxxxxxxxxxxxxxxx > > https://lists.freedesktop.org/mailman/listinfo/intel-gfx > > -- > Daniel Vetter > Software Engineer, Intel Corporation > http://blog.ffwll.ch -- Daniel Vetter Software Engineer, Intel Corporation http://blog.ffwll.ch