> -----Original Message----- > From: Intel-gfx <intel-gfx-bounces@xxxxxxxxxxxxxxxxxxxxx> On Behalf Of Chris > Wilson > Sent: Friday, April 24, 2020 1:33 AM > To: intel-gfx@xxxxxxxxxxxxxxxxxxxxx > Subject: [CI 1/2] drm/i915/selftests: Verify context isolation > > No unprivileged context should ever be allowed to modify logical state that is > visible to another; there should be no backchannels available or control > leakage for malicious actors. > > This test tries to write to a set of random registers using non-privileged > instructions (ala userspace). It should only be allowed to write into its > context state, and all writes should not be visible to a second context. To > verify this, we store the value of the register before writing to it in context A > (as this should be the default value inherited from the golden context state) > and do a read back from context B of the same register. The reads from both > contexts should be identical, the default value, except for a few free running > counters (either global or local). > > Signed-off-by: Chris Wilson <chris@xxxxxxxxxxxxxxxxxx> > Cc: Mika Kuoppala <mika.kuoppala@xxxxxxxxxxxxxxx> > --- > .../drm/i915/gem/selftests/i915_gem_context.c | 441 ++++++++++++++++++ > 1 file changed, 441 insertions(+) > > diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c > b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c > index f4f933240b39..c5c3433174dc 100644 > --- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c > +++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c > @@ -1865,6 +1865,446 @@ static int igt_vm_isolation(void *arg) > return err; > } > > +static struct i915_vma *create_vma(struct i915_address_space *vm, > +size_t sz) { > + struct drm_i915_gem_object *obj; > + struct i915_vma *vma; > + > + obj = i915_gem_object_create_internal(vm->i915, sz); > + if (IS_ERR(obj)) > + return ERR_CAST(obj); > + > + vma = i915_vma_instance(obj, vm, NULL); > + if (IS_ERR(vma)) > + i915_gem_object_put(obj); > + > + return vma; > +} > + > +struct iso_details { > + unsigned long count; > +}; > + > +enum { > + ISO_REG = 0, > + ISO_POISON, > + ISO_BEFORE, > + ISO_AFTER, > + __ISO__ > +}; > + > +static int iso_write(struct i915_gem_context *ctx, > + struct intel_engine_cs *engine, > + struct drm_i915_gem_object *obj, > + const struct iso_details *iso, > + u32 *ctl) > +{ > + struct i915_vma *batch, *vma; > + struct intel_context *ce; > + struct i915_request *rq; > + u32 *cs; > + int err; > + int i; > + > + ce = i915_gem_context_get_engine(ctx, engine->legacy_idx); > + if (IS_ERR(ce)) > + return PTR_ERR(ce); > + > + batch = create_vma(ce->vm, PAGE_ALIGN(16 * iso->count + 4)); > + if (IS_ERR(batch)) { > + err = PTR_ERR(batch); > + goto err_ce; > + } > + > + vma = i915_vma_instance(obj, ce->vm, NULL); > + if (IS_ERR(vma)) { > + err = PTR_ERR(vma); > + goto err_batch; > + } > + > + err = i915_vma_pin(batch, 0, 0, PIN_USER); > + if (err) > + goto err_batch; > + > + err = i915_vma_pin(vma, 0, 0, PIN_USER); > + if (err) > + goto err_unpin_batch; > + > + cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC); > + if (IS_ERR(cs)) { > + err = PTR_ERR(cs); > + goto err_vma; > + } > + > + for (i = 0; i < iso->count; i++) { > + *cs++ = MI_LOAD_REGISTER_IMM(1); > + *cs++ = ctl[i * __ISO__ + ISO_REG]; > + *cs++ = ctl[i * __ISO__ + ISO_POISON]; > + } > + *cs++ = MI_BATCH_BUFFER_END; > + > + i915_gem_object_flush_map(batch->obj); > + i915_gem_object_unpin_map(batch->obj); > + > + rq = intel_context_create_request(ce); > + if (IS_ERR(rq)) { > + err = PTR_ERR(rq); > + goto err_vma; > + } > + > + i915_vma_lock(vma); > + err = i915_request_await_object(rq, vma->obj, true); > + if (err == 0) > + err = i915_vma_move_to_active(vma, rq, > EXEC_OBJECT_WRITE); > + i915_vma_unlock(vma); > + if (err) > + goto err_rq; > + > + i915_vma_lock(batch); > + err = i915_request_await_object(rq, batch->obj, false); > + if (err == 0) > + err = i915_vma_move_to_active(batch, rq, 0); > + i915_vma_unlock(batch); > + if (err) > + goto err_rq; > + > + err = engine->emit_bb_start(rq, batch->node.start, batch->node.size, > +0); > + > +err_rq: > + i915_request_add(rq); > +err_vma: > + i915_vma_unpin(vma); > +err_unpin_batch: > + i915_vma_unpin(batch); > +err_batch: > + i915_vma_put(batch); > +err_ce: > + intel_context_put(ce); > + return err; > +} > + > +static int iso_read(struct i915_gem_context *ctx, > + struct intel_engine_cs *engine, > + struct drm_i915_gem_object *obj, > + const struct iso_details *iso, > + u32 *ctl, > + int idx) > +{ > + struct i915_vma *batch, *vma; > + struct intel_context *ce; > + struct i915_request *rq; > + u32 *cs; > + int err; > + int i; > + > + ce = i915_gem_context_get_engine(ctx, engine->legacy_idx); > + if (IS_ERR(ce)) > + return PTR_ERR(ce); > + > + batch = create_vma(ce->vm, PAGE_ALIGN(16 * iso->count + 4)); > + if (IS_ERR(batch)) { > + err = PTR_ERR(batch); > + goto err_ce; > + } > + > + vma = i915_vma_instance(obj, ce->vm, NULL); > + if (IS_ERR(vma)) { > + err = PTR_ERR(vma); > + goto err_batch; > + } > + > + err = i915_vma_pin(batch, 0, 0, PIN_USER); > + if (err) > + goto err_batch; > + > + err = i915_vma_pin(vma, 0, 0, PIN_USER); > + if (err) > + goto err_unpin_batch; > + > + cs = i915_gem_object_pin_map(batch->obj, I915_MAP_WC); > + if (IS_ERR(cs)) { > + err = PTR_ERR(cs); > + goto err_vma; > + } > + > + for (i = 0; i < iso->count; i++) { > + u64 addr = vma->node.start + (i * __ISO__ + idx) * > sizeof(u32); > + > + *cs++ = MI_STORE_REGISTER_MEM_GEN8; > + *cs++ = ctl[i * __ISO__ + ISO_REG]; > + *cs++ = lower_32_bits(addr); > + *cs++ = upper_32_bits(addr); > + } > + *cs++ = MI_BATCH_BUFFER_END; > + > + i915_gem_object_flush_map(batch->obj); > + i915_gem_object_unpin_map(batch->obj); > + > + rq = intel_context_create_request(ce); > + if (IS_ERR(rq)) { > + err = PTR_ERR(rq); > + goto err_vma; > + } > + > + i915_vma_lock(vma); > + err = i915_request_await_object(rq, vma->obj, true); > + if (err == 0) > + err = i915_vma_move_to_active(vma, rq, > EXEC_OBJECT_WRITE); > + i915_vma_unlock(vma); > + if (err) > + goto err_rq; > + > + i915_vma_lock(batch); > + err = i915_request_await_object(rq, batch->obj, false); > + if (err == 0) > + err = i915_vma_move_to_active(batch, rq, 0); > + i915_vma_unlock(batch); > + if (err) > + goto err_rq; > + > + err = engine->emit_bb_start(rq, batch->node.start, batch->node.size, > +0); > + > +err_rq: > + i915_request_add(rq); > +err_vma: > + i915_vma_unpin(vma); > +err_unpin_batch: > + i915_vma_unpin(batch); > +err_batch: > + i915_vma_put(batch); > +err_ce: > + intel_context_put(ce); > + return err; > +} > + > +static bool is_timestamp(struct drm_i915_private *i915, u32 x) { > + struct intel_engine_cs *engine; > + > + for_each_uabi_engine(engine, i915) { > + if (x == engine->mmio_base + 0x358) What is the register offset 0x38, and below offsets (35c and 3a8) represent? I know that you know, but for many people on the same journey with you, please define macro for things like these... > + return true; > + > + if (x == engine->mmio_base + 0x35c) > + return true; > + > + if (x == engine->mmio_base + 0x3a8) > + return true; > + } > + > + return false; > +} > + > +static bool is_whitelist(struct drm_i915_private *i915, u32 x) { > + struct intel_engine_cs *engine; > + > + for_each_uabi_engine(engine, i915) { > + const struct i915_wa_list *w = &engine->whitelist; > + int i; > + > + for (i = 0; i < w->count; i++) { > + if (x == i915_mmio_reg_offset(w->list[i].reg)) > + return true; > + } > + } > + > + return false; > +} > + > +static u32 random_reg(struct drm_i915_private *i915, > + struct rnd_state *prng) > +{ > + u32 x; > + > + /* > + * Pick any u32 aligned value that is not known to a free running > + * counter (e.g. a timestamp). > + */ > + do { > + x = prandom_u32_state(prng) % 10000 * sizeof(u32); > + if (!is_timestamp(i915, x) && !is_whitelist(i915, x)) > + return x; > + } while (1); > +} > + > +static void hexdump(const void *buf, size_t len) { > + const size_t rowsize = 8 * sizeof(u32); > + const void *prev = NULL; > + bool skip = false; > + size_t pos; > + > + for (pos = 0; pos < len; pos += rowsize) { > + char line[128]; > + > + if (prev && !memcmp(prev, buf + pos, rowsize)) { > + if (!skip) { > + pr_info("*\n"); > + skip = true; > + } > + continue; > + } > + > + WARN_ON_ONCE(hex_dump_to_buffer(buf + pos, len - pos, > + rowsize, sizeof(u32), > + line, sizeof(line), > + false) >= sizeof(line)); > + pr_info("[%04zx] %s\n", pos, line); > + > + prev = buf + pos; > + skip = false; > + } > +} > + > +static bool skip_isolation(const struct intel_engine_cs *engine) { > + if (engine->class == COPY_ENGINE_CLASS && INTEL_GEN(engine- > >i915) == 9) > + return true; > + > + if (engine->class == RENDER_CLASS && INTEL_GEN(engine->i915) < > 12) > + return true; > + > + if (IS_BROADWELL(engine->i915)) > + return true; > + Any reason for skipping context isolation check for those engines and device? Little comment will make code understandable. > + return false; > +} > + > +static int igt_reg_isolation(void *arg) { > + const struct iso_details iso = { .count = 1024 }; > + struct drm_i915_private *i915 = arg; > + struct i915_gem_context *ctx_a, *ctx_b; > + struct drm_i915_gem_object *obj; > + struct intel_engine_cs *engine; > + struct igt_live_test t; > + I915_RND_STATE(prng); > + struct file *file; > + unsigned long sz; > + u32 *ctl; > + int err; > + int i; > + > + if (INTEL_GEN(i915) < 8) /* for LRM/SRM */ > + return 0; > + > + /* > + * No state that we can write to from our context should be > + * observable by another. > + */ > + > + file = mock_file(i915); > + if (IS_ERR(file)) > + return PTR_ERR(file); > + > + err = igt_live_test_begin(&t, i915, __func__, ""); > + if (err) > + goto out_file; > + > + ctx_a = live_context(i915, file); > + if (IS_ERR(ctx_a)) { > + err = PTR_ERR(ctx_a); > + goto out_file; > + } > + > + ctx_b = live_context(i915, file); > + if (IS_ERR(ctx_b)) { > + err = PTR_ERR(ctx_b); > + goto out_file; > + } > + > + sz = PAGE_ALIGN(__ISO__ * iso.count * sizeof(u32)); > + obj = i915_gem_object_create_internal(i915, sz); > + if (IS_ERR(obj)) { > + err = PTR_ERR(obj); > + goto out_file; > + } > + > + ctl = i915_gem_object_pin_map(obj, I915_MAP_WC); > + if (IS_ERR(ctl)) { > + err = PTR_ERR(ctl); > + goto out_obj; > + } > + > + for (i = 0; i < iso.count; i++) { > + ctl[i * __ISO__ + ISO_REG] = random_reg(i915, &prng); > + ctl[i * __ISO__ + ISO_POISON] = prandom_u32_state(&prng); > + } > + > + for_each_uabi_engine(engine, i915) { > + if (!IS_ENABLED(CONFIG_DRM_I915_SELFTEST_BROKEN) && > + skip_isolation(engine)) > + continue; /* Just don't even ask */ > + > + for (i = 0; i < iso.count; i++) { > + ctl[i * __ISO__ + ISO_BEFORE] = 0xdeadbeef; > + ctl[i * __ISO__ + ISO_AFTER] = 0x00c0ffee; > + } > + i915_gem_object_flush_map(obj); > + > + err = iso_read(ctx_b, engine, obj, &iso, ctl, ISO_BEFORE); > + if (err) > + break; > + > + err = iso_read(ctx_a, engine, obj, &iso, ctl, ISO_BEFORE); > + if (err) > + break; > + > + /* Twice to record after the first pristine context save */ > + err = iso_read(ctx_b, engine, obj, &iso, ctl, ISO_BEFORE); > + if (err) > + break; > + > + err = iso_write(ctx_a, engine, obj, &iso, ctl); > + if (err) > + break; > + > + err = iso_read(ctx_b, engine, obj, &iso, ctl, ISO_AFTER); > + if (err) > + break; > + > + err = i915_gem_object_wait(obj, > + I915_WAIT_ALL | > + I915_WAIT_INTERRUPTIBLE, > + HZ / 2); > + if (err) > + break; > + > + for (i = 0; i < iso.count; i++) { > + const u32 *result = &ctl[__ISO__ * i]; > + > + if (result[ISO_BEFORE] == result[ISO_AFTER]) > + continue; > + > + pr_err("%s: poison 0x%x with %08x, in:%08x, > out:%08x\n", > + engine->name, > + result[ISO_REG], > + result[ISO_POISON], > + result[ISO_BEFORE], > + result[ISO_AFTER]); > + > + /* If we read back the poison in its entirety, whoops > */ > + if (result[ISO_AFTER] == result[ISO_POISON]) > + err = -EINVAL; > + } > + if (err) { > + hexdump(ctl, __ISO__ * iso.count * sizeof(u32)); > + break; > + } > + } > + > + i915_gem_object_unpin_map(obj); > + if (igt_live_test_end(&t)) > + err = -EIO; > + > +out_obj: > + i915_gem_object_put(obj); > +out_file: > + fput(file); > + return err; > +} > + > static bool skip_unused_engines(struct intel_context *ce, void *data) { > return !ce->state; > @@ -2000,6 +2440,7 @@ int i915_gem_context_live_selftests(struct > drm_i915_private *i915) > SUBTEST(igt_ctx_sseu), > SUBTEST(igt_shared_ctx_exec), > SUBTEST(igt_vm_isolation), > + SUBTEST(igt_reg_isolation), Is this reg isolation or context isolation? Thank you, ~Akeem > }; > > if (intel_gt_is_wedged(&i915->gt)) > -- > 2.20.1 > > _______________________________________________ > Intel-gfx mailing list > Intel-gfx@xxxxxxxxxxxxxxxxxxxxx > https://lists.freedesktop.org/mailman/listinfo/intel-gfx _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx