Quoting Daniele Ceraolo Spurio (2021-02-06 02:09:19) > From: "Huang, Sean Z" <sean.z.huang@xxxxxxxxx> > > Teardown is triggered when the display topology changes and no > long meets the secure playback requirement, and hardware trashes > all the encryption keys for display. Additionally, we want to emit a > teardown operation to make sure we're clean on boot and resume > > Signed-off-by: Huang, Sean Z <sean.z.huang@xxxxxxxxx> > Signed-off-by: Daniele Ceraolo Spurio <daniele.ceraolospurio@xxxxxxxxx> > --- > drivers/gpu/drm/i915/Makefile | 1 + > drivers/gpu/drm/i915/pxp/intel_pxp_cmd.c | 227 +++++++++++++++++++ > drivers/gpu/drm/i915/pxp/intel_pxp_cmd.h | 15 ++ > drivers/gpu/drm/i915/pxp/intel_pxp_session.c | 40 ++++ > drivers/gpu/drm/i915/pxp/intel_pxp_session.h | 1 + > drivers/gpu/drm/i915/pxp/intel_pxp_tee.c | 5 +- > 6 files changed, 288 insertions(+), 1 deletion(-) > create mode 100644 drivers/gpu/drm/i915/pxp/intel_pxp_cmd.c > create mode 100644 drivers/gpu/drm/i915/pxp/intel_pxp_cmd.h > > diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile > index 8519abcf6515..9698fec810ae 100644 > --- a/drivers/gpu/drm/i915/Makefile > +++ b/drivers/gpu/drm/i915/Makefile > @@ -271,6 +271,7 @@ i915-y += i915_perf.o > # Protected execution platform (PXP) support > i915-$(CONFIG_DRM_I915_PXP) += \ > pxp/intel_pxp.o \ > + pxp/intel_pxp_cmd.o \ > pxp/intel_pxp_session.o \ > pxp/intel_pxp_tee.o > > diff --git a/drivers/gpu/drm/i915/pxp/intel_pxp_cmd.c b/drivers/gpu/drm/i915/pxp/intel_pxp_cmd.c > new file mode 100644 > index 000000000000..3e2c3580cb1b > --- /dev/null > +++ b/drivers/gpu/drm/i915/pxp/intel_pxp_cmd.c > @@ -0,0 +1,227 @@ > +// SPDX-License-Identifier: MIT > +/* > + * Copyright(c) 2020, Intel Corporation. All rights reserved. > + */ > + > +#include "intel_pxp.h" > +#include "intel_pxp_session.h" > +#include "gt/intel_context.h" > +#include "gt/intel_engine_pm.h" > +#include "gt/intel_gpu_commands.h" > +#include "gt/intel_gt_buffer_pool.h" > + > +/* PXP GPU command definitions */ > + > +/* MI_SET_APPID */ > +#define MI_SET_APPID_SESSION_ID(x) ((x) << 0) > + > +/* MI_FLUSH_DW */ > +#define MI_FLUSH_DW_DW0_PROTECTED_MEMORY_ENABLE BIT(22) > + > +/* MI_WAIT */ > +#define MFX_WAIT_DW0_PXP_SYNC_CONTROL_FLAG BIT(9) > +#define MFX_WAIT_DW0_MFX_SYNC_CONTROL_FLAG BIT(8) > + > +/* CRYPTO_KEY_EXCHANGE */ > +#define CRYPTO_KEY_EXCHANGE ((0x3 << 29) | (0x01609 << 16)) > + > +static struct i915_vma *intel_pxp_get_batch(struct intel_context *ce, > + struct i915_gem_ww_ctx *ww, > + u32 size) > +{ > + struct intel_gt_buffer_pool_node *pool; > + struct i915_vma *batch; > + int err; > + > + intel_engine_pm_get(ce->engine); > + > +retry: > + err = intel_context_pin_ww(ce, ww); > + if (err) > + goto out; > + > + pool = intel_gt_get_buffer_pool(ce->engine->gt, size, I915_MAP_WC); > + if (IS_ERR(pool)) { > + err = PTR_ERR(pool); > + goto out_ctx; > + } > + > + batch = i915_vma_instance(pool->obj, ce->vm, NULL); > + if (IS_ERR(batch)) { > + err = PTR_ERR(batch); > + goto out_put; > + } > + > + err = i915_vma_pin_ww(batch, ww, 0, 0, PIN_USER); > + if (unlikely(err)) > + goto out_put; > + > + err = i915_gem_object_lock(pool->obj, ww); > + if (err) > + goto out_unpin; > + > + batch->private = pool; > + > + return batch; > + > +out_unpin: > + i915_vma_unpin(batch); > +out_put: > + intel_gt_buffer_pool_put(pool); > +out_ctx: > + intel_context_unpin(ce); > +out: > + if (err == -EDEADLK) { > + err = i915_gem_ww_ctx_backoff(ww); > + if (!err) > + goto retry; > + } > + intel_engine_pm_put(ce->engine); > + return ERR_PTR(err); > +} > + > +static void intel_pxp_put_batch(struct intel_context *ce, > + struct i915_vma *batch) > +{ > + i915_vma_unpin(batch); > + intel_gt_buffer_pool_put(batch->private); > + intel_context_unpin(ce); > + intel_engine_pm_put(ce->engine); > +} > + > +static int intel_pxp_submit_batch(struct intel_context *ce, > + struct i915_vma *batch) > +{ > + struct i915_request *rq; > + int err; > + > + rq = i915_request_create(ce); > + if (IS_ERR(rq)) > + return PTR_ERR(rq); > + > + err = i915_request_await_object(rq, batch->obj, false); > + if (!err) > + err = i915_vma_move_to_active(batch, rq, 0); > + if (err) > + goto out_rq; > + > + err = intel_gt_buffer_pool_mark_active(batch->private, rq); > + if (err) > + goto out_rq; > + > + if (ce->engine->emit_init_breadcrumb) { > + err = ce->engine->emit_init_breadcrumb(rq); > + if (err) > + goto out_rq; > + } > + > + err = ce->engine->emit_bb_start(rq, batch->node.start, > + batch->node.size, 0); > + if (err) > + goto out_rq; > + > +out_rq: > + i915_request_get(rq); > + > + if (unlikely(err)) > + i915_request_set_error_once(rq, err); > + > + i915_request_add(rq); > + > + if (!err && i915_request_wait(rq, 0, HZ / 5) < 0) > + err = -ETIME; > + > + i915_request_put(rq); > + > + return err; > +} > + > +/* stall until prior PXP and MFX/HCP/HUC objects are cmopleted */ > +#define MFX_WAIT_PXP \ > + MFX_WAIT | \ > + MFX_WAIT_DW0_PXP_SYNC_CONTROL_FLAG | \ > + MFX_WAIT_DW0_MFX_SYNC_CONTROL_FLAG; > + > +static u32 *pxp_emit_session_selection(u32 *cmd, u32 idx) > +{ > + *cmd++ = MFX_WAIT_PXP; > + > + /* pxp off */ > + *cmd++ = MI_FLUSH_DW; > + *cmd++ = 0; > + *cmd++ = 0; > + > + /* select session */ > + *cmd++ = MI_SET_APPID | MI_SET_APPID_SESSION_ID(idx); > + > + *cmd++ = MFX_WAIT_PXP; > + > + /* pxp on */ > + *cmd++ = MI_FLUSH_DW | MI_FLUSH_DW_DW0_PROTECTED_MEMORY_ENABLE; > + *cmd++ = 0; > + *cmd++ = 0; > + > + *cmd++ = MFX_WAIT_PXP; > + > + return cmd; > +} > + > +static u32 *pxp_emit_inline_termination(u32 *cmd) > +{ > + /* session inline termination */ > + *cmd++ = CRYPTO_KEY_EXCHANGE; > + *cmd++ = 0; > + > + return cmd; > +} > + > +static u32 *pxp_emit_batch_end(u32 *cmd) > +{ > + /* wait for cmds to go through */ > + *cmd++ = MFX_WAIT_PXP; > + > + *cmd++ = MI_BATCH_BUFFER_END; > + > + return cmd; > +} > + > +int intel_pxp_submit_session_termination(struct intel_pxp *pxp, u32 id) > +{ > + struct i915_vma *batch; > + struct i915_gem_ww_ctx ww; > + u32 *cmd; > + int err; > + > + if (!intel_pxp_is_enabled(pxp)) > + return 0; > + > + i915_gem_ww_ctx_init(&ww, false); > + > + batch = intel_pxp_get_batch(pxp->ce, &ww, PAGE_SIZE); Is there any reason at all to use the batch and not just emit directly into the ring? The command sequence is short. And you probably want to disable arbitration. Does this invalidation need priority for immediate execution? -Chris _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx