to improve efficiency and reduce the complexsity of vgpu ggtt support, vgpu ggtt page table operations are implemented in pv fashion and implemented pv version of bind/unbind for ggtt vma ops. The pv version of ggtt vma ops use the CTB protocol to communicate pv ggtt command along with data struct pv_vma from guest to GVT and then GVT will implement command handler of PV_CMD_BIND_GGTT and PV_CMD_UBIND_gGTT to support vgpu GGTT feature. new PV_GGTT pv_cap is used to control this level of pv support in both guest and host side. Signed-off-by: Xiaolin Zhang <xiaolin.zhang@xxxxxxxxx> --- drivers/gpu/drm/i915/i915_gem.c | 4 +++- drivers/gpu/drm/i915/i915_vgpu.c | 37 ++++++++++++++++++++++++++++++++++++- drivers/gpu/drm/i915/i915_vgpu.h | 3 +++ 3 files changed, 42 insertions(+), 2 deletions(-) diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c index bb0c129..77cd09b 100644 --- a/drivers/gpu/drm/i915/i915_gem.c +++ b/drivers/gpu/drm/i915/i915_gem.c @@ -1129,9 +1129,11 @@ int i915_gem_init(struct drm_i915_private *dev_priv) int ret; /* We need to fallback to 4K pages if host doesn't support huge gtt. */ - if (intel_vgpu_active(dev_priv) && !intel_vgpu_has_huge_gtt(dev_priv)) + if (intel_vgpu_active(dev_priv)) { mkwrite_device_info(dev_priv)->page_sizes = I915_GTT_PAGE_SIZE_4K; + intel_vgpu_config_pv_caps(dev_priv, PV_GGTT, &dev_priv->ggtt); + } ret = i915_gem_init_userptr(dev_priv); if (ret) diff --git a/drivers/gpu/drm/i915/i915_vgpu.c b/drivers/gpu/drm/i915/i915_vgpu.c index 9875e2f..4e50694 100644 --- a/drivers/gpu/drm/i915/i915_vgpu.c +++ b/drivers/gpu/drm/i915/i915_vgpu.c @@ -100,7 +100,7 @@ void intel_vgpu_detect(struct drm_i915_private *dev_priv) mutex_init(&dev_priv->vgpu.lock); /* guest driver PV capability */ - dev_priv->vgpu.pv_caps = PV_PPGTT; + dev_priv->vgpu.pv_caps = PV_PPGTT | PV_GGTT; if (!intel_vgpu_detect_pv_caps(dev_priv, shared_area)) { DRM_INFO("Virtual GPU for Intel GVT-g detected.\n"); @@ -458,6 +458,34 @@ static void ppgtt_unbind_vma_pv(struct i915_address_space *vm, vgpu_pv_vma_action(vm, vma, PV_CMD_UNBIND_PPGTT, 0, 0); } +static void ggtt_bind_vma_pv(struct i915_address_space *vm, + struct i915_vm_pt_stash *stash, + struct i915_vma *vma, + enum i915_cache_level cache_level, + u32 flags) +{ + struct drm_i915_gem_object *obj = vma->obj; + u32 pte_flags; + + if (i915_vma_is_bound(vma, ~flags & I915_VMA_BIND_MASK)) + return; + + /* Applicable to VLV (gen8+ do not support RO in the GGTT) */ + pte_flags = 0; + if (i915_gem_object_is_readonly(obj)) + pte_flags |= PTE_READ_ONLY; + + pte_flags = vma->vm->pte_encode(0, cache_level, 0); + vgpu_pv_vma_action(vm, vma, PV_CMD_BIND_GGTT, 0, pte_flags); + vma->page_sizes.gtt = I915_GTT_PAGE_SIZE; +} + +static void ggtt_unbind_vma_pv_nop(struct i915_address_space *vm, + struct i915_vma *vma) +{ + +} + /** * wait_for_desc_update - Wait for the command buffer descriptor update. * @desc: buffer descriptor @@ -733,6 +761,7 @@ void intel_vgpu_config_pv_caps(struct drm_i915_private *i915, enum pv_caps cap, void *data) { struct i915_ppgtt *ppgtt; + struct i915_ggtt *ggtt; if (!intel_vgpu_check_pv_cap(i915, cap)) return; @@ -742,6 +771,12 @@ void intel_vgpu_config_pv_caps(struct drm_i915_private *i915, ppgtt->vm.vma_ops.bind_vma = ppgtt_bind_vma_pv; ppgtt->vm.vma_ops.unbind_vma = ppgtt_unbind_vma_pv; } + + if (cap == PV_GGTT) { + ggtt = (struct i915_ggtt *)data; + ggtt->vm.vma_ops.bind_vma = ggtt_bind_vma_pv; + ggtt->vm.vma_ops.unbind_vma = ggtt_unbind_vma_pv_nop; + } } /** diff --git a/drivers/gpu/drm/i915/i915_vgpu.h b/drivers/gpu/drm/i915/i915_vgpu.h index 7e4ea99..588e361 100644 --- a/drivers/gpu/drm/i915/i915_vgpu.h +++ b/drivers/gpu/drm/i915/i915_vgpu.h @@ -38,6 +38,7 @@ struct i915_ggtt; enum pv_caps { PV_NONE = 0, PV_PPGTT = BIT(0), + PV_GGTT = BIT(1), }; /* vgpu PV commands */ @@ -45,6 +46,8 @@ enum intel_vgpu_pv_cmd { PV_CMD_DEFAULT = 0x0, PV_CMD_BIND_PPGTT, PV_CMD_UNBIND_PPGTT, + PV_CMD_BIND_GGTT, + PV_CMD_UNBIND_GGTT, }; /* A common shared page(4KB) between GVTg and vgpu allocated by guest */ -- 2.7.4 _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx