Re: [PATCH 05/16] drm/i915/vm_bind: Implement bind and unbind of object

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, Sep 29, 2022 at 06:28:39PM +0100, Matthew Auld wrote:
On 29/09/2022 17:38, Niranjana Vishwanathapura wrote:
On Thu, Sep 29, 2022 at 11:49:30AM +0100, Matthew Auld wrote:
On 28/09/2022 07:19, Niranjana Vishwanathapura wrote:
Add uapi and implement support for bind and unbind of an
object at the specified GPU virtual addresses.

The vm_bind mode is not supported in legacy execbuf2 ioctl.
It will be supported only in the newer execbuf3 ioctl.

Signed-off-by: Niranjana Vishwanathapura <niranjana.vishwanathapura@xxxxxxxxx>
Signed-off-by: Prathap Kumar Valsan <prathap.kumar.valsan@xxxxxxxxx>
Signed-off-by: Andi Shyti <andi.shyti@xxxxxxxxxxxxxxx>
---
 drivers/gpu/drm/i915/Makefile                 |   1 +
 .../gpu/drm/i915/gem/i915_gem_execbuffer.c    |   5 +
 drivers/gpu/drm/i915/gem/i915_gem_vm_bind.h   |  26 ++
 .../drm/i915/gem/i915_gem_vm_bind_object.c    | 306 ++++++++++++++++++
 drivers/gpu/drm/i915/gt/intel_gtt.c           |  10 +
 drivers/gpu/drm/i915/gt/intel_gtt.h           |  17 +
 drivers/gpu/drm/i915/i915_driver.c            |   3 +
 drivers/gpu/drm/i915/i915_vma.c               |   1 +
 drivers/gpu/drm/i915/i915_vma_types.h         |  14 +
 include/uapi/drm/i915_drm.h                   | 112 +++++++
 10 files changed, 495 insertions(+)
 create mode 100644 drivers/gpu/drm/i915/gem/i915_gem_vm_bind.h
 create mode 100644 drivers/gpu/drm/i915/gem/i915_gem_vm_bind_object.c

diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
index a26edcdadc21..9bf939ef18ea 100644
--- a/drivers/gpu/drm/i915/Makefile
+++ b/drivers/gpu/drm/i915/Makefile
@@ -166,6 +166,7 @@ gem-y += \
     gem/i915_gem_ttm_move.o \
     gem/i915_gem_ttm_pm.o \
     gem/i915_gem_userptr.o \
+    gem/i915_gem_vm_bind_object.o \
     gem/i915_gem_wait.o \
     gem/i915_gemfs.o
 i915-y += \
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
index cd75b0ca2555..f85f10cf9c34 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
@@ -781,6 +781,11 @@ static int eb_select_context(struct i915_execbuffer *eb)
     if (unlikely(IS_ERR(ctx)))
         return PTR_ERR(ctx);
+    if (ctx->vm->vm_bind_mode) {
+        i915_gem_context_put(ctx);
+        return -EOPNOTSUPP;
+    }
+
     eb->gem_context = ctx;
     if (i915_gem_context_has_full_ppgtt(ctx))
         eb->invalid_flags |= EXEC_OBJECT_NEEDS_GTT;
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_vm_bind.h b/drivers/gpu/drm/i915/gem/i915_gem_vm_bind.h
new file mode 100644
index 000000000000..36262a6357b5
--- /dev/null
+++ b/drivers/gpu/drm/i915/gem/i915_gem_vm_bind.h
@@ -0,0 +1,26 @@
+/* SPDX-License-Identifier: MIT */
+/*
+ * Copyright © 2022 Intel Corporation
+ */
+
+#ifndef __I915_GEM_VM_BIND_H
+#define __I915_GEM_VM_BIND_H
+
+#include <linux/types.h>
+
+struct drm_device;
+struct drm_file;
+struct i915_address_space;
+struct i915_vma;
+
+struct i915_vma *
+i915_gem_vm_bind_lookup_vma(struct i915_address_space *vm, u64 va);
+
+int i915_gem_vm_bind_ioctl(struct drm_device *dev, void *data,
+               struct drm_file *file);
+int i915_gem_vm_unbind_ioctl(struct drm_device *dev, void *data,
+                 struct drm_file *file);
+
+void i915_gem_vm_unbind_all(struct i915_address_space *vm);
+
+#endif /* __I915_GEM_VM_BIND_H */
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_vm_bind_object.c b/drivers/gpu/drm/i915/gem/i915_gem_vm_bind_object.c
new file mode 100644
index 000000000000..e529162abd2c
--- /dev/null
+++ b/drivers/gpu/drm/i915/gem/i915_gem_vm_bind_object.c
@@ -0,0 +1,306 @@
+// SPDX-License-Identifier: MIT
+/*
+ * Copyright © 2022 Intel Corporation
+ */
+
+#include <uapi/drm/i915_drm.h>
+
+#include <linux/interval_tree_generic.h>
+
+#include "gem/i915_gem_context.h"
+#include "gem/i915_gem_vm_bind.h"
+
+#include "gt/intel_gpu_commands.h"
+
+#define START(node) ((node)->start)
+#define LAST(node) ((node)->last)
+
+INTERVAL_TREE_DEFINE(struct i915_vma, rb, u64, __subtree_last,
+             START, LAST, static inline, i915_vm_bind_it)
+
+#undef START
+#undef LAST
+
+/**
+ * DOC: VM_BIND/UNBIND ioctls
+ *
+ * DRM_I915_GEM_VM_BIND/UNBIND ioctls allows UMD to bind/unbind GEM buffer + * objects (BOs) or sections of a BOs at specified GPU virtual addresses on a + * specified address space (VM). Multiple mappings can map to the same physical + * pages of an object (aliasing). These mappings (also referred to as persistent + * mappings) will be persistent across multiple GPU submissions (execbuf calls) + * issued by the UMD, without user having to provide a list of all required
+ * mappings during each submission (as required by older execbuf mode).
+ *
+ * The VM_BIND/UNBIND calls allow UMDs to request a timeline out fence for
+ * signaling the completion of bind/unbind operation.
+ *
+ * VM_BIND feature is advertised to user via I915_PARAM_VM_BIND_VERSION. + * User has to opt-in for VM_BIND mode of binding for an address space (VM) + * during VM creation time via I915_VM_CREATE_FLAGS_USE_VM_BIND extension.
+ *
+ * VM_BIND/UNBIND ioctl calls executed on different CPU threads concurrently + * are not ordered. Furthermore, parts of the VM_BIND/UNBIND operations can be
+ * done asynchronously, when valid out fence is specified.
+ *
+ * VM_BIND locking order is as below.
+ *
+ * 1) vm_bind_lock mutex will protect vm_bind lists. This lock is taken in + *    vm_bind/vm_unbind ioctl calls, in the execbuf path and while releasing the
+ *    mapping.
+ *
+ *    In future, when GPU page faults are supported, we can potentially use a + *    rwsem instead, so that multiple page fault handlers can take the read
+ *    side lock to lookup the mapping and hence can run in parallel.
+ *    The older execbuf mode of binding do not need this lock.
+ *
+ * 2) The object's dma-resv lock will protect i915_vma state and needs
+ *    to be held while binding/unbinding a vma in the async worker and while + *    updating dma-resv fence list of an object. Note that private BOs of a VM
+ *    will all share a dma-resv object.
+ *
+ * 3) Spinlock/s to protect some of the VM's lists like the list of
+ *    invalidated vmas (due to eviction and userptr invalidation) etc.
+ */
+
+/**
+ * i915_gem_vm_bind_lookup_vma() - lookup for persistent vma mapped at a
+ * specified address
+ * @vm: virtual address space to look for persistent vma
+ * @va: starting address where vma is mapped
+ *
+ * Retrieves the persistent vma mapped address @va from the @vm's vma tree.
+ *
+ * Returns vma pointer on success, NULL on failure.
+ */
+struct i915_vma *
+i915_gem_vm_bind_lookup_vma(struct i915_address_space *vm, u64 va)
+{
+    lockdep_assert_held(&vm->vm_bind_lock);
+
+    return i915_vm_bind_it_iter_first(&vm->va, va, va);
+}
+
+static void i915_gem_vm_bind_remove(struct i915_vma *vma, bool release_obj)
+{
+    lockdep_assert_held(&vma->vm->vm_bind_lock);
+
+    list_del_init(&vma->vm_bind_link);
+    i915_vm_bind_it_remove(vma, &vma->vm->va);
+
+    /* Release object */
+    if (release_obj)
+        i915_gem_object_put(vma->obj);
+}
+
+static int i915_gem_vm_unbind_vma(struct i915_address_space *vm,
+                  struct drm_i915_gem_vm_unbind *va)
+{
+    struct drm_i915_gem_object *obj;
+    struct i915_vma *vma;
+    int ret;
+
+    ret = mutex_lock_interruptible(&vm->vm_bind_lock);
+    if (ret)
+        return ret;
+
+    va->start = gen8_noncanonical_addr(va->start);
+    vma = i915_gem_vm_bind_lookup_vma(vm, va->start);
+
+    if (!vma)
+        ret = -ENOENT;
+    else if (vma->size != va->length)
+        ret = -EINVAL;
+
+    if (ret) {
+        mutex_unlock(&vm->vm_bind_lock);
+        return ret;
+    }
+
+    i915_gem_vm_bind_remove(vma, false);
+
+    mutex_unlock(&vm->vm_bind_lock);
+
+    /* Destroy vma and then release object */
+    obj = vma->obj;

No locks held here when touching vma->obj? Maybe a small comment explaining the safety would be good.


We do holding the object reference, so we should be good.
It is documented in vm_bind call when we take object reference.

Yeah, so that stops the object from disappearing, but what about the vma itself, when chasing the pointer? Can that only happen if the entire vm is nuked? So is it actually only safe since we are holding the vm ref here, and not the object ref?


Yah, a persistent vma can only be destroyed by either by vm_unbind
call or when the VM itself is released (as we hold object reference).
As we hold the VM reference here, it is safe to access the vma.

Regards,
Niranjana


+    ret = i915_gem_object_lock(obj, NULL);

lock_interruptible() ?

Actually, we shouldn't be checking for return error here.
We are not taking interruptible lock so that we don't have to
deal with having an error.


+    if (ret)
+        return ret;

IIUC we have already called bind_remove() above, so vm_bind_lookup() can't find it any more? Is the user meant to call this again, if we can't grab the object lock? Are we not leaking the below object_put() until the entire driver fd is nuked? Can we not grab the vm_bind_lock and then object_lock, before calling the bind_remove()? And then on lock error userspace can just call this again? Sorry if I'm misunderstanding something here.

As mentioned above, we don't expect any error here and I will remove
this error check here. That should fix the problem.
Yah, I guess we can take the object lock also before calling bind_remove()
as well, but I think it is not needed.


+
+    i915_vma_destroy(vma);
+    i915_gem_object_unlock(obj);
+
+    i915_gem_object_put(obj);
+
+    return 0;
+}
+
+/**
+ * i915_gem_vm_unbind_all() - unbind all persistent mappings from an
+ * address space
+ * @vm: Address spece to remove persistent mappings from
+ *
+ * Unbind all userspace requested vm_bind mappings from @vm.
+ */
+void i915_gem_vm_unbind_all(struct i915_address_space *vm)
+{
+    struct i915_vma *vma, *t;
+
+    mutex_lock(&vm->vm_bind_lock);
+    list_for_each_entry_safe(vma, t, &vm->vm_bind_list, vm_bind_link)
+        i915_gem_vm_bind_remove(vma, true);
+    list_for_each_entry_safe(vma, t, &vm->vm_bound_list, vm_bind_link)
+        i915_gem_vm_bind_remove(vma, true);
+    mutex_unlock(&vm->vm_bind_lock);
+}
+
+static struct i915_vma *vm_bind_get_vma(struct i915_address_space *vm,
+                    struct drm_i915_gem_object *obj,
+                    struct drm_i915_gem_vm_bind *va)
+{
+    struct i915_gtt_view view;
+    struct i915_vma *vma;
+
+    va->start = gen8_noncanonical_addr(va->start);
+    vma = i915_gem_vm_bind_lookup_vma(vm, va->start);
+    if (vma)
+        return ERR_PTR(-EEXIST);
+
+    view.type = I915_GTT_VIEW_PARTIAL;
+    view.partial.offset = va->offset >> PAGE_SHIFT;
+    view.partial.size = va->length >> PAGE_SHIFT;
+    vma = i915_vma_create_persistent(obj, vm, &view);
+    if (IS_ERR(vma))
+        return vma;
+
+    vma->start = va->start;
+    vma->last = va->start + va->length - 1;
+
+    return vma;
+}
+
+static int i915_gem_vm_bind_obj(struct i915_address_space *vm,
+                struct drm_i915_gem_vm_bind *va,
+                struct drm_file *file)
+{
+    struct drm_i915_gem_object *obj;
+    struct i915_vma *vma = NULL;
+    struct i915_gem_ww_ctx ww;
+    u64 pin_flags;
+    int ret = 0;
+
+    if (!vm->vm_bind_mode)
+        return -EOPNOTSUPP;

Also need to check MBZ on the extensions somewhere.


Ok, will do.

+
+    obj = i915_gem_object_lookup(file, va->handle);
+    if (!obj)
+        return -ENOENT;
+
+    if (!va->length ||
+        !IS_ALIGNED(va->offset | va->length,
+            i915_gem_object_max_page_size(obj->mm.placements,
+                              obj->mm.n_placements)) ||
+        range_overflows_t(u64, va->offset, va->length, obj->base.size)) {

Nit: could maybe split this into multiple different checks for better readability.


Ok, let me see.

+        ret = -EINVAL;
+        goto put_obj;
+    }
+
+    ret = mutex_lock_interruptible(&vm->vm_bind_lock);
+    if (ret)
+        goto put_obj;
+
+    vma = vm_bind_get_vma(vm, obj, va);
+    if (IS_ERR(vma)) {
+        ret = PTR_ERR(vma);
+        goto unlock_vm;
+    }
+
+    pin_flags = va->start | PIN_OFFSET_FIXED | PIN_USER;

va->start for sure needs some sanity checking first. Evil user could be setting the low bits here (0-11), which would cause havoc with our vma flags.


Ok, will do.

Regards,
Niranjana

+
+    for_i915_gem_ww(&ww, ret, true) {
+        ret = i915_gem_object_lock(vma->obj, &ww);
+        if (ret)
+            continue;
+
+        ret = i915_vma_pin_ww(vma, &ww, 0, 0, pin_flags);
+        if (ret)
+            continue;
+
+        /* Make it evictable */
+        __i915_vma_unpin(vma);
+
+        list_add_tail(&vma->vm_bind_link, &vm->vm_bound_list);
+        i915_vm_bind_it_insert(vma, &vm->va);
+
+        /* Hold object reference until vm_unbind */
+        i915_gem_object_get(vma->obj);
+    }
+
+    if (ret)
+        i915_vma_destroy(vma);
+unlock_vm:
+    mutex_unlock(&vm->vm_bind_lock);
+put_obj:
+    i915_gem_object_put(obj);
+
+    return ret;
+}
+
+/**
+ * i915_gem_vm_bind_ioctl() - ioctl function for binding a section of object
+ * at a specified virtual address
+ * @dev: drm_device pointer
+ * @data: ioctl data structure
+ * @file: drm_file pointer
+ *
+ * Adds the specified persistent mapping (virtual address to a section of an
+ * object) and binds it in the device page table.
+ *
+ * Returns 0 on success, error code on failure.
+ */
+int i915_gem_vm_bind_ioctl(struct drm_device *dev, void *data,
+               struct drm_file *file)
+{
+    struct drm_i915_gem_vm_bind *args = data;
+    struct i915_address_space *vm;
+    int ret;
+
+    vm = i915_gem_vm_lookup(file->driver_priv, args->vm_id);
+    if (unlikely(!vm))
+        return -ENOENT;
+
+    ret = i915_gem_vm_bind_obj(vm, args, file);
+
+    i915_vm_put(vm);
+    return ret;
+}
+
+/**
+ * i915_gem_vm_unbind_ioctl() - ioctl function for unbinding a mapping at a
+ * specified virtual address
+ * @dev: drm_device pointer
+ * @data: ioctl data structure
+ * @file: drm_file pointer
+ *
+ * Removes the persistent mapping at the specified address and unbinds it
+ * from the device page table.
+ *
+ * Returns 0 on success, error code on failure. -ENOENT is returned if the
+ * specified mapping is not found.
+ */
+int i915_gem_vm_unbind_ioctl(struct drm_device *dev, void *data,
+                 struct drm_file *file)
+{
+    struct drm_i915_gem_vm_unbind *args = data;
+    struct i915_address_space *vm;
+    int ret;
+
+    vm = i915_gem_vm_lookup(file->driver_priv, args->vm_id);
+    if (unlikely(!vm))
+        return -ENOENT;
+
+    ret = i915_gem_vm_unbind_vma(vm, args);
+
+    i915_vm_put(vm);
+    return ret;
+}
diff --git a/drivers/gpu/drm/i915/gt/intel_gtt.c b/drivers/gpu/drm/i915/gt/intel_gtt.c
index b67831833c9a..0daa70c6ed0d 100644
--- a/drivers/gpu/drm/i915/gt/intel_gtt.c
+++ b/drivers/gpu/drm/i915/gt/intel_gtt.c
@@ -12,6 +12,7 @@
 #include "gem/i915_gem_internal.h"
 #include "gem/i915_gem_lmem.h"
+#include "gem/i915_gem_vm_bind.h"
 #include "i915_trace.h"
 #include "i915_utils.h"
 #include "intel_gt.h"
@@ -176,6 +177,8 @@ int i915_vm_lock_objects(struct i915_address_space *vm,
 void i915_address_space_fini(struct i915_address_space *vm)
 {
     drm_mm_takedown(&vm->mm);
+    GEM_BUG_ON(!RB_EMPTY_ROOT(&vm->va.rb_root));
+    mutex_destroy(&vm->vm_bind_lock);
 }
 /**
@@ -202,6 +205,8 @@ static void __i915_vm_release(struct work_struct *work)
     struct i915_address_space *vm =
         container_of(work, struct i915_address_space, release_work);
+    i915_gem_vm_unbind_all(vm);
+
     __i915_vm_close(vm);
     /* Synchronize async unbinds. */
@@ -282,6 +287,11 @@ void i915_address_space_init(struct i915_address_space *vm, int subclass)
     INIT_LIST_HEAD(&vm->bound_list);
     INIT_LIST_HEAD(&vm->unbound_list);
+
+    vm->va = RB_ROOT_CACHED;
+    INIT_LIST_HEAD(&vm->vm_bind_list);
+    INIT_LIST_HEAD(&vm->vm_bound_list);
+    mutex_init(&vm->vm_bind_lock);
 }
 void *__px_vaddr(struct drm_i915_gem_object *p)
diff --git a/drivers/gpu/drm/i915/gt/intel_gtt.h b/drivers/gpu/drm/i915/gt/intel_gtt.h
index c0ca53cba9f0..b52061858161 100644
--- a/drivers/gpu/drm/i915/gt/intel_gtt.h
+++ b/drivers/gpu/drm/i915/gt/intel_gtt.h
@@ -259,6 +259,23 @@ struct i915_address_space {
      */
     struct list_head unbound_list;
+    /**
+     * @vm_bind_mode: flag to indicate vm_bind method of binding
+     *
+     * True: allow only vm_bind method of binding.
+     * False: allow only legacy execbuff method of binding.
+     */
+    bool vm_bind_mode:1;
+
+    /** @vm_bind_lock: Mutex to protect @vm_bind_list and @vm_bound_list */
+    struct mutex vm_bind_lock;
+    /** @vm_bind_list: List of vm_binding in process */
+    struct list_head vm_bind_list;
+    /** @vm_bound_list: List of vm_binding completed */
+    struct list_head vm_bound_list;
+    /* @va: tree of persistent vmas */
+    struct rb_root_cached va;
+
     /* Global GTT */
     bool is_ggtt:1;
diff --git a/drivers/gpu/drm/i915/i915_driver.c b/drivers/gpu/drm/i915/i915_driver.c
index fb3826dabe8b..c3a9a5031cdb 100644
--- a/drivers/gpu/drm/i915/i915_driver.c
+++ b/drivers/gpu/drm/i915/i915_driver.c
@@ -69,6 +69,7 @@
 #include "gem/i915_gem_ioctls.h"
 #include "gem/i915_gem_mman.h"
 #include "gem/i915_gem_pm.h"
+#include "gem/i915_gem_vm_bind.h"
 #include "gt/intel_gt.h"
 #include "gt/intel_gt_pm.h"
 #include "gt/intel_rc6.h"
@@ -1893,6 +1894,8 @@ static const struct drm_ioctl_desc i915_ioctls[] = {
     DRM_IOCTL_DEF_DRV(I915_QUERY, i915_query_ioctl, DRM_RENDER_ALLOW),
     DRM_IOCTL_DEF_DRV(I915_GEM_VM_CREATE, i915_gem_vm_create_ioctl, DRM_RENDER_ALLOW),      DRM_IOCTL_DEF_DRV(I915_GEM_VM_DESTROY, i915_gem_vm_destroy_ioctl, DRM_RENDER_ALLOW), +    DRM_IOCTL_DEF_DRV(I915_GEM_VM_BIND, i915_gem_vm_bind_ioctl, DRM_RENDER_ALLOW), +    DRM_IOCTL_DEF_DRV(I915_GEM_VM_UNBIND, i915_gem_vm_unbind_ioctl, DRM_RENDER_ALLOW),
 };
 /*
diff --git a/drivers/gpu/drm/i915/i915_vma.c b/drivers/gpu/drm/i915/i915_vma.c
index 5839e1f55f00..33f910473263 100644
--- a/drivers/gpu/drm/i915/i915_vma.c
+++ b/drivers/gpu/drm/i915/i915_vma.c
@@ -239,6 +239,7 @@ vma_create(struct drm_i915_gem_object *obj,
     spin_unlock(&obj->vma.lock);
     mutex_unlock(&vm->mutex);
+    INIT_LIST_HEAD(&vma->vm_bind_link);
     return vma;
 err_unlock:
diff --git a/drivers/gpu/drm/i915/i915_vma_types.h b/drivers/gpu/drm/i915/i915_vma_types.h
index 2200f1f103ba..f56ac07c6cfa 100644
--- a/drivers/gpu/drm/i915/i915_vma_types.h
+++ b/drivers/gpu/drm/i915/i915_vma_types.h
@@ -296,6 +296,20 @@ struct i915_vma {
     /** This object's place on the active/inactive lists */
     struct list_head vm_link;
+    /** @vm_bind_link: node for the vm_bind related lists of vm */
+    struct list_head vm_bind_link;
+
+    /** Interval tree structures for persistent vma */
+
+    /** @rb: node for the interval tree of vm for persistent vmas */
+    struct rb_node rb;
+    /** @start: start endpoint of the rb node */
+    u64 start;
+    /** @last: Last endpoint of the rb node */
+    u64 last;
+    /** @__subtree_last: last in subtree */
+    u64 __subtree_last;
+
     struct list_head obj_link; /* Link in the object's VMA list */
     struct rb_node obj_node;
     struct hlist_node obj_hash;
diff --git a/include/uapi/drm/i915_drm.h b/include/uapi/drm/i915_drm.h
index 520ad2691a99..9760564b4693 100644
--- a/include/uapi/drm/i915_drm.h
+++ b/include/uapi/drm/i915_drm.h
@@ -470,6 +470,8 @@ typedef struct _drm_i915_sarea {
 #define DRM_I915_GEM_VM_CREATE        0x3a
 #define DRM_I915_GEM_VM_DESTROY        0x3b
 #define DRM_I915_GEM_CREATE_EXT        0x3c
+#define DRM_I915_GEM_VM_BIND        0x3d
+#define DRM_I915_GEM_VM_UNBIND        0x3e
 /* Must be kept compact -- no holes */
 #define DRM_IOCTL_I915_INIT        DRM_IOW( DRM_COMMAND_BASE + DRM_I915_INIT, drm_i915_init_t)
@@ -534,6 +536,8 @@ typedef struct _drm_i915_sarea {
 #define DRM_IOCTL_I915_QUERY            DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_QUERY, struct drm_i915_query)  #define DRM_IOCTL_I915_GEM_VM_CREATE    DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_CREATE, struct drm_i915_gem_vm_control)  #define DRM_IOCTL_I915_GEM_VM_DESTROY    DRM_IOW (DRM_COMMAND_BASE + DRM_I915_GEM_VM_DESTROY, struct drm_i915_gem_vm_control) +#define DRM_IOCTL_I915_GEM_VM_BIND    DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_BIND, struct drm_i915_gem_vm_bind) +#define DRM_IOCTL_I915_GEM_VM_UNBIND    DRM_IOWR(DRM_COMMAND_BASE + DRM_I915_GEM_VM_UNBIND, struct drm_i915_gem_vm_unbind)
 /* Allow drivers to submit batchbuffers directly to hardware, relying
  * on the security mechanisms provided by hardware.
@@ -3717,6 +3721,114 @@ struct drm_i915_gem_create_ext_protected_content {  /* ID of the protected content session managed by i915 when PXP is active */
 #define I915_PROTECTED_CONTENT_DEFAULT_SESSION 0xf
+/**
+ * struct drm_i915_gem_vm_bind - VA to object mapping to bind.
+ *
+ * This structure is passed to VM_BIND ioctl and specifies the mapping of GPU + * virtual address (VA) range to the section of an object that should be bound
+ * in the device page table of the specified address space (VM).
+ * The VA range specified must be unique (ie., not currently bound) and can + * be mapped to whole object or a section of the object (partial binding). + * Multiple VA mappings can be created to the same section of the object
+ * (aliasing).
+ *
+ * The @start, @offset and @length must be 4K page aligned. However the DG2 + * and XEHPSDV has 64K page size for device local memory and has compact page + * table. On those platforms, for binding device local-memory objects, the + * @start, @offset and @length must be 64K aligned. Also, UMDs should not mix + * the local memory 64K page and the system memory 4K page bindings in the same
+ * 2M range.
+ *
+ * Error code -EINVAL will be returned if @start, @offset and @length are not + * properly aligned. In version 1 (See I915_PARAM_VM_BIND_VERSION), error code + * -ENOSPC will be returned if the VA range specified can't be reserved.
+ *
+ * VM_BIND/UNBIND ioctl calls executed on different CPU threads concurrently + * are not ordered. Furthermore, parts of the VM_BIND operation can be done
+ * asynchronously, if valid @fence is specified.
+ */
+struct drm_i915_gem_vm_bind {
+    /** @vm_id: VM (address space) id to bind */
+    __u32 vm_id;
+
+    /** @handle: Object handle */
+    __u32 handle;
+
+    /** @start: Virtual Address start to bind */
+    __u64 start;
+
+    /** @offset: Offset in object to bind */
+    __u64 offset;
+
+    /** @length: Length of mapping to bind */
+    __u64 length;
+
+    /**
+     * @flags: Currently reserved, MBZ.
+     *
+     * Note that @fence carries its own flags.
+     */
+    __u64 flags;
+
+    /** @rsvd: Reserved, MBZ */
+    __u64 rsvd[2];
+
+    /**
+     * @extensions: Zero-terminated chain of extensions.
+     *
+     * For future extensions. See struct i915_user_extension.
+     */
+    __u64 extensions;
+};
+
+/**
+ * struct drm_i915_gem_vm_unbind - VA to object mapping to unbind.
+ *
+ * This structure is passed to VM_UNBIND ioctl and specifies the GPU virtual + * address (VA) range that should be unbound from the device page table of the + * specified address space (VM). VM_UNBIND will force unbind the specified + * range from device page table without waiting for any GPU job to complete. + * It is UMDs responsibility to ensure the mapping is no longer in use before
+ * calling VM_UNBIND.
+ *
+ * If the specified mapping is not found, the ioctl will simply return without
+ * any error.
+ *
+ * VM_BIND/UNBIND ioctl calls executed on different CPU threads concurrently + * are not ordered. Furthermore, parts of the VM_UNBIND operation can be done
+ * asynchronously, if valid @fence is specified.
+ */
+struct drm_i915_gem_vm_unbind {
+    /** @vm_id: VM (address space) id to bind */
+    __u32 vm_id;
+
+    /** @rsvd: Reserved, MBZ */
+    __u32 rsvd;
+
+    /** @start: Virtual Address start to unbind */
+    __u64 start;
+
+    /** @length: Length of mapping to unbind */
+    __u64 length;
+
+    /**
+     * @flags: Currently reserved, MBZ.
+     *
+     * Note that @fence carries its own flags.
+     */
+    __u64 flags;
+
+    /** @rsvd2: Reserved, MBZ */
+    __u64 rsvd2[2];
+
+    /**
+     * @extensions: Zero-terminated chain of extensions.
+     *
+     * For future extensions. See struct i915_user_extension.
+     */
+    __u64 extensions;
+};
+
 #if defined(__cplusplus)
 }
 #endif



[Index of Archives]     [AMD Graphics]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux