With eudebug event interface, user space debugger process (like gdb) is able to keep track of resources created by another process (debuggee using drm/xe) and act upon these resources. For example, debugger can find a client vm which contains isa/elf for a particular shader/eu-kernel and then inspect and modify it (for example installing a breakpoint). Debugger first opens a connection to xe with a drm ioctl specifying target pid to connect. This returns an anon fd handle that can then be used to listen for events with dedicated ioctl. This patch introduces eudebug connection and event queuing, adding client create/destroy and vm create/destroy events as a baseline. More events for full debugger operation are needed and those will be introduced in follow up patches. The resource tracking parts are inspired by the work of Maciej Patelczyk on resource handling for i915. Chris Wilson suggested improvement of two ways mapping which makes it easy to use resource map as a definitive bookkeep of what resources are played to debugger in the discovery phase (on follow up patch). v2: - Kconfig support (Matthew) - ptraced access control (Lucas) - pass expected event length to user (Zbigniew) - only track long running VMs - checkpatch (Tilak) - include order (Andrzej) - 32bit fixes (Andrzej) - cleaner get_task_struct - remove xa_array and use clients.list for tracking (Mika) v3: - adapt to removal of clients.lock (Mika) - create_event cleanup (Christoph) Cc: Maarten Lankhorst <maarten.lankhorst@xxxxxxxxxxxxxxx> Cc: Lucas De Marchi <lucas.demarchi@xxxxxxxxx> Cc: Dominik Grzegorzek <dominik.grzegorzek@xxxxxxxxx> Cc: Andi Shyti <andi.shyti@xxxxxxxxxxxxxxx> Cc: Matt Roper <matthew.d.roper@xxxxxxxxx> Cc: Matthew Brost <matthew.brost@xxxxxxxxx> Cc: Zbigniew Kempczyński <zbigniew.kempczynski@xxxxxxxxx> Cc: Andrzej Hajda <andrzej.hajda@xxxxxxxxx> Signed-off-by: Mika Kuoppala <mika.kuoppala@xxxxxxxxxxxxxxx> Signed-off-by: Maciej Patelczyk <maciej.patelczyk@xxxxxxxxx> Signed-off-by: Dominik Grzegorzek <dominik.grzegorzek@xxxxxxxxx> Signed-off-by: Jonathan Cavitt <jonathan.cavitt@xxxxxxxxx> Signed-off-by: Christoph Manszewski <christoph.manszewski@xxxxxxxxx> --- drivers/gpu/drm/xe/Kconfig | 10 + drivers/gpu/drm/xe/Makefile | 2 + drivers/gpu/drm/xe/xe_device.c | 10 + drivers/gpu/drm/xe/xe_device_types.h | 35 + drivers/gpu/drm/xe/xe_eudebug.c | 1118 +++++++++++++++++++++++++ drivers/gpu/drm/xe/xe_eudebug.h | 46 + drivers/gpu/drm/xe/xe_eudebug_types.h | 169 ++++ drivers/gpu/drm/xe/xe_vm.c | 7 +- include/uapi/drm/xe_drm.h | 21 + include/uapi/drm/xe_drm_eudebug.h | 56 ++ 10 files changed, 1473 insertions(+), 1 deletion(-) create mode 100644 drivers/gpu/drm/xe/xe_eudebug.c create mode 100644 drivers/gpu/drm/xe/xe_eudebug.h create mode 100644 drivers/gpu/drm/xe/xe_eudebug_types.h create mode 100644 include/uapi/drm/xe_drm_eudebug.h diff --git a/drivers/gpu/drm/xe/Kconfig b/drivers/gpu/drm/xe/Kconfig index b51a2bde73e2..ed97730b1af3 100644 --- a/drivers/gpu/drm/xe/Kconfig +++ b/drivers/gpu/drm/xe/Kconfig @@ -87,6 +87,16 @@ config DRM_XE_FORCE_PROBE Use "!*" to block the probe of the driver for all known devices. +config DRM_XE_EUDEBUG + bool "Enable gdb debugger support (eudebug)" + depends on DRM_XE + default y + help + Choose this option if you want to add support for debugger (gdb) to + attach into process using Xe and debug the gpu/gpgpu programs. + With debugger support, Xe will provide interface for a debugger to + process to track, inspect and modify resources. + menu "drm/Xe Debugging" depends on DRM_XE depends on EXPERT diff --git a/drivers/gpu/drm/xe/Makefile b/drivers/gpu/drm/xe/Makefile index 7730e0596299..deabcdd3ea52 100644 --- a/drivers/gpu/drm/xe/Makefile +++ b/drivers/gpu/drm/xe/Makefile @@ -116,6 +116,8 @@ xe-y += xe_bb.o \ xe_wa.o \ xe_wopcm.o +xe-$(CONFIG_DRM_XE_EUDEBUG) += xe_eudebug.o + xe-$(CONFIG_HMM_MIRROR) += xe_hmm.o # graphics hardware monitoring (HWMON) support diff --git a/drivers/gpu/drm/xe/xe_device.c b/drivers/gpu/drm/xe/xe_device.c index d6fccea1e083..9ed0de1eba0b 100644 --- a/drivers/gpu/drm/xe/xe_device.c +++ b/drivers/gpu/drm/xe/xe_device.c @@ -28,6 +28,7 @@ #include "xe_dma_buf.h" #include "xe_drm_client.h" #include "xe_drv.h" +#include "xe_eudebug.h" #include "xe_exec.h" #include "xe_exec_queue.h" #include "xe_force_wake.h" @@ -100,6 +101,8 @@ static int xe_file_open(struct drm_device *dev, struct drm_file *file) put_task_struct(task); } + xe_eudebug_file_open(xef); + return 0; } @@ -153,6 +156,8 @@ static void xe_file_close(struct drm_device *dev, struct drm_file *file) xe_pm_runtime_get(xe); + xe_eudebug_file_close(xef); + /* * No need for exec_queue.lock here as there is no contention for it * when FD is closing as IOCTLs presumably can't be modifying the @@ -191,6 +196,7 @@ static const struct drm_ioctl_desc xe_ioctls[] = { DRM_IOCTL_DEF_DRV(XE_WAIT_USER_FENCE, xe_wait_user_fence_ioctl, DRM_RENDER_ALLOW), DRM_IOCTL_DEF_DRV(XE_OBSERVATION, xe_observation_ioctl, DRM_RENDER_ALLOW), + DRM_IOCTL_DEF_DRV(XE_EUDEBUG_CONNECT, xe_eudebug_connect_ioctl, DRM_RENDER_ALLOW), }; static long xe_drm_ioctl(struct file *file, unsigned int cmd, unsigned long arg) @@ -281,6 +287,8 @@ static void xe_device_destroy(struct drm_device *dev, void *dummy) { struct xe_device *xe = to_xe_device(dev); + xe_eudebug_fini(xe); + if (xe->preempt_fence_wq) destroy_workqueue(xe->preempt_fence_wq); @@ -352,6 +360,8 @@ struct xe_device *xe_device_create(struct pci_dev *pdev, INIT_LIST_HEAD(&xe->pinned.external_vram); INIT_LIST_HEAD(&xe->pinned.evicted); + xe_eudebug_init(xe); + xe->preempt_fence_wq = alloc_ordered_workqueue("xe-preempt-fence-wq", WQ_MEM_RECLAIM); xe->ordered_wq = alloc_ordered_workqueue("xe-ordered-wq", 0); diff --git a/drivers/gpu/drm/xe/xe_device_types.h b/drivers/gpu/drm/xe/xe_device_types.h index 1373a222f5a5..9f04e6476195 100644 --- a/drivers/gpu/drm/xe/xe_device_types.h +++ b/drivers/gpu/drm/xe/xe_device_types.h @@ -383,6 +383,17 @@ struct xe_device { struct workqueue_struct *wq; } sriov; +#if IS_ENABLED(CONFIG_DRM_XE_EUDEBUG) + /** @clients: eudebug clients info */ + struct { + /** @clients.lock: Protects client list */ + spinlock_t lock; + + /** @xa: client list for eudebug discovery */ + struct list_head list; + } clients; +#endif + /** @usm: unified memory state */ struct { /** @usm.asid: convert a ASID to VM */ @@ -525,6 +536,23 @@ struct xe_device { u8 vm_inject_error_position; #endif +#if IS_ENABLED(CONFIG_DRM_XE_EUDEBUG) + /** @debugger connection list and globals for device */ + struct { + /** @lock: protects the list of connections */ + spinlock_t lock; + + /** @list: list of connections, aka debuggers */ + struct list_head list; + + /** @session_count: session counter to track connections */ + u64 session_count; + + /** @available: is the debugging functionality available */ + bool available; + } eudebug; +#endif + /* private: */ #if IS_ENABLED(CONFIG_DRM_XE_DISPLAY) @@ -642,6 +670,13 @@ struct xe_file { /** @refcount: ref count of this xe file */ struct kref refcount; + +#if IS_ENABLED(CONFIG_DRM_XE_EUDEBUG) + struct { + /** @client_link: list entry in xe_device.clients.list */ + struct list_head client_link; + } eudebug; +#endif }; #endif diff --git a/drivers/gpu/drm/xe/xe_eudebug.c b/drivers/gpu/drm/xe/xe_eudebug.c new file mode 100644 index 000000000000..bbb5f1e81bb8 --- /dev/null +++ b/drivers/gpu/drm/xe/xe_eudebug.c @@ -0,0 +1,1118 @@ +// SPDX-License-Identifier: MIT +/* + * Copyright © 2023 Intel Corporation + */ + +#include <linux/anon_inodes.h> +#include <linux/delay.h> +#include <linux/poll.h> +#include <linux/uaccess.h> + +#include <drm/drm_managed.h> + +#include "xe_assert.h" +#include "xe_device.h" +#include "xe_eudebug.h" +#include "xe_eudebug_types.h" +#include "xe_macros.h" +#include "xe_vm.h" + +/* + * If there is no detected event read by userspace, during this period, assume + * userspace problem and disconnect debugger to allow forward progress. + */ +#define XE_EUDEBUG_NO_READ_DETECTED_TIMEOUT_MS (25 * 1000) + +#define for_each_debugger_rcu(debugger, head) \ + list_for_each_entry_rcu((debugger), (head), connection_link) +#define for_each_debugger(debugger, head) \ + list_for_each_entry((debugger), (head), connection_link) + +#define cast_event(T, event) container_of((event), typeof(*(T)), base) + +#define XE_EUDEBUG_DBG_STR "eudbg: %lld:%lu:%s (%d/%d) -> (%d/%d): " +#define XE_EUDEBUG_DBG_ARGS(d) (d)->session, \ + atomic_long_read(&(d)->events.seqno), \ + READ_ONCE(d->connection.status) <= 0 ? "disconnected" : "", \ + current->pid, \ + task_tgid_nr(current), \ + (d)->target_task->pid, \ + task_tgid_nr((d)->target_task) + +#define eu_err(d, fmt, ...) drm_err(&(d)->xe->drm, XE_EUDEBUG_DBG_STR # fmt, \ + XE_EUDEBUG_DBG_ARGS(d), ##__VA_ARGS__) +#define eu_warn(d, fmt, ...) drm_warn(&(d)->xe->drm, XE_EUDEBUG_DBG_STR # fmt, \ + XE_EUDEBUG_DBG_ARGS(d), ##__VA_ARGS__) +#define eu_dbg(d, fmt, ...) drm_dbg(&(d)->xe->drm, XE_EUDEBUG_DBG_STR # fmt, \ + XE_EUDEBUG_DBG_ARGS(d), ##__VA_ARGS__) + +#define xe_eudebug_assert(d, ...) xe_assert((d)->xe, ##__VA_ARGS__) + +#define struct_member(T, member) (((T *)0)->member) + +/* Keep 1:1 parity with uapi events */ +#define write_member(T_out, ptr, member, value) { \ + BUILD_BUG_ON(sizeof(*ptr) != sizeof(T_out)); \ + BUILD_BUG_ON(offsetof(typeof(*ptr), member) != \ + offsetof(typeof(T_out), member)); \ + BUILD_BUG_ON(sizeof(ptr->member) != sizeof(value)); \ + BUILD_BUG_ON(sizeof(struct_member(T_out, member)) != sizeof(value)); \ + BUILD_BUG_ON(!typecheck(typeof((ptr)->member), value)); \ + (ptr)->member = (value); \ + } + +static struct xe_eudebug_event * +event_fifo_pending(struct xe_eudebug *d) +{ + struct xe_eudebug_event *event; + + if (kfifo_peek(&d->events.fifo, &event)) + return event; + + return NULL; +} + +/* + * This is racy as we dont take the lock for read but all the + * callsites can handle the race so we can live without lock. + */ +__no_kcsan +static unsigned int +event_fifo_num_events_peek(const struct xe_eudebug * const d) +{ + return kfifo_len(&d->events.fifo); +} + +static bool +xe_eudebug_detached(struct xe_eudebug *d) +{ + int status; + + spin_lock(&d->connection.lock); + status = d->connection.status; + spin_unlock(&d->connection.lock); + + return status <= 0; +} + +static int +xe_eudebug_error(const struct xe_eudebug * const d) +{ + const int status = READ_ONCE(d->connection.status); + + return status <= 0 ? status : 0; +} + +static unsigned int +event_fifo_has_events(struct xe_eudebug *d) +{ + if (xe_eudebug_detached(d)) + return 1; + + return event_fifo_num_events_peek(d); +} + +static const struct rhashtable_params rhash_res = { + .head_offset = offsetof(struct xe_eudebug_handle, rh_head), + .key_len = sizeof_field(struct xe_eudebug_handle, key), + .key_offset = offsetof(struct xe_eudebug_handle, key), + .automatic_shrinking = true, +}; + +static struct xe_eudebug_resource * +resource_from_type(struct xe_eudebug_resources * const res, const int t) +{ + return &res->rt[t]; +} + +static struct xe_eudebug_resources * +xe_eudebug_resources_alloc(void) +{ + struct xe_eudebug_resources *res; + int err; + int i; + + res = kzalloc(sizeof(*res), GFP_ATOMIC); + if (!res) + return ERR_PTR(-ENOMEM); + + mutex_init(&res->lock); + + for (i = 0; i < XE_EUDEBUG_RES_TYPE_COUNT; i++) { + xa_init_flags(&res->rt[i].xa, XA_FLAGS_ALLOC1); + err = rhashtable_init(&res->rt[i].rh, &rhash_res); + + if (err) + break; + } + + if (err) { + while (i--) { + xa_destroy(&res->rt[i].xa); + rhashtable_destroy(&res->rt[i].rh); + } + + kfree(res); + return ERR_PTR(err); + } + + return res; +} + +static void res_free_fn(void *ptr, void *arg) +{ + XE_WARN_ON(ptr); + kfree(ptr); +} + +static void +xe_eudebug_destroy_resources(struct xe_eudebug *d) +{ + struct xe_eudebug_resources *res = d->res; + struct xe_eudebug_handle *h; + unsigned long j; + int i; + int err; + + mutex_lock(&res->lock); + for (i = 0; i < XE_EUDEBUG_RES_TYPE_COUNT; i++) { + struct xe_eudebug_resource *r = &res->rt[i]; + + xa_for_each(&r->xa, j, h) { + struct xe_eudebug_handle *t; + + err = rhashtable_remove_fast(&r->rh, + &h->rh_head, + rhash_res); + xe_eudebug_assert(d, !err); + t = xa_erase(&r->xa, h->id); + xe_eudebug_assert(d, t == h); + kfree(t); + } + } + mutex_unlock(&res->lock); + + for (i = 0; i < XE_EUDEBUG_RES_TYPE_COUNT; i++) { + struct xe_eudebug_resource *r = &res->rt[i]; + + rhashtable_free_and_destroy(&r->rh, res_free_fn, NULL); + xe_eudebug_assert(d, xa_empty(&r->xa)); + xa_destroy(&r->xa); + } + + mutex_destroy(&res->lock); + + kfree(res); +} + +static void xe_eudebug_free(struct kref *ref) +{ + struct xe_eudebug *d = container_of(ref, typeof(*d), ref); + struct xe_eudebug_event *event; + + while (kfifo_get(&d->events.fifo, &event)) + kfree(event); + + xe_eudebug_destroy_resources(d); + put_task_struct(d->target_task); + + xe_eudebug_assert(d, !kfifo_len(&d->events.fifo)); + + kfree_rcu(d, rcu); +} + +static void xe_eudebug_put(struct xe_eudebug *d) +{ + kref_put(&d->ref, xe_eudebug_free); +} + +static struct task_struct *find_get_target(const pid_t nr) +{ + struct task_struct *task; + + rcu_read_lock(); + task = pid_task(find_pid_ns(nr, task_active_pid_ns(current)), PIDTYPE_PID); + if (task) + get_task_struct(task); + rcu_read_unlock(); + + return task; +} + +static int +xe_eudebug_attach(struct xe_device *xe, struct xe_eudebug *d, + const pid_t pid_nr) +{ + struct task_struct *target; + struct xe_eudebug *iter; + int ret = 0; + + target = find_get_target(pid_nr); + if (!target) + return -ENOENT; + + if (!ptrace_may_access(target, PTRACE_MODE_READ_REALCREDS)) { + put_task_struct(target); + return -EACCES; + } + + XE_WARN_ON(d->connection.status != 0); + + spin_lock(&xe->eudebug.lock); + for_each_debugger(iter, &xe->eudebug.list) { + if (!same_thread_group(iter->target_task, target)) + continue; + + ret = -EBUSY; + } + + if (!ret && xe->eudebug.session_count + 1 == 0) + ret = -ENOSPC; + + if (!ret) { + d->connection.status = XE_EUDEBUG_STATUS_CONNECTED; + d->xe = xe; + d->target_task = get_task_struct(target); + d->session = ++xe->eudebug.session_count; + kref_get(&d->ref); + list_add_tail_rcu(&d->connection_link, &xe->eudebug.list); + } + spin_unlock(&xe->eudebug.lock); + + put_task_struct(target); + + return ret; +} + +static bool xe_eudebug_detach(struct xe_device *xe, + struct xe_eudebug *d, + const int err) +{ + bool detached = false; + + XE_WARN_ON(err > 0); + + spin_lock(&d->connection.lock); + if (d->connection.status == XE_EUDEBUG_STATUS_CONNECTED) { + d->connection.status = err; + detached = true; + } + spin_unlock(&d->connection.lock); + + if (!detached) + return false; + + spin_lock(&xe->eudebug.lock); + list_del_rcu(&d->connection_link); + spin_unlock(&xe->eudebug.lock); + + eu_dbg(d, "session %lld detached with %d", d->session, err); + + /* Our ref with the connection_link */ + xe_eudebug_put(d); + + return true; +} + +static int _xe_eudebug_disconnect(struct xe_eudebug *d, + const int err) +{ + wake_up_all(&d->events.write_done); + wake_up_all(&d->events.read_done); + + return xe_eudebug_detach(d->xe, d, err); +} + +#define xe_eudebug_disconnect(_d, _err) ({ \ + if (_xe_eudebug_disconnect((_d), (_err))) { \ + if ((_err) == 0 || (_err) == -ETIMEDOUT) \ + eu_dbg(d, "Session closed (%d)", (_err)); \ + else \ + eu_err(d, "Session disconnected, err = %d (%s:%d)", \ + (_err), __func__, __LINE__); \ + } \ +}) + +static int xe_eudebug_release(struct inode *inode, struct file *file) +{ + struct xe_eudebug *d = file->private_data; + + xe_eudebug_disconnect(d, 0); + xe_eudebug_put(d); + + return 0; +} + +static __poll_t xe_eudebug_poll(struct file *file, poll_table *wait) +{ + struct xe_eudebug * const d = file->private_data; + __poll_t ret = 0; + + poll_wait(file, &d->events.write_done, wait); + + if (xe_eudebug_detached(d)) { + ret |= EPOLLHUP; + if (xe_eudebug_error(d)) + ret |= EPOLLERR; + } + + if (event_fifo_num_events_peek(d)) + ret |= EPOLLIN; + + return ret; +} + +static ssize_t xe_eudebug_read(struct file *file, + char __user *buf, + size_t count, + loff_t *ppos) +{ + return -EINVAL; +} + +static struct xe_eudebug * +xe_eudebug_for_task_get(struct xe_device *xe, + struct task_struct *task) +{ + struct xe_eudebug *d, *iter; + + d = NULL; + + rcu_read_lock(); + for_each_debugger_rcu(iter, &xe->eudebug.list) { + if (!same_thread_group(iter->target_task, task)) + continue; + + if (kref_get_unless_zero(&iter->ref)) + d = iter; + + break; + } + rcu_read_unlock(); + + return d; +} + +static struct task_struct *find_task_get(struct xe_file *xef) +{ + struct task_struct *task; + struct pid *pid; + + rcu_read_lock(); + pid = rcu_dereference(xef->drm->pid); + task = pid_task(pid, PIDTYPE_PID); + if (task) + get_task_struct(task); + rcu_read_unlock(); + + return task; +} + +static struct xe_eudebug * +xe_eudebug_get(struct xe_file *xef) +{ + struct task_struct *task; + struct xe_eudebug *d; + + d = NULL; + task = find_task_get(xef); + if (task) { + d = xe_eudebug_for_task_get(to_xe_device(xef->drm->minor->dev), + task); + put_task_struct(task); + } + + if (!d) + return NULL; + + if (xe_eudebug_detached(d)) { + xe_eudebug_put(d); + return NULL; + } + + return d; +} + +static int xe_eudebug_queue_event(struct xe_eudebug *d, + struct xe_eudebug_event *event) +{ + const u64 wait_jiffies = msecs_to_jiffies(1000); + u64 last_read_detected_ts, last_head_seqno, start_ts; + + xe_eudebug_assert(d, event->len > sizeof(struct xe_eudebug_event)); + xe_eudebug_assert(d, event->type); + xe_eudebug_assert(d, event->type != DRM_XE_EUDEBUG_EVENT_READ); + + start_ts = ktime_get(); + last_read_detected_ts = start_ts; + last_head_seqno = 0; + + do { + struct xe_eudebug_event *head; + u64 head_seqno; + bool was_queued; + + if (xe_eudebug_detached(d)) + break; + + spin_lock(&d->events.lock); + head = event_fifo_pending(d); + if (head) + head_seqno = event->seqno; + else + head_seqno = 0; + + was_queued = kfifo_in(&d->events.fifo, &event, 1); + spin_unlock(&d->events.lock); + + wake_up_all(&d->events.write_done); + + if (was_queued) { + event = NULL; + break; + } + + XE_WARN_ON(!head_seqno); + + /* If we detect progress, restart timeout */ + if (last_head_seqno != head_seqno) + last_read_detected_ts = ktime_get(); + + last_head_seqno = head_seqno; + + wait_event_interruptible_timeout(d->events.read_done, + !kfifo_is_full(&d->events.fifo), + wait_jiffies); + + } while (ktime_ms_delta(ktime_get(), last_read_detected_ts) < + XE_EUDEBUG_NO_READ_DETECTED_TIMEOUT_MS); + + if (event) { + eu_dbg(d, + "event %llu queue failed (blocked %lld ms, avail %d)", + event ? event->seqno : 0, + ktime_ms_delta(ktime_get(), start_ts), + kfifo_avail(&d->events.fifo)); + + kfree(event); + + return -ETIMEDOUT; + } + + return 0; +} + +static struct xe_eudebug_handle * +alloc_handle(const int type, const u64 key) +{ + struct xe_eudebug_handle *h; + + h = kzalloc(sizeof(*h), GFP_ATOMIC); + if (!h) + return NULL; + + h->key = key; + + return h; +} + +static struct xe_eudebug_handle * +__find_handle(struct xe_eudebug_resource *r, + const u64 key) +{ + struct xe_eudebug_handle *h; + + h = rhashtable_lookup_fast(&r->rh, + &key, + rhash_res); + return h; +} + +static int find_handle(struct xe_eudebug_resources *res, + const int type, + const void *p) +{ + const u64 key = (uintptr_t)p; + struct xe_eudebug_resource *r; + struct xe_eudebug_handle *h; + int id; + + if (XE_WARN_ON(!key)) + return -EINVAL; + + r = resource_from_type(res, type); + + mutex_lock(&res->lock); + h = __find_handle(r, key); + id = h ? h->id : -ENOENT; + mutex_unlock(&res->lock); + + return id; +} + +static int _xe_eudebug_add_handle(struct xe_eudebug *d, + int type, + void *p, + u64 *seqno, + int *handle) +{ + const u64 key = (uintptr_t)p; + struct xe_eudebug_resource *r; + struct xe_eudebug_handle *h, *o; + int err; + + if (XE_WARN_ON(!p)) + return -EINVAL; + + if (xe_eudebug_detached(d)) + return -ENOTCONN; + + h = alloc_handle(type, key); + if (!h) + return -ENOMEM; + + r = resource_from_type(d->res, type); + + mutex_lock(&d->res->lock); + o = __find_handle(r, key); + if (!o) { + err = xa_alloc(&r->xa, &h->id, h, xa_limit_31b, GFP_KERNEL); + + if (h->id >= INT_MAX) { + xa_erase(&r->xa, h->id); + err = -ENOSPC; + } + + if (!err) + err = rhashtable_insert_fast(&r->rh, + &h->rh_head, + rhash_res); + + if (err) { + xa_erase(&r->xa, h->id); + } else { + if (seqno) + *seqno = atomic_long_inc_return(&d->events.seqno); + } + } else { + xe_eudebug_assert(d, o->id); + err = -EEXIST; + } + mutex_unlock(&d->res->lock); + + if (handle) + *handle = o ? o->id : h->id; + + if (err) { + kfree(h); + XE_WARN_ON(err > 0); + return err; + } + + xe_eudebug_assert(d, h->id); + + return h->id; +} + +static int xe_eudebug_add_handle(struct xe_eudebug *d, + int type, + void *p, + u64 *seqno) +{ + int ret; + + ret = _xe_eudebug_add_handle(d, type, p, seqno, NULL); + if (ret == -EEXIST || ret == -ENOTCONN) { + eu_dbg(d, "%d on adding %d", ret, type); + return 0; + } + + if (ret < 0) + xe_eudebug_disconnect(d, ret); + + return ret; +} + +static int _xe_eudebug_remove_handle(struct xe_eudebug *d, int type, void *p, + u64 *seqno) +{ + const u64 key = (uintptr_t)p; + struct xe_eudebug_resource *r; + struct xe_eudebug_handle *h, *xa_h; + int ret; + + if (XE_WARN_ON(!key)) + return -EINVAL; + + if (xe_eudebug_detached(d)) + return -ENOTCONN; + + r = resource_from_type(d->res, type); + + mutex_lock(&d->res->lock); + h = __find_handle(r, key); + if (h) { + ret = rhashtable_remove_fast(&r->rh, + &h->rh_head, + rhash_res); + xe_eudebug_assert(d, !ret); + xa_h = xa_erase(&r->xa, h->id); + xe_eudebug_assert(d, xa_h == h); + if (!ret) { + ret = h->id; + if (seqno) + *seqno = atomic_long_inc_return(&d->events.seqno); + } + } else { + ret = -ENOENT; + } + mutex_unlock(&d->res->lock); + + kfree(h); + + xe_eudebug_assert(d, ret); + + return ret; +} + +static int xe_eudebug_remove_handle(struct xe_eudebug *d, int type, void *p, + u64 *seqno) +{ + int ret; + + ret = _xe_eudebug_remove_handle(d, type, p, seqno); + if (ret == -ENOENT || ret == -ENOTCONN) { + eu_dbg(d, "%d on removing %d", ret, type); + return 0; + } + + if (ret < 0) + xe_eudebug_disconnect(d, ret); + + return ret; +} + +static struct xe_eudebug_event * +xe_eudebug_create_event(struct xe_eudebug *d, u16 type, u64 seqno, u16 flags, + u32 len) +{ + const u16 max_event = DRM_XE_EUDEBUG_EVENT_VM; + const u16 known_flags = + DRM_XE_EUDEBUG_EVENT_CREATE | + DRM_XE_EUDEBUG_EVENT_DESTROY | + DRM_XE_EUDEBUG_EVENT_STATE_CHANGE | + DRM_XE_EUDEBUG_EVENT_NEED_ACK; + struct xe_eudebug_event *event; + + BUILD_BUG_ON(type > max_event); + + xe_eudebug_assert(d, type <= max_event); + xe_eudebug_assert(d, !(~known_flags & flags)); + xe_eudebug_assert(d, len > sizeof(*event)); + + event = kzalloc(len, GFP_KERNEL); + if (!event) + return NULL; + + event->len = len; + event->type = type; + event->flags = flags; + event->seqno = seqno; + + return event; +} + +static long xe_eudebug_read_event(struct xe_eudebug *d, + const u64 arg, + const bool wait) +{ + struct xe_device *xe = d->xe; + struct drm_xe_eudebug_event __user * const user_orig = + u64_to_user_ptr(arg); + struct drm_xe_eudebug_event user_event; + struct xe_eudebug_event *event; + const unsigned int max_event = DRM_XE_EUDEBUG_EVENT_VM; + long ret = 0; + + if (XE_IOCTL_DBG(xe, copy_from_user(&user_event, user_orig, sizeof(user_event)))) + return -EFAULT; + + if (XE_IOCTL_DBG(xe, !user_event.type)) + return -EINVAL; + + if (XE_IOCTL_DBG(xe, user_event.type > max_event)) + return -EINVAL; + + if (XE_IOCTL_DBG(xe, user_event.type != DRM_XE_EUDEBUG_EVENT_READ)) + return -EINVAL; + + if (XE_IOCTL_DBG(xe, user_event.len < sizeof(*user_orig))) + return -EINVAL; + + if (XE_IOCTL_DBG(xe, user_event.flags)) + return -EINVAL; + + if (XE_IOCTL_DBG(xe, user_event.reserved)) + return -EINVAL; + + /* XXX: define wait time in connect arguments ? */ + if (wait) { + ret = wait_event_interruptible_timeout(d->events.write_done, + event_fifo_has_events(d), + msecs_to_jiffies(5 * 1000)); + + if (XE_IOCTL_DBG(xe, ret < 0)) + return ret; + } + + ret = 0; + spin_lock(&d->events.lock); + event = event_fifo_pending(d); + if (event) { + if (user_event.len < event->len) { + ret = -EMSGSIZE; + } else if (!kfifo_out(&d->events.fifo, &event, 1)) { + eu_warn(d, "internal fifo corruption"); + ret = -ENOTCONN; + } + } + spin_unlock(&d->events.lock); + + wake_up_all(&d->events.read_done); + + if (ret == -EMSGSIZE && put_user(event->len, &user_orig->len)) + ret = -EFAULT; + + if (XE_IOCTL_DBG(xe, ret)) + return ret; + + if (!event) { + if (xe_eudebug_detached(d)) + return -ENOTCONN; + if (!wait) + return -EAGAIN; + + return -ENOENT; + } + + if (copy_to_user(user_orig, event, event->len)) + ret = -EFAULT; + else + eu_dbg(d, "event read: type=%u, flags=0x%x, seqno=%llu", event->type, + event->flags, event->seqno); + + kfree(event); + + return ret; +} + +static long xe_eudebug_ioctl(struct file *file, + unsigned int cmd, + unsigned long arg) +{ + struct xe_eudebug * const d = file->private_data; + long ret; + + switch (cmd) { + case DRM_XE_EUDEBUG_IOCTL_READ_EVENT: + ret = xe_eudebug_read_event(d, arg, + !(file->f_flags & O_NONBLOCK)); + break; + + default: + ret = -EINVAL; + } + + return ret; +} + +static const struct file_operations fops = { + .owner = THIS_MODULE, + .release = xe_eudebug_release, + .poll = xe_eudebug_poll, + .read = xe_eudebug_read, + .unlocked_ioctl = xe_eudebug_ioctl, +}; + +static int +xe_eudebug_connect(struct xe_device *xe, + struct drm_xe_eudebug_connect *param) +{ + const u64 known_open_flags = 0; + unsigned long f_flags = 0; + struct xe_eudebug *d; + int fd, err; + + if (param->extensions) + return -EINVAL; + + if (!param->pid) + return -EINVAL; + + if (param->flags & ~known_open_flags) + return -EINVAL; + + if (param->version && param->version != DRM_XE_EUDEBUG_VERSION) + return -EINVAL; + + param->version = DRM_XE_EUDEBUG_VERSION; + + if (!xe->eudebug.available) + return -EOPNOTSUPP; + + d = kzalloc(sizeof(*d), GFP_KERNEL); + if (!d) + return -ENOMEM; + + kref_init(&d->ref); + spin_lock_init(&d->connection.lock); + init_waitqueue_head(&d->events.write_done); + init_waitqueue_head(&d->events.read_done); + + spin_lock_init(&d->events.lock); + INIT_KFIFO(d->events.fifo); + + d->res = xe_eudebug_resources_alloc(); + if (IS_ERR(d->res)) { + err = PTR_ERR(d->res); + goto err_free; + } + + err = xe_eudebug_attach(xe, d, param->pid); + if (err) + goto err_free_res; + + fd = anon_inode_getfd("[xe_eudebug]", &fops, d, f_flags); + if (fd < 0) { + err = fd; + goto err_detach; + } + + eu_dbg(d, "connected session %lld", d->session); + + return fd; + +err_detach: + xe_eudebug_detach(xe, d, err); +err_free_res: + xe_eudebug_destroy_resources(d); +err_free: + kfree(d); + + return err; +} + +int xe_eudebug_connect_ioctl(struct drm_device *dev, + void *data, + struct drm_file *file) +{ + struct xe_device *xe = to_xe_device(dev); + struct drm_xe_eudebug_connect * const param = data; + int ret = 0; + + ret = xe_eudebug_connect(xe, param); + + return ret; +} + +void xe_eudebug_init(struct xe_device *xe) +{ + spin_lock_init(&xe->eudebug.lock); + INIT_LIST_HEAD(&xe->eudebug.list); + + spin_lock_init(&xe->clients.lock); + INIT_LIST_HEAD(&xe->clients.list); + + xe->eudebug.available = true; +} + +void xe_eudebug_fini(struct xe_device *xe) +{ + xe_assert(xe, list_empty_careful(&xe->eudebug.list)); +} + +static int send_open_event(struct xe_eudebug *d, u32 flags, const u64 handle, + const u64 seqno) +{ + struct xe_eudebug_event *event; + struct xe_eudebug_event_open *eo; + + if (!handle) + return -EINVAL; + + if (XE_WARN_ON((long)handle >= INT_MAX)) + return -EINVAL; + + event = xe_eudebug_create_event(d, DRM_XE_EUDEBUG_EVENT_OPEN, seqno, + flags, sizeof(*eo)); + if (!event) + return -ENOMEM; + + eo = cast_event(eo, event); + + write_member(struct drm_xe_eudebug_event_client, eo, + client_handle, handle); + + return xe_eudebug_queue_event(d, event); +} + +static int client_create_event(struct xe_eudebug *d, struct xe_file *xef) +{ + u64 seqno; + int ret; + + ret = xe_eudebug_add_handle(d, XE_EUDEBUG_RES_TYPE_CLIENT, xef, &seqno); + if (ret > 0) + ret = send_open_event(d, DRM_XE_EUDEBUG_EVENT_CREATE, + ret, seqno); + + return ret; +} + +static int client_destroy_event(struct xe_eudebug *d, struct xe_file *xef) +{ + u64 seqno; + int ret; + + ret = xe_eudebug_remove_handle(d, XE_EUDEBUG_RES_TYPE_CLIENT, + xef, &seqno); + if (ret > 0) + ret = send_open_event(d, DRM_XE_EUDEBUG_EVENT_DESTROY, + ret, seqno); + + return ret; +} + +#define xe_eudebug_event_put(_d, _err) ({ \ + if ((_err)) \ + xe_eudebug_disconnect((_d), (_err)); \ + xe_eudebug_put((_d)); \ + }) + +void xe_eudebug_file_open(struct xe_file *xef) +{ + struct xe_eudebug *d; + + INIT_LIST_HEAD(&xef->eudebug.client_link); + spin_lock(&xef->xe->clients.lock); + list_add_tail(&xef->eudebug.client_link, &xef->xe->clients.list); + spin_unlock(&xef->xe->clients.lock); + + d = xe_eudebug_get(xef); + if (!d) + return; + + xe_eudebug_event_put(d, client_create_event(d, xef)); +} + +void xe_eudebug_file_close(struct xe_file *xef) +{ + struct xe_eudebug *d; + + d = xe_eudebug_get(xef); + if (d) + xe_eudebug_event_put(d, client_destroy_event(d, xef)); + + spin_lock(&xef->xe->clients.lock); + list_del_init(&xef->eudebug.client_link); + spin_unlock(&xef->xe->clients.lock); +} + +static int send_vm_event(struct xe_eudebug *d, u32 flags, + const u64 client_handle, + const u64 vm_handle, + const u64 seqno) +{ + struct xe_eudebug_event *event; + struct xe_eudebug_event_vm *e; + + event = xe_eudebug_create_event(d, DRM_XE_EUDEBUG_EVENT_VM, + seqno, flags, sizeof(*e)); + if (!event) + return -ENOMEM; + + e = cast_event(e, event); + + write_member(struct drm_xe_eudebug_event_vm, e, client_handle, client_handle); + write_member(struct drm_xe_eudebug_event_vm, e, vm_handle, vm_handle); + + return xe_eudebug_queue_event(d, event); +} + +static int vm_create_event(struct xe_eudebug *d, + struct xe_file *xef, struct xe_vm *vm) +{ + int h_c, h_vm; + u64 seqno; + int ret; + + if (!xe_vm_in_lr_mode(vm)) + return 0; + + h_c = find_handle(d->res, XE_EUDEBUG_RES_TYPE_CLIENT, xef); + if (h_c < 0) + return h_c; + + xe_eudebug_assert(d, h_c); + + h_vm = xe_eudebug_add_handle(d, XE_EUDEBUG_RES_TYPE_VM, vm, &seqno); + if (h_vm <= 0) + return h_vm; + + ret = send_vm_event(d, DRM_XE_EUDEBUG_EVENT_CREATE, h_c, h_vm, seqno); + + return ret; +} + +static int vm_destroy_event(struct xe_eudebug *d, + struct xe_file *xef, struct xe_vm *vm) +{ + int h_c, h_vm; + u64 seqno; + + if (!xe_vm_in_lr_mode(vm)) + return 0; + + h_c = find_handle(d->res, XE_EUDEBUG_RES_TYPE_CLIENT, xef); + if (h_c < 0) { + XE_WARN_ON("no client found for vm"); + eu_warn(d, "no client found for vm"); + return h_c; + } + + xe_eudebug_assert(d, h_c); + + h_vm = xe_eudebug_remove_handle(d, XE_EUDEBUG_RES_TYPE_VM, vm, &seqno); + if (h_vm <= 0) + return h_vm; + + return send_vm_event(d, DRM_XE_EUDEBUG_EVENT_DESTROY, h_c, h_vm, seqno); +} + +void xe_eudebug_vm_create(struct xe_file *xef, struct xe_vm *vm) +{ + struct xe_eudebug *d; + + if (!xe_vm_in_lr_mode(vm)) + return; + + d = xe_eudebug_get(xef); + if (!d) + return; + + xe_eudebug_event_put(d, vm_create_event(d, xef, vm)); +} + +void xe_eudebug_vm_destroy(struct xe_file *xef, struct xe_vm *vm) +{ + struct xe_eudebug *d; + + if (!xe_vm_in_lr_mode(vm)) + return; + + d = xe_eudebug_get(xef); + if (!d) + return; + + xe_eudebug_event_put(d, vm_destroy_event(d, xef, vm)); +} diff --git a/drivers/gpu/drm/xe/xe_eudebug.h b/drivers/gpu/drm/xe/xe_eudebug.h new file mode 100644 index 000000000000..e3247365f72f --- /dev/null +++ b/drivers/gpu/drm/xe/xe_eudebug.h @@ -0,0 +1,46 @@ +/* SPDX-License-Identifier: MIT */ +/* + * Copyright © 2023 Intel Corporation + */ + +#ifndef _XE_EUDEBUG_H_ + +struct drm_device; +struct drm_file; +struct xe_device; +struct xe_file; +struct xe_vm; + +#if IS_ENABLED(CONFIG_DRM_XE_EUDEBUG) + +int xe_eudebug_connect_ioctl(struct drm_device *dev, + void *data, + struct drm_file *file); + +void xe_eudebug_init(struct xe_device *xe); +void xe_eudebug_fini(struct xe_device *xe); + +void xe_eudebug_file_open(struct xe_file *xef); +void xe_eudebug_file_close(struct xe_file *xef); + +void xe_eudebug_vm_create(struct xe_file *xef, struct xe_vm *vm); +void xe_eudebug_vm_destroy(struct xe_file *xef, struct xe_vm *vm); + +#else + +static inline int xe_eudebug_connect_ioctl(struct drm_device *dev, + void *data, + struct drm_file *file) { return 0; } + +static inline void xe_eudebug_init(struct xe_device *xe) { } +static inline void xe_eudebug_fini(struct xe_device *xe) { } + +static inline void xe_eudebug_file_open(struct xe_file *xef) { } +static inline void xe_eudebug_file_close(struct xe_file *xef) { } + +static inline void xe_eudebug_vm_create(struct xe_file *xef, struct xe_vm *vm) { } +static inline void xe_eudebug_vm_destroy(struct xe_file *xef, struct xe_vm *vm) { } + +#endif /* CONFIG_DRM_XE_EUDEBUG */ + +#endif diff --git a/drivers/gpu/drm/xe/xe_eudebug_types.h b/drivers/gpu/drm/xe/xe_eudebug_types.h new file mode 100644 index 000000000000..a5185f18f640 --- /dev/null +++ b/drivers/gpu/drm/xe/xe_eudebug_types.h @@ -0,0 +1,169 @@ +/* SPDX-License-Identifier: MIT */ +/* + * Copyright © 2023 Intel Corporation + */ + +#ifndef __XE_EUDEBUG_TYPES_H_ + +#include <linux/completion.h> +#include <linux/kfifo.h> +#include <linux/kref.h> +#include <linux/mutex.h> +#include <linux/rbtree.h> +#include <linux/rhashtable.h> +#include <linux/wait.h> +#include <linux/xarray.h> + +#include <uapi/drm/xe_drm.h> + +struct xe_device; +struct task_struct; +struct xe_eudebug_event; + +#define CONFIG_DRM_XE_DEBUGGER_EVENT_QUEUE_SIZE 64 + +/** + * struct xe_eudebug_handle - eudebug resource handle + */ +struct xe_eudebug_handle { + /** @key: key value in rhashtable <key:id> */ + u64 key; + + /** @id: opaque handle id for xarray <id:key> */ + int id; + + /** @rh_head: rhashtable head */ + struct rhash_head rh_head; +}; + +/** + * struct xe_eudebug_resource - Resource map for one resource + */ +struct xe_eudebug_resource { + /** @xa: xarrays for <id->key> */ + struct xarray xa; + + /** @rh rhashtable for <key->id> */ + struct rhashtable rh; +}; + +#define XE_EUDEBUG_RES_TYPE_CLIENT 0 +#define XE_EUDEBUG_RES_TYPE_VM 1 +#define XE_EUDEBUG_RES_TYPE_COUNT (XE_EUDEBUG_RES_TYPE_VM + 1) + +/** + * struct xe_eudebug_resources - eudebug resources for all types + */ +struct xe_eudebug_resources { + /** @lock: guards access into rt */ + struct mutex lock; + + /** @rt: resource maps for all types */ + struct xe_eudebug_resource rt[XE_EUDEBUG_RES_TYPE_COUNT]; +}; + +/** + * struct xe_eudebug - Top level struct for eudebug: the connection + */ +struct xe_eudebug { + /** @ref: kref counter for this struct */ + struct kref ref; + + /** @rcu: rcu_head for rcu destruction */ + struct rcu_head rcu; + + /** @connection_link: our link into the xe_device:eudebug.list */ + struct list_head connection_link; + + struct { + /** @status: connected = 1, disconnected = error */ +#define XE_EUDEBUG_STATUS_CONNECTED 1 + int status; + + /** @lock: guards access to status */ + spinlock_t lock; + } connection; + + /** @xe: the parent device we are serving */ + struct xe_device *xe; + + /** @target_task: the task that we are debugging */ + struct task_struct *target_task; + + /** @res: the resource maps we track for target_task */ + struct xe_eudebug_resources *res; + + /** @session: session number for this connection (for logs) */ + u64 session; + + /** @events: kfifo queue of to-be-delivered events */ + struct { + /** @lock: guards access to fifo */ + spinlock_t lock; + + /** @fifo: queue of events pending */ + DECLARE_KFIFO(fifo, + struct xe_eudebug_event *, + CONFIG_DRM_XE_DEBUGGER_EVENT_QUEUE_SIZE); + + /** @write_done: waitqueue for signalling write to fifo */ + wait_queue_head_t write_done; + + /** @read_done: waitqueue for signalling read from fifo */ + wait_queue_head_t read_done; + + /** @event_seqno: seqno counter to stamp events for fifo */ + atomic_long_t seqno; + } events; + +}; + +/** + * struct xe_eudebug_event - Internal base event struct for eudebug + */ +struct xe_eudebug_event { + /** @len: length of this event, including payload */ + u32 len; + + /** @type: message type */ + u16 type; + + /** @flags: message flags */ + u16 flags; + + /** @seqno: sequence number for ordering */ + u64 seqno; + + /** @reserved: reserved field MBZ */ + u64 reserved; + + /** @data: payload bytes */ + u8 data[]; +}; + +/** + * struct xe_eudebug_event_open - Internal event for client open/close + */ +struct xe_eudebug_event_open { + /** @base: base event */ + struct xe_eudebug_event base; + + /** @client_handle: opaque handle for client */ + u64 client_handle; +}; + +/** + * struct xe_eudebug_event_vm - Internal event for vm open/close + */ +struct xe_eudebug_event_vm { + /** @base: base event */ + struct xe_eudebug_event base; + + /** @client_handle: client containing the vm open/close */ + u64 client_handle; + + /** @vm_handle: vm handle it's open/close */ + u64 vm_handle; +}; + +#endif diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c index 7788680da4e5..6f16049f4f6e 100644 --- a/drivers/gpu/drm/xe/xe_vm.c +++ b/drivers/gpu/drm/xe/xe_vm.c @@ -25,6 +25,7 @@ #include "xe_bo.h" #include "xe_device.h" #include "xe_drm_client.h" +#include "xe_eudebug.h" #include "xe_exec_queue.h" #include "xe_gt_pagefault.h" #include "xe_gt_tlb_invalidation.h" @@ -1797,6 +1798,8 @@ int xe_vm_create_ioctl(struct drm_device *dev, void *data, args->vm_id = id; + xe_eudebug_vm_create(xef, vm); + return 0; err_close_and_put: @@ -1828,8 +1831,10 @@ int xe_vm_destroy_ioctl(struct drm_device *dev, void *data, xa_erase(&xef->vm.xa, args->vm_id); mutex_unlock(&xef->vm.lock); - if (!err) + if (!err) { + xe_eudebug_vm_destroy(xef, vm); xe_vm_close_and_put(vm); + } return err; } diff --git a/include/uapi/drm/xe_drm.h b/include/uapi/drm/xe_drm.h index 4a8a4a63e99c..78479100a0b6 100644 --- a/include/uapi/drm/xe_drm.h +++ b/include/uapi/drm/xe_drm.h @@ -102,6 +102,7 @@ extern "C" { #define DRM_XE_EXEC 0x09 #define DRM_XE_WAIT_USER_FENCE 0x0a #define DRM_XE_OBSERVATION 0x0b +#define DRM_XE_EUDEBUG_CONNECT 0x0c /* Must be kept compact -- no holes */ @@ -117,6 +118,7 @@ extern "C" { #define DRM_IOCTL_XE_EXEC DRM_IOW(DRM_COMMAND_BASE + DRM_XE_EXEC, struct drm_xe_exec) #define DRM_IOCTL_XE_WAIT_USER_FENCE DRM_IOWR(DRM_COMMAND_BASE + DRM_XE_WAIT_USER_FENCE, struct drm_xe_wait_user_fence) #define DRM_IOCTL_XE_OBSERVATION DRM_IOW(DRM_COMMAND_BASE + DRM_XE_OBSERVATION, struct drm_xe_observation_param) +#define DRM_IOCTL_XE_EUDEBUG_CONNECT DRM_IOWR(DRM_COMMAND_BASE + DRM_XE_EUDEBUG_CONNECT, struct drm_xe_eudebug_connect) /** * DOC: Xe IOCTL Extensions @@ -1713,6 +1715,25 @@ struct drm_xe_oa_stream_info { __u64 reserved[3]; }; +/* + * Debugger ABI (ioctl and events) Version History: + * 0 - No debugger available + * 1 - Initial version + */ +#define DRM_XE_EUDEBUG_VERSION 1 + +struct drm_xe_eudebug_connect { + /** @extensions: Pointer to the first extension struct, if any */ + __u64 extensions; + + __u64 pid; /* input: Target process ID */ + __u32 flags; /* MBZ */ + + __u32 version; /* output: current ABI (ioctl / events) version */ +}; + +#include "xe_drm_eudebug.h" + #if defined(__cplusplus) } #endif diff --git a/include/uapi/drm/xe_drm_eudebug.h b/include/uapi/drm/xe_drm_eudebug.h new file mode 100644 index 000000000000..acf6071c82bf --- /dev/null +++ b/include/uapi/drm/xe_drm_eudebug.h @@ -0,0 +1,56 @@ +/* SPDX-License-Identifier: MIT */ +/* + * Copyright © 2023 Intel Corporation + */ + +#ifndef _UAPI_XE_DRM_EUDEBUG_H_ +#define _UAPI_XE_DRM_EUDEBUG_H_ + +#if defined(__cplusplus) +extern "C" { +#endif + +/** + * Do a eudebug event read for a debugger connection. + * + * This ioctl is available in debug version 1. + */ +#define DRM_XE_EUDEBUG_IOCTL_READ_EVENT _IO('j', 0x0) + +/* XXX: Document events to match their internal counterparts when moved to xe_drm.h */ +struct drm_xe_eudebug_event { + __u32 len; + + __u16 type; +#define DRM_XE_EUDEBUG_EVENT_NONE 0 +#define DRM_XE_EUDEBUG_EVENT_READ 1 +#define DRM_XE_EUDEBUG_EVENT_OPEN 2 +#define DRM_XE_EUDEBUG_EVENT_VM 3 + + __u16 flags; +#define DRM_XE_EUDEBUG_EVENT_CREATE (1 << 0) +#define DRM_XE_EUDEBUG_EVENT_DESTROY (1 << 1) +#define DRM_XE_EUDEBUG_EVENT_STATE_CHANGE (1 << 2) +#define DRM_XE_EUDEBUG_EVENT_NEED_ACK (1 << 3) + __u64 seqno; + __u64 reserved; +}; + +struct drm_xe_eudebug_event_client { + struct drm_xe_eudebug_event base; + + __u64 client_handle; /* This is unique per debug connection */ +}; + +struct drm_xe_eudebug_event_vm { + struct drm_xe_eudebug_event base; + + __u64 client_handle; + __u64 vm_handle; +}; + +#if defined(__cplusplus) +} +#endif + +#endif -- 2.43.0