On Thu, Jul 11, 2024 at 12:52 AM Daniel Almeida <daniel.almeida@xxxxxxxxxxxxx> wrote: > > Dump the state of the GPU. This feature is useful for debugging purposes. > --- > Hi everybody! > > For those looking for a branch instead, see [0]. > > I know this patch has (possibly many) issues. It is meant as a > discussion around the GEM abstractions for now. In particular, I am > aware of the series introducing Rust support for vmalloc and friends - > that is some very nice work! :) > > Danilo, as we've spoken before, I find it hard to work with `rust: drm: > gem: Add GEM object abstraction`. My patch is based on v1, but IIUC > the issue remains in v2: it is not possible to build a gem::ObjectRef > from a bindings::drm_gem_object*. > > Furthermore, gem::IntoGEMObject contains a Driver: drv::Driver > associated type: > > ``` > +/// Trait that represents a GEM object subtype > +pub trait IntoGEMObject: Sized + crate::private::Sealed { > + /// Owning driver for this type > + type Driver: drv::Driver; > + > ``` > > While this does work for Asahi and Nova - two drivers that are written > entirely in Rust - it is a blocker for any partially-converted drivers. > This is because there is no drv::Driver at all, only Rust functions that > are called from an existing C driver. > > IMHO, are unlikely to see full rewrites of any existing C code. But > partial convertions allows companies to write new features entirely in > Rust, or to migrate to Rust in small steps. For this reason, I think we > should strive to treat partially-converted drivers as first-class > citizens. > > [0]: https://gitlab.collabora.com/dwlsalmeida/for-upstream/-/tree/panthor-devcoredump?ref_type=heads > > drivers/gpu/drm/panthor/Kconfig | 13 ++ > drivers/gpu/drm/panthor/Makefile | 2 + > drivers/gpu/drm/panthor/dump.rs | 294 ++++++++++++++++++++++++ > drivers/gpu/drm/panthor/lib.rs | 10 + > drivers/gpu/drm/panthor/panthor_mmu.c | 39 ++++ > drivers/gpu/drm/panthor/panthor_mmu.h | 3 + > drivers/gpu/drm/panthor/panthor_rs.h | 40 ++++ > drivers/gpu/drm/panthor/panthor_sched.c | 28 ++- > drivers/gpu/drm/panthor/regs.rs | 264 +++++++++++++++++++++ > rust/bindings/bindings_helper.h | 3 + > 10 files changed, 695 insertions(+), 1 deletion(-) > create mode 100644 drivers/gpu/drm/panthor/dump.rs > create mode 100644 drivers/gpu/drm/panthor/lib.rs > create mode 100644 drivers/gpu/drm/panthor/panthor_rs.h > create mode 100644 drivers/gpu/drm/panthor/regs.rs > > diff --git a/drivers/gpu/drm/panthor/Kconfig b/drivers/gpu/drm/panthor/Kconfig > index 55b40ad07f3b..78d34e516f5b 100644 > --- a/drivers/gpu/drm/panthor/Kconfig > +++ b/drivers/gpu/drm/panthor/Kconfig > @@ -21,3 +21,16 @@ config DRM_PANTHOR > > Note that the Mali-G68 and Mali-G78, while Valhall architecture, will > be supported with the panfrost driver as they are not CSF GPUs. > + > +config DRM_PANTHOR_RS > + bool "Panthor Rust components" > + depends on DRM_PANTHOR > + depends on RUST > + help > + Enable Panthor's Rust components > + > +config DRM_PANTHOR_COREDUMP > + bool "Panthor devcoredump support" > + depends on DRM_PANTHOR_RS > + help > + Dump the GPU state through devcoredump for debugging purposes > \ No newline at end of file > diff --git a/drivers/gpu/drm/panthor/Makefile b/drivers/gpu/drm/panthor/Makefile > index 15294719b09c..10387b02cd69 100644 > --- a/drivers/gpu/drm/panthor/Makefile > +++ b/drivers/gpu/drm/panthor/Makefile > @@ -11,4 +11,6 @@ panthor-y := \ > panthor_mmu.o \ > panthor_sched.o > > +panthor-$(CONFIG_DRM_PANTHOR_RS) += lib.o > obj-$(CONFIG_DRM_PANTHOR) += panthor.o > + > diff --git a/drivers/gpu/drm/panthor/dump.rs b/drivers/gpu/drm/panthor/dump.rs > new file mode 100644 > index 000000000000..77fe5f420300 > --- /dev/null > +++ b/drivers/gpu/drm/panthor/dump.rs > @@ -0,0 +1,294 @@ > +// SPDX-License-Identifier: GPL-2.0 > +// SPDX-FileCopyrightText: Copyright Collabora 2024 > + > +//! Dump the GPU state to a file, so we can figure out what went wrong if it > +//! crashes. > +//! > +//! The dump is comprised of the following sections: > +//! > +//! Registers, > +//! Firmware interface (TODO) > +//! Buffer objects (the whole VM) > +//! > +//! Each section is preceded by a header that describes it. Most importantly, > +//! each header starts with a magic number that should be used by userspace to > +//! when decoding. > +//! > + > +use alloc::DumpAllocator; > +use kernel::bindings; > +use kernel::prelude::*; > + > +use crate::regs; > +use crate::regs::GpuRegister; > + > +// PANT > +const MAGIC: u32 = 0x544e4150; > + > +#[derive(Copy, Clone)] > +#[repr(u32)] > +enum HeaderType { > + /// A register dump > + Registers, > + /// The VM data, > + Vm, > + /// A dump of the firmware interface > + _FirmwareInterface, > +} > + > +#[repr(C)] > +pub(crate) struct DumpArgs { > + dev: *mut bindings::device, > + /// The slot for the job > + slot: i32, > + /// The active buffer objects > + bos: *mut *mut bindings::drm_gem_object, > + /// The number of active buffer objects > + bo_count: usize, > + /// The base address of the registers to use when reading. > + reg_base_addr: *mut core::ffi::c_void, > +} > + > +#[repr(C)] > +pub(crate) struct Header { > + magic: u32, > + ty: HeaderType, > + header_size: u32, > + data_size: u32, > +} > + > +#[repr(C)] > +#[derive(Clone, Copy)] > +pub(crate) struct RegisterDump { > + register: GpuRegister, > + value: u32, > +} > + > +/// The registers to dump > +const REGISTERS: [GpuRegister; 18] = [ > + regs::SHADER_READY_LO, > + regs::SHADER_READY_HI, > + regs::TILER_READY_LO, > + regs::TILER_READY_HI, > + regs::L2_READY_LO, > + regs::L2_READY_HI, > + regs::JOB_INT_MASK, > + regs::JOB_INT_STAT, > + regs::MMU_INT_MASK, > + regs::MMU_INT_STAT, > + regs::as_transtab_lo(0), > + regs::as_transtab_hi(0), > + regs::as_memattr_lo(0), > + regs::as_memattr_hi(0), > + regs::as_faultstatus(0), > + regs::as_faultaddress_lo(0), > + regs::as_faultaddress_hi(0), > + regs::as_status(0), > +]; > + > +mod alloc { > + use core::ptr::NonNull; > + > + use kernel::bindings; > + use kernel::prelude::*; > + > + use crate::dump::Header; > + use crate::dump::HeaderType; > + use crate::dump::MAGIC; > + > + pub(crate) struct DumpAllocator { > + mem: NonNull<core::ffi::c_void>, > + pos: usize, > + capacity: usize, > + } > + > + impl DumpAllocator { > + pub(crate) fn new(size: usize) -> Result<Self> { > + if isize::try_from(size).unwrap() == isize::MAX { > + return Err(EINVAL); > + } > + > + // Let's cheat a bit here, since there is no Rust vmalloc allocator > + // for the time being. > + // > + // Safety: just a FFI call to alloc memory > + let mem = NonNull::new(unsafe { > + bindings::__vmalloc_noprof( > + size.try_into().unwrap(), > + bindings::GFP_KERNEL | bindings::GFP_NOWAIT | 1 << bindings::___GFP_NORETRY_BIT, > + ) > + }); > + > + let mem = match mem { > + Some(buffer) => buffer, > + None => return Err(ENOMEM), > + }; > + > + // Ssfety: just a FFI call to zero out the memory. Mem and size were > + // used to allocate the memory above. > + unsafe { core::ptr::write_bytes(mem.as_ptr(), 0, size) }; > + Ok(Self { > + mem, > + pos: 0, > + capacity: size, > + }) > + } > + > + fn alloc_mem(&mut self, size: usize) -> Option<*mut u8> { > + assert!(size % 8 == 0, "Allocation size must be 8-byte aligned"); > + if isize::try_from(size).unwrap() == isize::MAX { > + return None; > + } else if self.pos + size > self.capacity { > + kernel::pr_debug!("DumpAllocator out of memory"); > + None > + } else { > + let offset = self.pos; > + self.pos += size; > + > + // Safety: we know that this is a valid allocation, so > + // dereferencing is safe. We don't ever return two pointers to > + // the same address, so we adhere to the aliasing rules. We make > + // sure that the memory is zero-initialized before being handed > + // out (this happens when the allocator is first created) and we > + // enforce a 8 byte alignment rule. > + Some(unsafe { self.mem.as_ptr().offset(offset as isize) as *mut u8 }) > + } > + } > + > + pub(crate) fn alloc<T>(&mut self) -> Option<&mut T> { > + let mem = self.alloc_mem(core::mem::size_of::<T>())? as *mut T; > + // Safety: we uphold safety guarantees in alloc_mem(), so this is > + // safe to dereference. This code doesn't properly handle when T requires a large alignment. > + Some(unsafe { &mut *mem }) > + } > + > + pub(crate) fn alloc_bytes(&mut self, num_bytes: usize) -> Option<&mut [u8]> { > + let mem = self.alloc_mem(num_bytes)?; > + > + // Safety: we uphold safety guarantees in alloc_mem(), so this is > + // safe to build a slice > + Some(unsafe { core::slice::from_raw_parts_mut(mem, num_bytes) }) > + } Using references for functions that allocate is generally wrong. References imply that you don't have ownership of the memory, but allocator functions would normally return ownership of the allocation. As-is, the code seems to leak these allocations. > + pub(crate) fn alloc_header(&mut self, ty: HeaderType, data_size: u32) -> &mut Header { > + let hdr: &mut Header = self.alloc().unwrap(); > + hdr.magic = MAGIC; > + hdr.ty = ty; > + hdr.header_size = core::mem::size_of::<Header>() as u32; > + hdr.data_size = data_size; > + hdr > + } > + > + pub(crate) fn is_end(&self) -> bool { > + self.pos == self.capacity > + } > + > + pub(crate) fn dump(self) -> (NonNull<core::ffi::c_void>, usize) { > + (self.mem, self.capacity) > + } > + } > +} > + > +fn dump_registers(alloc: &mut DumpAllocator, args: &DumpArgs) { > + let sz = core::mem::size_of_val(®ISTERS); > + alloc.alloc_header(HeaderType::Registers, sz.try_into().unwrap()); > + > + for reg in ®ISTERS { > + let dumped_reg: &mut RegisterDump = alloc.alloc().unwrap(); > + dumped_reg.register = *reg; > + dumped_reg.value = reg.read(args.reg_base_addr); > + } > +} > + > +fn dump_bo(alloc: &mut DumpAllocator, bo: &mut bindings::drm_gem_object) { > + let mut map = bindings::iosys_map::default(); > + > + // Safety: we trust the kernel to provide a valid BO. > + let ret = unsafe { bindings::drm_gem_vmap_unlocked(bo, &mut map as _) }; > + if ret != 0 { > + pr_warn!("Failed to map BO"); > + return; > + } > + > + let sz = bo.size; > + > + // Safety: we know that the vaddr is valid and we know the BO size. > + let mapped_bo: &mut [u8] = > + unsafe { core::slice::from_raw_parts_mut(map.__bindgen_anon_1.vaddr as *mut _, sz) }; You don't write to this memory, so I would avoid the mutable reference. > + alloc.alloc_header(HeaderType::Vm, sz as u32); > + > + let bo_data = alloc.alloc_bytes(sz).unwrap(); > + bo_data.copy_from_slice(&mapped_bo[..]); > + > + // Safety: BO is valid and was previously mapped. > + unsafe { bindings::drm_gem_vunmap_unlocked(bo, &mut map as _) }; You don't need `as _` here. You can just pass a mutable reference and Rust will automatically cast it to raw pointer. > +} > + > +/// Dumps the current state of the GPU to a file > +/// > +/// # Safety > +/// > +/// `Args` must be aligned and non-null. > +/// All fields of `DumpArgs` must be valid. > +#[no_mangle] > +pub(crate) extern "C" fn panthor_core_dump(args: *const DumpArgs) -> core::ffi::c_int { > + assert!(!args.is_null()); > + // Safety: we checked whether the pointer was null. It is assumed to be > + // aligned as per the safety requirements. > + let args = unsafe { &*args }; Creating a reference requires that it isn't dangling, so the safety requirements should require that. Also, panthor_core_dump should be unsafe. > + // > + // TODO: Ideally, we would use the safe GEM abstraction from the kernel > + // crate, but I see no way to create a drm::gem::ObjectRef from a > + // bindings::drm_gem_object. drm::gem::IntoGEMObject is only implemented for > + // drm::gem::Object, which means that new references can only be created > + // from a Rust-owned GEM object. > + // > + // It also has a has a `type Driver: drv::Driver` associated type, from > + // which it can access the `File` associated type. But not all GEM functions > + // take a file, though. For example, `drm_gem_vmap_unlocked` (used here) > + // does not. > + // > + // This associated type is a blocker here, because there is no actual > + // drv::Driver. We're only implementing a few functions in Rust. > + let mut bos = match Vec::with_capacity(args.bo_count, GFP_KERNEL) { > + Ok(bos) => bos, > + Err(_) => return ENOMEM.to_errno(), > + }; > + for i in 0..args.bo_count { > + // Safety: `args` is assumed valid as per the safety requirements. > + // `bos` is a valid pointer to a valid array of valid pointers. > + let bo = unsafe { &mut **args.bos.add(i) }; > + bos.push(bo, GFP_KERNEL).unwrap(); > + } > + > + let mut sz = core::mem::size_of::<Header>(); > + sz += REGISTERS.len() * core::mem::size_of::<RegisterDump>(); > + > + for bo in &mut *bos { > + sz += core::mem::size_of::<Header>(); > + sz += bo.size; > + } > + > + // Everything must fit within this allocation, otherwise it was miscomputed. > + let mut alloc = match DumpAllocator::new(sz) { > + Ok(alloc) => alloc, > + Err(e) => return e.to_errno(), > + }; > + > + dump_registers(&mut alloc, &args); > + for bo in bos { > + dump_bo(&mut alloc, bo); > + } > + > + if !alloc.is_end() { > + pr_warn!("DumpAllocator: wrong allocation size"); > + } > + > + let (mem, size) = alloc.dump(); > + > + // Safety: `mem` is a valid pointer to a valid allocation of `size` bytes. > + unsafe { bindings::dev_coredumpv(args.dev, mem.as_ptr(), size, bindings::GFP_KERNEL) }; > + > + 0 > +} > diff --git a/drivers/gpu/drm/panthor/lib.rs b/drivers/gpu/drm/panthor/lib.rs > new file mode 100644 > index 000000000000..faef8662d0f5 > --- /dev/null > +++ b/drivers/gpu/drm/panthor/lib.rs > @@ -0,0 +1,10 @@ > +// SPDX-License-Identifier: GPL-2.0 > +// SPDX-FileCopyrightText: Copyright Collabora 2024 > + > +//! The Rust components of the Panthor driver > + > +#[cfg(CONFIG_DRM_PANTHOR_COREDUMP)] > +mod dump; > +mod regs; > + > +const __LOG_PREFIX: &[u8] = b"panthor\0"; > diff --git a/drivers/gpu/drm/panthor/panthor_mmu.c b/drivers/gpu/drm/panthor/panthor_mmu.c > index fa0a002b1016..f8934de41ffa 100644 > --- a/drivers/gpu/drm/panthor/panthor_mmu.c > +++ b/drivers/gpu/drm/panthor/panthor_mmu.c > @@ -2,6 +2,8 @@ > /* Copyright 2019 Linaro, Ltd, Rob Herring <robh@xxxxxxxxxx> */ > /* Copyright 2023 Collabora ltd. */ > > +#include "drm/drm_gem.h" > +#include "linux/gfp_types.h" > #include <drm/drm_debugfs.h> > #include <drm/drm_drv.h> > #include <drm/drm_exec.h> > @@ -2619,6 +2621,43 @@ int panthor_vm_prepare_mapped_bos_resvs(struct drm_exec *exec, struct panthor_vm > return drm_gpuvm_prepare_objects(&vm->base, exec, slot_count); > } > > +/** > + * panthor_vm_bo_dump() - Dump the VM BOs for debugging purposes. > + * > + * > + * @vm: VM targeted by the GPU job. > + * @count: The number of BOs returned > + * > + * Return: an array of pointers to the BOs backing the whole VM. > + */ > +struct drm_gem_object ** > +panthor_vm_dump(struct panthor_vm *vm, u32 *count) > +{ > + struct drm_gpuva *va, *next; > + struct drm_gem_object **objs; > + *count = 0; > + u32 i = 0; > + > + mutex_lock(&vm->op_lock); > + drm_gpuvm_for_each_va_safe(va, next, &vm->base) { > + (*count)++; > + } > + > + objs = kcalloc(*count, sizeof(struct drm_gem_object *), GFP_KERNEL); > + if (!objs) { > + mutex_unlock(&vm->op_lock); > + return ERR_PTR(-ENOMEM); > + } > + > + drm_gpuvm_for_each_va_safe(va, next, &vm->base) { > + objs[i] = va->gem.obj; > + i++; > + } > + mutex_unlock(&vm->op_lock); > + > + return objs; > +} > + > /** > * panthor_mmu_unplug() - Unplug the MMU logic > * @ptdev: Device. > diff --git a/drivers/gpu/drm/panthor/panthor_mmu.h b/drivers/gpu/drm/panthor/panthor_mmu.h > index f3c1ed19f973..e9369c19e5b5 100644 > --- a/drivers/gpu/drm/panthor/panthor_mmu.h > +++ b/drivers/gpu/drm/panthor/panthor_mmu.h > @@ -50,6 +50,9 @@ int panthor_vm_add_bos_resvs_deps_to_job(struct panthor_vm *vm, > void panthor_vm_add_job_fence_to_bos_resvs(struct panthor_vm *vm, > struct drm_sched_job *job); > > +struct drm_gem_object ** > +panthor_vm_dump(struct panthor_vm *vm, u32 *count); > + > struct dma_resv *panthor_vm_resv(struct panthor_vm *vm); > struct drm_gem_object *panthor_vm_root_gem(struct panthor_vm *vm); > > diff --git a/drivers/gpu/drm/panthor/panthor_rs.h b/drivers/gpu/drm/panthor/panthor_rs.h > new file mode 100644 > index 000000000000..024db09be9a1 > --- /dev/null > +++ b/drivers/gpu/drm/panthor/panthor_rs.h > @@ -0,0 +1,40 @@ > +// SPDX-License-Identifier: GPL-2.0 > +// SPDX-FileCopyrightText: Copyright Collabora 2024 > + > +#include <drm/drm_gem.h> > + > +struct PanthorDumpArgs { > + struct device *dev; > + /** > + * The slot for the job > + */ > + s32 slot; > + /** > + * The active buffer objects > + */ > + struct drm_gem_object **bos; > + /** > + * The number of active buffer objects > + */ > + size_t bo_count; > + /** > + * The base address of the registers to use when reading. > + */ > + void *reg_base_addr; > +}; > + > +/** > + * Dumps the current state of the GPU to a file > + * > + * # Safety > + * > + * All fields of `DumpArgs` must be valid. > + */ > +#ifdef CONFIG_DRM_PANTHOR_RS > +int panthor_core_dump(const struct PanthorDumpArgs *args); > +#else > +inline int panthor_core_dump(const struct PanthorDumpArgs *args) > +{ > + return 0; > +} > +#endif > diff --git a/drivers/gpu/drm/panthor/panthor_sched.c b/drivers/gpu/drm/panthor/panthor_sched.c > index 79ffcbc41d78..39e1654d930e 100644 > --- a/drivers/gpu/drm/panthor/panthor_sched.c > +++ b/drivers/gpu/drm/panthor/panthor_sched.c > @@ -1,6 +1,9 @@ > // SPDX-License-Identifier: GPL-2.0 or MIT > /* Copyright 2023 Collabora ltd. */ > > +#include "drm/drm_gem.h" > +#include "linux/gfp_types.h" > +#include "linux/slab.h" > #include <drm/drm_drv.h> > #include <drm/drm_exec.h> > #include <drm/drm_gem_shmem_helper.h> > @@ -31,6 +34,7 @@ > #include "panthor_mmu.h" > #include "panthor_regs.h" > #include "panthor_sched.h" > +#include "panthor_rs.h" > > /** > * DOC: Scheduler > @@ -2805,6 +2809,27 @@ static void group_sync_upd_work(struct work_struct *work) > group_put(group); > } > > +static void dump_job(struct panthor_device *dev, struct panthor_job *job) > +{ > + struct panthor_vm *vm = job->group->vm; > + struct drm_gem_object **objs; > + u32 count; > + > + objs = panthor_vm_dump(vm, &count); > + > + if (!IS_ERR(objs)) { > + struct PanthorDumpArgs args = { > + .dev = job->group->ptdev->base.dev, > + .bos = objs, > + .bo_count = count, > + .reg_base_addr = dev->iomem, > + }; > + panthor_core_dump(&args); > + kfree(objs); > + } > +} > + > + > static struct dma_fence * > queue_run_job(struct drm_sched_job *sched_job) > { > @@ -2929,7 +2954,7 @@ queue_run_job(struct drm_sched_job *sched_job) > } > > done_fence = dma_fence_get(job->done_fence); > - > + dump_job(ptdev, job); > out_unlock: > mutex_unlock(&sched->lock); > pm_runtime_mark_last_busy(ptdev->base.dev); > @@ -2950,6 +2975,7 @@ queue_timedout_job(struct drm_sched_job *sched_job) > drm_warn(&ptdev->base, "job timeout\n"); > > drm_WARN_ON(&ptdev->base, atomic_read(&sched->reset.in_progress)); > + dump_job(ptdev, job); > > queue_stop(queue, job); > > diff --git a/drivers/gpu/drm/panthor/regs.rs b/drivers/gpu/drm/panthor/regs.rs > new file mode 100644 > index 000000000000..514bc9ee2856 > --- /dev/null > +++ b/drivers/gpu/drm/panthor/regs.rs > @@ -0,0 +1,264 @@ > +// SPDX-License-Identifier: GPL-2.0 > +// SPDX-FileCopyrightText: Copyright Collabora 2024 > +// SPDX-FileCopyrightText: (C) COPYRIGHT 2010-2022 ARM Limited. All rights reserved. > + > +//! The registers for Panthor, extracted from panthor_regs.h > + > +#![allow(unused_macros, unused_imports, dead_code)] > + > +use kernel::bindings; > + > +use core::ops::Add; > +use core::ops::Shl; > +use core::ops::Shr; > + > +#[repr(transparent)] > +#[derive(Clone, Copy)] > +pub(crate) struct GpuRegister(u64); > + > +impl GpuRegister { > + pub(crate) fn read(&self, iomem: *const core::ffi::c_void) -> u32 { > + // Safety: `reg` represents a valid address > + unsafe { > + let addr = iomem.offset(self.0 as isize); > + bindings::readl_relaxed(addr as *const _) > + } > + } > +} > + > +pub(crate) const fn bit(index: u64) -> u64 { > + 1 << index > +} > +pub(crate) const fn genmask(high: u64, low: u64) -> u64 { > + ((1 << (high - low + 1)) - 1) << low > +} > + > +pub(crate) const GPU_ID: GpuRegister = GpuRegister(0x0); > +pub(crate) const fn gpu_arch_major(x: u64) -> GpuRegister { > + GpuRegister((x) >> 28) > +} > +pub(crate) const fn gpu_arch_minor(x: u64) -> GpuRegister { > + GpuRegister((x) & genmask(27, 24) >> 24) > +} > +pub(crate) const fn gpu_arch_rev(x: u64) -> GpuRegister { > + GpuRegister((x) & genmask(23, 20) >> 20) > +} > +pub(crate) const fn gpu_prod_major(x: u64) -> GpuRegister { > + GpuRegister((x) & genmask(19, 16) >> 16) > +} > +pub(crate) const fn gpu_ver_major(x: u64) -> GpuRegister { > + GpuRegister((x) & genmask(15, 12) >> 12) > +} > +pub(crate) const fn gpu_ver_minor(x: u64) -> GpuRegister { > + GpuRegister((x) & genmask(11, 4) >> 4) > +} > +pub(crate) const fn gpu_ver_status(x: u64) -> GpuRegister { > + GpuRegister(x & genmask(3, 0)) > +} > +pub(crate) const GPU_L2_FEATURES: GpuRegister = GpuRegister(0x4); > +pub(crate) const fn gpu_l2_features_line_size(x: u64) -> GpuRegister { > + GpuRegister(1 << ((x) & genmask(7, 0))) > +} > +pub(crate) const GPU_CORE_FEATURES: GpuRegister = GpuRegister(0x8); > +pub(crate) const GPU_TILER_FEATURES: GpuRegister = GpuRegister(0xc); > +pub(crate) const GPU_MEM_FEATURES: GpuRegister = GpuRegister(0x10); > +pub(crate) const GROUPS_L2_COHERENT: GpuRegister = GpuRegister(bit(0)); > +pub(crate) const GPU_MMU_FEATURES: GpuRegister = GpuRegister(0x14); > +pub(crate) const fn gpu_mmu_features_va_bits(x: u64) -> GpuRegister { > + GpuRegister((x) & genmask(7, 0)) > +} > +pub(crate) const fn gpu_mmu_features_pa_bits(x: u64) -> GpuRegister { > + GpuRegister(((x) >> 8) & genmask(7, 0)) > +} > +pub(crate) const GPU_AS_PRESENT: GpuRegister = GpuRegister(0x18); > +pub(crate) const GPU_CSF_ID: GpuRegister = GpuRegister(0x1c); > +pub(crate) const GPU_INT_RAWSTAT: GpuRegister = GpuRegister(0x20); > +pub(crate) const GPU_INT_CLEAR: GpuRegister = GpuRegister(0x24); > +pub(crate) const GPU_INT_MASK: GpuRegister = GpuRegister(0x28); > +pub(crate) const GPU_INT_STAT: GpuRegister = GpuRegister(0x2c); > +pub(crate) const GPU_IRQ_FAULT: GpuRegister = GpuRegister(bit(0)); > +pub(crate) const GPU_IRQ_PROTM_FAULT: GpuRegister = GpuRegister(bit(1)); > +pub(crate) const GPU_IRQ_RESET_COMPLETED: GpuRegister = GpuRegister(bit(8)); > +pub(crate) const GPU_IRQ_POWER_CHANGED: GpuRegister = GpuRegister(bit(9)); > +pub(crate) const GPU_IRQ_POWER_CHANGED_ALL: GpuRegister = GpuRegister(bit(10)); > +pub(crate) const GPU_IRQ_CLEAN_CACHES_COMPLETED: GpuRegister = GpuRegister(bit(17)); > +pub(crate) const GPU_IRQ_DOORBELL_MIRROR: GpuRegister = GpuRegister(bit(18)); > +pub(crate) const GPU_IRQ_MCU_STATUS_CHANGED: GpuRegister = GpuRegister(bit(19)); > +pub(crate) const GPU_CMD: GpuRegister = GpuRegister(0x30); > +const fn gpu_cmd_def(ty: u64, payload: u64) -> u64 { > + (ty) | ((payload) << 8) > +} > +pub(crate) const fn gpu_soft_reset() -> GpuRegister { > + GpuRegister(gpu_cmd_def(1, 1)) > +} > +pub(crate) const fn gpu_hard_reset() -> GpuRegister { > + GpuRegister(gpu_cmd_def(1, 2)) > +} > +pub(crate) const CACHE_CLEAN: GpuRegister = GpuRegister(bit(0)); > +pub(crate) const CACHE_INV: GpuRegister = GpuRegister(bit(1)); > +pub(crate) const GPU_STATUS: GpuRegister = GpuRegister(0x34); > +pub(crate) const GPU_STATUS_ACTIVE: GpuRegister = GpuRegister(bit(0)); > +pub(crate) const GPU_STATUS_PWR_ACTIVE: GpuRegister = GpuRegister(bit(1)); > +pub(crate) const GPU_STATUS_PAGE_FAULT: GpuRegister = GpuRegister(bit(4)); > +pub(crate) const GPU_STATUS_PROTM_ACTIVE: GpuRegister = GpuRegister(bit(7)); > +pub(crate) const GPU_STATUS_DBG_ENABLED: GpuRegister = GpuRegister(bit(8)); > +pub(crate) const GPU_FAULT_STATUS: GpuRegister = GpuRegister(0x3c); > +pub(crate) const GPU_FAULT_ADDR_LO: GpuRegister = GpuRegister(0x40); > +pub(crate) const GPU_FAULT_ADDR_HI: GpuRegister = GpuRegister(0x44); > +pub(crate) const GPU_PWR_KEY: GpuRegister = GpuRegister(0x50); > +pub(crate) const GPU_PWR_KEY_UNLOCK: GpuRegister = GpuRegister(0x2968a819); > +pub(crate) const GPU_PWR_OVERRIDE0: GpuRegister = GpuRegister(0x54); > +pub(crate) const GPU_PWR_OVERRIDE1: GpuRegister = GpuRegister(0x58); > +pub(crate) const GPU_TIMESTAMP_OFFSET_LO: GpuRegister = GpuRegister(0x88); > +pub(crate) const GPU_TIMESTAMP_OFFSET_HI: GpuRegister = GpuRegister(0x8c); > +pub(crate) const GPU_CYCLE_COUNT_LO: GpuRegister = GpuRegister(0x90); > +pub(crate) const GPU_CYCLE_COUNT_HI: GpuRegister = GpuRegister(0x94); > +pub(crate) const GPU_TIMESTAMP_LO: GpuRegister = GpuRegister(0x98); > +pub(crate) const GPU_TIMESTAMP_HI: GpuRegister = GpuRegister(0x9c); > +pub(crate) const GPU_THREAD_MAX_THREADS: GpuRegister = GpuRegister(0xa0); > +pub(crate) const GPU_THREAD_MAX_WORKGROUP_SIZE: GpuRegister = GpuRegister(0xa4); > +pub(crate) const GPU_THREAD_MAX_BARRIER_SIZE: GpuRegister = GpuRegister(0xa8); > +pub(crate) const GPU_THREAD_FEATURES: GpuRegister = GpuRegister(0xac); > +pub(crate) const fn gpu_texture_features(n: u64) -> GpuRegister { > + GpuRegister(0xB0 + ((n) * 4)) > +} > +pub(crate) const GPU_SHADER_PRESENT_LO: GpuRegister = GpuRegister(0x100); > +pub(crate) const GPU_SHADER_PRESENT_HI: GpuRegister = GpuRegister(0x104); > +pub(crate) const GPU_TILER_PRESENT_LO: GpuRegister = GpuRegister(0x110); > +pub(crate) const GPU_TILER_PRESENT_HI: GpuRegister = GpuRegister(0x114); > +pub(crate) const GPU_L2_PRESENT_LO: GpuRegister = GpuRegister(0x120); > +pub(crate) const GPU_L2_PRESENT_HI: GpuRegister = GpuRegister(0x124); > +pub(crate) const SHADER_READY_LO: GpuRegister = GpuRegister(0x140); > +pub(crate) const SHADER_READY_HI: GpuRegister = GpuRegister(0x144); > +pub(crate) const TILER_READY_LO: GpuRegister = GpuRegister(0x150); > +pub(crate) const TILER_READY_HI: GpuRegister = GpuRegister(0x154); > +pub(crate) const L2_READY_LO: GpuRegister = GpuRegister(0x160); > +pub(crate) const L2_READY_HI: GpuRegister = GpuRegister(0x164); > +pub(crate) const SHADER_PWRON_LO: GpuRegister = GpuRegister(0x180); > +pub(crate) const SHADER_PWRON_HI: GpuRegister = GpuRegister(0x184); > +pub(crate) const TILER_PWRON_LO: GpuRegister = GpuRegister(0x190); > +pub(crate) const TILER_PWRON_HI: GpuRegister = GpuRegister(0x194); > +pub(crate) const L2_PWRON_LO: GpuRegister = GpuRegister(0x1a0); > +pub(crate) const L2_PWRON_HI: GpuRegister = GpuRegister(0x1a4); > +pub(crate) const SHADER_PWROFF_LO: GpuRegister = GpuRegister(0x1c0); > +pub(crate) const SHADER_PWROFF_HI: GpuRegister = GpuRegister(0x1c4); > +pub(crate) const TILER_PWROFF_LO: GpuRegister = GpuRegister(0x1d0); > +pub(crate) const TILER_PWROFF_HI: GpuRegister = GpuRegister(0x1d4); > +pub(crate) const L2_PWROFF_LO: GpuRegister = GpuRegister(0x1e0); > +pub(crate) const L2_PWROFF_HI: GpuRegister = GpuRegister(0x1e4); > +pub(crate) const SHADER_PWRTRANS_LO: GpuRegister = GpuRegister(0x200); > +pub(crate) const SHADER_PWRTRANS_HI: GpuRegister = GpuRegister(0x204); > +pub(crate) const TILER_PWRTRANS_LO: GpuRegister = GpuRegister(0x210); > +pub(crate) const TILER_PWRTRANS_HI: GpuRegister = GpuRegister(0x214); > +pub(crate) const L2_PWRTRANS_LO: GpuRegister = GpuRegister(0x220); > +pub(crate) const L2_PWRTRANS_HI: GpuRegister = GpuRegister(0x224); > +pub(crate) const SHADER_PWRACTIVE_LO: GpuRegister = GpuRegister(0x240); > +pub(crate) const SHADER_PWRACTIVE_HI: GpuRegister = GpuRegister(0x244); > +pub(crate) const TILER_PWRACTIVE_LO: GpuRegister = GpuRegister(0x250); > +pub(crate) const TILER_PWRACTIVE_HI: GpuRegister = GpuRegister(0x254); > +pub(crate) const L2_PWRACTIVE_LO: GpuRegister = GpuRegister(0x260); > +pub(crate) const L2_PWRACTIVE_HI: GpuRegister = GpuRegister(0x264); > +pub(crate) const GPU_REVID: GpuRegister = GpuRegister(0x280); > +pub(crate) const GPU_COHERENCY_FEATURES: GpuRegister = GpuRegister(0x300); > +pub(crate) const GPU_COHERENCY_PROTOCOL: GpuRegister = GpuRegister(0x304); > +pub(crate) const GPU_COHERENCY_ACE: GpuRegister = GpuRegister(0); > +pub(crate) const GPU_COHERENCY_ACE_LITE: GpuRegister = GpuRegister(1); > +pub(crate) const GPU_COHERENCY_NONE: GpuRegister = GpuRegister(31); > +pub(crate) const MCU_CONTROL: GpuRegister = GpuRegister(0x700); > +pub(crate) const MCU_CONTROL_ENABLE: GpuRegister = GpuRegister(1); > +pub(crate) const MCU_CONTROL_AUTO: GpuRegister = GpuRegister(2); > +pub(crate) const MCU_CONTROL_DISABLE: GpuRegister = GpuRegister(0); > +pub(crate) const MCU_STATUS: GpuRegister = GpuRegister(0x704); > +pub(crate) const MCU_STATUS_DISABLED: GpuRegister = GpuRegister(0); > +pub(crate) const MCU_STATUS_ENABLED: GpuRegister = GpuRegister(1); > +pub(crate) const MCU_STATUS_HALT: GpuRegister = GpuRegister(2); > +pub(crate) const MCU_STATUS_FATAL: GpuRegister = GpuRegister(3); > +pub(crate) const JOB_INT_RAWSTAT: GpuRegister = GpuRegister(0x1000); > +pub(crate) const JOB_INT_CLEAR: GpuRegister = GpuRegister(0x1004); > +pub(crate) const JOB_INT_MASK: GpuRegister = GpuRegister(0x1008); > +pub(crate) const JOB_INT_STAT: GpuRegister = GpuRegister(0x100c); > +pub(crate) const JOB_INT_GLOBAL_IF: GpuRegister = GpuRegister(bit(31)); > +pub(crate) const fn job_int_csg_if(x: u64) -> GpuRegister { > + GpuRegister(bit(x)) > +} > +pub(crate) const MMU_INT_RAWSTAT: GpuRegister = GpuRegister(0x2000); > +pub(crate) const MMU_INT_CLEAR: GpuRegister = GpuRegister(0x2004); > +pub(crate) const MMU_INT_MASK: GpuRegister = GpuRegister(0x2008); > +pub(crate) const MMU_INT_STAT: GpuRegister = GpuRegister(0x200c); > +pub(crate) const MMU_BASE: GpuRegister = GpuRegister(0x2400); > +pub(crate) const MMU_AS_SHIFT: GpuRegister = GpuRegister(6); > +const fn mmu_as(as_: u64) -> u64 { > + MMU_BASE.0 + ((as_) << MMU_AS_SHIFT.0) > +} > +pub(crate) const fn as_transtab_lo(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x0) > +} > +pub(crate) const fn as_transtab_hi(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x4) > +} > +pub(crate) const fn as_memattr_lo(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x8) > +} > +pub(crate) const fn as_memattr_hi(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0xC) > +} > +pub(crate) const fn as_memattr_aarch64_inner_alloc_expl(w: u64, r: u64) -> GpuRegister { > + GpuRegister((3 << 2) | (if w > 0 { bit(0) } else { 0 } | (if r > 0 { bit(1) } else { 0 }))) > +} > +pub(crate) const fn as_lockaddr_lo(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x10) > +} > +pub(crate) const fn as_lockaddr_hi(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x14) > +} > +pub(crate) const fn as_command(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x18) > +} > +pub(crate) const AS_COMMAND_NOP: GpuRegister = GpuRegister(0); > +pub(crate) const AS_COMMAND_UPDATE: GpuRegister = GpuRegister(1); > +pub(crate) const AS_COMMAND_LOCK: GpuRegister = GpuRegister(2); > +pub(crate) const AS_COMMAND_UNLOCK: GpuRegister = GpuRegister(3); > +pub(crate) const AS_COMMAND_FLUSH_PT: GpuRegister = GpuRegister(4); > +pub(crate) const AS_COMMAND_FLUSH_MEM: GpuRegister = GpuRegister(5); > +pub(crate) const fn as_faultstatus(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x1C) > +} > +pub(crate) const fn as_faultaddress_lo(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x20) > +} > +pub(crate) const fn as_faultaddress_hi(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x24) > +} > +pub(crate) const fn as_status(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x28) > +} > +pub(crate) const AS_STATUS_AS_ACTIVE: GpuRegister = GpuRegister(bit(0)); > +pub(crate) const fn as_transcfg_lo(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x30) > +} > +pub(crate) const fn as_transcfg_hi(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x34) > +} > +pub(crate) const fn as_transcfg_ina_bits(x: u64) -> GpuRegister { > + GpuRegister((x) << 6) > +} > +pub(crate) const fn as_transcfg_outa_bits(x: u64) -> GpuRegister { > + GpuRegister((x) << 14) > +} > +pub(crate) const AS_TRANSCFG_SL_CONCAT: GpuRegister = GpuRegister(bit(22)); > +pub(crate) const AS_TRANSCFG_PTW_RA: GpuRegister = GpuRegister(bit(30)); > +pub(crate) const AS_TRANSCFG_DISABLE_HIER_AP: GpuRegister = GpuRegister(bit(33)); > +pub(crate) const AS_TRANSCFG_DISABLE_AF_FAULT: GpuRegister = GpuRegister(bit(34)); > +pub(crate) const AS_TRANSCFG_WXN: GpuRegister = GpuRegister(bit(35)); > +pub(crate) const AS_TRANSCFG_XREADABLE: GpuRegister = GpuRegister(bit(36)); > +pub(crate) const fn as_faultextra_lo(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x38) > +} > +pub(crate) const fn as_faultextra_hi(as_: u64) -> GpuRegister { > + GpuRegister(mmu_as(as_) + 0x3C) > +} > +pub(crate) const CSF_GPU_LATEST_FLUSH_ID: GpuRegister = GpuRegister(0x10000); > +pub(crate) const fn csf_doorbell(i: u64) -> GpuRegister { > + GpuRegister(0x80000 + ((i) * 0x10000)) > +} > +pub(crate) const CSF_GLB_DOORBELL_ID: GpuRegister = GpuRegister(0); > diff --git a/rust/bindings/bindings_helper.h b/rust/bindings/bindings_helper.h > index b245db8d5a87..4ee4b97e7930 100644 > --- a/rust/bindings/bindings_helper.h > +++ b/rust/bindings/bindings_helper.h > @@ -12,15 +12,18 @@ > #include <drm/drm_gem.h> > #include <drm/drm_ioctl.h> > #include <kunit/test.h> > +#include <linux/devcoredump.h> > #include <linux/errname.h> > #include <linux/ethtool.h> > #include <linux/jiffies.h> > +#include <linux/iosys-map.h> > #include <linux/mdio.h> > #include <linux/pci.h> > #include <linux/phy.h> > #include <linux/refcount.h> > #include <linux/sched.h> > #include <linux/slab.h> > +#include <linux/vmalloc.h> > #include <linux/wait.h> > #include <linux/workqueue.h> > > -- > 2.45.2 > >