On Thu, 5 Dec 2024 23:39:07 +0000 Adrián Larumbe <adrian.larumbe@xxxxxxxxxxxxx> wrote: > This will display the sizes of kenrel BO's bound to an open file, which are > otherwise not exposed to UM through a handle. > > The sizes recorded are as follows: > - Per group: suspend buffer, protm-suspend buffer, syncobjcs > - Per queue: ringbuffer, profiling slots, firmware interface > - For all heaps in all heap pools across all VM's bound to an open file, > record size of all heap chuks, and for each pool the gpu_context BO too. > > This does not record the size of FW regions, as these aren't bound to a > specific open file and remain active through the whole life of the driver. > > Signed-off-by: Adrián Larumbe <adrian.larumbe@xxxxxxxxxxxxx> > --- > drivers/gpu/drm/panthor/panthor_drv.c | 14 +++++++++- > drivers/gpu/drm/panthor/panthor_heap.c | 26 ++++++++++++++++++ > drivers/gpu/drm/panthor/panthor_heap.h | 2 ++ > drivers/gpu/drm/panthor/panthor_mmu.c | 35 +++++++++++++++++++++++++ > drivers/gpu/drm/panthor/panthor_mmu.h | 4 +++ > drivers/gpu/drm/panthor/panthor_sched.c | 26 ++++++++++++++++++ > drivers/gpu/drm/panthor/panthor_sched.h | 4 +++ > 7 files changed, 110 insertions(+), 1 deletion(-) > > diff --git a/drivers/gpu/drm/panthor/panthor_drv.c b/drivers/gpu/drm/panthor/panthor_drv.c > index ac7e53f6e3f0..94f1d5f16e35 100644 > --- a/drivers/gpu/drm/panthor/panthor_drv.c > +++ b/drivers/gpu/drm/panthor/panthor_drv.c > @@ -876,7 +876,7 @@ static int panthor_ioctl_vm_create(struct drm_device *ddev, void *data, > if (!drm_dev_enter(ddev, &cookie)) > return -ENODEV; > > - ret = panthor_vm_pool_create_vm(ptdev, pfile->vms, args); > + ret = panthor_vm_pool_create_vm(ptdev, pfile->vms, args); Looks like an unrelated formatting fix. Can you move it to its own commit? > if (ret >= 0) { > args->id = ret; > ret = 0; > @@ -1457,12 +1457,24 @@ static void panthor_gpu_show_fdinfo(struct panthor_device *ptdev, > drm_printf(p, "drm-curfreq-panthor:\t%lu Hz\n", ptdev->current_frequency); > } > > +static void panthor_show_internal_memory_stats(struct drm_printer *p, struct drm_file *file) > +{ > + struct panthor_file *pfile = file->driver_priv; > + struct drm_memory_stats status = {0}; > + > + panthor_group_internal_sizes(pfile, &status); > + panthor_vm_heaps_size(pfile, &status); > + > + drm_print_memory_stats(p, &status, DRM_GEM_OBJECT_RESIDENT, "internal"); > +} > + > static void panthor_show_fdinfo(struct drm_printer *p, struct drm_file *file) > { > struct drm_device *dev = file->minor->dev; > struct panthor_device *ptdev = container_of(dev, struct panthor_device, base); > > panthor_gpu_show_fdinfo(ptdev, file->driver_priv, p); > + panthor_show_internal_memory_stats(p, file); > > drm_show_memory_stats(p, file); > } > diff --git a/drivers/gpu/drm/panthor/panthor_heap.c b/drivers/gpu/drm/panthor/panthor_heap.c > index 3796a9eb22af..e4464c5e93ef 100644 > --- a/drivers/gpu/drm/panthor/panthor_heap.c > +++ b/drivers/gpu/drm/panthor/panthor_heap.c > @@ -603,3 +603,29 @@ void panthor_heap_pool_destroy(struct panthor_heap_pool *pool) > > panthor_heap_pool_put(pool); > } > + > +/** > + * panthor_heap_pool_size() - Calculate size of all chunks across all heaps in a pool > + * @pool: Pool whose total chunk size to calculate. > + * > + * This function adds the size of all heap chunks across all heaps in the > + * argument pool. It also adds the size of the gpu contexts kernel bo. > + * It is meant to be used by fdinfo for displaying the size of internal > + * driver BO's that aren't exposed to userspace through a GEM handle. > + * > + */ > +size_t panthor_heap_pool_size(struct panthor_heap_pool *pool) > +{ > + struct panthor_heap *heap; > + unsigned long i; > + size_t size = 0; > + > + down_write(&pool->lock); > + xa_for_each(&pool->xa, i, heap) > + size += heap->chunk_size * heap->chunk_count; > + up_write(&pool->lock); > + > + size += pool->gpu_contexts->obj->size; > + > + return size; > +} > diff --git a/drivers/gpu/drm/panthor/panthor_heap.h b/drivers/gpu/drm/panthor/panthor_heap.h > index 25a5f2bba445..e3358d4e8edb 100644 > --- a/drivers/gpu/drm/panthor/panthor_heap.h > +++ b/drivers/gpu/drm/panthor/panthor_heap.h > @@ -27,6 +27,8 @@ struct panthor_heap_pool * > panthor_heap_pool_get(struct panthor_heap_pool *pool); > void panthor_heap_pool_put(struct panthor_heap_pool *pool); > > +size_t panthor_heap_pool_size(struct panthor_heap_pool *pool); > + > int panthor_heap_grow(struct panthor_heap_pool *pool, > u64 heap_gpu_va, > u32 renderpasses_in_flight, > diff --git a/drivers/gpu/drm/panthor/panthor_mmu.c b/drivers/gpu/drm/panthor/panthor_mmu.c > index 7ba8470a7543..e2f27a1667c3 100644 > --- a/drivers/gpu/drm/panthor/panthor_mmu.c > +++ b/drivers/gpu/drm/panthor/panthor_mmu.c > @@ -1937,6 +1937,41 @@ struct panthor_heap_pool *panthor_vm_get_heap_pool(struct panthor_vm *vm, bool c > return pool; > } > > +/** > + * panthor_vm_heaps_size() - Calculate size of all heap chunks across all > + * heaps over all the heap pools in a VM > + * @pfile: File. > + * @status: Memory status to be updated. > + * > + * Calculate all heap chunk sizes in all heap pools bound to a VM. If the VM > + * is active, record the size as active as well. > + */ > +void panthor_vm_heaps_size(struct panthor_file *pfile, struct drm_memory_stats *status) > +{ > + struct panthor_vm *vm; > + unsigned long i; > + > + if (!pfile->vms) > + return; > + > + xa_for_each(&pfile->vms->xa, i, vm) { > + size_t size; > + > + mutex_lock(&vm->heaps.lock); > + if (!vm->heaps.pool) { > + mutex_unlock(&vm->heaps.lock); > + continue; > + } > + size = panthor_heap_pool_size(vm->heaps.pool); > + mutex_unlock(&vm->heaps.lock); > + > + status->resident += size; > + status->private += size; > + if (vm->as.id >= 0) > + status->active += size; > + } > +} > + > static u64 mair_to_memattr(u64 mair, bool coherent) > { > u64 memattr = 0; > diff --git a/drivers/gpu/drm/panthor/panthor_mmu.h b/drivers/gpu/drm/panthor/panthor_mmu.h > index 8d21e83d8aba..25f7aea39ed9 100644 > --- a/drivers/gpu/drm/panthor/panthor_mmu.h > +++ b/drivers/gpu/drm/panthor/panthor_mmu.h > @@ -5,10 +5,12 @@ > #ifndef __PANTHOR_MMU_H__ > #define __PANTHOR_MMU_H__ > > +#include <linux/types.h> > #include <linux/dma-resv.h> > > struct drm_exec; > struct drm_sched_job; > +struct drm_memory_stats; > struct panthor_gem_object; > struct panthor_heap_pool; > struct panthor_vm; > @@ -37,6 +39,8 @@ int panthor_vm_flush_all(struct panthor_vm *vm); > struct panthor_heap_pool * > panthor_vm_get_heap_pool(struct panthor_vm *vm, bool create); > > +void panthor_vm_heaps_size(struct panthor_file *pfile, struct drm_memory_stats *status); > + > struct panthor_vm *panthor_vm_get(struct panthor_vm *vm); > void panthor_vm_put(struct panthor_vm *vm); > struct panthor_vm *panthor_vm_create(struct panthor_device *ptdev, bool for_mcu, > diff --git a/drivers/gpu/drm/panthor/panthor_sched.c b/drivers/gpu/drm/panthor/panthor_sched.c > index ef4bec7ff9c7..6a4d5f63c86b 100644 > --- a/drivers/gpu/drm/panthor/panthor_sched.c > +++ b/drivers/gpu/drm/panthor/panthor_sched.c > @@ -582,6 +582,9 @@ struct panthor_group { > /** @queues: Queues owned by this group. */ > struct panthor_queue *queues[MAX_CS_PER_CSG]; > > + /** @bo_sizes: Aggregate size of internal kernel BO's held by the group. */ > + size_t kbo_sizes; Given fdinfo_show() is not a hot-path, I'd do the sum of all kbos in a helper function taking a group as an argument. > + > /** > * @csg_id: ID of the FW group slot. > * > @@ -3305,6 +3308,7 @@ group_create_queue(struct panthor_group *group, > ret = PTR_ERR(queue->ringbuf); > goto err_free_queue; > } > + group->kbo_sizes += queue->ringbuf->obj->size; > > ret = panthor_kernel_bo_vmap(queue->ringbuf); > if (ret) > @@ -3319,6 +3323,7 @@ group_create_queue(struct panthor_group *group, > ret = PTR_ERR(queue->iface.mem); > goto err_free_queue; > } > + group->kbo_sizes += queue->iface.mem->obj->size; > > queue->profiling.slot_count = > calc_profiling_ringbuf_num_slots(group->ptdev, args->ringbuf_size); > @@ -3336,6 +3341,7 @@ group_create_queue(struct panthor_group *group, > ret = PTR_ERR(queue->profiling.slots); > goto err_free_queue; > } > + group->kbo_sizes += queue->profiling.slots->obj->size; > > ret = panthor_kernel_bo_vmap(queue->profiling.slots); > if (ret) > @@ -3433,6 +3439,7 @@ int panthor_group_create(struct panthor_file *pfile, > group->suspend_buf = NULL; > goto err_put_group; > } > + group->kbo_sizes += group->suspend_buf->obj->size; > > suspend_size = csg_iface->control->protm_suspend_size; > group->protm_suspend_buf = panthor_fw_alloc_suspend_buf_mem(ptdev, suspend_size); > @@ -3441,6 +3448,7 @@ int panthor_group_create(struct panthor_file *pfile, > group->protm_suspend_buf = NULL; > goto err_put_group; > } > + group->kbo_sizes += group->protm_suspend_buf->obj->size; > > group->syncobjs = panthor_kernel_bo_create(ptdev, group->vm, > group_args->queues.count * > @@ -3453,6 +3461,7 @@ int panthor_group_create(struct panthor_file *pfile, > ret = PTR_ERR(group->syncobjs); > goto err_put_group; > } > + group->kbo_sizes += group->syncobjs->obj->size; > > ret = panthor_kernel_bo_vmap(group->syncobjs); > if (ret) > @@ -3606,6 +3615,23 @@ void panthor_group_pool_destroy(struct panthor_file *pfile) > pfile->groups = NULL; > } > > +void panthor_group_internal_sizes(struct panthor_file *pfile, struct drm_memory_stats *status) > +{ > + struct panthor_group_pool *gpool = pfile->groups; > + struct panthor_group *group; > + unsigned long i; > + > + if (IS_ERR_OR_NULL(gpool)) > + return; > + > + xa_for_each(&gpool->xa, i, group) { > + status->resident += group->kbo_sizes; > + status->private += group->kbo_sizes; > + if (group->csg_id >= 0) > + status->active += group->kbo_sizes; > + } > +} > + > static void job_release(struct kref *ref) > { > struct panthor_job *job = container_of(ref, struct panthor_job, refcount); > diff --git a/drivers/gpu/drm/panthor/panthor_sched.h b/drivers/gpu/drm/panthor/panthor_sched.h > index 5ae6b4bde7c5..e17c56a40d9c 100644 > --- a/drivers/gpu/drm/panthor/panthor_sched.h > +++ b/drivers/gpu/drm/panthor/panthor_sched.h > @@ -4,11 +4,14 @@ > #ifndef __PANTHOR_SCHED_H__ > #define __PANTHOR_SCHED_H__ > > +#include <linux/types.h> > + > struct drm_exec; > struct dma_fence; > struct drm_file; > struct drm_gem_object; > struct drm_sched_job; > +struct drm_memory_stats; > struct drm_panthor_group_create; > struct drm_panthor_queue_create; > struct drm_panthor_group_get_state; > @@ -36,6 +39,7 @@ void panthor_job_update_resvs(struct drm_exec *exec, struct drm_sched_job *job); > > int panthor_group_pool_create(struct panthor_file *pfile); > void panthor_group_pool_destroy(struct panthor_file *pfile); > +void panthor_group_internal_sizes(struct panthor_file *pfile, struct drm_memory_stats *status); s/panthor_group_internal_sizes/panthor_group_kbo_sizes/, as I find the term internal a bit vague. This looks good otherwise, and I certainly prefer this version over the previous one involving a global lock and per-file kernel-BO registration.