On Thu, Jun 06, 2024 at 03:18:14PM +0200, Christian König wrote: > Am 06.06.24 um 15:06 schrieb Pierre-Eric Pelloux-Prayer: > > When tracing is enabled, being able to identify which device is sending > > events is useful; for this the next commit will extend events to include > > drm_device::primary::index. > > That sounds like a rather bad idea since the primary index is really just an > arbitrary number and not defined for all devices. > > Why not use the device name instead? This way you don't need this change in > the first place. > +1. Matt > Regards, > Christian. > > > > > Since the device member is only used in the drm_* log macros, we can > > replace it by a drm_device pointer. > > > > Signed-off-by: Pierre-Eric Pelloux-Prayer <pierre-eric.pelloux-prayer@xxxxxxx> > > --- > > drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 2 +- > > drivers/gpu/drm/etnaviv/etnaviv_sched.c | 2 +- > > drivers/gpu/drm/imagination/pvr_queue.c | 2 +- > > drivers/gpu/drm/lima/lima_sched.c | 2 +- > > drivers/gpu/drm/msm/msm_ringbuffer.c | 2 +- > > drivers/gpu/drm/nouveau/nouveau_sched.c | 2 +- > > drivers/gpu/drm/panfrost/panfrost_job.c | 2 +- > > drivers/gpu/drm/panthor/panthor_mmu.c | 2 +- > > drivers/gpu/drm/panthor/panthor_sched.c | 2 +- > > drivers/gpu/drm/scheduler/sched_entity.c | 2 +- > > drivers/gpu/drm/scheduler/sched_main.c | 26 +++++++++++----------- > > drivers/gpu/drm/v3d/v3d_sched.c | 12 +++++----- > > drivers/gpu/drm/xe/xe_execlist.c | 2 +- > > drivers/gpu/drm/xe/xe_gpu_scheduler.c | 2 +- > > drivers/gpu/drm/xe/xe_gpu_scheduler.h | 2 +- > > drivers/gpu/drm/xe/xe_guc_submit.c | 2 +- > > include/drm/gpu_scheduler.h | 4 ++-- > > 17 files changed, 35 insertions(+), 35 deletions(-) > > > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > > index 932dc93b2e63..7f2a68ad8034 100644 > > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c > > @@ -2738,7 +2738,7 @@ static int amdgpu_device_init_schedulers(struct amdgpu_device *adev) > > ring->num_hw_submission, 0, > > timeout, adev->reset_domain->wq, > > ring->sched_score, ring->name, > > - adev->dev); > > + &adev->ddev); > > if (r) { > > DRM_ERROR("Failed to create scheduler on ring %s.\n", > > ring->name); > > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_sched.c b/drivers/gpu/drm/etnaviv/etnaviv_sched.c > > index c4b04b0dee16..c4345b68a51f 100644 > > --- a/drivers/gpu/drm/etnaviv/etnaviv_sched.c > > +++ b/drivers/gpu/drm/etnaviv/etnaviv_sched.c > > @@ -138,7 +138,7 @@ int etnaviv_sched_init(struct etnaviv_gpu *gpu) > > DRM_SCHED_PRIORITY_COUNT, > > etnaviv_hw_jobs_limit, etnaviv_job_hang_limit, > > msecs_to_jiffies(500), NULL, NULL, > > - dev_name(gpu->dev), gpu->dev); > > + dev_name(gpu->dev), gpu->drm); > > if (ret) > > return ret; > > diff --git a/drivers/gpu/drm/imagination/pvr_queue.c b/drivers/gpu/drm/imagination/pvr_queue.c > > index 5ed9c98fb599..cdbb6c01e952 100644 > > --- a/drivers/gpu/drm/imagination/pvr_queue.c > > +++ b/drivers/gpu/drm/imagination/pvr_queue.c > > @@ -1287,7 +1287,7 @@ struct pvr_queue *pvr_queue_create(struct pvr_context *ctx, > > pvr_dev->sched_wq, 1, 64 * 1024, 1, > > msecs_to_jiffies(500), > > pvr_dev->sched_wq, NULL, "pvr-queue", > > - pvr_dev->base.dev); > > + &pvr_dev->base); > > if (err) > > goto err_release_ufo; > > diff --git a/drivers/gpu/drm/lima/lima_sched.c b/drivers/gpu/drm/lima/lima_sched.c > > index bbf3f8feab94..db6ee7650468 100644 > > --- a/drivers/gpu/drm/lima/lima_sched.c > > +++ b/drivers/gpu/drm/lima/lima_sched.c > > @@ -526,7 +526,7 @@ int lima_sched_pipe_init(struct lima_sched_pipe *pipe, const char *name) > > 1, > > lima_job_hang_limit, > > msecs_to_jiffies(timeout), NULL, > > - NULL, name, pipe->ldev->dev); > > + NULL, name, pipe->ldev->ddev); > > } > > void lima_sched_pipe_fini(struct lima_sched_pipe *pipe) > > diff --git a/drivers/gpu/drm/msm/msm_ringbuffer.c b/drivers/gpu/drm/msm/msm_ringbuffer.c > > index 9d6655f96f0c..3a4b3816f2c9 100644 > > --- a/drivers/gpu/drm/msm/msm_ringbuffer.c > > +++ b/drivers/gpu/drm/msm/msm_ringbuffer.c > > @@ -101,7 +101,7 @@ struct msm_ringbuffer *msm_ringbuffer_new(struct msm_gpu *gpu, int id, > > ret = drm_sched_init(&ring->sched, &msm_sched_ops, NULL, > > DRM_SCHED_PRIORITY_COUNT, > > num_hw_submissions, 0, sched_timeout, > > - NULL, NULL, to_msm_bo(ring->bo)->name, gpu->dev->dev); > > + NULL, NULL, to_msm_bo(ring->bo)->name, gpu->dev); > > if (ret) { > > goto fail; > > } > > diff --git a/drivers/gpu/drm/nouveau/nouveau_sched.c b/drivers/gpu/drm/nouveau/nouveau_sched.c > > index 32fa2e273965..386839bed8a2 100644 > > --- a/drivers/gpu/drm/nouveau/nouveau_sched.c > > +++ b/drivers/gpu/drm/nouveau/nouveau_sched.c > > @@ -419,7 +419,7 @@ nouveau_sched_init(struct nouveau_sched *sched, struct nouveau_drm *drm, > > ret = drm_sched_init(drm_sched, &nouveau_sched_ops, wq, > > NOUVEAU_SCHED_PRIORITY_COUNT, > > credit_limit, 0, job_hang_limit, > > - NULL, NULL, "nouveau_sched", drm->dev->dev); > > + NULL, NULL, "nouveau_sched", drm->dev); > > if (ret) > > goto fail_wq; > > diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c b/drivers/gpu/drm/panfrost/panfrost_job.c > > index a61ef0af9a4e..28c7680a8dbf 100644 > > --- a/drivers/gpu/drm/panfrost/panfrost_job.c > > +++ b/drivers/gpu/drm/panfrost/panfrost_job.c > > @@ -875,7 +875,7 @@ int panfrost_job_init(struct panfrost_device *pfdev) > > nentries, 0, > > msecs_to_jiffies(JOB_TIMEOUT_MS), > > pfdev->reset.wq, > > - NULL, "pan_js", pfdev->dev); > > + NULL, "pan_js", pfdev->ddev); > > if (ret) { > > dev_err(pfdev->dev, "Failed to create scheduler: %d.", ret); > > goto err_sched; > > diff --git a/drivers/gpu/drm/panthor/panthor_mmu.c b/drivers/gpu/drm/panthor/panthor_mmu.c > > index fa0a002b1016..b9c5b500b7d1 100644 > > --- a/drivers/gpu/drm/panthor/panthor_mmu.c > > +++ b/drivers/gpu/drm/panthor/panthor_mmu.c > > @@ -2293,7 +2293,7 @@ panthor_vm_create(struct panthor_device *ptdev, bool for_mcu, > > ret = drm_sched_init(&vm->sched, &panthor_vm_bind_ops, ptdev->mmu->vm.wq, > > 1, 1, 0, > > MAX_SCHEDULE_TIMEOUT, NULL, NULL, > > - "panthor-vm-bind", ptdev->base.dev); > > + "panthor-vm-bind", &ptdev->base); > > if (ret) > > goto err_free_io_pgtable; > > diff --git a/drivers/gpu/drm/panthor/panthor_sched.c b/drivers/gpu/drm/panthor/panthor_sched.c > > index 79ffcbc41d78..47e52f61571b 100644 > > --- a/drivers/gpu/drm/panthor/panthor_sched.c > > +++ b/drivers/gpu/drm/panthor/panthor_sched.c > > @@ -3043,7 +3043,7 @@ group_create_queue(struct panthor_group *group, > > args->ringbuf_size / (NUM_INSTRS_PER_SLOT * sizeof(u64)), > > 0, msecs_to_jiffies(JOB_TIMEOUT_MS), > > group->ptdev->reset.wq, > > - NULL, "panthor-queue", group->ptdev->base.dev); > > + NULL, "panthor-queue", &group->ptdev->base); > > if (ret) > > goto err_free_queue; > > diff --git a/drivers/gpu/drm/scheduler/sched_entity.c b/drivers/gpu/drm/scheduler/sched_entity.c > > index 58c8161289fe..194798b9ce09 100644 > > --- a/drivers/gpu/drm/scheduler/sched_entity.c > > +++ b/drivers/gpu/drm/scheduler/sched_entity.c > > @@ -92,7 +92,7 @@ int drm_sched_entity_init(struct drm_sched_entity *entity, > > * the lowest priority available. > > */ > > if (entity->priority >= sched_list[0]->num_rqs) { > > - drm_err(sched_list[0], "entity with out-of-bounds priority:%u num_rqs:%u\n", > > + drm_err(sched_list[0]->dev, "entity with out-of-bounds priority:%u num_rqs:%u\n", > > entity->priority, sched_list[0]->num_rqs); > > entity->priority = max_t(s32, (s32) sched_list[0]->num_rqs - 1, > > (s32) DRM_SCHED_PRIORITY_KERNEL); > > diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c > > index 7e90c9f95611..74a2fe51e653 100644 > > --- a/drivers/gpu/drm/scheduler/sched_main.c > > +++ b/drivers/gpu/drm/scheduler/sched_main.c > > @@ -103,9 +103,9 @@ static u32 drm_sched_available_credits(struct drm_gpu_scheduler *sched) > > { > > u32 credits; > > - drm_WARN_ON(sched, check_sub_overflow(sched->credit_limit, > > - atomic_read(&sched->credit_count), > > - &credits)); > > + drm_WARN_ON(sched->dev, check_sub_overflow(sched->credit_limit, > > + atomic_read(&sched->credit_count), > > + &credits)); > > return credits; > > } > > @@ -130,14 +130,14 @@ static bool drm_sched_can_queue(struct drm_gpu_scheduler *sched, > > if (sched->ops->update_job_credits) { > > s_job->credits = sched->ops->update_job_credits(s_job); > > - drm_WARN(sched, !s_job->credits, > > + drm_WARN(sched->dev, !s_job->credits, > > "Jobs with zero credits bypass job-flow control.\n"); > > } > > /* If a job exceeds the credit limit, truncate it to the credit limit > > * itself to guarantee forward progress. > > */ > > - if (drm_WARN(sched, s_job->credits > sched->credit_limit, > > + if (drm_WARN(sched->dev, s_job->credits > sched->credit_limit, > > "Jobs may not exceed the credit limit, truncate.\n")) > > s_job->credits = sched->credit_limit; > > @@ -701,7 +701,7 @@ void drm_sched_start(struct drm_gpu_scheduler *sched, bool full_recovery) > > if (r == -ENOENT) > > drm_sched_job_done(s_job, fence->error); > > else if (r) > > - DRM_DEV_ERROR(sched->dev, "fence add callback failed (%d)\n", > > + DRM_DEV_ERROR(sched->dev->dev, "fence add callback failed (%d)\n", > > r); > > } else > > drm_sched_job_done(s_job, -ECANCELED); > > @@ -797,7 +797,7 @@ int drm_sched_job_init(struct drm_sched_job *job, > > * or worse--a blank screen--leave a trail in the > > * logs, so this can be debugged easier. > > */ > > - drm_err(job->sched, "%s: entity has no rq!\n", __func__); > > + drm_err(job->sched->dev, "%s: entity has no rq!\n", __func__); > > return -ENOENT; > > } > > @@ -1215,7 +1215,7 @@ static void drm_sched_run_job_work(struct work_struct *w) > > if (r == -ENOENT) > > drm_sched_job_done(sched_job, fence->error); > > else if (r) > > - DRM_DEV_ERROR(sched->dev, "fence add callback failed (%d)\n", r); > > + DRM_DEV_ERROR(sched->dev->dev, "fence add callback failed (%d)\n", r); > > } else { > > drm_sched_job_done(sched_job, IS_ERR(fence) ? > > PTR_ERR(fence) : 0); > > @@ -1240,7 +1240,7 @@ static void drm_sched_run_job_work(struct work_struct *w) > > * used > > * @score: optional score atomic shared with other schedulers > > * @name: name used for debugging > > - * @dev: target &struct device > > + * @dev: target &struct drm_device > > * > > * Return 0 on success, otherwise error code. > > */ > > @@ -1249,7 +1249,7 @@ int drm_sched_init(struct drm_gpu_scheduler *sched, > > struct workqueue_struct *submit_wq, > > u32 num_rqs, u32 credit_limit, unsigned int hang_limit, > > long timeout, struct workqueue_struct *timeout_wq, > > - atomic_t *score, const char *name, struct device *dev) > > + atomic_t *score, const char *name, struct drm_device *dev) > > { > > int i; > > @@ -1265,7 +1265,7 @@ int drm_sched_init(struct drm_gpu_scheduler *sched, > > if (num_rqs > DRM_SCHED_PRIORITY_COUNT) { > > /* This is a gross violation--tell drivers what the problem is. > > */ > > - drm_err(sched, "%s: num_rqs cannot be greater than DRM_SCHED_PRIORITY_COUNT\n", > > + drm_err(dev, "%s: num_rqs cannot be greater than DRM_SCHED_PRIORITY_COUNT\n", > > __func__); > > return -EINVAL; > > } else if (sched->sched_rq) { > > @@ -1273,7 +1273,7 @@ int drm_sched_init(struct drm_gpu_scheduler *sched, > > * fine-tune their DRM calling order, and return all > > * is good. > > */ > > - drm_warn(sched, "%s: scheduler already initialized!\n", __func__); > > + drm_warn(dev, "%s: scheduler already initialized!\n", __func__); > > return 0; > > } > > @@ -1322,7 +1322,7 @@ int drm_sched_init(struct drm_gpu_scheduler *sched, > > Out_check_own: > > if (sched->own_submit_wq) > > destroy_workqueue(sched->submit_wq); > > - drm_err(sched, "%s: Failed to setup GPU scheduler--out of memory\n", __func__); > > + drm_err(dev, "%s: Failed to setup GPU scheduler--out of memory\n", __func__); > > return -ENOMEM; > > } > > EXPORT_SYMBOL(drm_sched_init); > > diff --git a/drivers/gpu/drm/v3d/v3d_sched.c b/drivers/gpu/drm/v3d/v3d_sched.c > > index 7cd8c335cd9b..73383b6ef9bb 100644 > > --- a/drivers/gpu/drm/v3d/v3d_sched.c > > +++ b/drivers/gpu/drm/v3d/v3d_sched.c > > @@ -740,7 +740,7 @@ v3d_sched_init(struct v3d_dev *v3d) > > DRM_SCHED_PRIORITY_COUNT, > > hw_jobs_limit, job_hang_limit, > > msecs_to_jiffies(hang_limit_ms), NULL, > > - NULL, "v3d_bin", v3d->drm.dev); > > + NULL, "v3d_bin", &v3d->drm); > > if (ret) > > return ret; > > @@ -749,7 +749,7 @@ v3d_sched_init(struct v3d_dev *v3d) > > DRM_SCHED_PRIORITY_COUNT, > > hw_jobs_limit, job_hang_limit, > > msecs_to_jiffies(hang_limit_ms), NULL, > > - NULL, "v3d_render", v3d->drm.dev); > > + NULL, "v3d_render", &v3d->drm); > > if (ret) > > goto fail; > > @@ -758,7 +758,7 @@ v3d_sched_init(struct v3d_dev *v3d) > > DRM_SCHED_PRIORITY_COUNT, > > hw_jobs_limit, job_hang_limit, > > msecs_to_jiffies(hang_limit_ms), NULL, > > - NULL, "v3d_tfu", v3d->drm.dev); > > + NULL, "v3d_tfu", &v3d->drm); > > if (ret) > > goto fail; > > @@ -768,7 +768,7 @@ v3d_sched_init(struct v3d_dev *v3d) > > DRM_SCHED_PRIORITY_COUNT, > > hw_jobs_limit, job_hang_limit, > > msecs_to_jiffies(hang_limit_ms), NULL, > > - NULL, "v3d_csd", v3d->drm.dev); > > + NULL, "v3d_csd", &v3d->drm); > > if (ret) > > goto fail; > > @@ -777,7 +777,7 @@ v3d_sched_init(struct v3d_dev *v3d) > > DRM_SCHED_PRIORITY_COUNT, > > hw_jobs_limit, job_hang_limit, > > msecs_to_jiffies(hang_limit_ms), NULL, > > - NULL, "v3d_cache_clean", v3d->drm.dev); > > + NULL, "v3d_cache_clean", &v3d->drm); > > if (ret) > > goto fail; > > } > > @@ -787,7 +787,7 @@ v3d_sched_init(struct v3d_dev *v3d) > > DRM_SCHED_PRIORITY_COUNT, > > 1, job_hang_limit, > > msecs_to_jiffies(hang_limit_ms), NULL, > > - NULL, "v3d_cpu", v3d->drm.dev); > > + NULL, "v3d_cpu", &v3d->drm); > > if (ret) > > goto fail; > > diff --git a/drivers/gpu/drm/xe/xe_execlist.c b/drivers/gpu/drm/xe/xe_execlist.c > > index dece2785933c..dc81e9f39727 100644 > > --- a/drivers/gpu/drm/xe/xe_execlist.c > > +++ b/drivers/gpu/drm/xe/xe_execlist.c > > @@ -336,7 +336,7 @@ static int execlist_exec_queue_init(struct xe_exec_queue *q) > > q->lrc[0].ring.size / MAX_JOB_SIZE_BYTES, > > XE_SCHED_HANG_LIMIT, XE_SCHED_JOB_TIMEOUT, > > NULL, NULL, q->hwe->name, > > - gt_to_xe(q->gt)->drm.dev); > > + >_to_xe(q->gt)->drm); > > if (err) > > goto err_free; > > diff --git a/drivers/gpu/drm/xe/xe_gpu_scheduler.c b/drivers/gpu/drm/xe/xe_gpu_scheduler.c > > index e4ad1d6ce1d5..66d36cac82a0 100644 > > --- a/drivers/gpu/drm/xe/xe_gpu_scheduler.c > > +++ b/drivers/gpu/drm/xe/xe_gpu_scheduler.c > > @@ -61,7 +61,7 @@ int xe_sched_init(struct xe_gpu_scheduler *sched, > > uint32_t hw_submission, unsigned hang_limit, > > long timeout, struct workqueue_struct *timeout_wq, > > atomic_t *score, const char *name, > > - struct device *dev) > > + struct drm_device *dev) > > { > > sched->ops = xe_ops; > > INIT_LIST_HEAD(&sched->msgs); > > diff --git a/drivers/gpu/drm/xe/xe_gpu_scheduler.h b/drivers/gpu/drm/xe/xe_gpu_scheduler.h > > index 10c6bb9c9386..9a75457813f2 100644 > > --- a/drivers/gpu/drm/xe/xe_gpu_scheduler.h > > +++ b/drivers/gpu/drm/xe/xe_gpu_scheduler.h > > @@ -16,7 +16,7 @@ int xe_sched_init(struct xe_gpu_scheduler *sched, > > uint32_t hw_submission, unsigned hang_limit, > > long timeout, struct workqueue_struct *timeout_wq, > > atomic_t *score, const char *name, > > - struct device *dev); > > + struct drm_device *dev); > > void xe_sched_fini(struct xe_gpu_scheduler *sched); > > void xe_sched_submission_start(struct xe_gpu_scheduler *sched); > > diff --git a/drivers/gpu/drm/xe/xe_guc_submit.c b/drivers/gpu/drm/xe/xe_guc_submit.c > > index e4e3658e6a13..b9c114f2c715 100644 > > --- a/drivers/gpu/drm/xe/xe_guc_submit.c > > +++ b/drivers/gpu/drm/xe/xe_guc_submit.c > > @@ -1208,7 +1208,7 @@ static int guc_exec_queue_init(struct xe_exec_queue *q) > > get_submit_wq(guc), > > q->lrc[0].ring.size / MAX_JOB_SIZE_BYTES, 64, > > timeout, guc_to_gt(guc)->ordered_wq, NULL, > > - q->name, gt_to_xe(q->gt)->drm.dev); > > + q->name, >_to_xe(q->gt)->drm); > > if (err) > > goto err_free; > > diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h > > index 5acc64954a88..0ba8716ec069 100644 > > --- a/include/drm/gpu_scheduler.h > > +++ b/include/drm/gpu_scheduler.h > > @@ -539,7 +539,7 @@ struct drm_gpu_scheduler { > > bool free_guilty; > > bool pause_submit; > > bool own_submit_wq; > > - struct device *dev; > > + struct drm_device *dev; > > }; > > int drm_sched_init(struct drm_gpu_scheduler *sched, > > @@ -547,7 +547,7 @@ int drm_sched_init(struct drm_gpu_scheduler *sched, > > struct workqueue_struct *submit_wq, > > u32 num_rqs, u32 credit_limit, unsigned int hang_limit, > > long timeout, struct workqueue_struct *timeout_wq, > > - atomic_t *score, const char *name, struct device *dev); > > + atomic_t *score, const char *name, struct drm_device *dev); > > void drm_sched_fini(struct drm_gpu_scheduler *sched); > > int drm_sched_job_init(struct drm_sched_job *job, >