Am 17.04.19 um 19:59 schrieb Christian König: > Am 17.04.19 um 19:53 schrieb Grodzovsky, Andrey: >> On 4/17/19 1:17 PM, Christian König wrote: >>> I can't review this patch, since I'm one of the authors of it, but in >>> general your changes look good to me now. >>> >>> For patch #5 I think it might be cleaner if we move incrementing of >>> the hw_rq_count while starting the scheduler again. >> >> But the increment of hw_rq_count is conditional on if the guilty job >> was signaled, moving it into drm_sched_start will also force me to pass >> 'job_signaled' flag into drm_sched_start which is against your original >> comment that we don't want to pass this logic around helper functions >> and keep it all in one place which is amdgpu_device_gpu_recover. > > Well I hope that incrementing hw_rq_count is conditional for signaled > jobs anyway, or otherwise we would seriously mess up the counter. > > E.g. in drm_sched_stop() we also only decrement it when we where able > to remove the callback. Ok, checking the code again we don't need any special handling here since all signaled jobs are already removed from the mirror_list. Christian. > > Christian. > >> >> Andrey >> >> >>> Regards, >>> Christian. >>> >>> Am 17.04.19 um 16:36 schrieb Grodzovsky, Andrey: >>>> Ping on this patch and patch 5. The rest already RBed. >>>> >>>> Andrey >>>> >>>> On 4/16/19 2:23 PM, Andrey Grodzovsky wrote: >>>>> From: Christian König <christian.koenig@xxxxxxx> >>>>> >>>>> We now destroy finished jobs from the worker thread to make sure that >>>>> we never destroy a job currently in timeout processing. >>>>> By this we avoid holding lock around ring mirror list in >>>>> drm_sched_stop >>>>> which should solve a deadlock reported by a user. >>>>> >>>>> v2: Remove unused variable. >>>>> v4: Move guilty job free into sched code. >>>>> >>>>> Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=109692 >>>>> >>>>> Signed-off-by: Christian König <christian.koenig@xxxxxxx> >>>>> Signed-off-by: Andrey Grodzovsky <andrey.grodzovsky@xxxxxxx> >>>>> --- >>>>> drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 9 +- >>>>> drivers/gpu/drm/etnaviv/etnaviv_dump.c | 4 - >>>>> drivers/gpu/drm/etnaviv/etnaviv_sched.c | 2 +- >>>>> drivers/gpu/drm/lima/lima_sched.c | 2 +- >>>>> drivers/gpu/drm/panfrost/panfrost_job.c | 2 +- >>>>> drivers/gpu/drm/scheduler/sched_main.c | 145 >>>>> +++++++++++++++++------------ >>>>> drivers/gpu/drm/v3d/v3d_sched.c | 2 +- >>>>> include/drm/gpu_scheduler.h | 6 +- >>>>> 8 files changed, 94 insertions(+), 78 deletions(-) >>>>> >>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>>> index 7cee269..a0e165c 100644 >>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c >>>>> @@ -3334,7 +3334,7 @@ static int amdgpu_device_pre_asic_reset(struct >>>>> amdgpu_device *adev, >>>>> if (!ring || !ring->sched.thread) >>>>> continue; >>>>> - drm_sched_stop(&ring->sched); >>>>> + drm_sched_stop(&ring->sched, &job->base); >>>>> /* after all hw jobs are reset, hw fence is >>>>> meaningless, so force_completion */ >>>>> amdgpu_fence_driver_force_completion(ring); >>>>> @@ -3343,8 +3343,6 @@ static int amdgpu_device_pre_asic_reset(struct >>>>> amdgpu_device *adev, >>>>> if(job) >>>>> drm_sched_increase_karma(&job->base); >>>>> - >>>>> - >>>>> if (!amdgpu_sriov_vf(adev)) { >>>>> if (!need_full_reset) >>>>> @@ -3482,8 +3480,7 @@ static int amdgpu_do_asic_reset(struct >>>>> amdgpu_hive_info *hive, >>>>> return r; >>>>> } >>>>> -static void amdgpu_device_post_asic_reset(struct amdgpu_device >>>>> *adev, >>>>> - struct amdgpu_job *job) >>>>> +static void amdgpu_device_post_asic_reset(struct amdgpu_device >>>>> *adev) >>>>> { >>>>> int i; >>>>> @@ -3623,7 +3620,7 @@ int amdgpu_device_gpu_recover(struct >>>>> amdgpu_device *adev, >>>>> /* Post ASIC reset for all devs .*/ >>>>> list_for_each_entry(tmp_adev, device_list_handle, >>>>> gmc.xgmi.head) { >>>>> - amdgpu_device_post_asic_reset(tmp_adev, tmp_adev == adev ? >>>>> job : NULL); >>>>> + amdgpu_device_post_asic_reset(tmp_adev); >>>>> if (r) { >>>>> /* bad news, how to tell it to userspace ? */ >>>>> diff --git a/drivers/gpu/drm/etnaviv/etnaviv_dump.c >>>>> b/drivers/gpu/drm/etnaviv/etnaviv_dump.c >>>>> index 33854c9..5778d9c 100644 >>>>> --- a/drivers/gpu/drm/etnaviv/etnaviv_dump.c >>>>> +++ b/drivers/gpu/drm/etnaviv/etnaviv_dump.c >>>>> @@ -135,13 +135,11 @@ void etnaviv_core_dump(struct etnaviv_gpu *gpu) >>>>> mmu_size + gpu->buffer.size; >>>>> /* Add in the active command buffers */ >>>>> - spin_lock_irqsave(&gpu->sched.job_list_lock, flags); >>>>> list_for_each_entry(s_job, &gpu->sched.ring_mirror_list, >>>>> node) { >>>>> submit = to_etnaviv_submit(s_job); >>>>> file_size += submit->cmdbuf.size; >>>>> n_obj++; >>>>> } >>>>> - spin_unlock_irqrestore(&gpu->sched.job_list_lock, flags); >>>>> /* Add in the active buffer objects */ >>>>> list_for_each_entry(vram, &gpu->mmu->mappings, mmu_node) { >>>>> @@ -183,14 +181,12 @@ void etnaviv_core_dump(struct etnaviv_gpu *gpu) >>>>> gpu->buffer.size, >>>>> etnaviv_cmdbuf_get_va(&gpu->buffer)); >>>>> - spin_lock_irqsave(&gpu->sched.job_list_lock, flags); >>>>> list_for_each_entry(s_job, &gpu->sched.ring_mirror_list, >>>>> node) { >>>>> submit = to_etnaviv_submit(s_job); >>>>> etnaviv_core_dump_mem(&iter, ETDUMP_BUF_CMD, >>>>> submit->cmdbuf.vaddr, submit->cmdbuf.size, >>>>> etnaviv_cmdbuf_get_va(&submit->cmdbuf)); >>>>> } >>>>> - spin_unlock_irqrestore(&gpu->sched.job_list_lock, flags); >>>>> /* Reserve space for the bomap */ >>>>> if (n_bomap_pages) { >>>>> diff --git a/drivers/gpu/drm/etnaviv/etnaviv_sched.c >>>>> b/drivers/gpu/drm/etnaviv/etnaviv_sched.c >>>>> index 6d24fea..a813c82 100644 >>>>> --- a/drivers/gpu/drm/etnaviv/etnaviv_sched.c >>>>> +++ b/drivers/gpu/drm/etnaviv/etnaviv_sched.c >>>>> @@ -109,7 +109,7 @@ static void etnaviv_sched_timedout_job(struct >>>>> drm_sched_job *sched_job) >>>>> } >>>>> /* block scheduler */ >>>>> - drm_sched_stop(&gpu->sched); >>>>> + drm_sched_stop(&gpu->sched, sched_job); >>>>> if(sched_job) >>>>> drm_sched_increase_karma(sched_job); >>>>> diff --git a/drivers/gpu/drm/lima/lima_sched.c >>>>> b/drivers/gpu/drm/lima/lima_sched.c >>>>> index 97bd9c1..df98931 100644 >>>>> --- a/drivers/gpu/drm/lima/lima_sched.c >>>>> +++ b/drivers/gpu/drm/lima/lima_sched.c >>>>> @@ -300,7 +300,7 @@ static struct dma_fence >>>>> *lima_sched_run_job(struct drm_sched_job *job) >>>>> static void lima_sched_handle_error_task(struct lima_sched_pipe >>>>> *pipe, >>>>> struct lima_sched_task *task) >>>>> { >>>>> - drm_sched_stop(&pipe->base); >>>>> + drm_sched_stop(&pipe->base, &task->base); >>>>> if (task) >>>>> drm_sched_increase_karma(&task->base); >>>>> diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c >>>>> b/drivers/gpu/drm/panfrost/panfrost_job.c >>>>> index 0a7ed04..c6336b7 100644 >>>>> --- a/drivers/gpu/drm/panfrost/panfrost_job.c >>>>> +++ b/drivers/gpu/drm/panfrost/panfrost_job.c >>>>> @@ -385,7 +385,7 @@ static void panfrost_job_timedout(struct >>>>> drm_sched_job *sched_job) >>>>> sched_job); >>>>> for (i = 0; i < NUM_JOB_SLOTS; i++) >>>>> - drm_sched_stop(&pfdev->js->queue[i].sched); >>>>> + drm_sched_stop(&pfdev->js->queue[i].sched, sched_job); >>>>> if (sched_job) >>>>> drm_sched_increase_karma(sched_job); >>>>> diff --git a/drivers/gpu/drm/scheduler/sched_main.c >>>>> b/drivers/gpu/drm/scheduler/sched_main.c >>>>> index 19fc601..21e8734 100644 >>>>> --- a/drivers/gpu/drm/scheduler/sched_main.c >>>>> +++ b/drivers/gpu/drm/scheduler/sched_main.c >>>>> @@ -265,32 +265,6 @@ void drm_sched_resume_timeout(struct >>>>> drm_gpu_scheduler *sched, >>>>> } >>>>> EXPORT_SYMBOL(drm_sched_resume_timeout); >>>>> -/* job_finish is called after hw fence signaled >>>>> - */ >>>>> -static void drm_sched_job_finish(struct work_struct *work) >>>>> -{ >>>>> - struct drm_sched_job *s_job = container_of(work, struct >>>>> drm_sched_job, >>>>> - finish_work); >>>>> - struct drm_gpu_scheduler *sched = s_job->sched; >>>>> - unsigned long flags; >>>>> - >>>>> - /* >>>>> - * Canceling the timeout without removing our job from the ring >>>>> mirror >>>>> - * list is safe, as we will only end up in this worker if our >>>>> jobs >>>>> - * finished fence has been signaled. So even if some another >>>>> worker >>>>> - * manages to find this job as the next job in the list, the >>>>> fence >>>>> - * signaled check below will prevent the timeout to be >>>>> restarted. >>>>> - */ >>>>> - cancel_delayed_work_sync(&sched->work_tdr); >>>>> - >>>>> - spin_lock_irqsave(&sched->job_list_lock, flags); >>>>> - /* queue TDR for next job */ >>>>> - drm_sched_start_timeout(sched); >>>>> - spin_unlock_irqrestore(&sched->job_list_lock, flags); >>>>> - >>>>> - sched->ops->free_job(s_job); >>>>> -} >>>>> - >>>>> static void drm_sched_job_begin(struct drm_sched_job *s_job) >>>>> { >>>>> struct drm_gpu_scheduler *sched = s_job->sched; >>>>> @@ -315,6 +289,13 @@ static void drm_sched_job_timedout(struct >>>>> work_struct *work) >>>>> if (job) >>>>> job->sched->ops->timedout_job(job); >>>>> + /* >>>>> + * Guilty job did complete and hence needs to be manually >>>>> removed >>>>> + * See drm_sched_stop doc. >>>>> + */ >>>>> + if (list_empty(&job->node)) >>>>> + job->sched->ops->free_job(job); >>>>> + >>>>> spin_lock_irqsave(&sched->job_list_lock, flags); >>>>> drm_sched_start_timeout(sched); >>>>> spin_unlock_irqrestore(&sched->job_list_lock, flags); >>>>> @@ -371,23 +352,26 @@ EXPORT_SYMBOL(drm_sched_increase_karma); >>>>> * @sched: scheduler instance >>>>> * @bad: bad scheduler job >>>>> * >>>>> + * Stop the scheduler and also removes and frees all completed jobs. >>>>> + * Note: bad job will not be freed as it might be used later and so >>>>> it's >>>>> + * callers responsibility to release it manually if it's not part >>>>> of the >>>>> + * mirror list any more. >>>>> + * >>>>> */ >>>>> -void drm_sched_stop(struct drm_gpu_scheduler *sched) >>>>> +void drm_sched_stop(struct drm_gpu_scheduler *sched, struct >>>>> drm_sched_job *bad) >>>>> { >>>>> - struct drm_sched_job *s_job; >>>>> + struct drm_sched_job *s_job, *tmp; >>>>> unsigned long flags; >>>>> - struct dma_fence *last_fence = NULL; >>>>> kthread_park(sched->thread); >>>>> /* >>>>> - * Verify all the signaled jobs in mirror list are removed from >>>>> the ring >>>>> - * by waiting for the latest job to enter the list. This should >>>>> insure that >>>>> - * also all the previous jobs that were in flight also already >>>>> singaled >>>>> - * and removed from the list. >>>>> + * Iterate the job list from later to earlier one and either >>>>> deactive >>>>> + * their HW callbacks or remove them from mirror list if they >>>>> already >>>>> + * signaled. >>>>> + * This iteration is thread safe as sched thread is stopped. >>>>> */ >>>>> - spin_lock_irqsave(&sched->job_list_lock, flags); >>>>> - list_for_each_entry_reverse(s_job, &sched->ring_mirror_list, >>>>> node) { >>>>> + list_for_each_entry_safe_reverse(s_job, tmp, >>>>> &sched->ring_mirror_list, node) { >>>>> if (s_job->s_fence->parent && >>>>> dma_fence_remove_callback(s_job->s_fence->parent, >>>>> &s_job->cb)) { >>>>> @@ -395,16 +379,30 @@ void drm_sched_stop(struct drm_gpu_scheduler >>>>> *sched) >>>>> s_job->s_fence->parent = NULL; >>>>> atomic_dec(&sched->hw_rq_count); >>>>> } else { >>>>> - last_fence = dma_fence_get(&s_job->s_fence->finished); >>>>> - break; >>>>> + /* >>>>> + * remove job from ring_mirror_list. >>>>> + * Locking here is for concurrent resume timeout >>>>> + */ >>>>> + spin_lock_irqsave(&sched->job_list_lock, flags); >>>>> + list_del_init(&s_job->node); >>>>> + spin_unlock_irqrestore(&sched->job_list_lock, flags); >>>>> + >>>>> + /* >>>>> + * Wait for job's HW fence callback to finish using >>>>> s_job >>>>> + * before releasing it. >>>>> + * >>>>> + * Job is still alive so fence refcount at least 1 >>>>> + */ >>>>> + dma_fence_wait(&s_job->s_fence->finished, false); >>>>> + >>>>> + /* >>>>> + * We must keep bad job alive for later use during >>>>> + * recovery by some of the drivers >>>>> + */ >>>>> + if (bad != s_job) >>>>> + sched->ops->free_job(s_job); >>>>> } >>>>> } >>>>> - spin_unlock_irqrestore(&sched->job_list_lock, flags); >>>>> - >>>>> - if (last_fence) { >>>>> - dma_fence_wait(last_fence, false); >>>>> - dma_fence_put(last_fence); >>>>> - } >>>>> } >>>>> EXPORT_SYMBOL(drm_sched_stop); >>>>> @@ -418,6 +416,7 @@ EXPORT_SYMBOL(drm_sched_stop); >>>>> void drm_sched_start(struct drm_gpu_scheduler *sched, bool >>>>> full_recovery) >>>>> { >>>>> struct drm_sched_job *s_job, *tmp; >>>>> + unsigned long flags; >>>>> int r; >>>>> if (!full_recovery) >>>>> @@ -425,9 +424,7 @@ void drm_sched_start(struct drm_gpu_scheduler >>>>> *sched, bool full_recovery) >>>>> /* >>>>> * Locking the list is not required here as the sched thread >>>>> is parked >>>>> - * so no new jobs are being pushed in to HW and in >>>>> drm_sched_stop we >>>>> - * flushed all the jobs who were still in mirror list but who >>>>> already >>>>> - * signaled and removed them self from the list. Also concurrent >>>>> + * so no new jobs are being inserted or removed. Also concurrent >>>>> * GPU recovers can't run in parallel. >>>>> */ >>>>> list_for_each_entry_safe(s_job, tmp, >>>>> &sched->ring_mirror_list, node) { >>>>> @@ -445,7 +442,9 @@ void drm_sched_start(struct drm_gpu_scheduler >>>>> *sched, bool full_recovery) >>>>> drm_sched_process_job(NULL, &s_job->cb); >>>>> } >>>>> + spin_lock_irqsave(&sched->job_list_lock, flags); >>>>> drm_sched_start_timeout(sched); >>>>> + spin_unlock_irqrestore(&sched->job_list_lock, flags); >>>>> unpark: >>>>> kthread_unpark(sched->thread); >>>>> @@ -464,7 +463,6 @@ void drm_sched_resubmit_jobs(struct >>>>> drm_gpu_scheduler *sched) >>>>> uint64_t guilty_context; >>>>> bool found_guilty = false; >>>>> - /*TODO DO we need spinlock here ? */ >>>>> list_for_each_entry_safe(s_job, tmp, >>>>> &sched->ring_mirror_list, node) { >>>>> struct drm_sched_fence *s_fence = s_job->s_fence; >>>>> @@ -514,7 +512,6 @@ int drm_sched_job_init(struct drm_sched_job >>>>> *job, >>>>> return -ENOMEM; >>>>> job->id = atomic64_inc_return(&sched->job_id_count); >>>>> - INIT_WORK(&job->finish_work, drm_sched_job_finish); >>>>> INIT_LIST_HEAD(&job->node); >>>>> return 0; >>>>> @@ -597,24 +594,53 @@ static void drm_sched_process_job(struct >>>>> dma_fence *f, struct dma_fence_cb *cb) >>>>> struct drm_sched_job *s_job = container_of(cb, struct >>>>> drm_sched_job, cb); >>>>> struct drm_sched_fence *s_fence = s_job->s_fence; >>>>> struct drm_gpu_scheduler *sched = s_fence->sched; >>>>> - unsigned long flags; >>>>> - >>>>> - cancel_delayed_work(&sched->work_tdr); >>>>> atomic_dec(&sched->hw_rq_count); >>>>> atomic_dec(&sched->num_jobs); >>>>> - spin_lock_irqsave(&sched->job_list_lock, flags); >>>>> - /* remove job from ring_mirror_list */ >>>>> - list_del_init(&s_job->node); >>>>> - spin_unlock_irqrestore(&sched->job_list_lock, flags); >>>>> + trace_drm_sched_process_job(s_fence); >>>>> drm_sched_fence_finished(s_fence); >>>>> - >>>>> - trace_drm_sched_process_job(s_fence); >>>>> wake_up_interruptible(&sched->wake_up_worker); >>>>> +} >>>>> + >>>>> +/** >>>>> + * drm_sched_cleanup_jobs - destroy finished jobs >>>>> + * >>>>> + * @sched: scheduler instance >>>>> + * >>>>> + * Remove all finished jobs from the mirror list and destroy them. >>>>> + */ >>>>> +static void drm_sched_cleanup_jobs(struct drm_gpu_scheduler *sched) >>>>> +{ >>>>> + unsigned long flags; >>>>> + >>>>> + /* Don't destroy jobs while the timeout worker is running */ >>>>> + if (!cancel_delayed_work(&sched->work_tdr)) >>>>> + return; >>>>> + >>>>> + >>>>> + while (!list_empty(&sched->ring_mirror_list)) { >>>>> + struct drm_sched_job *job; >>>>> + >>>>> + job = list_first_entry(&sched->ring_mirror_list, >>>>> + struct drm_sched_job, node); >>>>> + if (!dma_fence_is_signaled(&job->s_fence->finished)) >>>>> + break; >>>>> + >>>>> + spin_lock_irqsave(&sched->job_list_lock, flags); >>>>> + /* remove job from ring_mirror_list */ >>>>> + list_del_init(&job->node); >>>>> + spin_unlock_irqrestore(&sched->job_list_lock, flags); >>>>> + >>>>> + sched->ops->free_job(job); >>>>> + } >>>>> + >>>>> + /* queue timeout for next job */ >>>>> + spin_lock_irqsave(&sched->job_list_lock, flags); >>>>> + drm_sched_start_timeout(sched); >>>>> + spin_unlock_irqrestore(&sched->job_list_lock, flags); >>>>> - schedule_work(&s_job->finish_work); >>>>> } >>>>> /** >>>>> @@ -656,9 +682,10 @@ static int drm_sched_main(void *param) >>>>> struct dma_fence *fence; >>>>> wait_event_interruptible(sched->wake_up_worker, >>>>> + (drm_sched_cleanup_jobs(sched), >>>>> (!drm_sched_blocked(sched) && >>>>> (entity = >>>>> drm_sched_select_entity(sched))) || >>>>> - kthread_should_stop()); >>>>> + kthread_should_stop())); >>>>> if (!entity) >>>>> continue; >>>>> diff --git a/drivers/gpu/drm/v3d/v3d_sched.c >>>>> b/drivers/gpu/drm/v3d/v3d_sched.c >>>>> index e740f3b..1a4abe7 100644 >>>>> --- a/drivers/gpu/drm/v3d/v3d_sched.c >>>>> +++ b/drivers/gpu/drm/v3d/v3d_sched.c >>>>> @@ -232,7 +232,7 @@ v3d_gpu_reset_for_timeout(struct v3d_dev *v3d, >>>>> struct drm_sched_job *sched_job) >>>>> /* block scheduler */ >>>>> for (q = 0; q < V3D_MAX_QUEUES; q++) >>>>> - drm_sched_stop(&v3d->queue[q].sched); >>>>> + drm_sched_stop(&v3d->queue[q].sched, sched_job); >>>>> if (sched_job) >>>>> drm_sched_increase_karma(sched_job); >>>>> diff --git a/include/drm/gpu_scheduler.h >>>>> b/include/drm/gpu_scheduler.h >>>>> index 0daca4d..9ee0f27 100644 >>>>> --- a/include/drm/gpu_scheduler.h >>>>> +++ b/include/drm/gpu_scheduler.h >>>>> @@ -167,9 +167,6 @@ struct drm_sched_fence >>>>> *to_drm_sched_fence(struct dma_fence *f); >>>>> * @sched: the scheduler instance on which this job is scheduled. >>>>> * @s_fence: contains the fences for the scheduling of job. >>>>> * @finish_cb: the callback for the finished fence. >>>>> - * @finish_work: schedules the function @drm_sched_job_finish once >>>>> the job has >>>>> - * finished to remove the job from the >>>>> - * @drm_gpu_scheduler.ring_mirror_list. >>>>> * @node: used to append this struct to the >>>>> @drm_gpu_scheduler.ring_mirror_list. >>>>> * @id: a unique id assigned to each job scheduled on the >>>>> scheduler. >>>>> * @karma: increment on every hang caused by this job. If this >>>>> exceeds the hang >>>>> @@ -188,7 +185,6 @@ struct drm_sched_job { >>>>> struct drm_gpu_scheduler *sched; >>>>> struct drm_sched_fence *s_fence; >>>>> struct dma_fence_cb finish_cb; >>>>> - struct work_struct finish_work; >>>>> struct list_head node; >>>>> uint64_t id; >>>>> atomic_t karma; >>>>> @@ -296,7 +292,7 @@ int drm_sched_job_init(struct drm_sched_job *job, >>>>> void *owner); >>>>> void drm_sched_job_cleanup(struct drm_sched_job *job); >>>>> void drm_sched_wakeup(struct drm_gpu_scheduler *sched); >>>>> -void drm_sched_stop(struct drm_gpu_scheduler *sched); >>>>> +void drm_sched_stop(struct drm_gpu_scheduler *sched, struct >>>>> drm_sched_job *bad); >>>>> void drm_sched_start(struct drm_gpu_scheduler *sched, bool >>>>> full_recovery); >>>>> void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched); >>>>> void drm_sched_increase_karma(struct drm_sched_job *bad); >> _______________________________________________ >> amd-gfx mailing list >> amd-gfx@xxxxxxxxxxxxxxxxxxxxx >> https://lists.freedesktop.org/mailman/listinfo/amd-gfx > _______________________________________________ amd-gfx mailing list amd-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/amd-gfx