Re: [PATCH v4 3/5] drm/scheduler: rework job destruction

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Am 17.04.19 um 19:53 schrieb Grodzovsky, Andrey:
On 4/17/19 1:17 PM, Christian König wrote:
I can't review this patch, since I'm one of the authors of it, but in
general your changes look good to me now.

For patch #5 I think it might be cleaner if we move incrementing of
the hw_rq_count while starting the scheduler again.

But the increment of  hw_rq_count is conditional on if the guilty job
was signaled, moving it into drm_sched_start will also force me to pass
'job_signaled' flag into drm_sched_start which is against your original
comment that we don't want to pass this logic around helper functions
and keep it all in one place which is amdgpu_device_gpu_recover.

Well I hope that incrementing hw_rq_count is conditional for signaled jobs anyway, or otherwise we would seriously mess up the counter.

E.g. in drm_sched_stop() we also only decrement it when we where able to remove the callback.

Christian.


Andrey


Regards,
Christian.

Am 17.04.19 um 16:36 schrieb Grodzovsky, Andrey:
Ping on this patch and patch 5. The rest already RBed.

Andrey

On 4/16/19 2:23 PM, Andrey Grodzovsky wrote:
From: Christian König <christian.koenig@xxxxxxx>

We now destroy finished jobs from the worker thread to make sure that
we never destroy a job currently in timeout processing.
By this we avoid holding lock around ring mirror list in drm_sched_stop
which should solve a deadlock reported by a user.

v2: Remove unused variable.
v4: Move guilty job free into sched code.

Bugzilla: https://bugs.freedesktop.org/show_bug.cgi?id=109692

Signed-off-by: Christian König <christian.koenig@xxxxxxx>
Signed-off-by: Andrey Grodzovsky <andrey.grodzovsky@xxxxxxx>
---
    drivers/gpu/drm/amd/amdgpu/amdgpu_device.c |   9 +-
    drivers/gpu/drm/etnaviv/etnaviv_dump.c     |   4 -
    drivers/gpu/drm/etnaviv/etnaviv_sched.c    |   2 +-
    drivers/gpu/drm/lima/lima_sched.c          |   2 +-
    drivers/gpu/drm/panfrost/panfrost_job.c    |   2 +-
    drivers/gpu/drm/scheduler/sched_main.c     | 145
+++++++++++++++++------------
    drivers/gpu/drm/v3d/v3d_sched.c            |   2 +-
    include/drm/gpu_scheduler.h                |   6 +-
    8 files changed, 94 insertions(+), 78 deletions(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
index 7cee269..a0e165c 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
@@ -3334,7 +3334,7 @@ static int amdgpu_device_pre_asic_reset(struct
amdgpu_device *adev,
            if (!ring || !ring->sched.thread)
                continue;
    -        drm_sched_stop(&ring->sched);
+        drm_sched_stop(&ring->sched, &job->base);
               /* after all hw jobs are reset, hw fence is
meaningless, so force_completion */
            amdgpu_fence_driver_force_completion(ring);
@@ -3343,8 +3343,6 @@ static int amdgpu_device_pre_asic_reset(struct
amdgpu_device *adev,
        if(job)
            drm_sched_increase_karma(&job->base);
    -
-
        if (!amdgpu_sriov_vf(adev)) {
               if (!need_full_reset)
@@ -3482,8 +3480,7 @@ static int amdgpu_do_asic_reset(struct
amdgpu_hive_info *hive,
        return r;
    }
    -static void amdgpu_device_post_asic_reset(struct amdgpu_device
*adev,
-                      struct amdgpu_job *job)
+static void amdgpu_device_post_asic_reset(struct amdgpu_device *adev)
    {
        int i;
    @@ -3623,7 +3620,7 @@ int amdgpu_device_gpu_recover(struct
amdgpu_device *adev,
           /* Post ASIC reset for all devs .*/
        list_for_each_entry(tmp_adev, device_list_handle,
gmc.xgmi.head) {
-        amdgpu_device_post_asic_reset(tmp_adev, tmp_adev == adev ?
job : NULL);
+        amdgpu_device_post_asic_reset(tmp_adev);
               if (r) {
                /* bad news, how to tell it to userspace ? */
diff --git a/drivers/gpu/drm/etnaviv/etnaviv_dump.c
b/drivers/gpu/drm/etnaviv/etnaviv_dump.c
index 33854c9..5778d9c 100644
--- a/drivers/gpu/drm/etnaviv/etnaviv_dump.c
+++ b/drivers/gpu/drm/etnaviv/etnaviv_dump.c
@@ -135,13 +135,11 @@ void etnaviv_core_dump(struct etnaviv_gpu *gpu)
                mmu_size + gpu->buffer.size;
           /* Add in the active command buffers */
-    spin_lock_irqsave(&gpu->sched.job_list_lock, flags);
        list_for_each_entry(s_job, &gpu->sched.ring_mirror_list, node) {
            submit = to_etnaviv_submit(s_job);
            file_size += submit->cmdbuf.size;
            n_obj++;
        }
-    spin_unlock_irqrestore(&gpu->sched.job_list_lock, flags);
           /* Add in the active buffer objects */
        list_for_each_entry(vram, &gpu->mmu->mappings, mmu_node) {
@@ -183,14 +181,12 @@ void etnaviv_core_dump(struct etnaviv_gpu *gpu)
                      gpu->buffer.size,
etnaviv_cmdbuf_get_va(&gpu->buffer));
    -    spin_lock_irqsave(&gpu->sched.job_list_lock, flags);
        list_for_each_entry(s_job, &gpu->sched.ring_mirror_list, node) {
            submit = to_etnaviv_submit(s_job);
            etnaviv_core_dump_mem(&iter, ETDUMP_BUF_CMD,
                          submit->cmdbuf.vaddr, submit->cmdbuf.size,
etnaviv_cmdbuf_get_va(&submit->cmdbuf));
        }
-    spin_unlock_irqrestore(&gpu->sched.job_list_lock, flags);
           /* Reserve space for the bomap */
        if (n_bomap_pages) {
diff --git a/drivers/gpu/drm/etnaviv/etnaviv_sched.c
b/drivers/gpu/drm/etnaviv/etnaviv_sched.c
index 6d24fea..a813c82 100644
--- a/drivers/gpu/drm/etnaviv/etnaviv_sched.c
+++ b/drivers/gpu/drm/etnaviv/etnaviv_sched.c
@@ -109,7 +109,7 @@ static void etnaviv_sched_timedout_job(struct
drm_sched_job *sched_job)
        }
           /* block scheduler */
-    drm_sched_stop(&gpu->sched);
+    drm_sched_stop(&gpu->sched, sched_job);
           if(sched_job)
            drm_sched_increase_karma(sched_job);
diff --git a/drivers/gpu/drm/lima/lima_sched.c
b/drivers/gpu/drm/lima/lima_sched.c
index 97bd9c1..df98931 100644
--- a/drivers/gpu/drm/lima/lima_sched.c
+++ b/drivers/gpu/drm/lima/lima_sched.c
@@ -300,7 +300,7 @@ static struct dma_fence
*lima_sched_run_job(struct drm_sched_job *job)
    static void lima_sched_handle_error_task(struct lima_sched_pipe
*pipe,
                         struct lima_sched_task *task)
    {
-    drm_sched_stop(&pipe->base);
+    drm_sched_stop(&pipe->base, &task->base);
           if (task)
            drm_sched_increase_karma(&task->base);
diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c
b/drivers/gpu/drm/panfrost/panfrost_job.c
index 0a7ed04..c6336b7 100644
--- a/drivers/gpu/drm/panfrost/panfrost_job.c
+++ b/drivers/gpu/drm/panfrost/panfrost_job.c
@@ -385,7 +385,7 @@ static void panfrost_job_timedout(struct
drm_sched_job *sched_job)
            sched_job);
           for (i = 0; i < NUM_JOB_SLOTS; i++)
-        drm_sched_stop(&pfdev->js->queue[i].sched);
+        drm_sched_stop(&pfdev->js->queue[i].sched, sched_job);
           if (sched_job)
            drm_sched_increase_karma(sched_job);
diff --git a/drivers/gpu/drm/scheduler/sched_main.c
b/drivers/gpu/drm/scheduler/sched_main.c
index 19fc601..21e8734 100644
--- a/drivers/gpu/drm/scheduler/sched_main.c
+++ b/drivers/gpu/drm/scheduler/sched_main.c
@@ -265,32 +265,6 @@ void drm_sched_resume_timeout(struct
drm_gpu_scheduler *sched,
    }
    EXPORT_SYMBOL(drm_sched_resume_timeout);
    -/* job_finish is called after hw fence signaled
- */
-static void drm_sched_job_finish(struct work_struct *work)
-{
-    struct drm_sched_job *s_job = container_of(work, struct
drm_sched_job,
-                           finish_work);
-    struct drm_gpu_scheduler *sched = s_job->sched;
-    unsigned long flags;
-
-    /*
-     * Canceling the timeout without removing our job from the ring
mirror
-     * list is safe, as we will only end up in this worker if our jobs
-     * finished fence has been signaled. So even if some another
worker
-     * manages to find this job as the next job in the list, the fence
-     * signaled check below will prevent the timeout to be restarted.
-     */
-    cancel_delayed_work_sync(&sched->work_tdr);
-
-    spin_lock_irqsave(&sched->job_list_lock, flags);
-    /* queue TDR for next job */
-    drm_sched_start_timeout(sched);
-    spin_unlock_irqrestore(&sched->job_list_lock, flags);
-
-    sched->ops->free_job(s_job);
-}
-
    static void drm_sched_job_begin(struct drm_sched_job *s_job)
    {
        struct drm_gpu_scheduler *sched = s_job->sched;
@@ -315,6 +289,13 @@ static void drm_sched_job_timedout(struct
work_struct *work)
        if (job)
            job->sched->ops->timedout_job(job);
    +    /*
+     * Guilty job did complete and hence needs to be manually removed
+     * See drm_sched_stop doc.
+     */
+    if (list_empty(&job->node))
+        job->sched->ops->free_job(job);
+
        spin_lock_irqsave(&sched->job_list_lock, flags);
        drm_sched_start_timeout(sched);
        spin_unlock_irqrestore(&sched->job_list_lock, flags);
@@ -371,23 +352,26 @@ EXPORT_SYMBOL(drm_sched_increase_karma);
     * @sched: scheduler instance
     * @bad: bad scheduler job
     *
+ * Stop the scheduler and also removes and frees all completed jobs.
+ * Note: bad job will not be freed as it might be used later and so
it's
+ * callers responsibility to release it manually if it's not part
of the
+ * mirror list any more.
+ *
     */
-void drm_sched_stop(struct drm_gpu_scheduler *sched)
+void drm_sched_stop(struct drm_gpu_scheduler *sched, struct
drm_sched_job *bad)
    {
-    struct drm_sched_job *s_job;
+    struct drm_sched_job *s_job, *tmp;
        unsigned long flags;
-    struct dma_fence *last_fence =  NULL;
           kthread_park(sched->thread);
           /*
-     * Verify all the signaled jobs in mirror list are removed from
the ring
-     * by waiting for the latest job to enter the list. This should
insure that
-     * also all the previous jobs that were in flight also already
singaled
-     * and removed from the list.
+     * Iterate the job list from later to  earlier one and either
deactive
+     * their HW callbacks or remove them from mirror list if they
already
+     * signaled.
+     * This iteration is thread safe as sched thread is stopped.
         */
-    spin_lock_irqsave(&sched->job_list_lock, flags);
-    list_for_each_entry_reverse(s_job, &sched->ring_mirror_list,
node) {
+    list_for_each_entry_safe_reverse(s_job, tmp,
&sched->ring_mirror_list, node) {
            if (s_job->s_fence->parent &&
dma_fence_remove_callback(s_job->s_fence->parent,
                              &s_job->cb)) {
@@ -395,16 +379,30 @@ void drm_sched_stop(struct drm_gpu_scheduler
*sched)
                s_job->s_fence->parent = NULL;
                atomic_dec(&sched->hw_rq_count);
            } else {
-             last_fence = dma_fence_get(&s_job->s_fence->finished);
-             break;
+            /*
+             * remove job from ring_mirror_list.
+             * Locking here is for concurrent resume timeout
+             */
+            spin_lock_irqsave(&sched->job_list_lock, flags);
+            list_del_init(&s_job->node);
+ spin_unlock_irqrestore(&sched->job_list_lock, flags);
+
+            /*
+             * Wait for job's HW fence callback to finish using s_job
+             * before releasing it.
+             *
+             * Job is still alive so fence refcount at least 1
+             */
+ dma_fence_wait(&s_job->s_fence->finished, false);
+
+            /*
+             * We must keep bad job alive for later use during
+             * recovery by some of the drivers
+             */
+            if (bad != s_job)
+                sched->ops->free_job(s_job);
            }
        }
-    spin_unlock_irqrestore(&sched->job_list_lock, flags);
-
-    if (last_fence) {
-        dma_fence_wait(last_fence, false);
-        dma_fence_put(last_fence);
-    }
    }
       EXPORT_SYMBOL(drm_sched_stop);
@@ -418,6 +416,7 @@ EXPORT_SYMBOL(drm_sched_stop);
    void drm_sched_start(struct drm_gpu_scheduler *sched, bool
full_recovery)
    {
        struct drm_sched_job *s_job, *tmp;
+    unsigned long flags;
        int r;
           if (!full_recovery)
@@ -425,9 +424,7 @@ void drm_sched_start(struct drm_gpu_scheduler
*sched, bool full_recovery)
           /*
         * Locking the list is not required here as the sched thread
is parked
-     * so no new jobs are being pushed in to HW and in
drm_sched_stop we
-     * flushed all the jobs who were still in mirror list but who
already
-     * signaled and removed them self from the list. Also concurrent
+     * so no new jobs are being inserted or removed. Also concurrent
         * GPU recovers can't run in parallel.
         */
        list_for_each_entry_safe(s_job, tmp,
&sched->ring_mirror_list, node) {
@@ -445,7 +442,9 @@ void drm_sched_start(struct drm_gpu_scheduler
*sched, bool full_recovery)
                drm_sched_process_job(NULL, &s_job->cb);
        }
    +    spin_lock_irqsave(&sched->job_list_lock, flags);
        drm_sched_start_timeout(sched);
+    spin_unlock_irqrestore(&sched->job_list_lock, flags);
       unpark:
        kthread_unpark(sched->thread);
@@ -464,7 +463,6 @@ void drm_sched_resubmit_jobs(struct
drm_gpu_scheduler *sched)
        uint64_t guilty_context;
        bool found_guilty = false;
    -    /*TODO DO we need spinlock here ? */
        list_for_each_entry_safe(s_job, tmp,
&sched->ring_mirror_list, node) {
            struct drm_sched_fence *s_fence = s_job->s_fence;
    @@ -514,7 +512,6 @@ int drm_sched_job_init(struct drm_sched_job
*job,
            return -ENOMEM;
        job->id = atomic64_inc_return(&sched->job_id_count);
    -    INIT_WORK(&job->finish_work, drm_sched_job_finish);
        INIT_LIST_HEAD(&job->node);
           return 0;
@@ -597,24 +594,53 @@ static void drm_sched_process_job(struct
dma_fence *f, struct dma_fence_cb *cb)
        struct drm_sched_job *s_job = container_of(cb, struct
drm_sched_job, cb);
        struct drm_sched_fence *s_fence = s_job->s_fence;
        struct drm_gpu_scheduler *sched = s_fence->sched;
-    unsigned long flags;
-
-    cancel_delayed_work(&sched->work_tdr);
           atomic_dec(&sched->hw_rq_count);
        atomic_dec(&sched->num_jobs);
    -    spin_lock_irqsave(&sched->job_list_lock, flags);
-    /* remove job from ring_mirror_list */
-    list_del_init(&s_job->node);
-    spin_unlock_irqrestore(&sched->job_list_lock, flags);
+    trace_drm_sched_process_job(s_fence);
           drm_sched_fence_finished(s_fence);
-
-    trace_drm_sched_process_job(s_fence);
        wake_up_interruptible(&sched->wake_up_worker);
+}
+
+/**
+ * drm_sched_cleanup_jobs - destroy finished jobs
+ *
+ * @sched: scheduler instance
+ *
+ * Remove all finished jobs from the mirror list and destroy them.
+ */
+static void drm_sched_cleanup_jobs(struct drm_gpu_scheduler *sched)
+{
+    unsigned long flags;
+
+    /* Don't destroy jobs while the timeout worker is running */
+    if (!cancel_delayed_work(&sched->work_tdr))
+        return;
+
+
+    while (!list_empty(&sched->ring_mirror_list)) {
+        struct drm_sched_job *job;
+
+        job = list_first_entry(&sched->ring_mirror_list,
+                       struct drm_sched_job, node);
+        if (!dma_fence_is_signaled(&job->s_fence->finished))
+            break;
+
+        spin_lock_irqsave(&sched->job_list_lock, flags);
+        /* remove job from ring_mirror_list */
+        list_del_init(&job->node);
+        spin_unlock_irqrestore(&sched->job_list_lock, flags);
+
+        sched->ops->free_job(job);
+    }
+
+    /* queue timeout for next job */
+    spin_lock_irqsave(&sched->job_list_lock, flags);
+    drm_sched_start_timeout(sched);
+    spin_unlock_irqrestore(&sched->job_list_lock, flags);
    -    schedule_work(&s_job->finish_work);
    }
       /**
@@ -656,9 +682,10 @@ static int drm_sched_main(void *param)
            struct dma_fence *fence;
wait_event_interruptible(sched->wake_up_worker,
+                     (drm_sched_cleanup_jobs(sched),
                         (!drm_sched_blocked(sched) &&
                          (entity = drm_sched_select_entity(sched))) ||
-                     kthread_should_stop());
+                     kthread_should_stop()));
               if (!entity)
                continue;
diff --git a/drivers/gpu/drm/v3d/v3d_sched.c
b/drivers/gpu/drm/v3d/v3d_sched.c
index e740f3b..1a4abe7 100644
--- a/drivers/gpu/drm/v3d/v3d_sched.c
+++ b/drivers/gpu/drm/v3d/v3d_sched.c
@@ -232,7 +232,7 @@ v3d_gpu_reset_for_timeout(struct v3d_dev *v3d,
struct drm_sched_job *sched_job)
           /* block scheduler */
        for (q = 0; q < V3D_MAX_QUEUES; q++)
-        drm_sched_stop(&v3d->queue[q].sched);
+        drm_sched_stop(&v3d->queue[q].sched, sched_job);
           if (sched_job)
            drm_sched_increase_karma(sched_job);
diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h
index 0daca4d..9ee0f27 100644
--- a/include/drm/gpu_scheduler.h
+++ b/include/drm/gpu_scheduler.h
@@ -167,9 +167,6 @@ struct drm_sched_fence
*to_drm_sched_fence(struct dma_fence *f);
     * @sched: the scheduler instance on which this job is scheduled.
     * @s_fence: contains the fences for the scheduling of job.
     * @finish_cb: the callback for the finished fence.
- * @finish_work: schedules the function @drm_sched_job_finish once
the job has
- *               finished to remove the job from the
- *               @drm_gpu_scheduler.ring_mirror_list.
     * @node: used to append this struct to the
@drm_gpu_scheduler.ring_mirror_list.
     * @id: a unique id assigned to each job scheduled on the scheduler.
     * @karma: increment on every hang caused by this job. If this
exceeds the hang
@@ -188,7 +185,6 @@ struct drm_sched_job {
        struct drm_gpu_scheduler    *sched;
        struct drm_sched_fence        *s_fence;
        struct dma_fence_cb        finish_cb;
-    struct work_struct        finish_work;
        struct list_head        node;
        uint64_t            id;
        atomic_t            karma;
@@ -296,7 +292,7 @@ int drm_sched_job_init(struct drm_sched_job *job,
                   void *owner);
    void drm_sched_job_cleanup(struct drm_sched_job *job);
    void drm_sched_wakeup(struct drm_gpu_scheduler *sched);
-void drm_sched_stop(struct drm_gpu_scheduler *sched);
+void drm_sched_stop(struct drm_gpu_scheduler *sched, struct
drm_sched_job *bad);
    void drm_sched_start(struct drm_gpu_scheduler *sched, bool
full_recovery);
    void drm_sched_resubmit_jobs(struct drm_gpu_scheduler *sched);
    void drm_sched_increase_karma(struct drm_sched_job *bad);
_______________________________________________
amd-gfx mailing list
amd-gfx@xxxxxxxxxxxxxxxxxxxxx
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

_______________________________________________
amd-gfx mailing list
amd-gfx@xxxxxxxxxxxxxxxxxxxxx
https://lists.freedesktop.org/mailman/listinfo/amd-gfx




[Index of Archives]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux