On Wed, Oct 24, 2018 at 07:10:23PM +0530, Sharat Masetty wrote: > Hi, > > Can you please review this and share your thoughts on this approach. We > primarily need a clean way to release the finished fence in case the job > was not pushed to the queue. So, this patch adds the new API to clean up > the resources allocated in sched_job_init() > > Additionally I also move the fence_put(finished_fence) from the > scheduler to the drivers handler of free_job(). The drivers get to use > this new API. This is done so that the layer creating the sched object is > the one freeing up the resources as well. > > Signed-off-by: Sharat Masetty <smasetty@xxxxxxxxxxxxxx> > --- > drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c | 2 +- > drivers/gpu/drm/amd/amdgpu/amdgpu_job.c | 2 ++ > drivers/gpu/drm/etnaviv/etnaviv_sched.c | 2 ++ > drivers/gpu/drm/msm/msm_sched.c | 8 +++----- Just a nit - msm_sched.c isn't a thing yet. Can you re-generate a patch against drm-next so that if folks like this they can get it merged sooner? Jordan > drivers/gpu/drm/scheduler/gpu_scheduler.c | 11 +++++++++-- > drivers/gpu/drm/v3d/v3d_sched.c | 2 ++ > include/drm/gpu_scheduler.h | 1 + > 7 files changed, 20 insertions(+), 8 deletions(-) > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c > index 9c85a90..9a83152 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c > @@ -1198,7 +1198,7 @@ static int amdgpu_cs_submit(struct amdgpu_cs_parser *p, > r = amdgpu_ctx_add_fence(p->ctx, ring, p->fence, &seq); > if (r) { > dma_fence_put(p->fence); > - dma_fence_put(&job->base.s_fence->finished); > + drm_sched_job_cleanup(&job->base); > amdgpu_job_free(job); > amdgpu_mn_unlock(p->mn); > return r; > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c > index 2bd5676..5d252d4 100644 > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c > @@ -100,6 +100,8 @@ static void amdgpu_job_free_cb(struct drm_sched_job *s_job) > { > struct amdgpu_job *job = container_of(s_job, struct amdgpu_job, base); > > + drm_sched_job_cleanup(s_job); > + > amdgpu_ring_priority_put(job->ring, s_job->s_priority); > dma_fence_put(job->fence); > amdgpu_sync_free(&job->sync); > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_sched.c b/drivers/gpu/drm/etnaviv/etnaviv_sched.c > index 50d6b88..30398c5 100644 > --- a/drivers/gpu/drm/etnaviv/etnaviv_sched.c > +++ b/drivers/gpu/drm/etnaviv/etnaviv_sched.c > @@ -127,6 +127,8 @@ static void etnaviv_sched_free_job(struct drm_sched_job *sched_job) > { > struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); > > + drm_sched_job_cleanup(sched_job); > + > etnaviv_submit_put(submit); > } > > diff --git a/drivers/gpu/drm/msm/msm_sched.c b/drivers/gpu/drm/msm/msm_sched.c > index 70b7713..748acc6 100644 > --- a/drivers/gpu/drm/msm/msm_sched.c > +++ b/drivers/gpu/drm/msm/msm_sched.c > @@ -235,6 +235,8 @@ static void msm_sched_free_job(struct drm_sched_job *sched_job) > struct msm_gpu *gpu = submit->gpu; > int i; > > + drm_sched_job_cleanup(sched_job); > + > mutex_lock(&gpu->dev->struct_mutex); > > for (i = 0; i < submit->nr_bos; i++) { > @@ -300,11 +302,7 @@ int msm_sched_job_init(struct drm_sched_job *sched_job) > mutex_unlock(&ring->fence_idr_lock); > > if (submit->out_fence_id < 0) { > - /* > - * TODO: The scheduler's finished fence leaks here since the > - * job will not be pushed to the queue. Need to update scheduler > - * to fix this cleanly(?) > - */ > + drm_sched_job_cleanup(sched_job); > dma_fence_put(submit->out_fence); > submit->out_fence = NULL; > return -ENOMEM; > diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler.c b/drivers/gpu/drm/scheduler/gpu_scheduler.c > index f5534ff..bbf9315 100644 > --- a/drivers/gpu/drm/scheduler/gpu_scheduler.c > +++ b/drivers/gpu/drm/scheduler/gpu_scheduler.c > @@ -214,7 +214,6 @@ static void drm_sched_entity_kill_jobs_cb(struct dma_fence *f, > finish_cb); > drm_sched_fence_finished(job->s_fence); > WARN_ON(job->s_fence->parent); > - dma_fence_put(&job->s_fence->finished); > job->sched->ops->free_job(job); > } > > @@ -480,7 +479,6 @@ static void drm_sched_job_finish(struct work_struct *work) > drm_sched_queue_delayed_work(next); > } > spin_unlock(&sched->job_list_lock); > - dma_fence_put(&s_job->s_fence->finished); > sched->ops->free_job(s_job); > } > > @@ -636,6 +634,15 @@ int drm_sched_job_init(struct drm_sched_job *job, > EXPORT_SYMBOL(drm_sched_job_init); > > /** > + * Cleanup sched_job resources > + */ > +void drm_sched_job_cleanup(struct drm_sched_job *job) > +{ > + dma_fence_put(&job->s_fence->finished); > +} > +EXPORT_SYMBOL(drm_sched_job_cleanup); > + > +/** > * Return ture if we can push more jobs to the hw. > */ > static bool drm_sched_ready(struct drm_gpu_scheduler *sched) > diff --git a/drivers/gpu/drm/v3d/v3d_sched.c b/drivers/gpu/drm/v3d/v3d_sched.c > index b07bece..4abd73c 100644 > --- a/drivers/gpu/drm/v3d/v3d_sched.c > +++ b/drivers/gpu/drm/v3d/v3d_sched.c > @@ -35,6 +35,8 @@ > { > struct v3d_job *job = to_v3d_job(sched_job); > > + drm_sched_job_cleanup(sched_job); > + > v3d_exec_put(job->exec); > } > > diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h > index 5c59c38..2ec3ddf 100644 > --- a/include/drm/gpu_scheduler.h > +++ b/include/drm/gpu_scheduler.h > @@ -213,6 +213,7 @@ int drm_sched_job_init(struct drm_sched_job *job, > struct drm_gpu_scheduler *sched, > struct drm_sched_entity *entity, > void *owner); > +void drm_sched_job_cleanup(struct drm_sched_job *job); > void drm_sched_hw_job_reset(struct drm_gpu_scheduler *sched, > struct drm_sched_job *job); > void drm_sched_job_recovery(struct drm_gpu_scheduler *sched); > -- > 1.9.1 > -- The Qualcomm Innovation Center, Inc. is a member of Code Aurora Forum, a Linux Foundation Collaborative Project