Re: [RFC PATCH] drm/scheduler: Add drm_sched_job_cleanup

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Oct 24, 2018 at 07:10:23PM +0530, Sharat Masetty wrote:
> Hi,
> 
> Can you please review this and share your thoughts on this approach. We
> primarily need a clean way to release the finished fence in case the job
> was not pushed to the queue. So, this patch adds the new API to clean up
> the resources allocated in sched_job_init()
> 
> Additionally I also move the fence_put(finished_fence) from the
> scheduler to the drivers handler of free_job(). The drivers get to use
> this new API. This is done so that the layer creating the sched object is
> the one freeing up the resources as well.
> 
> Signed-off-by: Sharat Masetty <smasetty@xxxxxxxxxxxxxx>
> ---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c    |  2 +-
>  drivers/gpu/drm/amd/amdgpu/amdgpu_job.c   |  2 ++
>  drivers/gpu/drm/etnaviv/etnaviv_sched.c   |  2 ++
>  drivers/gpu/drm/msm/msm_sched.c           |  8 +++-----

Just a nit - msm_sched.c isn't a thing yet. Can you re-generate a patch against
drm-next so that if folks like this they can get it merged sooner?

Jordan
>  drivers/gpu/drm/scheduler/gpu_scheduler.c | 11 +++++++++--
>  drivers/gpu/drm/v3d/v3d_sched.c           |  2 ++
>  include/drm/gpu_scheduler.h               |  1 +
>  7 files changed, 20 insertions(+), 8 deletions(-)
> 
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> index 9c85a90..9a83152 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
> @@ -1198,7 +1198,7 @@ static int amdgpu_cs_submit(struct amdgpu_cs_parser *p,
>  	r = amdgpu_ctx_add_fence(p->ctx, ring, p->fence, &seq);
>  	if (r) {
>  		dma_fence_put(p->fence);
> -		dma_fence_put(&job->base.s_fence->finished);
> +		drm_sched_job_cleanup(&job->base);
>  		amdgpu_job_free(job);
>  		amdgpu_mn_unlock(p->mn);
>  		return r;
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
> index 2bd5676..5d252d4 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_job.c
> @@ -100,6 +100,8 @@ static void amdgpu_job_free_cb(struct drm_sched_job *s_job)
>  {
>  	struct amdgpu_job *job = container_of(s_job, struct amdgpu_job, base);
>  
> +	drm_sched_job_cleanup(s_job);
> +
>  	amdgpu_ring_priority_put(job->ring, s_job->s_priority);
>  	dma_fence_put(job->fence);
>  	amdgpu_sync_free(&job->sync);
> diff --git a/drivers/gpu/drm/etnaviv/etnaviv_sched.c b/drivers/gpu/drm/etnaviv/etnaviv_sched.c
> index 50d6b88..30398c5 100644
> --- a/drivers/gpu/drm/etnaviv/etnaviv_sched.c
> +++ b/drivers/gpu/drm/etnaviv/etnaviv_sched.c
> @@ -127,6 +127,8 @@ static void etnaviv_sched_free_job(struct drm_sched_job *sched_job)
>  {
>  	struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job);
>  
> +	drm_sched_job_cleanup(sched_job);
> +
>  	etnaviv_submit_put(submit);
>  }
>  
> diff --git a/drivers/gpu/drm/msm/msm_sched.c b/drivers/gpu/drm/msm/msm_sched.c
> index 70b7713..748acc6 100644
> --- a/drivers/gpu/drm/msm/msm_sched.c
> +++ b/drivers/gpu/drm/msm/msm_sched.c
> @@ -235,6 +235,8 @@ static void msm_sched_free_job(struct drm_sched_job *sched_job)
>  	struct msm_gpu *gpu = submit->gpu;
>  	int i;
>  
> +	drm_sched_job_cleanup(sched_job);
> +
>  	mutex_lock(&gpu->dev->struct_mutex);
>  
>  	for (i = 0; i < submit->nr_bos; i++) {
> @@ -300,11 +302,7 @@ int msm_sched_job_init(struct drm_sched_job *sched_job)
>  	mutex_unlock(&ring->fence_idr_lock);
>  
>  	if (submit->out_fence_id < 0) {
> -		/*
> -		 * TODO: The scheduler's finished fence leaks here since the
> -		 * job will not be pushed to the queue. Need to update scheduler
> -		 * to fix this cleanly(?)
> -		 */
> +		drm_sched_job_cleanup(sched_job);
>  		dma_fence_put(submit->out_fence);
>  		submit->out_fence = NULL;
>  		return -ENOMEM;
> diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler.c b/drivers/gpu/drm/scheduler/gpu_scheduler.c
> index f5534ff..bbf9315 100644
> --- a/drivers/gpu/drm/scheduler/gpu_scheduler.c
> +++ b/drivers/gpu/drm/scheduler/gpu_scheduler.c
> @@ -214,7 +214,6 @@ static void drm_sched_entity_kill_jobs_cb(struct dma_fence *f,
>  						 finish_cb);
>  	drm_sched_fence_finished(job->s_fence);
>  	WARN_ON(job->s_fence->parent);
> -	dma_fence_put(&job->s_fence->finished);
>  	job->sched->ops->free_job(job);
>  }
>  
> @@ -480,7 +479,6 @@ static void drm_sched_job_finish(struct work_struct *work)
>  			drm_sched_queue_delayed_work(next);
>  	}
>  	spin_unlock(&sched->job_list_lock);
> -	dma_fence_put(&s_job->s_fence->finished);
>  	sched->ops->free_job(s_job);
>  }
>  
> @@ -636,6 +634,15 @@ int drm_sched_job_init(struct drm_sched_job *job,
>  EXPORT_SYMBOL(drm_sched_job_init);
>  
>  /**
> + * Cleanup sched_job resources
> + */
> +void drm_sched_job_cleanup(struct drm_sched_job *job)
> +{
> +	dma_fence_put(&job->s_fence->finished);
> +}
> +EXPORT_SYMBOL(drm_sched_job_cleanup);
> +
> +/**
>   * Return ture if we can push more jobs to the hw.
>   */
>  static bool drm_sched_ready(struct drm_gpu_scheduler *sched)
> diff --git a/drivers/gpu/drm/v3d/v3d_sched.c b/drivers/gpu/drm/v3d/v3d_sched.c
> index b07bece..4abd73c 100644
> --- a/drivers/gpu/drm/v3d/v3d_sched.c
> +++ b/drivers/gpu/drm/v3d/v3d_sched.c
> @@ -35,6 +35,8 @@
>  {
>  	struct v3d_job *job = to_v3d_job(sched_job);
>  
> +	drm_sched_job_cleanup(sched_job);
> +
>  	v3d_exec_put(job->exec);
>  }
>  
> diff --git a/include/drm/gpu_scheduler.h b/include/drm/gpu_scheduler.h
> index 5c59c38..2ec3ddf 100644
> --- a/include/drm/gpu_scheduler.h
> +++ b/include/drm/gpu_scheduler.h
> @@ -213,6 +213,7 @@ int drm_sched_job_init(struct drm_sched_job *job,
>  		       struct drm_gpu_scheduler *sched,
>  		       struct drm_sched_entity *entity,
>  		       void *owner);
> +void drm_sched_job_cleanup(struct drm_sched_job *job);
>  void drm_sched_hw_job_reset(struct drm_gpu_scheduler *sched,
>  			    struct drm_sched_job *job);
>  void drm_sched_job_recovery(struct drm_gpu_scheduler *sched);
> -- 
> 1.9.1
> 

-- 
The Qualcomm Innovation Center, Inc. is a member of Code Aurora Forum,
a Linux Foundation Collaborative Project
_______________________________________________
dri-devel mailing list
dri-devel@xxxxxxxxxxxxxxxxxxxxx
https://lists.freedesktop.org/mailman/listinfo/dri-devel




[Index of Archives]     [Linux DRI Users]     [Linux Intel Graphics]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [XFree86]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux