On Fri, 2021-10-01 at 09:37 +0100, Melissa Wen wrote: > On 10/01, Iago Toral wrote: > > On Thu, 2021-09-30 at 17:19 +0100, Melissa Wen wrote: > > > Using the generic extension from the previous patch, a specific > > > multisync > > > extension enables more than one in/out binary syncobj per job > > > submission. > > > Arrays of syncobjs are set in struct drm_v3d_multisync, that also > > > cares > > > of determining the stage for sync (wait deps) according to the > > > job > > > queue. > > > > > > v2: > > > - subclass the generic extension struct (Daniel) > > > - simplify adding dependency conditions to make understandable > > > (Iago) > > > > > > v3: > > > - fix conditions to consider single or multiples in/out_syncs > > > (Iago) > > > - remove irrelevant comment (Iago) > > > > > > Signed-off-by: Melissa Wen <mwen@xxxxxxxxxx> > > > --- > > > drivers/gpu/drm/v3d/v3d_drv.c | 6 +- > > > drivers/gpu/drm/v3d/v3d_drv.h | 24 +++-- > > > drivers/gpu/drm/v3d/v3d_gem.c | 185 > > > ++++++++++++++++++++++++++++++ > > > ---- > > > include/uapi/drm/v3d_drm.h | 49 ++++++++- > > > 4 files changed, 232 insertions(+), 32 deletions(-) > > > > > > diff --git a/drivers/gpu/drm/v3d/v3d_drv.c > > > b/drivers/gpu/drm/v3d/v3d_drv.c > > > index 3d6b9bcce2f7..bd46396a1ae0 100644 > > > --- a/drivers/gpu/drm/v3d/v3d_drv.c > > > +++ b/drivers/gpu/drm/v3d/v3d_drv.c > > > @@ -96,6 +96,9 @@ static int v3d_get_param_ioctl(struct > > > drm_device > > > > (...) > > > > > @@ -516,9 +536,11 @@ > > > v3d_attach_fences_and_unlock_reservation(struct > > > drm_file *file_priv, > > > struct v3d_job *job, > > > struct ww_acquire_ctx > > > *acquire_ctx, > > > u32 out_sync, > > > + struct v3d_submit_ext *se, > > > struct dma_fence *done_fence) > > > { > > > struct drm_syncobj *sync_out; > > > + bool has_multisync = se && (se->flags & > > > > We always pass the 'se' pointer from a local variable allocated in > > the > > stack of the caller so it is never going to be NULL. > > > > I am happy to keep the NULL check if you want to protect against > > future > > changes just in case, but if we do that then... > > > > > DRM_V3D_EXT_ID_MULTI_SYNC); > > > int i; > > > > > > for (i = 0; i < job->bo_count; i++) { > > > > (...) > > > > > +static void > > > +v3d_put_multisync_post_deps(struct v3d_submit_ext *se) > > > +{ > > > + unsigned int i; > > > + > > > + if (!se->out_sync_count) > > > > ...we should also check for NULL here for consistency. > yes, consistency makes sense here. > > Also, I think there is another problem in the code: we always call > > v3d_job_cleanup for failed jobs, but only call v3d_job_put for > > successful jobs. However, reading the docs for drm_sched_job_init: > > > > "Drivers must make sure drm_sched_job_cleanup() if this function > > returns successfully, even when @job is aborted before > > drm_sched_job_arm() is called." > > > > So my understanding is that we should call v3d_job_cleanup instead > > of > > v3d_job_put for successful jobs or we would be leaking sched > > resources > > on every successful job, no? > > When job_init is successful, v3d_job_cleanup is called by scheduler > when > job is completed. drm_sched_job_cleanup describes how it works after > drm_sched_job_arm: > > " After that point of no return @job is committed to be executed by > the > scheduler, and this function should be called from the > &drm_sched_backend_ops.free_job callback." > > On v3d_sched.c, .free_job points to v3d_sched_job_free(), which in > turn > calls v3d_job_cleanup() (and then drm_sched_job_cleanup). So, it > looks > ok. > > Also, we can say that the very last v3d_job_put() is in charge to > decrement refcount initialized (set 1) in v3d_job_init(); while the > v3d_job_cleanup() from v3d_sched_job_free() callback decrements > refcount > that was incremented when v3d_job_push() pushed the job to the > scheduler. > > So, refcount pairs seem consistent, I mean, get and put. And about > drm_sched_job_cleanup, it is explicitly called when job_init fails or > after that by the scheduler. > A. Ah ok, thanks for explaining this! With the consistency issue discussed above fixed, for the series: Reviewed-by: Iago Toral Quiroga <itoral@xxxxxxxxxx> Iago