On Tue, Jun 25, 2019 at 03:55:06PM +0200, Christian König wrote: > The ww_mutex framework allows for detecting deadlocks when multiple > threads try to acquire the same set of locks in different order. > > The problem is that handling those deadlocks was the burden of the user of > the ww_mutex implementation and at least some users didn't got that right > on the first try. > > So introduce a new reservation_context object which can be used to > simplify the deadlock handling. This is done by tracking all locked > reservation objects in the context as well as the last contended > reservation object. > > When a deadlock occurse we now unlock all previously locked object and > acquire the contended lock in the slow path. After this is done -EDEADLK > is still returned to signal that all other locks now need to be > re-acquired again. > > Signed-off-by: Christian König <christian.koenig@xxxxxxx> > --- > drivers/dma-buf/reservation.c | 82 +++++++++++++++++++++++++++++++++++ > include/linux/reservation.h | 38 ++++++++++++++++ > 2 files changed, 120 insertions(+) > > diff --git a/drivers/dma-buf/reservation.c b/drivers/dma-buf/reservation.c > index 4d32e2c67862..9e53e42b053a 100644 > --- a/drivers/dma-buf/reservation.c > +++ b/drivers/dma-buf/reservation.c > @@ -55,6 +55,88 @@ EXPORT_SYMBOL(reservation_seqcount_class); > const char reservation_seqcount_string[] = "reservation_seqcount"; > EXPORT_SYMBOL(reservation_seqcount_string); > > +/** > + * reservation_context_init - initialize a reservation context > + * @ctx: the context to initialize > + * > + * Start using this reservation context to lock reservation objects for update. Bunch of hyperlinks here for more consistent story would be really nice in the kerneldoc. > + */ > +void reservation_context_init(struct reservation_context *ctx) > +{ > + ww_acquire_init(&ctx->ctx, &reservation_ww_class); > + init_llist_head(&ctx->locked); > + ctx->contended = NULL; > +} > +EXPORT_SYMBOL(reservation_context_init); > + > +/** > + * reservation_context_unlock_all - unlock all reservation objects > + * @ctx: the context which holds the reservation objects > + * > + * Unlocks all reservation objects locked with this context. > + */ > +void reservation_context_unlock_all(struct reservation_context *ctx) I'd just call this reservation_unlock_all or so. Feel free to ignore the bikeshed. > +{ > + struct reservation_object *obj, *next; > + > + if (ctx->contended) > + ww_mutex_unlock(&ctx->contended->lock); > + ctx->contended = NULL; > + > + llist_for_each_entry_safe(obj, next, ctx->locked.first, locked) > + ww_mutex_unlock(&obj->lock); > + init_llist_head(&ctx->locked); > +} > +EXPORT_SYMBOL(reservation_context_unlock_all); > + > +/** > + * reservation_context_lock - lock a reservation object with deadlock handling > + * @ctx: the context which should be used to lock the object > + * @obj: the object which needs to be locked > + * @interruptible: if we should wait interruptible or not > + * > + * Use @ctx to lock the reservation object. If a deadlock is detected we backoff > + * by releasing all locked objects and use the slow path to lock the reservation > + * object. After successfully locking in the slow path -EDEADLK is returned to > + * signal that all other locks must be re-taken as well. > + */ > +int reservation_context_lock(struct reservation_context *ctx, > + struct reservation_object *obj, > + bool interruptible) reservation_lock_ctx is what we generally used in drm_modeset_lock, I like that bikeshed a bit better. Also to stay in style I think the explicit set of functions is much better, i.e. reservation_lock_ctx, reservation_lock_interruptible_ctx and reservation_trylock_ctx (later useful for lru applications where you still want to drop the entire pile with resrvation_unlock_ctx). That's what all the other locking things do. ttm_bo_reserve has a long list of parameters, and I can never remember which is which. I don't think that's a great style. Another option for interruptible vs. not is to store that in the reservation_context and dtrt. Since generally interruptible or not is a propery of the top-level handler - you need be able to pass EDEADLCK all the way up anyway. > +{ > + int ret = 0; > + > + if (unlikely(ctx->contended == obj)) > + ctx->contended = NULL; Imo cleaner to handle that with EALREADY filtering from the ww_mutex_lock. > + else if (interruptible) > + ret = ww_mutex_lock_interruptible(&obj->lock, &ctx->ctx); > + else > + ret = ww_mutex_lock(&obj->lock, &ctx->ctx); > + > + if (likely(!ret)) { > + /* don't use llist_add here, we have separate locking */ > + obj->locked.next = ctx->locked.first; > + ctx->locked.first = &obj->locked; > + return 0; > + } > + if (unlikely(ret != -EDEADLK)) > + return ret; > + > + reservation_context_unlock_all(ctx); > + > + if (interruptible) { > + ret = ww_mutex_lock_slow_interruptible(&obj->lock, &ctx->ctx); > + if (unlikely(ret)) > + return ret; > + } else { > + ww_mutex_lock_slow(&obj->lock, &ctx->ctx); > + } > + > + ctx->contended = obj; > + return -EDEADLK; > +} > +EXPORT_SYMBOL(reservation_context_lock); > + > /** > * reservation_object_reserve_shared - Reserve space to add shared fences to > * a reservation_object. > diff --git a/include/linux/reservation.h b/include/linux/reservation.h > index ee750765cc94..a8a52e5d3e80 100644 > --- a/include/linux/reservation.h > +++ b/include/linux/reservation.h > @@ -44,11 +44,48 @@ > #include <linux/slab.h> > #include <linux/seqlock.h> > #include <linux/rcupdate.h> > +#include <linux/llist.h> > > extern struct ww_class reservation_ww_class; > extern struct lock_class_key reservation_seqcount_class; > extern const char reservation_seqcount_string[]; > > +/** > + * struct reservation_context - context to lock reservation objects > + * @ctx: ww_acquire_ctx used for deadlock detection > + * @locked: list of reservation objects locked in this context > + * @contended: contended reservation object > + */ > +struct reservation_context { > + struct ww_acquire_ctx ctx; > + struct llist_head locked; > + struct reservation_object *contended; > +}; > + > +/** > + * reservation_context_done - wrapper for ww_acquire_done > + * @ctx: the reservation context which is done with locking > + */ > +static inline void reservation_context_done(struct reservation_context *ctx) > +{ > + ww_acquire_done(&ctx->ctx); > +} > + > +/** > + * reservation_context_fini - wrapper for ww_acquire_fini > + * @ctx: the reservation context which is finished > + */ > +static inline void reservation_context_fini(struct reservation_context *ctx) > +{ > + ww_acquire_fini(&ctx->ctx); > +} > + > +void reservation_context_init(struct reservation_context *ctx); > +void reservation_context_unlock_all(struct reservation_context *ctx); > +int reservation_context_lock(struct reservation_context *ctx, > + struct reservation_object *obj, > + bool interruptible); Needs a __must_check. > + > /** > * struct reservation_object_list - a list of shared fences > * @rcu: for internal use > @@ -71,6 +108,7 @@ struct reservation_object_list { > */ > struct reservation_object { > struct ww_mutex lock; > + struct llist_node locked; > seqcount_t seq; > > struct dma_fence __rcu *fence_excl; Aside from the nits&bikesheds, I like. -Daniel > -- > 2.17.1 > -- Daniel Vetter Software Engineer, Intel Corporation http://blog.ffwll.ch _______________________________________________ dri-devel mailing list dri-devel@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/dri-devel