Sorry I burried myself in some other stuff ... On Sat, Aug 10, 2019 at 12:51:00PM +0200, Christian König wrote: > Am 07.08.19 um 16:17 schrieb Chris Wilson: > > Quoting Christian König (2019-08-07 14:53:12) > > > The only remaining use for this is to protect against setting a new exclusive > > > fence while we grab both exclusive and shared. That can also be archived by > > > looking if the exclusive fence has changed or not after completing the > > > operation. > > > > > > v2: switch setting excl fence to rcu_assign_pointer > > > > > > Signed-off-by: Christian König <christian.koenig@xxxxxxx> > > > --- > > > drivers/dma-buf/reservation.c | 24 +++++------------------- > > > include/linux/reservation.h | 9 ++------- > > > 2 files changed, 7 insertions(+), 26 deletions(-) > > > > > > diff --git a/drivers/dma-buf/reservation.c b/drivers/dma-buf/reservation.c > > > index 90bc6ef03598..f7f4a0858c2a 100644 > > > --- a/drivers/dma-buf/reservation.c > > > +++ b/drivers/dma-buf/reservation.c > > > @@ -49,12 +49,6 @@ > > > DEFINE_WD_CLASS(reservation_ww_class); > > > EXPORT_SYMBOL(reservation_ww_class); > > > -struct lock_class_key reservation_seqcount_class; > > > -EXPORT_SYMBOL(reservation_seqcount_class); > > > - > > > -const char reservation_seqcount_string[] = "reservation_seqcount"; > > > -EXPORT_SYMBOL(reservation_seqcount_string); > > > - > > > /** > > > * reservation_object_list_alloc - allocate fence list > > > * @shared_max: number of fences we need space for > > > @@ -103,9 +97,6 @@ static void reservation_object_list_free(struct reservation_object_list *list) > > > void reservation_object_init(struct reservation_object *obj) > > > { > > > ww_mutex_init(&obj->lock, &reservation_ww_class); > > > - > > > - __seqcount_init(&obj->seq, reservation_seqcount_string, > > > - &reservation_seqcount_class); > > > RCU_INIT_POINTER(obj->fence, NULL); > > > RCU_INIT_POINTER(obj->fence_excl, NULL); > > > } > > > @@ -282,12 +273,10 @@ void reservation_object_add_excl_fence(struct reservation_object *obj, > > > dma_fence_get(fence); > > > preempt_disable(); > > > - write_seqcount_begin(&obj->seq); > > > - /* write_seqcount_begin provides the necessary memory barrier */ > > > - RCU_INIT_POINTER(obj->fence_excl, fence); > > > + rcu_assign_pointer(obj->fence_excl, fence); > > > + /* pointer update must be visible before we modify the shared_count */ Pls add a "see reservation_object_fence()" here or similar. > > > if (old) > > > - old->shared_count = 0; > > > - write_seqcount_end(&obj->seq); > > > + smp_store_mb(old->shared_count, 0); So your comment and the kerneldoc don't match up. Quoting Documentation/memory-barriers.txt: This assigns the value to the variable and then inserts a full memory barrier after it. It isn't guaranteed to insert anything more than a compiler barrier in a UP compilation. So order is 1. store 2. fence, but your comment suggests you want it the other way round. > > > preempt_enable(); > > > /* inplace update, no shared fences */ > > > @@ -368,11 +357,8 @@ int reservation_object_copy_fences(struct reservation_object *dst, > > > old = reservation_object_get_excl(dst); > > > preempt_disable(); > > > - write_seqcount_begin(&dst->seq); > > > - /* write_seqcount_begin provides the necessary memory barrier */ > > > - RCU_INIT_POINTER(dst->fence_excl, new); > > > - RCU_INIT_POINTER(dst->fence, dst_list); > > > - write_seqcount_end(&dst->seq); > > > + rcu_assign_pointer(dst->fence_excl, new); > > > + rcu_assign_pointer(dst->fence, dst_list); > > > preempt_enable(); > > > reservation_object_list_free(src_list); > > > diff --git a/include/linux/reservation.h b/include/linux/reservation.h > > > index 044a5cd4af50..fd29baad0be3 100644 > > > --- a/include/linux/reservation.h > > > +++ b/include/linux/reservation.h > > > @@ -46,8 +46,6 @@ > > > #include <linux/rcupdate.h> > > > extern struct ww_class reservation_ww_class; > > > -extern struct lock_class_key reservation_seqcount_class; > > > -extern const char reservation_seqcount_string[]; > > > /** > > > * struct reservation_object_list - a list of shared fences > > > @@ -71,7 +69,6 @@ struct reservation_object_list { > > > */ > > > struct reservation_object { > > > struct ww_mutex lock; > > > - seqcount_t seq; > > > struct dma_fence __rcu *fence_excl; > > > struct reservation_object_list __rcu *fence; > > > @@ -156,14 +153,12 @@ reservation_object_fences(struct reservation_object *obj, > > > struct reservation_object_list **list, > > > u32 *shared_count) > > > { > > > - unsigned int seq; > > > - > > > do { > > > - seq = read_seqcount_begin(&obj->seq); > > > *excl = rcu_dereference(obj->fence_excl); I think you need a barrier between this and the read of shared_count below. But rcu_derefence only gives you a dependent barrier, i.e. only stuff that's accesses through this pointer is ordered. Which means the access to ->shared_count, which goes through another pointer, isn't actually ordered. I think the implementation is that it is an unconditional compiler barrier (but that might change), but you're definitely missing the cpu barrier, so a cpue might speculate the entire thing out of order. I think you need another smb_rmb(); here > > > *list = rcu_dereference(obj->fence); > > > *shared_count = *list ? (*list)->shared_count : 0; > > > - } while (read_seqcount_retry(&obj->seq, seq)); > > > + smp_rmb(); /* See reservation_object_add_excl_fence */ This fence here I think prevents the re-reading of ->fence_excl from getting hoisted above the critical reads. So this is just the open-coded seqlock retry loop. > > > + } while (rcu_access_pointer(obj->fence_excl) != *excl); What if someone is real fast (like really real fast) and recycles the exclusive fence so you read the same pointer twice, but everything else changed? reused fence pointer is a lot more likely than seqlock wrapping around. > > > } > > Reviewed-by: Chris Wilson <chris@xxxxxxxxxxxxxxxxxx> > > > > I think this is correct. Now see if we can convince Daniel! > > Daniel any objections to this? IGTs look good as well, so if not I'm going > to push it. Not really convinced. Also haven't looked at the entire thing yet, this is just from staring at this patch in isolation and poking at it. -Daniel -- Daniel Vetter Software Engineer, Intel Corporation http://blog.ffwll.ch _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx