Am 29.08.2018 um 12:44 schrieb Chunming Zhou: > VK_KHR_timeline_semaphore: > This extension introduces a new type of semaphore that has an integer payload > identifying a point in a timeline. Such timeline semaphores support the > following operations: > * CPU query - A host operation that allows querying the payload of the > timeline semaphore. > * CPU wait - A host operation that allows a blocking wait for a > timeline semaphore to reach a specified value. > * Device wait - A device operation that allows waiting for a > timeline semaphore to reach a specified value. > * Device signal - A device operation that allows advancing the > timeline semaphore to a specified value. > > Since it's a timeline, that means the front time point(PT) always is signaled before the late PT. > a. signal PT design: > Signal PT fence N depends on PT[N-1] fence and signal opertion fence, when PT[N] fence is signaled, > the timeline will increase to value of PT[N]. > b. wait PT design: > Wait PT fence is signaled by reaching timeline point value, when timeline is increasing, will compare > wait PTs value with new timeline value, if PT value is lower than timeline value, then wait PT will be > signaled, otherwise keep in list. semaphore wait operation can wait on any point of timeline, > so need a RB tree to order them. And wait PT could ahead of signal PT, we need a sumission fence to > perform that. > > v2: > 1. remove unused DRM_SYNCOBJ_CREATE_TYPE_NORMAL. (Christian) > 2. move unexposed denitions to .c file. (Daniel Vetter) > 3. split up the change to drm_syncobj_find_fence() in a separate patch. (Christian) > 4. split up the change to drm_syncobj_replace_fence() in a separate patch. > 5. drop the submission_fence implementation and instead use wait_event() for that. (Christian) > 6. WARN_ON(point != 0) for NORMAL type syncobj case. (Daniel Vetter) > > v3: > 1. replace normal syncobj with timeline implemenation. (Vetter and Christian) > a. normal syncobj signal op will create a signal PT to tail of signal pt list. > b. normal syncobj wait op will create a wait pt with last signal point, and this wait PT is only signaled by related signal point PT. > 2. some bug fix and clean up > 3. tested by ./deqp-vk -n dEQP-VK*semaphore* for normal syncobj > > TODO: > 1. CPU query and wait on timeline semaphore. > > Signed-off-by: Chunming Zhou <david1.zhou at amd.com> > Cc: Christian Konig <christian.koenig at amd.com> > Cc: Dave Airlie <airlied at redhat.com> > Cc: Daniel Rakos <Daniel.Rakos at amd.com> > Cc: Daniel Vetter <daniel at ffwll.ch> > --- > drivers/gpu/drm/drm_syncobj.c | 593 ++++++++++++++++----- > drivers/gpu/drm/i915/i915_gem_execbuffer.c | 4 +- > include/drm/drm_syncobj.h | 78 +-- > include/uapi/drm/drm.h | 1 + > 4 files changed, 505 insertions(+), 171 deletions(-) > > diff --git a/drivers/gpu/drm/drm_syncobj.c b/drivers/gpu/drm/drm_syncobj.c > index ab43559398d0..f701d9ef1b81 100644 > --- a/drivers/gpu/drm/drm_syncobj.c > +++ b/drivers/gpu/drm/drm_syncobj.c > @@ -56,6 +56,50 @@ > #include "drm_internal.h" > #include <drm/drm_syncobj.h> > > +/* merge normal syncobj to timeline syncobj, the point interval is 1 */ > +#define DRM_SYNCOBJ_NORMAL_POINT 1 > + > +struct drm_syncobj_stub_fence { > + struct dma_fence base; > + spinlock_t lock; > +}; > + > +static const char *drm_syncobj_stub_fence_get_name(struct dma_fence *fence) > +{ > + return "syncobjstub"; > +} > + > +static bool drm_syncobj_stub_fence_enable_signaling(struct dma_fence *fence) > +{ > + return !dma_fence_is_signaled(fence); > +} > +static void drm_syncobj_stub_fence_release(struct dma_fence *f) > +{ > + kfree(f); > +} > +static const struct dma_fence_ops drm_syncobj_stub_fence_ops = { > + .get_driver_name = drm_syncobj_stub_fence_get_name, > + .get_timeline_name = drm_syncobj_stub_fence_get_name, > + .enable_signaling = drm_syncobj_stub_fence_enable_signaling, > + .release = drm_syncobj_stub_fence_release, > +}; Do we really need to move that around? Could reduce the size of the patch quite a bit if we don't. > + > +struct drm_syncobj_wait_pt { > + struct drm_syncobj_stub_fence base; > + u64 value; > + struct rb_node node; > +}; > +struct drm_syncobj_signal_pt { > + struct drm_syncobj_stub_fence base; > + struct dma_fence *signal_fence; > + struct dma_fence *pre_pt_base; > + struct dma_fence_cb signal_cb; > + struct dma_fence_cb pre_pt_cb; > + struct drm_syncobj *syncobj; > + u64 value; > + struct list_head list; > +}; What are those two structures good for and why is the stub fence their base? > + > /** > * drm_syncobj_find - lookup and reference a sync object. > * @file_private: drm file private pointer > @@ -82,59 +126,247 @@ struct drm_syncobj *drm_syncobj_find(struct drm_file *file_private, > } > EXPORT_SYMBOL(drm_syncobj_find); > > -static void drm_syncobj_add_callback_locked(struct drm_syncobj *syncobj, > - struct drm_syncobj_cb *cb, > - drm_syncobj_func_t func) > +static void drm_syncobj_timeline_init(struct drm_syncobj_timeline > + *syncobj_timeline) > { > - cb->func = func; > - list_add_tail(&cb->node, &syncobj->cb_list); > + syncobj_timeline->timeline_context = dma_fence_context_alloc(1); > + syncobj_timeline->timeline = 0; > + syncobj_timeline->signal_point = 0; > + init_waitqueue_head(&syncobj_timeline->wq); > + > + syncobj_timeline->wait_pt_tree = RB_ROOT; > + INIT_LIST_HEAD(&syncobj_timeline->signal_pt_list); > } > > -static int drm_syncobj_fence_get_or_add_callback(struct drm_syncobj *syncobj, > - struct dma_fence **fence, > - struct drm_syncobj_cb *cb, > - drm_syncobj_func_t func) > +static void drm_syncobj_timeline_fini(struct drm_syncobj *syncobj, > + struct drm_syncobj_timeline *syncobj_timeline) > { > - int ret; > + struct rb_node *node = NULL; > + struct drm_syncobj_wait_pt *wait_pt = NULL; > + struct drm_syncobj_signal_pt *signal_pt = NULL, *tmp; > + > + spin_lock(&syncobj->lock); > + for(node = rb_first(&syncobj_timeline->wait_pt_tree); > + node != NULL; ) { Better use rbtree_postorder_for_each_entry_safe() for this. Regards, Christian. > + wait_pt = rb_entry(node, struct drm_syncobj_wait_pt, node); > + node = rb_next(node); > + rb_erase(&wait_pt->node, > + &syncobj_timeline->wait_pt_tree); > + RB_CLEAR_NODE(&wait_pt->node); > + spin_unlock(&syncobj->lock); > + dma_fence_wait(&wait_pt->base.base, true); > + spin_lock(&syncobj->lock); > + /* kfree(wait_pt) is excuted by fence put */ > + dma_fence_put(&wait_pt->base.base); > + } > + list_for_each_entry_safe(signal_pt, tmp, > + &syncobj_timeline->signal_pt_list, list) { > + list_del(&signal_pt->list); > + if (signal_pt->signal_fence) { > + dma_fence_remove_callback(signal_pt->signal_fence, > + &signal_pt->signal_cb); > + dma_fence_put(signal_pt->signal_fence); > + } > + if (signal_pt->pre_pt_base) { > + dma_fence_remove_callback(signal_pt->pre_pt_base, > + &signal_pt->pre_pt_cb); > + dma_fence_put(signal_pt->pre_pt_base); > + } > + dma_fence_put(&signal_pt->base.base); > + } > + spin_unlock(&syncobj->lock); > +} > + > > - *fence = drm_syncobj_fence_get(syncobj); > - if (*fence) > - return 1; > +static bool drm_syncobj_normal_signal_wait_pts(struct drm_syncobj *syncobj, > + u64 signal_pt) > +{ > + struct rb_node *node = NULL; > + struct drm_syncobj_wait_pt *wait_pt = NULL; > > spin_lock(&syncobj->lock); > - /* We've already tried once to get a fence and failed. Now that we > - * have the lock, try one more time just to be sure we don't add a > - * callback when a fence has already been set. > - */ > - if (syncobj->fence) { > - *fence = dma_fence_get(rcu_dereference_protected(syncobj->fence, > - lockdep_is_held(&syncobj->lock))); > - ret = 1; > - } else { > - *fence = NULL; > - drm_syncobj_add_callback_locked(syncobj, cb, func); > - ret = 0; > + for(node = rb_first(&syncobj->syncobj_timeline.wait_pt_tree); > + node != NULL; ) { > + wait_pt = rb_entry(node, struct drm_syncobj_wait_pt, node); > + node = rb_next(node); > + /* for normal syncobj */ > + if (wait_pt->value == signal_pt) { > + dma_fence_signal(&wait_pt->base.base); > + rb_erase(&wait_pt->node, > + &syncobj->syncobj_timeline.wait_pt_tree); > + RB_CLEAR_NODE(&wait_pt->node); > + /* kfree(wait_pt) is excuted by fence put */ > + dma_fence_put(&wait_pt->base.base); > + spin_unlock(&syncobj->lock); > + return true; > + } > } > spin_unlock(&syncobj->lock); > + return false; > +} > > - return ret; > +static void drm_syncobj_timeline_signal_wait_pts(struct drm_syncobj *syncobj) > +{ > + struct rb_node *node = NULL; > + struct drm_syncobj_wait_pt *wait_pt = NULL; > + > + spin_lock(&syncobj->lock); > + for(node = rb_first(&syncobj->syncobj_timeline.wait_pt_tree); > + node != NULL; ) { > + wait_pt = rb_entry(node, struct drm_syncobj_wait_pt, node); > + node = rb_next(node); > + if (wait_pt->value <= syncobj->syncobj_timeline.timeline) { > + dma_fence_signal(&wait_pt->base.base); > + rb_erase(&wait_pt->node, > + &syncobj->syncobj_timeline.wait_pt_tree); > + RB_CLEAR_NODE(&wait_pt->node); > + /* kfree(wait_pt) is excuted by fence put */ > + dma_fence_put(&wait_pt->base.base); > + } else { > + /* the loop is from left to right, the later entry value is > + * bigger, so don't need to check any more */ > + break; > + } > + } > + spin_unlock(&syncobj->lock); > } > > -void drm_syncobj_add_callback(struct drm_syncobj *syncobj, > - struct drm_syncobj_cb *cb, > - drm_syncobj_func_t func) > + > + > +static void pt_fence_cb(struct drm_syncobj_signal_pt *signal_pt) > { > + struct dma_fence *fence = NULL; > + struct drm_syncobj *syncobj; > + struct drm_syncobj_signal_pt *tail_pt; > + u64 pt_value = signal_pt->value; > + > + dma_fence_signal(&signal_pt->base.base); > + fence = signal_pt->signal_fence; > + signal_pt->signal_fence = NULL; > + dma_fence_put(fence); > + fence = signal_pt->pre_pt_base; > + signal_pt->pre_pt_base = NULL; > + dma_fence_put(fence); > + > + syncobj = signal_pt->syncobj; > spin_lock(&syncobj->lock); > - drm_syncobj_add_callback_locked(syncobj, cb, func); > + syncobj->syncobj_timeline.timeline = pt_value; > + tail_pt = list_last_entry(&syncobj->syncobj_timeline.signal_pt_list, > + struct drm_syncobj_signal_pt, list); > + if ((syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL && signal_pt != tail_pt) > + || syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) { > + list_del(&signal_pt->list); > + /* kfree(signal_pt) will be executed by below fence put */ > + dma_fence_put(&signal_pt->base.base); > + } > spin_unlock(&syncobj->lock); > + if (syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL) > + drm_syncobj_normal_signal_wait_pts(syncobj, pt_value); > + else > + drm_syncobj_timeline_signal_wait_pts(syncobj); > } > +static void pt_signal_fence_func(struct dma_fence *fence, > + struct dma_fence_cb *cb) > +{ > + struct drm_syncobj_signal_pt *signal_pt = > + container_of(cb, struct drm_syncobj_signal_pt, signal_cb); > > -void drm_syncobj_remove_callback(struct drm_syncobj *syncobj, > - struct drm_syncobj_cb *cb) > + if (signal_pt->pre_pt_base && > + !dma_fence_is_signaled(signal_pt->pre_pt_base)) > + return; > + > + pt_fence_cb(signal_pt); > +} > +static void pt_pre_fence_func(struct dma_fence *fence, > + struct dma_fence_cb *cb) > +{ > + struct drm_syncobj_signal_pt *signal_pt = > + container_of(cb, struct drm_syncobj_signal_pt, pre_pt_cb); > + > + if (signal_pt->signal_fence && > + !dma_fence_is_signaled(signal_pt->pre_pt_base)) > + return; > + > + pt_fence_cb(signal_pt); > +} > + > +static int drm_syncobj_timeline_create_signal_pt(struct drm_syncobj *syncobj, > + struct dma_fence *fence, > + u64 point) > { > + struct drm_syncobj_signal_pt *signal_pt = > + kzalloc(sizeof(struct drm_syncobj_signal_pt), GFP_KERNEL); > + struct drm_syncobj_stub_fence *base; > + struct drm_syncobj_signal_pt *tail_pt; > + struct dma_fence *tail_pt_fence = NULL; > + int ret = 0; > + > + if (!signal_pt) > + return -ENOMEM; > + if (syncobj->syncobj_timeline.signal_point >= point) { > + DRM_WARN("A later signal is ready!"); > + goto out; > + } > + if (!fence) > + goto out; > + dma_fence_get(fence); > spin_lock(&syncobj->lock); > - list_del_init(&cb->node); > + base = &signal_pt->base; > + spin_lock_init(&base->lock); > + dma_fence_init(&base->base, > + &drm_syncobj_stub_fence_ops, > + &base->lock, > + syncobj->syncobj_timeline.timeline_context, point); > + signal_pt->signal_fence = fence; > + /* timeline syncobj must take this dependency */ > + if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) { > + if (!list_empty(&syncobj->syncobj_timeline.signal_pt_list)) { > + tail_pt = list_last_entry(&syncobj->syncobj_timeline.signal_pt_list, > + struct drm_syncobj_signal_pt, list); > + tail_pt_fence = &tail_pt->base.base; > + if (dma_fence_is_signaled(tail_pt_fence)) > + tail_pt_fence = NULL; > + else > + signal_pt->pre_pt_base = > + dma_fence_get(tail_pt_fence); > + } > + } > + > + signal_pt->value = point; > + syncobj->syncobj_timeline.signal_point = point; > + signal_pt->syncobj = syncobj; > + INIT_LIST_HEAD(&signal_pt->list); > + list_add_tail(&signal_pt->list, &syncobj->syncobj_timeline.signal_pt_list); > spin_unlock(&syncobj->lock); > + wake_up_all(&syncobj->syncobj_timeline.wq); > + /** > + * Every pt is depending on signal fence and previous pt fence, add > + * callbacks to them > + */ > + ret = dma_fence_add_callback(signal_pt->signal_fence, > + &signal_pt->signal_cb, > + pt_signal_fence_func); > + > + if (signal_pt->pre_pt_base) { > + ret = dma_fence_add_callback(signal_pt->pre_pt_base, > + &signal_pt->pre_pt_cb, > + pt_pre_fence_func); > + if (ret == -ENOENT) > + pt_pre_fence_func(signal_pt->pre_pt_base, > + &signal_pt->pre_pt_cb); > + else if (ret) > + goto out; > + } else if (ret == -ENOENT) { > + pt_signal_fence_func(signal_pt->signal_fence, > + &signal_pt->signal_cb); > + } else if (ret) { > + goto out; > + } > + > + return 0; > +out: > + dma_fence_put(&signal_pt->base.base); > + return ret; > } > > /** > @@ -149,53 +381,30 @@ void drm_syncobj_replace_fence(struct drm_syncobj *syncobj, > u64 point, > struct dma_fence *fence) > { > - struct dma_fence *old_fence; > - struct drm_syncobj_cb *cur, *tmp; > - > - if (fence) > - dma_fence_get(fence); > - > - spin_lock(&syncobj->lock); > - > - old_fence = rcu_dereference_protected(syncobj->fence, > - lockdep_is_held(&syncobj->lock)); > - rcu_assign_pointer(syncobj->fence, fence); > - > - if (fence != old_fence) { > - list_for_each_entry_safe(cur, tmp, &syncobj->cb_list, node) { > - list_del_init(&cur->node); > - cur->func(syncobj, cur); > + if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) { > + if (fence) > + drm_syncobj_timeline_create_signal_pt(syncobj, fence, > + point); > + return; > + } else if (syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL) { > + u64 pt_value; > + > + if (!fence) { > + drm_syncobj_timeline_fini(syncobj, > + &syncobj->syncobj_timeline); > + drm_syncobj_timeline_init(&syncobj->syncobj_timeline); > + return; > } > + pt_value = syncobj->syncobj_timeline.signal_point + > + DRM_SYNCOBJ_NORMAL_POINT; > + drm_syncobj_timeline_create_signal_pt(syncobj, fence, pt_value); > + return; > + } else { > + DRM_ERROR("the syncobj type isn't support\n"); > } > - > - spin_unlock(&syncobj->lock); > - > - dma_fence_put(old_fence); > } > EXPORT_SYMBOL(drm_syncobj_replace_fence); > > -struct drm_syncobj_stub_fence { > - struct dma_fence base; > - spinlock_t lock; > -}; > - > -static const char *drm_syncobj_stub_fence_get_name(struct dma_fence *fence) > -{ > - return "syncobjstub"; > -} > - > -static bool drm_syncobj_stub_fence_enable_signaling(struct dma_fence *fence) > -{ > - return !dma_fence_is_signaled(fence); > -} > - > -static const struct dma_fence_ops drm_syncobj_stub_fence_ops = { > - .get_driver_name = drm_syncobj_stub_fence_get_name, > - .get_timeline_name = drm_syncobj_stub_fence_get_name, > - .enable_signaling = drm_syncobj_stub_fence_enable_signaling, > - .release = NULL, > -}; > - > static int drm_syncobj_assign_null_handle(struct drm_syncobj *syncobj) > { > struct drm_syncobj_stub_fence *fence; > @@ -215,6 +424,143 @@ static int drm_syncobj_assign_null_handle(struct drm_syncobj *syncobj) > return 0; > } > > +static struct drm_syncobj_wait_pt * > +drm_syncobj_timeline_lookup_wait_pt_fence(struct drm_syncobj *syncobj, > + u64 point, > + struct dma_fence **fence) > +{ > + struct rb_node *node = syncobj->syncobj_timeline.wait_pt_tree.rb_node; > + struct drm_syncobj_wait_pt *wait_pt = NULL; > + > + > + spin_lock(&syncobj->lock); > + while(node) { > + int result; > + > + wait_pt = rb_entry(node, struct drm_syncobj_wait_pt, node); > + result = point - wait_pt->value; > + if (result < 0) { > + node = node->rb_left; > + } else if (result > 0) { > + node = node->rb_right; > + } else { > + if (fence) > + *fence = dma_fence_get(&wait_pt->base.base); > + break; > + } > + } > + spin_unlock(&syncobj->lock); > + > + return wait_pt; > +} > + > +static struct drm_syncobj_wait_pt * > +drm_syncobj_timeline_create_wait_pt_and_return_fence(struct drm_syncobj *syncobj, > + u64 point, > + struct dma_fence **fence) > +{ > + struct drm_syncobj_wait_pt *wait_pt = kzalloc(sizeof(*wait_pt), > + GFP_KERNEL); > + struct drm_syncobj_stub_fence *base; > + struct rb_node **new = &(syncobj->syncobj_timeline.wait_pt_tree.rb_node), *parent = NULL; > + struct drm_syncobj_signal_pt *tail_signal_pt = > + list_last_entry(&syncobj->syncobj_timeline.signal_pt_list, > + struct drm_syncobj_signal_pt, list); > + > + if (!wait_pt) > + return NULL; > + base = &wait_pt->base; > + spin_lock_init(&base->lock); > + dma_fence_init(&base->base, > + &drm_syncobj_stub_fence_ops, > + &base->lock, > + syncobj->syncobj_timeline.timeline_context, point); > + wait_pt->value = point; > + > + /* wait pt must be in an order, so that we can easily lookup and signal > + * it */ > + spin_lock(&syncobj->lock); > + if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE && > + point <= syncobj->syncobj_timeline.timeline) > + dma_fence_signal(&wait_pt->base.base); > + if ((syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL) && > + (point == tail_signal_pt->value) && > + dma_fence_is_signaled(&tail_signal_pt->base.base)) > + dma_fence_signal(&wait_pt->base.base); > + while(*new) { > + struct drm_syncobj_wait_pt *this = > + rb_entry(*new, struct drm_syncobj_wait_pt, node); > + int result = wait_pt->value - this->value; > + > + parent = *new; > + if (result < 0) > + new = &((*new)->rb_left); > + else if (result > 0) > + new = &((*new)->rb_right); > + else > + goto exist; > + } > + if (fence) > + *fence = dma_fence_get(&wait_pt->base.base); > + rb_link_node(&wait_pt->node, parent, new); > + rb_insert_color(&wait_pt->node, &syncobj->syncobj_timeline.wait_pt_tree); > + spin_unlock(&syncobj->lock); > + return wait_pt; > +exist: > + spin_unlock(&syncobj->lock); > + dma_fence_put(&wait_pt->base.base); > + wait_pt = drm_syncobj_timeline_lookup_wait_pt_fence(syncobj, point, > + fence); > + return wait_pt; > +} > + > +static struct dma_fence * > +drm_syncobj_timeline_point_get(struct drm_syncobj *syncobj, u64 point, u64 flags) > +{ > + struct drm_syncobj_wait_pt *wait_pt; > + struct dma_fence *fence = NULL; > + > + /* already signaled, simply return a signaled stub fence */ > + if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE && > + point <= syncobj->syncobj_timeline.timeline) { > + struct drm_syncobj_stub_fence *fence; > + > + fence = kzalloc(sizeof(*fence), GFP_KERNEL); > + if (fence == NULL) > + return NULL; > + > + spin_lock_init(&fence->lock); > + dma_fence_init(&fence->base, &drm_syncobj_stub_fence_ops, > + &fence->lock, 0, 0); > + dma_fence_signal(&fence->base); > + return &fence->base; > + } > + > + /* check if the wait pt exists */ > + wait_pt = drm_syncobj_timeline_lookup_wait_pt_fence(syncobj, point, > + &fence); > + if (!fence) { > + /* This is a new wait pt, so create it */ > + wait_pt = drm_syncobj_timeline_create_wait_pt_and_return_fence(syncobj, point, > + &fence); > + if (!fence) > + return NULL; > + } > + if (wait_pt) { > + int ret = 0; > + > + if (flags & DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT) { > + ret = wait_event_interruptible_timeout(syncobj->syncobj_timeline.wq, > + wait_pt->value <= syncobj->syncobj_timeline.signal_point, > + msecs_to_jiffies(10000)); /* wait 10s */ > + if (ret <= 0) > + return NULL; > + } > + return fence; > + } > + return NULL; > +} > + > /** > * drm_syncobj_find_fence - lookup and reference the fence in a sync object > * @file_private: drm file private pointer > @@ -229,20 +575,45 @@ static int drm_syncobj_assign_null_handle(struct drm_syncobj *syncobj) > * contains a reference to the fence, which must be released by calling > * dma_fence_put(). > */ > -int drm_syncobj_find_fence(struct drm_file *file_private, > - u32 handle, u64 point, > - struct dma_fence **fence) > +int drm_syncobj_search_fence(struct drm_syncobj *syncobj, u64 point, > + u64 flags, struct dma_fence **fence) > { > - struct drm_syncobj *syncobj = drm_syncobj_find(file_private, handle); > int ret = 0; > > if (!syncobj) > return -ENOENT; > > - *fence = drm_syncobj_fence_get(syncobj); > + if (syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL) { > + /*NORMAL syncobj always wait on last pt */ > + u64 tail_pt_value = syncobj->syncobj_timeline.signal_point; > + > + if (tail_pt_value == 0) > + tail_pt_value += DRM_SYNCOBJ_NORMAL_POINT; > + /* NORMAL syncobj doesn't care point value */ > + WARN_ON(point != 0); > + *fence = drm_syncobj_timeline_point_get(syncobj, tail_pt_value, > + flags); > + } else if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) { > + *fence = drm_syncobj_timeline_point_get(syncobj, point, > + flags); > + } else { > + DRM_ERROR("Don't support this type syncobj\n"); > + *fence = NULL; > + } > if (!*fence) { > ret = -EINVAL; > } > + return ret; > +} > +EXPORT_SYMBOL(drm_syncobj_search_fence); > +int drm_syncobj_find_fence(struct drm_file *file_private, > + u32 handle, u64 point, > + struct dma_fence **fence) { > + struct drm_syncobj *syncobj = drm_syncobj_find(file_private, handle); > + > + int ret = drm_syncobj_search_fence(syncobj, point, > + DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT, > + fence); > drm_syncobj_put(syncobj); > return ret; > } > @@ -259,7 +630,7 @@ void drm_syncobj_free(struct kref *kref) > struct drm_syncobj *syncobj = container_of(kref, > struct drm_syncobj, > refcount); > - drm_syncobj_replace_fence(syncobj, 0, NULL); > + drm_syncobj_timeline_fini(syncobj, &syncobj->syncobj_timeline); > kfree(syncobj); > } > EXPORT_SYMBOL(drm_syncobj_free); > @@ -287,8 +658,12 @@ int drm_syncobj_create(struct drm_syncobj **out_syncobj, uint32_t flags, > return -ENOMEM; > > kref_init(&syncobj->refcount); > - INIT_LIST_HEAD(&syncobj->cb_list); > spin_lock_init(&syncobj->lock); > + if (flags & DRM_SYNCOBJ_CREATE_TYPE_TIMELINE) > + syncobj->type = DRM_SYNCOBJ_TYPE_TIMELINE; > + else > + syncobj->type = DRM_SYNCOBJ_TYPE_NORMAL; > + drm_syncobj_timeline_init(&syncobj->syncobj_timeline); > > if (flags & DRM_SYNCOBJ_CREATE_SIGNALED) { > ret = drm_syncobj_assign_null_handle(syncobj); > @@ -646,7 +1021,6 @@ struct syncobj_wait_entry { > struct task_struct *task; > struct dma_fence *fence; > struct dma_fence_cb fence_cb; > - struct drm_syncobj_cb syncobj_cb; > }; > > static void syncobj_wait_fence_func(struct dma_fence *fence, > @@ -658,18 +1032,6 @@ static void syncobj_wait_fence_func(struct dma_fence *fence, > wake_up_process(wait->task); > } > > -static void syncobj_wait_syncobj_func(struct drm_syncobj *syncobj, > - struct drm_syncobj_cb *cb) > -{ > - struct syncobj_wait_entry *wait = > - container_of(cb, struct syncobj_wait_entry, syncobj_cb); > - > - /* This happens inside the syncobj lock */ > - wait->fence = dma_fence_get(rcu_dereference_protected(syncobj->fence, > - lockdep_is_held(&syncobj->lock))); > - wake_up_process(wait->task); > -} > - > static signed long drm_syncobj_array_wait_timeout(struct drm_syncobj **syncobjs, > uint32_t count, > uint32_t flags, > @@ -693,14 +1055,11 @@ static signed long drm_syncobj_array_wait_timeout(struct drm_syncobj **syncobjs, > signaled_count = 0; > for (i = 0; i < count; ++i) { > entries[i].task = current; > - entries[i].fence = drm_syncobj_fence_get(syncobjs[i]); > + ret = drm_syncobj_search_fence(syncobjs[i], 0, 0, > + &entries[i].fence); > if (!entries[i].fence) { > - if (flags & DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT) { > - continue; > - } else { > - ret = -EINVAL; > - goto cleanup_entries; > - } > + ret = -EINVAL; > + goto cleanup_entries; > } > > if (dma_fence_is_signaled(entries[i].fence)) { > @@ -728,15 +1087,6 @@ static signed long drm_syncobj_array_wait_timeout(struct drm_syncobj **syncobjs, > * fallthough and try a 0 timeout wait! > */ > > - if (flags & DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT) { > - for (i = 0; i < count; ++i) { > - drm_syncobj_fence_get_or_add_callback(syncobjs[i], > - &entries[i].fence, > - &entries[i].syncobj_cb, > - syncobj_wait_syncobj_func); > - } > - } > - > do { > set_current_state(TASK_INTERRUPTIBLE); > > @@ -784,13 +1134,10 @@ static signed long drm_syncobj_array_wait_timeout(struct drm_syncobj **syncobjs, > > cleanup_entries: > for (i = 0; i < count; ++i) { > - if (entries[i].syncobj_cb.func) > - drm_syncobj_remove_callback(syncobjs[i], > - &entries[i].syncobj_cb); > + dma_fence_put(entries[i].fence); > if (entries[i].fence_cb.func) > dma_fence_remove_callback(entries[i].fence, > &entries[i].fence_cb); > - dma_fence_put(entries[i].fence); > } > kfree(entries); > > @@ -965,12 +1312,20 @@ drm_syncobj_reset_ioctl(struct drm_device *dev, void *data, > if (ret < 0) > return ret; > > - for (i = 0; i < args->count_handles; i++) > - drm_syncobj_replace_fence(syncobjs[i], 0, NULL); > - > + for (i = 0; i < args->count_handles; i++) { > + if (syncobjs[i]->type == DRM_SYNCOBJ_TYPE_TIMELINE) { > + DRM_ERROR("timeline syncobj cannot reset!\n"); > + ret = -EINVAL; > + goto out; > + } > + drm_syncobj_timeline_fini(syncobjs[i], > + &syncobjs[i]->syncobj_timeline); > + drm_syncobj_timeline_init(&syncobjs[i]->syncobj_timeline); > + } > +out: > drm_syncobj_array_free(syncobjs, args->count_handles); > > - return 0; > + return ret; > } > > int > diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/i915_gem_execbuffer.c > index 7209dd832d39..bb20d318c9d6 100644 > --- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c > +++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c > @@ -2182,7 +2182,9 @@ await_fence_array(struct i915_execbuffer *eb, > if (!(flags & I915_EXEC_FENCE_WAIT)) > continue; > > - fence = drm_syncobj_fence_get(syncobj); > + drm_syncobj_search_fence(syncobj, 0, > + DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT, > + &fence); > if (!fence) > return -EINVAL; > > diff --git a/include/drm/drm_syncobj.h b/include/drm/drm_syncobj.h > index 425432b85a87..657c97dc25ec 100644 > --- a/include/drm/drm_syncobj.h > +++ b/include/drm/drm_syncobj.h > @@ -30,6 +30,25 @@ > > struct drm_syncobj_cb; > > +enum drm_syncobj_type { > + DRM_SYNCOBJ_TYPE_NORMAL, > + DRM_SYNCOBJ_TYPE_TIMELINE > +}; > + > +struct drm_syncobj_timeline { > + wait_queue_head_t wq; > + u64 timeline_context; > + /** > + * @timeline: syncobj timeline > + */ > + u64 timeline; > + u64 signal_point; > + > + > + struct rb_root wait_pt_tree; > + struct list_head signal_pt_list; > +}; > + > /** > * struct drm_syncobj - sync object. > * > @@ -41,19 +60,16 @@ struct drm_syncobj { > */ > struct kref refcount; > /** > - * @fence: > - * NULL or a pointer to the fence bound to this object. > - * > - * This field should not be used directly. Use drm_syncobj_fence_get() > - * and drm_syncobj_replace_fence() instead. > + * @type: indicate syncobj type > */ > - struct dma_fence __rcu *fence; > + enum drm_syncobj_type type; > /** > - * @cb_list: List of callbacks to call when the &fence gets replaced. > + * @syncobj_timeline: timeline > */ > - struct list_head cb_list; > + struct drm_syncobj_timeline syncobj_timeline; > + > /** > - * @lock: Protects &cb_list and write-locks &fence. > + * @lock: Protects timeline list and write-locks &fence. > */ > spinlock_t lock; > /** > @@ -62,25 +78,6 @@ struct drm_syncobj { > struct file *file; > }; > > -typedef void (*drm_syncobj_func_t)(struct drm_syncobj *syncobj, > - struct drm_syncobj_cb *cb); > - > -/** > - * struct drm_syncobj_cb - callback for drm_syncobj_add_callback > - * @node: used by drm_syncob_add_callback to append this struct to > - * &drm_syncobj.cb_list > - * @func: drm_syncobj_func_t to call > - * > - * This struct will be initialized by drm_syncobj_add_callback, additional > - * data can be passed along by embedding drm_syncobj_cb in another struct. > - * The callback will get called the next time drm_syncobj_replace_fence is > - * called. > - */ > -struct drm_syncobj_cb { > - struct list_head node; > - drm_syncobj_func_t func; > -}; > - > void drm_syncobj_free(struct kref *kref); > > /** > @@ -106,29 +103,6 @@ drm_syncobj_put(struct drm_syncobj *obj) > kref_put(&obj->refcount, drm_syncobj_free); > } > > -/** > - * drm_syncobj_fence_get - get a reference to a fence in a sync object > - * @syncobj: sync object. > - * > - * This acquires additional reference to &drm_syncobj.fence contained in @obj, > - * if not NULL. It is illegal to call this without already holding a reference. > - * No locks required. > - * > - * Returns: > - * Either the fence of @obj or NULL if there's none. > - */ > -static inline struct dma_fence * > -drm_syncobj_fence_get(struct drm_syncobj *syncobj) > -{ > - struct dma_fence *fence; > - > - rcu_read_lock(); > - fence = dma_fence_get_rcu_safe(&syncobj->fence); > - rcu_read_unlock(); > - > - return fence; > -} > - > struct drm_syncobj *drm_syncobj_find(struct drm_file *file_private, > u32 handle); > void drm_syncobj_replace_fence(struct drm_syncobj *syncobj, u64 point, > @@ -142,5 +116,7 @@ int drm_syncobj_create(struct drm_syncobj **out_syncobj, uint32_t flags, > int drm_syncobj_get_handle(struct drm_file *file_private, > struct drm_syncobj *syncobj, u32 *handle); > int drm_syncobj_get_fd(struct drm_syncobj *syncobj, int *p_fd); > +int drm_syncobj_search_fence(struct drm_syncobj *syncobj, u64 point, > + u64 flags, struct dma_fence **fence); > > #endif > diff --git a/include/uapi/drm/drm.h b/include/uapi/drm/drm.h > index 300f336633f2..cebdb2541eb7 100644 > --- a/include/uapi/drm/drm.h > +++ b/include/uapi/drm/drm.h > @@ -717,6 +717,7 @@ struct drm_prime_handle { > struct drm_syncobj_create { > __u32 handle; > #define DRM_SYNCOBJ_CREATE_SIGNALED (1 << 0) > +#define DRM_SYNCOBJ_CREATE_TYPE_TIMELINE (1 << 1) > __u32 flags; > }; >