[PATCH 5/5] [RFC]drm: add syncobj timeline support v3

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On 2018å¹´08æ??29æ?¥ 19:42, Christian König wrote:
> Am 29.08.2018 um 12:44 schrieb Chunming Zhou:
>> VK_KHR_timeline_semaphore:
>> This extension introduces a new type of semaphore that has an integer 
>> payload
>> identifying a point in a timeline. Such timeline semaphores support the
>> following operations:
>>     * CPU query - A host operation that allows querying the payload 
>> of the
>>       timeline semaphore.
>>     * CPU wait - A host operation that allows a blocking wait for a
>>       timeline semaphore to reach a specified value.
>>     * Device wait - A device operation that allows waiting for a
>>       timeline semaphore to reach a specified value.
>>     * Device signal - A device operation that allows advancing the
>>       timeline semaphore to a specified value.
>>
>> Since it's a timeline, that means the front time point(PT) always is 
>> signaled before the late PT.
>> a. signal PT design:
>> Signal PT fence N depends on PT[N-1] fence and signal opertion fence, 
>> when PT[N] fence is signaled,
>> the timeline will increase to value of PT[N].
>> b. wait PT design:
>> Wait PT fence is signaled by reaching timeline point value, when 
>> timeline is increasing, will compare
>> wait PTs value with new timeline value, if PT value is lower than 
>> timeline value, then wait PT will be
>> signaled, otherwise keep in list. semaphore wait operation can wait 
>> on any point of timeline,
>> so need a RB tree to order them. And wait PT could ahead of signal 
>> PT, we need a sumission fence to
>> perform that.
>>
>> v2:
>> 1. remove unused DRM_SYNCOBJ_CREATE_TYPE_NORMAL. (Christian)
>> 2. move unexposed denitions to .c file. (Daniel Vetter)
>> 3. split up the change to drm_syncobj_find_fence() in a separate 
>> patch. (Christian)
>> 4. split up the change to drm_syncobj_replace_fence() in a separate 
>> patch.
>> 5. drop the submission_fence implementation and instead use 
>> wait_event() for that. (Christian)
>> 6. WARN_ON(point != 0) for NORMAL type syncobj case. (Daniel Vetter)
>>
>> v3:
>> 1. replace normal syncobj with timeline implemenation. (Vetter and 
>> Christian)
>>      a. normal syncobj signal op will create a signal PT to tail of 
>> signal pt list.
>>      b. normal syncobj wait op will create a wait pt with last signal 
>> point, and this wait PT is only signaled by related signal point PT.
>> 2. some bug fix and clean up
>> 3. tested by ./deqp-vk -n dEQP-VK*semaphore* for normal syncobj
>>
>> TODO:
>> 1. CPU query and wait on timeline semaphore.
>>
>> Signed-off-by: Chunming Zhou <david1.zhou at amd.com>
>> Cc: Christian Konig <christian.koenig at amd.com>
>> Cc: Dave Airlie <airlied at redhat.com>
>> Cc: Daniel Rakos <Daniel.Rakos at amd.com>
>> Cc: Daniel Vetter <daniel at ffwll.ch>
>> ---
>>   drivers/gpu/drm/drm_syncobj.c              | 593 ++++++++++++++++-----
>>   drivers/gpu/drm/i915/i915_gem_execbuffer.c |   4 +-
>>   include/drm/drm_syncobj.h                  |  78 +--
>>   include/uapi/drm/drm.h                     |   1 +
>>   4 files changed, 505 insertions(+), 171 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/drm_syncobj.c 
>> b/drivers/gpu/drm/drm_syncobj.c
>> index ab43559398d0..f701d9ef1b81 100644
>> --- a/drivers/gpu/drm/drm_syncobj.c
>> +++ b/drivers/gpu/drm/drm_syncobj.c
>> @@ -56,6 +56,50 @@
>>   #include "drm_internal.h"
>>   #include <drm/drm_syncobj.h>
>>   +/* merge normal syncobj to timeline syncobj, the point interval is 
>> 1 */
>> +#define DRM_SYNCOBJ_NORMAL_POINT 1
>> +
>> +struct drm_syncobj_stub_fence {
>> +    struct dma_fence base;
>> +    spinlock_t lock;
>> +};
>> +
>> +static const char *drm_syncobj_stub_fence_get_name(struct dma_fence 
>> *fence)
>> +{
>> +        return "syncobjstub";
>> +}
>> +
>> +static bool drm_syncobj_stub_fence_enable_signaling(struct dma_fence 
>> *fence)
>> +{
>> +    return !dma_fence_is_signaled(fence);
>> +}
>> +static void drm_syncobj_stub_fence_release(struct dma_fence *f)
>> +{
>> +    kfree(f);
>> +}
>> +static const struct dma_fence_ops drm_syncobj_stub_fence_ops = {
>> +    .get_driver_name = drm_syncobj_stub_fence_get_name,
>> +    .get_timeline_name = drm_syncobj_stub_fence_get_name,
>> +    .enable_signaling = drm_syncobj_stub_fence_enable_signaling,
>> +    .release = drm_syncobj_stub_fence_release,
>> +};
>
> Do we really need to move that around? Could reduce the size of the 
> patch quite a bit if we don't.

stub fence is used widely in both normal and timeline syncobj, if you 
think which increase patch size, I can do a separate patch for that.

>
>> +
>> +struct drm_syncobj_wait_pt {
>> +    struct drm_syncobj_stub_fence base;
>> +    u64    value;
>> +    struct rb_node   node;
>> +};
>> +struct drm_syncobj_signal_pt {
>> +    struct drm_syncobj_stub_fence base;
>> +    struct dma_fence *signal_fence;
>> +    struct dma_fence *pre_pt_base;
>> +    struct dma_fence_cb signal_cb;
>> +    struct dma_fence_cb pre_pt_cb;
>> +    struct drm_syncobj *syncobj;
>> +    u64    value;
>> +    struct list_head list;
>> +};
>
> What are those two structures good for
They are used to record wait ops points and signal ops points.
For timeline, they are connected by timeline value, works like:
     a. signal pt increase timeline value to signal_pt value when 
signal_pt is signaled. signal_pt is depending on previous pt fence and 
itself signal fence from signal ops.
     b. wait pt tree checks if timeline value over itself value.

For normal, works like:
     a. signal pt increase 1 for syncobj_timeline->signal_point every 
time when signal ops is performed.
     b. when wait ops is comming, wait pt will fetch above last signal 
pt value as its wait point. wait pt will be only signaled by equal point 
value signal_pt.


> and why is the stub fence their base?
Good question, I tried to kzalloc for them as well when I debug them, I 
encountered a problem:
I lookup/find wait_pt or signal_pt successfully, but when I tried to use 
them, they are freed sometimes, and results in NULL point.
and generally, when lookup them, we often need their stub fence as well, 
and in the meantime,  their lifecycle are same.
Above reason, I make stub fence as their base.

>
>> +
>>   /**
>>    * drm_syncobj_find - lookup and reference a sync object.
>>    * @file_private: drm file private pointer
>> @@ -82,59 +126,247 @@ struct drm_syncobj *drm_syncobj_find(struct 
>> drm_file *file_private,
>>   }
>>   EXPORT_SYMBOL(drm_syncobj_find);
>>   -static void drm_syncobj_add_callback_locked(struct drm_syncobj 
>> *syncobj,
>> -                        struct drm_syncobj_cb *cb,
>> -                        drm_syncobj_func_t func)
>> +static void drm_syncobj_timeline_init(struct drm_syncobj_timeline
>> +                      *syncobj_timeline)
>>   {
>> -    cb->func = func;
>> -    list_add_tail(&cb->node, &syncobj->cb_list);
>> +    syncobj_timeline->timeline_context = dma_fence_context_alloc(1);
>> +    syncobj_timeline->timeline = 0;
>> +    syncobj_timeline->signal_point = 0;
>> +    init_waitqueue_head(&syncobj_timeline->wq);
>> +
>> +    syncobj_timeline->wait_pt_tree = RB_ROOT;
>> +    INIT_LIST_HEAD(&syncobj_timeline->signal_pt_list);
>>   }
>>   -static int drm_syncobj_fence_get_or_add_callback(struct 
>> drm_syncobj *syncobj,
>> -                         struct dma_fence **fence,
>> -                         struct drm_syncobj_cb *cb,
>> -                         drm_syncobj_func_t func)
>> +static void drm_syncobj_timeline_fini(struct drm_syncobj *syncobj,
>> +                      struct drm_syncobj_timeline *syncobj_timeline)
>>   {
>> -    int ret;
>> +    struct rb_node *node = NULL;
>> +    struct drm_syncobj_wait_pt *wait_pt = NULL;
>> +    struct drm_syncobj_signal_pt *signal_pt = NULL, *tmp;
>> +
>> +    spin_lock(&syncobj->lock);
>> +    for(node = rb_first(&syncobj_timeline->wait_pt_tree);
>> +        node != NULL; ) {
>
> Better use rbtree_postorder_for_each_entry_safe() for this.
 From the comments, seems we cannot use it here, we need erase node here.
Comments:
  * Note, however, that it cannot handle other modifications that 
re-order the
  * rbtree it is iterating over. This includes calling rb_erase() on 
@pos, as
  * rb_erase() may rebalance the tree, causing us to miss some nodes.
  */

Thanks,
David Zhou
>
> Regards,
> Christian.
>
>> +        wait_pt = rb_entry(node, struct drm_syncobj_wait_pt, node);
>> +        node = rb_next(node);
>> +        rb_erase(&wait_pt->node,
>> +             &syncobj_timeline->wait_pt_tree);
>> +        RB_CLEAR_NODE(&wait_pt->node);
>> +        spin_unlock(&syncobj->lock);
>> +        dma_fence_wait(&wait_pt->base.base, true);
>> +        spin_lock(&syncobj->lock);
>> +        /* kfree(wait_pt) is excuted by fence put */
>> +        dma_fence_put(&wait_pt->base.base);
>> +    }
>> +    list_for_each_entry_safe(signal_pt, tmp,
>> +                 &syncobj_timeline->signal_pt_list, list) {
>> +        list_del(&signal_pt->list);
>> +        if (signal_pt->signal_fence) {
>> + dma_fence_remove_callback(signal_pt->signal_fence,
>> +                          &signal_pt->signal_cb);
>> +            dma_fence_put(signal_pt->signal_fence);
>> +        }
>> +        if (signal_pt->pre_pt_base) {
>> + dma_fence_remove_callback(signal_pt->pre_pt_base,
>> +                          &signal_pt->pre_pt_cb);
>> +            dma_fence_put(signal_pt->pre_pt_base);
>> +        }
>> +        dma_fence_put(&signal_pt->base.base);
>> +    }
>> +    spin_unlock(&syncobj->lock);
>> +}
>> +
>>   -    *fence = drm_syncobj_fence_get(syncobj);
>> -    if (*fence)
>> -        return 1;
>> +static bool drm_syncobj_normal_signal_wait_pts(struct drm_syncobj 
>> *syncobj,
>> +                           u64 signal_pt)
>> +{
>> +    struct rb_node *node = NULL;
>> +    struct drm_syncobj_wait_pt *wait_pt = NULL;
>>         spin_lock(&syncobj->lock);
>> -    /* We've already tried once to get a fence and failed.  Now that we
>> -     * have the lock, try one more time just to be sure we don't add a
>> -     * callback when a fence has already been set.
>> -     */
>> -    if (syncobj->fence) {
>> -        *fence = 
>> dma_fence_get(rcu_dereference_protected(syncobj->fence,
>> - lockdep_is_held(&syncobj->lock)));
>> -        ret = 1;
>> -    } else {
>> -        *fence = NULL;
>> -        drm_syncobj_add_callback_locked(syncobj, cb, func);
>> -        ret = 0;
>> +    for(node = rb_first(&syncobj->syncobj_timeline.wait_pt_tree);
>> +        node != NULL; ) {
>> +        wait_pt = rb_entry(node, struct drm_syncobj_wait_pt, node);
>> +        node = rb_next(node);
>> +        /* for normal syncobj */
>> +        if (wait_pt->value == signal_pt) {
>> +            dma_fence_signal(&wait_pt->base.base);
>> +            rb_erase(&wait_pt->node,
>> + &syncobj->syncobj_timeline.wait_pt_tree);
>> +            RB_CLEAR_NODE(&wait_pt->node);
>> +            /* kfree(wait_pt) is excuted by fence put */
>> +            dma_fence_put(&wait_pt->base.base);
>> +            spin_unlock(&syncobj->lock);
>> +            return true;
>> +        }
>>       }
>>       spin_unlock(&syncobj->lock);
>> +    return false;
>> +}
>>   -    return ret;
>> +static void drm_syncobj_timeline_signal_wait_pts(struct drm_syncobj 
>> *syncobj)
>> +{
>> +    struct rb_node *node = NULL;
>> +    struct drm_syncobj_wait_pt *wait_pt = NULL;
>> +
>> +    spin_lock(&syncobj->lock);
>> +    for(node = rb_first(&syncobj->syncobj_timeline.wait_pt_tree);
>> +        node != NULL; ) {
>> +        wait_pt = rb_entry(node, struct drm_syncobj_wait_pt, node);
>> +        node = rb_next(node);
>> +        if (wait_pt->value <= syncobj->syncobj_timeline.timeline) {
>> +            dma_fence_signal(&wait_pt->base.base);
>> +            rb_erase(&wait_pt->node,
>> + &syncobj->syncobj_timeline.wait_pt_tree);
>> +            RB_CLEAR_NODE(&wait_pt->node);
>> +            /* kfree(wait_pt) is excuted by fence put */
>> +            dma_fence_put(&wait_pt->base.base);
>> +        } else {
>> +            /* the loop is from left to right, the later entry value is
>> +             * bigger, so don't need to check any more */
>> +            break;
>> +        }
>> +    }
>> +    spin_unlock(&syncobj->lock);
>>   }
>>   -void drm_syncobj_add_callback(struct drm_syncobj *syncobj,
>> -                  struct drm_syncobj_cb *cb,
>> -                  drm_syncobj_func_t func)
>> +
>> +
>> +static void pt_fence_cb(struct drm_syncobj_signal_pt *signal_pt)
>>   {
>> +    struct dma_fence *fence = NULL;
>> +    struct drm_syncobj *syncobj;
>> +    struct drm_syncobj_signal_pt *tail_pt;
>> +    u64 pt_value = signal_pt->value;
>> +
>> +    dma_fence_signal(&signal_pt->base.base);
>> +    fence = signal_pt->signal_fence;
>> +    signal_pt->signal_fence = NULL;
>> +    dma_fence_put(fence);
>> +    fence = signal_pt->pre_pt_base;
>> +    signal_pt->pre_pt_base = NULL;
>> +    dma_fence_put(fence);
>> +
>> +    syncobj = signal_pt->syncobj;
>>       spin_lock(&syncobj->lock);
>> -    drm_syncobj_add_callback_locked(syncobj, cb, func);
>> +    syncobj->syncobj_timeline.timeline = pt_value;
>> +    tail_pt = 
>> list_last_entry(&syncobj->syncobj_timeline.signal_pt_list,
>> +                  struct drm_syncobj_signal_pt, list);
>> +    if ((syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL && signal_pt != 
>> tail_pt)
>> +        || syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) {
>> +        list_del(&signal_pt->list);
>> +        /* kfree(signal_pt) will be  executed by below fence put */
>> +        dma_fence_put(&signal_pt->base.base);
>> +    }
>>       spin_unlock(&syncobj->lock);
>> +    if (syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL)
>> +        drm_syncobj_normal_signal_wait_pts(syncobj, pt_value);
>> +    else
>> +        drm_syncobj_timeline_signal_wait_pts(syncobj);
>>   }
>> +static void pt_signal_fence_func(struct dma_fence *fence,
>> +                 struct dma_fence_cb *cb)
>> +{
>> +    struct drm_syncobj_signal_pt *signal_pt =
>> +        container_of(cb, struct drm_syncobj_signal_pt, signal_cb);
>>   -void drm_syncobj_remove_callback(struct drm_syncobj *syncobj,
>> -                 struct drm_syncobj_cb *cb)
>> +    if (signal_pt->pre_pt_base &&
>> +        !dma_fence_is_signaled(signal_pt->pre_pt_base))
>> +        return;
>> +
>> +    pt_fence_cb(signal_pt);
>> +}
>> +static void pt_pre_fence_func(struct dma_fence *fence,
>> +                 struct dma_fence_cb *cb)
>> +{
>> +    struct drm_syncobj_signal_pt *signal_pt =
>> +        container_of(cb, struct drm_syncobj_signal_pt, pre_pt_cb);
>> +
>> +    if (signal_pt->signal_fence &&
>> +        !dma_fence_is_signaled(signal_pt->pre_pt_base))
>> +        return;
>> +
>> +    pt_fence_cb(signal_pt);
>> +}
>> +
>> +static int drm_syncobj_timeline_create_signal_pt(struct drm_syncobj 
>> *syncobj,
>> +                         struct dma_fence *fence,
>> +                         u64 point)
>>   {
>> +    struct drm_syncobj_signal_pt *signal_pt =
>> +        kzalloc(sizeof(struct drm_syncobj_signal_pt), GFP_KERNEL);
>> +    struct drm_syncobj_stub_fence *base;
>> +    struct drm_syncobj_signal_pt *tail_pt;
>> +    struct dma_fence *tail_pt_fence = NULL;
>> +    int ret = 0;
>> +
>> +    if (!signal_pt)
>> +        return -ENOMEM;
>> +    if (syncobj->syncobj_timeline.signal_point >= point) {
>> +        DRM_WARN("A later signal is ready!");
>> +        goto out;
>> +    }
>> +    if (!fence)
>> +        goto out;
>> +    dma_fence_get(fence);
>>       spin_lock(&syncobj->lock);
>> -    list_del_init(&cb->node);
>> +    base = &signal_pt->base;
>> +    spin_lock_init(&base->lock);
>> +    dma_fence_init(&base->base,
>> +               &drm_syncobj_stub_fence_ops,
>> +               &base->lock,
>> +               syncobj->syncobj_timeline.timeline_context, point);
>> +    signal_pt->signal_fence = fence;
>> +    /* timeline syncobj must take this dependency */
>> +    if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) {
>> +        if (!list_empty(&syncobj->syncobj_timeline.signal_pt_list)) {
>> +            tail_pt = 
>> list_last_entry(&syncobj->syncobj_timeline.signal_pt_list,
>> +                          struct drm_syncobj_signal_pt, list);
>> +            tail_pt_fence = &tail_pt->base.base;
>> +            if (dma_fence_is_signaled(tail_pt_fence))
>> +                tail_pt_fence = NULL;
>> +            else
>> +                signal_pt->pre_pt_base =
>> +                    dma_fence_get(tail_pt_fence);
>> +        }
>> +    }
>> +
>> +    signal_pt->value = point;
>> +    syncobj->syncobj_timeline.signal_point = point;
>> +    signal_pt->syncobj = syncobj;
>> +    INIT_LIST_HEAD(&signal_pt->list);
>> +    list_add_tail(&signal_pt->list, 
>> &syncobj->syncobj_timeline.signal_pt_list);
>>       spin_unlock(&syncobj->lock);
>> +    wake_up_all(&syncobj->syncobj_timeline.wq);
>> +    /**
>> +     * Every pt is depending on signal fence and previous pt fence, add
>> +     * callbacks to them
>> +     */
>> +    ret = dma_fence_add_callback(signal_pt->signal_fence,
>> +                     &signal_pt->signal_cb,
>> +                     pt_signal_fence_func);
>> +
>> +    if (signal_pt->pre_pt_base) {
>> +        ret = dma_fence_add_callback(signal_pt->pre_pt_base,
>> +                         &signal_pt->pre_pt_cb,
>> +                         pt_pre_fence_func);
>> +        if (ret == -ENOENT)
>> +            pt_pre_fence_func(signal_pt->pre_pt_base,
>> +                      &signal_pt->pre_pt_cb);
>> +        else if (ret)
>> +            goto out;
>> +    } else if (ret == -ENOENT) {
>> +        pt_signal_fence_func(signal_pt->signal_fence,
>> +                     &signal_pt->signal_cb);
>> +    } else if (ret) {
>> +            goto out;
>> +    }
>> +
>> +    return 0;
>> +out:
>> +    dma_fence_put(&signal_pt->base.base);
>> +    return ret;
>>   }
>>     /**
>> @@ -149,53 +381,30 @@ void drm_syncobj_replace_fence(struct 
>> drm_syncobj *syncobj,
>>                      u64 point,
>>                      struct dma_fence *fence)
>>   {
>> -    struct dma_fence *old_fence;
>> -    struct drm_syncobj_cb *cur, *tmp;
>> -
>> -    if (fence)
>> -        dma_fence_get(fence);
>> -
>> -    spin_lock(&syncobj->lock);
>> -
>> -    old_fence = rcu_dereference_protected(syncobj->fence,
>> - lockdep_is_held(&syncobj->lock));
>> -    rcu_assign_pointer(syncobj->fence, fence);
>> -
>> -    if (fence != old_fence) {
>> -        list_for_each_entry_safe(cur, tmp, &syncobj->cb_list, node) {
>> -            list_del_init(&cur->node);
>> -            cur->func(syncobj, cur);
>> +    if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) {
>> +        if (fence)
>> +            drm_syncobj_timeline_create_signal_pt(syncobj, fence,
>> +                                  point);
>> +        return;
>> +    } else if (syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL) {
>> +        u64 pt_value;
>> +
>> +        if (!fence) {
>> +            drm_syncobj_timeline_fini(syncobj,
>> +                          &syncobj->syncobj_timeline);
>> + drm_syncobj_timeline_init(&syncobj->syncobj_timeline);
>> +            return;
>>           }
>> +        pt_value = syncobj->syncobj_timeline.signal_point +
>> +            DRM_SYNCOBJ_NORMAL_POINT;
>> +        drm_syncobj_timeline_create_signal_pt(syncobj, fence, 
>> pt_value);
>> +        return;
>> +    } else {
>> +        DRM_ERROR("the syncobj type isn't support\n");
>>       }
>> -
>> -    spin_unlock(&syncobj->lock);
>> -
>> -    dma_fence_put(old_fence);
>>   }
>>   EXPORT_SYMBOL(drm_syncobj_replace_fence);
>>   -struct drm_syncobj_stub_fence {
>> -    struct dma_fence base;
>> -    spinlock_t lock;
>> -};
>> -
>> -static const char *drm_syncobj_stub_fence_get_name(struct dma_fence 
>> *fence)
>> -{
>> -        return "syncobjstub";
>> -}
>> -
>> -static bool drm_syncobj_stub_fence_enable_signaling(struct dma_fence 
>> *fence)
>> -{
>> -    return !dma_fence_is_signaled(fence);
>> -}
>> -
>> -static const struct dma_fence_ops drm_syncobj_stub_fence_ops = {
>> -    .get_driver_name = drm_syncobj_stub_fence_get_name,
>> -    .get_timeline_name = drm_syncobj_stub_fence_get_name,
>> -    .enable_signaling = drm_syncobj_stub_fence_enable_signaling,
>> -    .release = NULL,
>> -};
>> -
>>   static int drm_syncobj_assign_null_handle(struct drm_syncobj *syncobj)
>>   {
>>       struct drm_syncobj_stub_fence *fence;
>> @@ -215,6 +424,143 @@ static int 
>> drm_syncobj_assign_null_handle(struct drm_syncobj *syncobj)
>>       return 0;
>>   }
>>   +static struct drm_syncobj_wait_pt *
>> +drm_syncobj_timeline_lookup_wait_pt_fence(struct drm_syncobj *syncobj,
>> +                      u64 point,
>> +                      struct dma_fence **fence)
>> +{
>> +    struct rb_node *node = 
>> syncobj->syncobj_timeline.wait_pt_tree.rb_node;
>> +    struct drm_syncobj_wait_pt *wait_pt = NULL;
>> +
>> +
>> +    spin_lock(&syncobj->lock);
>> +    while(node) {
>> +        int result;
>> +
>> +        wait_pt = rb_entry(node, struct drm_syncobj_wait_pt, node);
>> +        result = point - wait_pt->value;
>> +        if (result < 0) {
>> +            node = node->rb_left;
>> +        } else if (result > 0) {
>> +            node = node->rb_right;
>> +        } else {
>> +            if (fence)
>> +            *fence = dma_fence_get(&wait_pt->base.base);
>> +            break;
>> +        }
>> +    }
>> +    spin_unlock(&syncobj->lock);
>> +
>> +    return wait_pt;
>> +}
>> +
>> +static struct drm_syncobj_wait_pt *
>> +drm_syncobj_timeline_create_wait_pt_and_return_fence(struct 
>> drm_syncobj *syncobj,
>> +                             u64 point,
>> +                             struct dma_fence **fence)
>> +{
>> +    struct drm_syncobj_wait_pt *wait_pt = kzalloc(sizeof(*wait_pt),
>> +                              GFP_KERNEL);
>> +    struct drm_syncobj_stub_fence *base;
>> +    struct rb_node **new = 
>> &(syncobj->syncobj_timeline.wait_pt_tree.rb_node), *parent = NULL;
>> +    struct drm_syncobj_signal_pt *tail_signal_pt =
>> + list_last_entry(&syncobj->syncobj_timeline.signal_pt_list,
>> +                struct drm_syncobj_signal_pt, list);
>> +
>> +    if (!wait_pt)
>> +        return NULL;
>> +    base = &wait_pt->base;
>> +    spin_lock_init(&base->lock);
>> +    dma_fence_init(&base->base,
>> +               &drm_syncobj_stub_fence_ops,
>> +               &base->lock,
>> +               syncobj->syncobj_timeline.timeline_context, point);
>> +    wait_pt->value = point;
>> +
>> +    /* wait pt must be in an order, so that we can easily lookup and 
>> signal
>> +     * it */
>> +    spin_lock(&syncobj->lock);
>> +    if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE &&
>> +        point <= syncobj->syncobj_timeline.timeline)
>> +        dma_fence_signal(&wait_pt->base.base);
>> +    if ((syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL) &&
>> +        (point == tail_signal_pt->value) &&
>> + dma_fence_is_signaled(&tail_signal_pt->base.base))
>> +        dma_fence_signal(&wait_pt->base.base);
>> +    while(*new) {
>> +        struct drm_syncobj_wait_pt *this =
>> +            rb_entry(*new, struct drm_syncobj_wait_pt, node);
>> +        int result = wait_pt->value - this->value;
>> +
>> +        parent = *new;
>> +        if (result < 0)
>> +            new = &((*new)->rb_left);
>> +        else if (result > 0)
>> +            new = &((*new)->rb_right);
>> +        else
>> +            goto exist;
>> +    }
>> +    if (fence)
>> +        *fence = dma_fence_get(&wait_pt->base.base);
>> +    rb_link_node(&wait_pt->node, parent, new);
>> +    rb_insert_color(&wait_pt->node, 
>> &syncobj->syncobj_timeline.wait_pt_tree);
>> +    spin_unlock(&syncobj->lock);
>> +    return wait_pt;
>> +exist:
>> +    spin_unlock(&syncobj->lock);
>> +    dma_fence_put(&wait_pt->base.base);
>> +    wait_pt = drm_syncobj_timeline_lookup_wait_pt_fence(syncobj, point,
>> +                                fence);
>> +    return wait_pt;
>> +}
>> +
>> +static struct dma_fence *
>> +drm_syncobj_timeline_point_get(struct drm_syncobj *syncobj, u64 
>> point, u64 flags)
>> +{
>> +    struct drm_syncobj_wait_pt *wait_pt;
>> +    struct dma_fence *fence = NULL;
>> +
>> +    /* already signaled, simply return a signaled stub fence */
>> +    if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE &&
>> +        point <= syncobj->syncobj_timeline.timeline) {
>> +        struct drm_syncobj_stub_fence *fence;
>> +
>> +        fence = kzalloc(sizeof(*fence), GFP_KERNEL);
>> +        if (fence == NULL)
>> +            return NULL;
>> +
>> +        spin_lock_init(&fence->lock);
>> +        dma_fence_init(&fence->base, &drm_syncobj_stub_fence_ops,
>> +                   &fence->lock, 0, 0);
>> +        dma_fence_signal(&fence->base);
>> +        return &fence->base;
>> +    }
>> +
>> +    /* check if the wait pt exists */
>> +    wait_pt = drm_syncobj_timeline_lookup_wait_pt_fence(syncobj, point,
>> +                                &fence);
>> +    if (!fence) {
>> +        /* This is a new wait pt, so create it */
>> +        wait_pt = 
>> drm_syncobj_timeline_create_wait_pt_and_return_fence(syncobj, point,
>> +                                  &fence);
>> +        if (!fence)
>> +            return NULL;
>> +    }
>> +    if (wait_pt) {
>> +        int ret = 0;
>> +
>> +        if (flags & DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT) {
>> +            ret = 
>> wait_event_interruptible_timeout(syncobj->syncobj_timeline.wq,
>> +                                   wait_pt->value <= 
>> syncobj->syncobj_timeline.signal_point,
>> +                                   msecs_to_jiffies(10000)); /* wait 
>> 10s */
>> +            if (ret <= 0)
>> +                return NULL;
>> +        }
>> +        return fence;
>> +    }
>> +    return NULL;
>> +}
>> +
>>   /**
>>    * drm_syncobj_find_fence - lookup and reference the fence in a 
>> sync object
>>    * @file_private: drm file private pointer
>> @@ -229,20 +575,45 @@ static int 
>> drm_syncobj_assign_null_handle(struct drm_syncobj *syncobj)
>>    * contains a reference to the fence, which must be released by 
>> calling
>>    * dma_fence_put().
>>    */
>> -int drm_syncobj_find_fence(struct drm_file *file_private,
>> -               u32 handle, u64 point,
>> -               struct dma_fence **fence)
>> +int drm_syncobj_search_fence(struct drm_syncobj *syncobj, u64 point,
>> +                 u64 flags, struct dma_fence **fence)
>>   {
>> -    struct drm_syncobj *syncobj = drm_syncobj_find(file_private, 
>> handle);
>>       int ret = 0;
>>         if (!syncobj)
>>           return -ENOENT;
>>   -    *fence = drm_syncobj_fence_get(syncobj);
>> +    if (syncobj->type == DRM_SYNCOBJ_TYPE_NORMAL) {
>> +        /*NORMAL syncobj always wait on last pt */
>> +        u64 tail_pt_value = syncobj->syncobj_timeline.signal_point;
>> +
>> +        if (tail_pt_value == 0)
>> +            tail_pt_value += DRM_SYNCOBJ_NORMAL_POINT;
>> +        /* NORMAL syncobj doesn't care point value */
>> +        WARN_ON(point != 0);
>> +        *fence = drm_syncobj_timeline_point_get(syncobj, tail_pt_value,
>> +                            flags);
>> +    } else if (syncobj->type == DRM_SYNCOBJ_TYPE_TIMELINE) {
>> +        *fence = drm_syncobj_timeline_point_get(syncobj, point,
>> +                            flags);
>> +    } else {
>> +        DRM_ERROR("Don't support this type syncobj\n");
>> +        *fence = NULL;
>> +    }
>>       if (!*fence) {
>>           ret = -EINVAL;
>>       }
>> +    return ret;
>> +}
>> +EXPORT_SYMBOL(drm_syncobj_search_fence);
>> +int drm_syncobj_find_fence(struct drm_file *file_private,
>> +               u32 handle, u64 point,
>> +               struct dma_fence **fence) {
>> +    struct drm_syncobj *syncobj = drm_syncobj_find(file_private, 
>> handle);
>> +
>> +    int ret = drm_syncobj_search_fence(syncobj, point,
>> +                    DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT,
>> +                    fence);
>>       drm_syncobj_put(syncobj);
>>       return ret;
>>   }
>> @@ -259,7 +630,7 @@ void drm_syncobj_free(struct kref *kref)
>>       struct drm_syncobj *syncobj = container_of(kref,
>>                              struct drm_syncobj,
>>                              refcount);
>> -    drm_syncobj_replace_fence(syncobj, 0, NULL);
>> +    drm_syncobj_timeline_fini(syncobj, &syncobj->syncobj_timeline);
>>       kfree(syncobj);
>>   }
>>   EXPORT_SYMBOL(drm_syncobj_free);
>> @@ -287,8 +658,12 @@ int drm_syncobj_create(struct drm_syncobj 
>> **out_syncobj, uint32_t flags,
>>           return -ENOMEM;
>>         kref_init(&syncobj->refcount);
>> -    INIT_LIST_HEAD(&syncobj->cb_list);
>>       spin_lock_init(&syncobj->lock);
>> +    if (flags & DRM_SYNCOBJ_CREATE_TYPE_TIMELINE)
>> +        syncobj->type = DRM_SYNCOBJ_TYPE_TIMELINE;
>> +    else
>> +        syncobj->type = DRM_SYNCOBJ_TYPE_NORMAL;
>> + drm_syncobj_timeline_init(&syncobj->syncobj_timeline);
>>         if (flags & DRM_SYNCOBJ_CREATE_SIGNALED) {
>>           ret = drm_syncobj_assign_null_handle(syncobj);
>> @@ -646,7 +1021,6 @@ struct syncobj_wait_entry {
>>       struct task_struct *task;
>>       struct dma_fence *fence;
>>       struct dma_fence_cb fence_cb;
>> -    struct drm_syncobj_cb syncobj_cb;
>>   };
>>     static void syncobj_wait_fence_func(struct dma_fence *fence,
>> @@ -658,18 +1032,6 @@ static void syncobj_wait_fence_func(struct 
>> dma_fence *fence,
>>       wake_up_process(wait->task);
>>   }
>>   -static void syncobj_wait_syncobj_func(struct drm_syncobj *syncobj,
>> -                      struct drm_syncobj_cb *cb)
>> -{
>> -    struct syncobj_wait_entry *wait =
>> -        container_of(cb, struct syncobj_wait_entry, syncobj_cb);
>> -
>> -    /* This happens inside the syncobj lock */
>> -    wait->fence = 
>> dma_fence_get(rcu_dereference_protected(syncobj->fence,
>> - lockdep_is_held(&syncobj->lock)));
>> -    wake_up_process(wait->task);
>> -}
>> -
>>   static signed long drm_syncobj_array_wait_timeout(struct 
>> drm_syncobj **syncobjs,
>>                             uint32_t count,
>>                             uint32_t flags,
>> @@ -693,14 +1055,11 @@ static signed long 
>> drm_syncobj_array_wait_timeout(struct drm_syncobj **syncobjs,
>>       signaled_count = 0;
>>       for (i = 0; i < count; ++i) {
>>           entries[i].task = current;
>> -        entries[i].fence = drm_syncobj_fence_get(syncobjs[i]);
>> +        ret = drm_syncobj_search_fence(syncobjs[i], 0, 0,
>> +                           &entries[i].fence);
>>           if (!entries[i].fence) {
>> -            if (flags & DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT) {
>> -                continue;
>> -            } else {
>> -                ret = -EINVAL;
>> -                goto cleanup_entries;
>> -            }
>> +            ret = -EINVAL;
>> +            goto cleanup_entries;
>>           }
>>             if (dma_fence_is_signaled(entries[i].fence)) {
>> @@ -728,15 +1087,6 @@ static signed long 
>> drm_syncobj_array_wait_timeout(struct drm_syncobj **syncobjs,
>>        * fallthough and try a 0 timeout wait!
>>        */
>>   -    if (flags & DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT) {
>> -        for (i = 0; i < count; ++i) {
>> -            drm_syncobj_fence_get_or_add_callback(syncobjs[i],
>> -                                  &entries[i].fence,
>> -                                  &entries[i].syncobj_cb,
>> -                                  syncobj_wait_syncobj_func);
>> -        }
>> -    }
>> -
>>       do {
>>           set_current_state(TASK_INTERRUPTIBLE);
>>   @@ -784,13 +1134,10 @@ static signed long 
>> drm_syncobj_array_wait_timeout(struct drm_syncobj **syncobjs,
>>     cleanup_entries:
>>       for (i = 0; i < count; ++i) {
>> -        if (entries[i].syncobj_cb.func)
>> -            drm_syncobj_remove_callback(syncobjs[i],
>> -                            &entries[i].syncobj_cb);
>> +        dma_fence_put(entries[i].fence);
>>           if (entries[i].fence_cb.func)
>>               dma_fence_remove_callback(entries[i].fence,
>>                             &entries[i].fence_cb);
>> -        dma_fence_put(entries[i].fence);
>>       }
>>       kfree(entries);
>>   @@ -965,12 +1312,20 @@ drm_syncobj_reset_ioctl(struct drm_device 
>> *dev, void *data,
>>       if (ret < 0)
>>           return ret;
>>   -    for (i = 0; i < args->count_handles; i++)
>> -        drm_syncobj_replace_fence(syncobjs[i], 0, NULL);
>> -
>> +    for (i = 0; i < args->count_handles; i++) {
>> +        if (syncobjs[i]->type == DRM_SYNCOBJ_TYPE_TIMELINE) {
>> +            DRM_ERROR("timeline syncobj cannot reset!\n");
>> +            ret = -EINVAL;
>> +            goto out;
>> +        }
>> +        drm_syncobj_timeline_fini(syncobjs[i],
>> +                      &syncobjs[i]->syncobj_timeline);
>> + drm_syncobj_timeline_init(&syncobjs[i]->syncobj_timeline);
>> +    }
>> +out:
>>       drm_syncobj_array_free(syncobjs, args->count_handles);
>>   -    return 0;
>> +    return ret;
>>   }
>>     int
>> diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c 
>> b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
>> index 7209dd832d39..bb20d318c9d6 100644
>> --- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c
>> +++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c
>> @@ -2182,7 +2182,9 @@ await_fence_array(struct i915_execbuffer *eb,
>>           if (!(flags & I915_EXEC_FENCE_WAIT))
>>               continue;
>>   -        fence = drm_syncobj_fence_get(syncobj);
>> +        drm_syncobj_search_fence(syncobj, 0,
>> +                     DRM_SYNCOBJ_WAIT_FLAGS_WAIT_FOR_SUBMIT,
>> +                     &fence);
>>           if (!fence)
>>               return -EINVAL;
>>   diff --git a/include/drm/drm_syncobj.h b/include/drm/drm_syncobj.h
>> index 425432b85a87..657c97dc25ec 100644
>> --- a/include/drm/drm_syncobj.h
>> +++ b/include/drm/drm_syncobj.h
>> @@ -30,6 +30,25 @@
>>     struct drm_syncobj_cb;
>>   +enum drm_syncobj_type {
>> +    DRM_SYNCOBJ_TYPE_NORMAL,
>> +    DRM_SYNCOBJ_TYPE_TIMELINE
>> +};
>> +
>> +struct drm_syncobj_timeline {
>> +    wait_queue_head_t    wq;
>> +    u64 timeline_context;
>> +    /**
>> +     * @timeline: syncobj timeline
>> +     */
>> +    u64 timeline;
>> +    u64 signal_point;
>> +
>> +
>> +    struct rb_root wait_pt_tree;
>> +    struct list_head signal_pt_list;
>> +};
>> +
>>   /**
>>    * struct drm_syncobj - sync object.
>>    *
>> @@ -41,19 +60,16 @@ struct drm_syncobj {
>>        */
>>       struct kref refcount;
>>       /**
>> -     * @fence:
>> -     * NULL or a pointer to the fence bound to this object.
>> -     *
>> -     * This field should not be used directly. Use 
>> drm_syncobj_fence_get()
>> -     * and drm_syncobj_replace_fence() instead.
>> +     * @type: indicate syncobj type
>>        */
>> -    struct dma_fence __rcu *fence;
>> +    enum drm_syncobj_type type;
>>       /**
>> -     * @cb_list: List of callbacks to call when the &fence gets 
>> replaced.
>> +     * @syncobj_timeline: timeline
>>        */
>> -    struct list_head cb_list;
>> +    struct drm_syncobj_timeline syncobj_timeline;
>> +
>>       /**
>> -     * @lock: Protects &cb_list and write-locks &fence.
>> +     * @lock: Protects timeline list and write-locks &fence.
>>        */
>>       spinlock_t lock;
>>       /**
>> @@ -62,25 +78,6 @@ struct drm_syncobj {
>>       struct file *file;
>>   };
>>   -typedef void (*drm_syncobj_func_t)(struct drm_syncobj *syncobj,
>> -                   struct drm_syncobj_cb *cb);
>> -
>> -/**
>> - * struct drm_syncobj_cb - callback for drm_syncobj_add_callback
>> - * @node: used by drm_syncob_add_callback to append this struct to
>> - *      &drm_syncobj.cb_list
>> - * @func: drm_syncobj_func_t to call
>> - *
>> - * This struct will be initialized by drm_syncobj_add_callback, 
>> additional
>> - * data can be passed along by embedding drm_syncobj_cb in another 
>> struct.
>> - * The callback will get called the next time 
>> drm_syncobj_replace_fence is
>> - * called.
>> - */
>> -struct drm_syncobj_cb {
>> -    struct list_head node;
>> -    drm_syncobj_func_t func;
>> -};
>> -
>>   void drm_syncobj_free(struct kref *kref);
>>     /**
>> @@ -106,29 +103,6 @@ drm_syncobj_put(struct drm_syncobj *obj)
>>       kref_put(&obj->refcount, drm_syncobj_free);
>>   }
>>   -/**
>> - * drm_syncobj_fence_get - get a reference to a fence in a sync object
>> - * @syncobj: sync object.
>> - *
>> - * This acquires additional reference to &drm_syncobj.fence 
>> contained in @obj,
>> - * if not NULL. It is illegal to call this without already holding a 
>> reference.
>> - * No locks required.
>> - *
>> - * Returns:
>> - * Either the fence of @obj or NULL if there's none.
>> - */
>> -static inline struct dma_fence *
>> -drm_syncobj_fence_get(struct drm_syncobj *syncobj)
>> -{
>> -    struct dma_fence *fence;
>> -
>> -    rcu_read_lock();
>> -    fence = dma_fence_get_rcu_safe(&syncobj->fence);
>> -    rcu_read_unlock();
>> -
>> -    return fence;
>> -}
>> -
>>   struct drm_syncobj *drm_syncobj_find(struct drm_file *file_private,
>>                        u32 handle);
>>   void drm_syncobj_replace_fence(struct drm_syncobj *syncobj, u64 point,
>> @@ -142,5 +116,7 @@ int drm_syncobj_create(struct drm_syncobj 
>> **out_syncobj, uint32_t flags,
>>   int drm_syncobj_get_handle(struct drm_file *file_private,
>>                  struct drm_syncobj *syncobj, u32 *handle);
>>   int drm_syncobj_get_fd(struct drm_syncobj *syncobj, int *p_fd);
>> +int drm_syncobj_search_fence(struct drm_syncobj *syncobj, u64 point,
>> +                 u64 flags, struct dma_fence **fence);
>>     #endif
>> diff --git a/include/uapi/drm/drm.h b/include/uapi/drm/drm.h
>> index 300f336633f2..cebdb2541eb7 100644
>> --- a/include/uapi/drm/drm.h
>> +++ b/include/uapi/drm/drm.h
>> @@ -717,6 +717,7 @@ struct drm_prime_handle {
>>   struct drm_syncobj_create {
>>       __u32 handle;
>>   #define DRM_SYNCOBJ_CREATE_SIGNALED (1 << 0)
>> +#define DRM_SYNCOBJ_CREATE_TYPE_TIMELINE (1 << 1)
>>       __u32 flags;
>>   };
>



[Index of Archives]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux