On Fri, Feb 15, 2019 at 11:51 AM Christian König <ckoenig.leichtzumerken@xxxxxxxxx> wrote:
Am 15.02.19 um 17:49 schrieb Jason Ekstrand:
On Fri, Feb 15, 2019 at 9:52 AM Lionel Landwerlin via dri-devel <dri-devel@xxxxxxxxxxxxxxxxxxxxx> wrote:
On 15/02/2019 14:32, Koenig, Christian wrote:
> Am 15.02.19 um 15:23 schrieb Lionel Landwerlin:
>> Hi Christian, David,
>>
>> For timeline semaphore we need points to signaled in order.
>> I'm struggling to understand how this fence-chain implementation
>> preserves ordering of the seqnos.
>>
>> One of the scenario I can see an issue happening is when you have a
>> timeline with points 1 & 2 and userspace submits for 2 different
>> engines :
>> - first with let's say a blitter style engine on point 2
>> - then a 3d style engine on point 1
> Yeah, and where exactly is the problem?
>
> Seqno 1 will signal when the 3d style engine finishes work.
>
> And seqno 2 will signal when both seqno 1 is signaled and the blitter
> style engine has finished its work.
That's an interesting interpretation of the spec. I think it's legal and I could see that behavior may be desirable in some ways.
Well we actually had this discussion multiple times now, both internally as well as on the mailing list. Please also see the previous mails with Daniel on this topic.
I dug through dri-devel and read everything I could find with a search for "timeline semaphore" I didn't find all that much but this did come up once.
My initial suggestion was actually to exactly what Leonid suggested as well.
And following this I used a rather simple container for the implementation, e.g. just a ring buffer indexed by the sequence number. In this scenario userspace can specify on syncobj creation time how big the window for sequence numbers should be, e.g. in this implementation how big the ring buffer would be.
This was rejected by our guys who actually wrote a good part of the Vulkan specification. Daniel then has gone into the same direction during the public discussion.
I agree with whoever said that specifying a ringbuffer size is unacceptable. I'm not really sure how that's relevant though. Is a ringbuffer required to implement the behavior that is being suggested here? Genuine question; I'm trying to get back up to speed.
[SNIP]
I think what Christian is suggesting is a valid interpretation of the spec though it is rather unconventional. The Vulkan spec, as it stands today, requires that the application ensure that at the time of signaling, the timeline semaphore value increases. This means that all of the above possible cases are technically illegal in Vulkan and so it doesn't really matter what we do as long as we don't do anyting especially stupid.
And exactly that's the point. When an application does something stupid with its own submissions then this is not much of a problem.
But this interface is meant to be made for communication between processes, and here we want to be sure that nobody can do anything stupid.
My understanding of how this works on Windows is that a wait operation on 3 is a wait until x >= 3 where x is a 64-bit value and a signal operation is simply a write to x. This means that, in the above cases, waits on 1 will be triggered immediately when 2 is written but waits on 2 may or may not happen at all depending on whether the GPU write which overwrites x to 1 or the CPU (or potentially GPU in a different context) read gets there first such that the reader observes 2. If you mess this up and something isn't signaled, that's your fault.
Yeah and I think that this is actually not a good idea at all. Implementing it like this ultimately means that you can only use polling on the number.
Yeah, there are problems with it. I'm just putting it out there for reference and because it's what developers expect regardless of whether that's a good thing or not.
4. If you do get into a sticky situation, you can unblock an entire timeline by using the CPU signal ioctl to set it to a high value.
Well I think that this could be problematic as well. Keep in mind that main use case for this is sharing timelines between processes.
In other words you don't want applications to be able to mess with it to much.
Cross-process is exactly why you want it. Suppose you're a compositor and you have a timeline shared with another application and you've submitted work which waits on it. Then you get a notification somehow (SIGHUP?) that the client has died leaving you hanging. What do you do? You take the semaphore that's shared with you and the client and whack it to UINT64_MAX to unblock yourself. Of course, this can be abused and that's always the risk you take with timelines.
Of all these reasons, I think 1 and 2 carry the most weight. 2, in particular, is interesting if we one day want to implement the same behavior with a simple 64-bit value like Windows does. Immagine, for instance, a scenario where the GPU is doing it's own scheduling or command buffers are submitted ahead of the signal operation being available and told to just sit on the GPU until they see x >= 3. (Yes, there are issues here with residency, contention, etc. I'm asking you to use your immagination.) Assuming you can do 64-bit atomics (there are aparently issues here with PCIe that make things sticky), the behavior I'm suggesting is completely implementable in that way whereas the behavior Christian is suggesting is only implementable if you're maintaining a CPU-side list of fences. I don't think we want to paint ourselves into that corner.
Actually we already had such an implementation with radeon. And I can only say that it was a totally PAIN IN THE A* to maintain.
This is one of the reason why we are not using hardware semaphores any more with amdgpu.
Yeah, there are serious issues with just using a 64-bit integer and, to be honest, I haven't thought (or fought, for that matter) through them all.
--Jason
Regards,
Christian.
--Jason
>
> Regards,
> Christian.
>
>> -Lionel
>>
>> On 07/12/2018 09:55, Chunming Zhou wrote:
>>> From: Christian König <ckoenig.leichtzumerken@xxxxxxxxx>
>>>
>>> Lockless container implementation similar to a dma_fence_array, but with
>>> only two elements per node and automatic garbage collection.
>>>
>>> v2: properly document dma_fence_chain_for_each, add
>>> dma_fence_chain_find_seqno,
>>> drop prev reference during garbage collection if it's not a
>>> chain fence.
>>> v3: use head and iterator for dma_fence_chain_for_each
>>> v4: fix reference count in dma_fence_chain_enable_signaling
>>>
>>> Signed-off-by: Christian König <christian.koenig@xxxxxxx>
>>> ---
>>> drivers/dma-buf/Makefile | 3 +-
>>> drivers/dma-buf/dma-fence-chain.c | 241 ++++++++++++++++++++++++++++++
>>> include/linux/dma-fence-chain.h | 81 ++++++++++
>>> 3 files changed, 324 insertions(+), 1 deletion(-)
>>> create mode 100644 drivers/dma-buf/dma-fence-chain.c
>>> create mode 100644 include/linux/dma-fence-chain.h
>>>
>>> diff --git a/drivers/dma-buf/Makefile b/drivers/dma-buf/Makefile
>>> index 0913a6ccab5a..1f006e083eb9 100644
>>> --- a/drivers/dma-buf/Makefile
>>> +++ b/drivers/dma-buf/Makefile
>>> @@ -1,4 +1,5 @@
>>> -obj-y := dma-buf.o dma-fence.o dma-fence-array.o reservation.o
>>> seqno-fence.o
>>> +obj-y := dma-buf.o dma-fence.o dma-fence-array.o dma-fence-chain.o \
>>> + reservation.o seqno-fence.o
>>> obj-$(CONFIG_SYNC_FILE) += sync_file.o
>>> obj-$(CONFIG_SW_SYNC) += sw_sync.o sync_debug.o
>>> obj-$(CONFIG_UDMABUF) += udmabuf.o
>>> diff --git a/drivers/dma-buf/dma-fence-chain.c
>>> b/drivers/dma-buf/dma-fence-chain.c
>>> new file mode 100644
>>> index 000000000000..0c5e3c902fa0
>>> --- /dev/null
>>> +++ b/drivers/dma-buf/dma-fence-chain.c
>>> @@ -0,0 +1,241 @@
>>> +/*
>>> + * fence-chain: chain fences together in a timeline
>>> + *
>>> + * Copyright (C) 2018 Advanced Micro Devices, Inc.
>>> + * Authors:
>>> + * Christian König <christian.koenig@xxxxxxx>
>>> + *
>>> + * This program is free software; you can redistribute it and/or
>>> modify it
>>> + * under the terms of the GNU General Public License version 2 as
>>> published by
>>> + * the Free Software Foundation.
>>> + *
>>> + * This program is distributed in the hope that it will be useful,
>>> but WITHOUT
>>> + * ANY WARRANTY; without even the implied warranty of
>>> MERCHANTABILITY or
>>> + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public
>>> License for
>>> + * more details.
>>> + */
>>> +
>>> +#include <linux/dma-fence-chain.h>
>>> +
>>> +static bool dma_fence_chain_enable_signaling(struct dma_fence *fence);
>>> +
>>> +/**
>>> + * dma_fence_chain_get_prev - use RCU to get a reference to the
>>> previous fence
>>> + * @chain: chain node to get the previous node from
>>> + *
>>> + * Use dma_fence_get_rcu_safe to get a reference to the previous
>>> fence of the
>>> + * chain node.
>>> + */
>>> +static struct dma_fence *dma_fence_chain_get_prev(struct
>>> dma_fence_chain *chain)
>>> +{
>>> + struct dma_fence *prev;
>>> +
>>> + rcu_read_lock();
>>> + prev = dma_fence_get_rcu_safe(&chain->prev);
>>> + rcu_read_unlock();
>>> + return prev;
>>> +}
>>> +
>>> +/**
>>> + * dma_fence_chain_walk - chain walking function
>>> + * @fence: current chain node
>>> + *
>>> + * Walk the chain to the next node. Returns the next fence or NULL
>>> if we are at
>>> + * the end of the chain. Garbage collects chain nodes which are already
>>> + * signaled.
>>> + */
>>> +struct dma_fence *dma_fence_chain_walk(struct dma_fence *fence)
>>> +{
>>> + struct dma_fence_chain *chain, *prev_chain;
>>> + struct dma_fence *prev, *replacement, *tmp;
>>> +
>>> + chain = to_dma_fence_chain(fence);
>>> + if (!chain) {
>>> + dma_fence_put(fence);
>>> + return NULL;
>>> + }
>>> +
>>> + while ((prev = dma_fence_chain_get_prev(chain))) {
>>> +
>>> + prev_chain = to_dma_fence_chain(prev);
>>> + if (prev_chain) {
>>> + if (!dma_fence_is_signaled(prev_chain->fence))
>>> + break;
>>> +
>>> + replacement = dma_fence_chain_get_prev(prev_chain);
>>> + } else {
>>> + if (!dma_fence_is_signaled(prev))
>>> + break;
>>> +
>>> + replacement = NULL;
>>> + }
>>> +
>>> + tmp = cmpxchg(&chain->prev, prev, replacement);
>>> + if (tmp == prev)
>>> + dma_fence_put(tmp);
>>> + else
>>> + dma_fence_put(replacement);
>>> + dma_fence_put(prev);
>>> + }
>>> +
>>> + dma_fence_put(fence);
>>> + return prev;
>>> +}
>>> +EXPORT_SYMBOL(dma_fence_chain_walk);
>>> +
>>> +/**
>>> + * dma_fence_chain_find_seqno - find fence chain node by seqno
>>> + * @pfence: pointer to the chain node where to start
>>> + * @seqno: the sequence number to search for
>>> + *
>>> + * Advance the fence pointer to the chain node which will signal
>>> this sequence
>>> + * number. If no sequence number is provided then this is a no-op.
>>> + *
>>> + * Returns EINVAL if the fence is not a chain node or the sequence
>>> number has
>>> + * not yet advanced far enough.
>>> + */
>>> +int dma_fence_chain_find_seqno(struct dma_fence **pfence, uint64_t
>>> seqno)
>>> +{
>>> + struct dma_fence_chain *chain;
>>> +
>>> + if (!seqno)
>>> + return 0;
>>> +
>>> + chain = to_dma_fence_chain(*pfence);
>>> + if (!chain || chain->base.seqno < seqno)
>>> + return -EINVAL;
>>> +
>>> + dma_fence_chain_for_each(*pfence, &chain->base) {
>>> + if ((*pfence)->context != chain->base.context ||
>>> + to_dma_fence_chain(*pfence)->prev_seqno < seqno)
>>> + break;
>>> + }
>>> + dma_fence_put(&chain->base);
>>> +
>>> + return 0;
>>> +}
>>> +EXPORT_SYMBOL(dma_fence_chain_find_seqno);
>>> +
>>> +static const char *dma_fence_chain_get_driver_name(struct dma_fence
>>> *fence)
>>> +{
>>> + return "dma_fence_chain";
>>> +}
>>> +
>>> +static const char *dma_fence_chain_get_timeline_name(struct
>>> dma_fence *fence)
>>> +{
>>> + return "unbound";
>>> +}
>>> +
>>> +static void dma_fence_chain_irq_work(struct irq_work *work)
>>> +{
>>> + struct dma_fence_chain *chain;
>>> +
>>> + chain = container_of(work, typeof(*chain), work);
>>> +
>>> + /* Try to rearm the callback */
>>> + if (!dma_fence_chain_enable_signaling(&chain->base))
>>> + /* Ok, we are done. No more unsignaled fences left */
>>> + dma_fence_signal(&chain->base);
>>> + dma_fence_put(&chain->base);
>>> +}
>>> +
>>> +static void dma_fence_chain_cb(struct dma_fence *f, struct
>>> dma_fence_cb *cb)
>>> +{
>>> + struct dma_fence_chain *chain;
>>> +
>>> + chain = container_of(cb, typeof(*chain), cb);
>>> + irq_work_queue(&chain->work);
>>> + dma_fence_put(f);
>>> +}
>>> +
>>> +static bool dma_fence_chain_enable_signaling(struct dma_fence *fence)
>>> +{
>>> + struct dma_fence_chain *head = to_dma_fence_chain(fence);
>>> +
>>> + dma_fence_get(&head->base);
>>> + dma_fence_chain_for_each(fence, &head->base) {
>>> + struct dma_fence_chain *chain = to_dma_fence_chain(fence);
>>> + struct dma_fence *f = chain ? chain->fence : fence;
>>> +
>>> + dma_fence_get(f);
>>> + if (!dma_fence_add_callback(f, &head->cb,
>>> dma_fence_chain_cb)) {
>>> + dma_fence_put(fence);
>>> + return true;
>>> + }
>>> + dma_fence_put(f);
>>> + }
>>> + dma_fence_put(&head->base);
>>> + return false;
>>> +}
>>> +
>>> +static bool dma_fence_chain_signaled(struct dma_fence *fence)
>>> +{
>>> + dma_fence_chain_for_each(fence, fence) {
>>> + struct dma_fence_chain *chain = to_dma_fence_chain(fence);
>>> + struct dma_fence *f = chain ? chain->fence : fence;
>>> +
>>> + if (!dma_fence_is_signaled(f)) {
>>> + dma_fence_put(fence);
>>> + return false;
>>> + }
>>> + }
>>> +
>>> + return true;
>>> +}
>>> +
>>> +static void dma_fence_chain_release(struct dma_fence *fence)
>>> +{
>>> + struct dma_fence_chain *chain = to_dma_fence_chain(fence);
>>> +
>>> + dma_fence_put(chain->prev);
>>> + dma_fence_put(chain->fence);
>>> + dma_fence_free(fence);
>>> +}
>>> +
>>> +const struct dma_fence_ops dma_fence_chain_ops = {
>>> + .get_driver_name = dma_fence_chain_get_driver_name,
>>> + .get_timeline_name = dma_fence_chain_get_timeline_name,
>>> + .enable_signaling = dma_fence_chain_enable_signaling,
>>> + .signaled = dma_fence_chain_signaled,
>>> + .release = dma_fence_chain_release,
>>> +};
>>> +EXPORT_SYMBOL(dma_fence_chain_ops);
>>> +
>>> +/**
>>> + * dma_fence_chain_init - initialize a fence chain
>>> + * @chain: the chain node to initialize
>>> + * @prev: the previous fence
>>> + * @fence: the current fence
>>> + *
>>> + * Initialize a new chain node and either start a new chain or add
>>> the node to
>>> + * the existing chain of the previous fence.
>>> + */
>>> +void dma_fence_chain_init(struct dma_fence_chain *chain,
>>> + struct dma_fence *prev,
>>> + struct dma_fence *fence,
>>> + uint64_t seqno)
>>> +{
>>> + struct dma_fence_chain *prev_chain = to_dma_fence_chain(prev);
>>> + uint64_t context;
>>> +
>>> + spin_lock_init(&chain->lock);
>>> + chain->prev = prev;
>>> + chain->fence = fence;
>>> + chain->prev_seqno = 0;
>>> + init_irq_work(&chain->work, dma_fence_chain_irq_work);
>>> +
>>> + /* Try to reuse the context of the previous chain node. */
>>> + if (prev_chain && __dma_fence_is_later(seqno, prev->seqno)) {
>>> + context = prev->context;
>>> + chain->prev_seqno = prev->seqno;
>>> + } else {
>>> + context = dma_fence_context_alloc(1);
>>> + /* Make sure that we always have a valid sequence number. */
>>> + if (prev_chain)
>>> + seqno = max(prev->seqno, seqno);
>>> + }
>>> +
>>> + dma_fence_init(&chain->base, &dma_fence_chain_ops,
>>> + &chain->lock, context, seqno);
>>> +}
>>> +EXPORT_SYMBOL(dma_fence_chain_init);
>>> diff --git a/include/linux/dma-fence-chain.h
>>> b/include/linux/dma-fence-chain.h
>>> new file mode 100644
>>> index 000000000000..a5c2e8c6915c
>>> --- /dev/null
>>> +++ b/include/linux/dma-fence-chain.h
>>> @@ -0,0 +1,81 @@
>>> +/*
>>> + * fence-chain: chain fences together in a timeline
>>> + *
>>> + * Copyright (C) 2018 Advanced Micro Devices, Inc.
>>> + * Authors:
>>> + * Christian König <christian.koenig@xxxxxxx>
>>> + *
>>> + * This program is free software; you can redistribute it and/or
>>> modify it
>>> + * under the terms of the GNU General Public License version 2 as
>>> published by
>>> + * the Free Software Foundation.
>>> + *
>>> + * This program is distributed in the hope that it will be useful,
>>> but WITHOUT
>>> + * ANY WARRANTY; without even the implied warranty of
>>> MERCHANTABILITY or
>>> + * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public
>>> License for
>>> + * more details.
>>> + */
>>> +
>>> +#ifndef __LINUX_DMA_FENCE_CHAIN_H
>>> +#define __LINUX_DMA_FENCE_CHAIN_H
>>> +
>>> +#include <linux/dma-fence.h>
>>> +#include <linux/irq_work.h>
>>> +
>>> +/**
>>> + * struct dma_fence_chain - fence to represent an node of a fence chain
>>> + * @base: fence base class
>>> + * @lock: spinlock for fence handling
>>> + * @prev: previous fence of the chain
>>> + * @prev_seqno: original previous seqno before garbage collection
>>> + * @fence: encapsulated fence
>>> + * @cb: callback structure for signaling
>>> + * @work: irq work item for signaling
>>> + */
>>> +struct dma_fence_chain {
>>> + struct dma_fence base;
>>> + spinlock_t lock;
>>> + struct dma_fence *prev;
>>> + u64 prev_seqno;
>>> + struct dma_fence *fence;
>>> + struct dma_fence_cb cb;
>>> + struct irq_work work;
>>> +};
>>> +
>>> +extern const struct dma_fence_ops dma_fence_chain_ops;
>>> +
>>> +/**
>>> + * to_dma_fence_chain - cast a fence to a dma_fence_chain
>>> + * @fence: fence to cast to a dma_fence_array
>>> + *
>>> + * Returns NULL if the fence is not a dma_fence_chain,
>>> + * or the dma_fence_chain otherwise.
>>> + */
>>> +static inline struct dma_fence_chain *
>>> +to_dma_fence_chain(struct dma_fence *fence)
>>> +{
>>> + if (!fence || fence->ops != &dma_fence_chain_ops)
>>> + return NULL;
>>> +
>>> + return container_of(fence, struct dma_fence_chain, base);
>>> +}
>>> +
>>> +/**
>>> + * dma_fence_chain_for_each - iterate over all fences in chain
>>> + * @iter: current fence
>>> + * @head: starting point
>>> + *
>>> + * Iterate over all fences in the chain. We keep a reference to the
>>> current
>>> + * fence while inside the loop which must be dropped when breaking out.
>>> + */
>>> +#define dma_fence_chain_for_each(iter, head) \
>>> + for (iter = dma_fence_get(head); iter; \
>>> + iter = dma_fence_chain_walk(head))
>>> +
>>> +struct dma_fence *dma_fence_chain_walk(struct dma_fence *fence);
>>> +int dma_fence_chain_find_seqno(struct dma_fence **pfence, uint64_t
>>> seqno);
>>> +void dma_fence_chain_init(struct dma_fence_chain *chain,
>>> + struct dma_fence *prev,
>>> + struct dma_fence *fence,
>>> + uint64_t seqno);
>>> +
>>> +#endif /* __LINUX_DMA_FENCE_CHAIN_H */
>>
_______________________________________________
dri-devel mailing list
dri-devel@xxxxxxxxxxxxxxxxxxxxx
https://lists.freedesktop.org/mailman/listinfo/dri-devel
_______________________________________________ amd-gfx mailing list amd-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/amd-gfx