Hi, On Wed, 2019-02-13 at 09:59 +0100, Hans Verkuil wrote: > On 2/13/19 6:58 AM, Alexandre Courbot wrote: > > On Wed, Feb 13, 2019 at 2:53 PM Alexandre Courbot <acourbot@xxxxxxxxxxxx> wrote: > > > [snip] > > > +Buffers used as reference frames can be queued back to the ``CAPTURE`` queue as > > > +soon as all the frames they are affecting have been queued to the ``OUTPUT`` > > > +queue. The driver will refrain from using the reference buffer as a decoding > > > +target until all the frames depending on it are decoded. > > > > Just want to highlight this part in order to make sure that this is > > indeed what we agreed on. The recent changes to vb2_find_timestamp() > > suggest this, but maybe I misunderstood the intent. It makes the > > kernel responsible for tracking referenced buffers and not using them > > until all the dependent frames are decoded, something the client could > > also do. > > I don't think this is quite right. Once this patch https://patchwork.linuxtv.org/patch/54275/ > is in the vb2 core will track when a buffer can no longer be used as a > reference buffer because the underlying memory might have disappeared. > > The core does not check if it makes sense to use a buffer as a reference > frame, just that it is valid memory. > > So the driver has to check that the timestamp refers to an existing > buffer, but userspace has to check that it queues everything in the > right order and that the reference buffer won't be overwritten > before the last output buffer using that reference buffer has been > decoded. > > So I would say that the second sentence in your paragraph is wrong. > > The first sentence isn't quite right either, but I am not really sure how > to phrase it. It is possible to queue a reference buffer even if > not all output buffers referring to it have been decoded, provided > that by the time the driver starts to use this buffer this actually > has happened. Is there a way we can guarantee this? Looking at the rest of the spec, it says that capture buffers "are returned in decode order" but that doesn't imply that they are picked up in the order they are queued. It seems quite troublesome for the core to check whether each queued capture buffer is used as a reference for one of the queued requests to decide whether to pick it up or not. > But this is an optimization and theoretically it can depend on the > driver behavior. It is always safe to only queue a reference frame > when all frames depending on it have been decoded. So I am leaning > towards not complicating matters and keeping your first sentence > as-is. Yes, I believe it would be much simpler to require userspace to only queue capture buffers once they are no longer needed as references. This also means that the dmabuf fd can't be changed so we are guaranteed that memory will still be there. Cheers, Paul -- Paul Kocialkowski, Bootlin Embedded Linux and kernel engineering https://bootlin.com