Re: Proposed updates and guidelines for MPEG-2, H.264 and H.265 stateless support

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

Le samedi 18 mai 2019 à 12:04 +0200, Jernej Škrabec a écrit :
> Dne sobota, 18. maj 2019 ob 11:50:37 CEST je Paul Kocialkowski napisal(a):
> > Hi,
> > 
> > On Fri, 2019-05-17 at 16:43 -0400, Nicolas Dufresne wrote:
> > > Le jeudi 16 mai 2019 à 20:45 +0200, Paul Kocialkowski a écrit :
> > > > Hi,
> > > > 
> > > > Le jeudi 16 mai 2019 à 14:24 -0400, Nicolas Dufresne a écrit :
> > > > > Le mercredi 15 mai 2019 à 22:59 +0200, Paul Kocialkowski a écrit :
> > > > > > Hi,
> > > > > > 
> > > > > > Le mercredi 15 mai 2019 à 14:54 -0400, Nicolas Dufresne a écrit :
> > > > > > > Le mercredi 15 mai 2019 à 19:42 +0200, Paul Kocialkowski a écrit :
> > > > > > > > Hi,
> > > > > > > > 
> > > > > > > > Le mercredi 15 mai 2019 à 10:42 -0400, Nicolas Dufresne a écrit 
> :
> > > > > > > > > Le mercredi 15 mai 2019 à 12:09 +0200, Paul Kocialkowski a 
> écrit :
> > > > > > > > > > Hi,
> > > > > > > > > > 
> > > > > > > > > > With the Rockchip stateless VPU driver in the works, we now
> > > > > > > > > > have a
> > > > > > > > > > better idea of what the situation is like on platforms other
> > > > > > > > > > than
> > > > > > > > > > Allwinner. This email shares my conclusions about the
> > > > > > > > > > situation and how
> > > > > > > > > > we should update the MPEG-2, H.264 and H.265 controls
> > > > > > > > > > accordingly.
> > > > > > > > > > 
> > > > > > > > > > - Per-slice decoding
> > > > > > > > > > 
> > > > > > > > > > We've discussed this one already[0] and Hans has submitted a
> > > > > > > > > > patch[1]
> > > > > > > > > > to implement the required core bits. When we agree it looks
> > > > > > > > > > good, we
> > > > > > > > > > should lift the restriction that all slices must be
> > > > > > > > > > concatenated and
> > > > > > > > > > have them submitted as individual requests.
> > > > > > > > > > 
> > > > > > > > > > One question is what to do about other controls. I feel like
> > > > > > > > > > it would
> > > > > > > > > > make sense to always pass all the required controls for
> > > > > > > > > > decoding the
> > > > > > > > > > slice, including the ones that don't change across slices.
> > > > > > > > > > But there
> > > > > > > > > > may be no particular advantage to this and only downsides.
> > > > > > > > > > Not doing it
> > > > > > > > > > and relying on the "control cache" can work, but we need to
> > > > > > > > > > specify
> > > > > > > > > > that only a single stream can be decoded per opened instance
> > > > > > > > > > of the
> > > > > > > > > > v4l2 device. This is the assumption we're going with for
> > > > > > > > > > handling
> > > > > > > > > > multi-slice anyway, so it shouldn't be an issue.
> > > > > > > > > 
> > > > > > > > > My opinion on this is that the m2m instance is a state, and
> > > > > > > > > the driver
> > > > > > > > > should be responsible of doing time-division multiplexing
> > > > > > > > > across
> > > > > > > > > multiple m2m instance jobs. Doing the time-division
> > > > > > > > > multiplexing in
> > > > > > > > > userspace would require some sort of daemon to work properly
> > > > > > > > > across
> > > > > > > > > processes. I also think the kernel is better place for doing
> > > > > > > > > resource
> > > > > > > > > access scheduling in general.
> > > > > > > > 
> > > > > > > > I agree with that yes. We always have a single m2m context and
> > > > > > > > specific
> > > > > > > > controls per opened device so keeping cached values works out
> > > > > > > > well.
> > > > > > > > 
> > > > > > > > So maybe we shall explicitly require that the request with the
> > > > > > > > first
> > > > > > > > slice for a frame also contains the per-frame controls.
> > > > > > > > 
> > > > > > > > > > - Annex-B formats
> > > > > > > > > > 
> > > > > > > > > > I don't think we have really reached a conclusion on the
> > > > > > > > > > pixel formats
> > > > > > > > > > we want to expose. The main issue is how to deal with codecs
> > > > > > > > > > that need
> > > > > > > > > > the full slice NALU with start code, where the slice_header
> > > > > > > > > > is
> > > > > > > > > > duplicated in raw bitstream, when others are fine with just
> > > > > > > > > > the encoded
> > > > > > > > > > slice data and the parsed slice header control.
> > > > > > > > > > 
> > > > > > > > > > My initial thinking was that we'd need 3 formats:
> > > > > > > > > > - One that only takes only the slice compressed data
> > > > > > > > > > (without raw slice
> > > > > > > > > > header and start code);
> > > > > > > > > > - One that takes both the NALU data (including start code,
> > > > > > > > > > raw header
> > > > > > > > > > and compressed data) and slice header controls;
> > > > > > > > > > - One that takes the NALU data but no slice header.
> > > > > > > > > > 
> > > > > > > > > > But I no longer think the latter really makes sense in the
> > > > > > > > > > context of
> > > > > > > > > > stateless video decoding.
> > > > > > > > > > 
> > > > > > > > > > A side-note: I think we should definitely have data offsets
> > > > > > > > > > in every
> > > > > > > > > > case, so that implementations can just push the whole NALU
> > > > > > > > > > regardless
> > > > > > > > > > of the format if they're lazy.
> > > > > > > > > 
> > > > > > > > > I realize that I didn't share our latest research on the
> > > > > > > > > subject. So a
> > > > > > > > > slice in the original bitstream is formed of the following
> > > > > > > > > blocks
> > > > > > > > > 
> > > > > > > > > (simplified):
> > > > > > > > >   [nal_header][nal_type][slice_header][slice]
> > > > > > > > 
> > > > > > > > Thanks for the details!
> > > > > > > > 
> > > > > > > > > nal_header:
> > > > > > > > > This one is a header used to locate the start and the end of
> > > > > > > > > the of a
> > > > > > > > > NAL. There is two standard forms, the ANNEX B / start code, a
> > > > > > > > > sequence
> > > > > > > > > of 3 bytes 0x00 0x00 0x01, you'll often see 4 bytes, the first
> > > > > > > > > byte
> > > > > > > > > would be a leading 0 from the previous NAL padding, but this
> > > > > > > > > is also
> > > > > > > > > totally valid start code. The second form is the AVC form,
> > > > > > > > > notably used
> > > > > > > > > in ISOMP4 container. It simply is the size of the NAL. You
> > > > > > > > > must keep
> > > > > > > > > your buffer aligned to NALs in this case as you cannot scan
> > > > > > > > > from random
> > > > > > > > > location.
> > > > > > > > > 
> > > > > > > > > nal_type:
> > > > > > > > > It's a bit more then just the type, but it contains at least
> > > > > > > > > the
> > > > > > > > > information of the nal type. This has different size on H.264
> > > > > > > > > and HEVC
> > > > > > > > > but I know it's size is in bytes.
> > > > > > > > > 
> > > > > > > > > slice_header:
> > > > > > > > > This contains per slice parameters, like the modification
> > > > > > > > > lists to
> > > > > > > > > apply on the references. This one has a size in bits, not in
> > > > > > > > > bytes.
> > > > > > > > > 
> > > > > > > > > slice:
> > > > > > > > > I don't really know what is in it exactly, but this is the
> > > > > > > > > data used to
> > > > > > > > > decode. This bit has a special coding called the
> > > > > > > > > anti-emulation, which
> > > > > > > > > prevents a start-code from appearing in it. This coding is
> > > > > > > > > present in
> > > > > > > > > both forms, ANNEX-B or AVC (in GStreamer and some reference
> > > > > > > > > manual they
> > > > > > > > > call ANNEX-B the bytestream format).
> > > > > > > > > 
> > > > > > > > > So, what we notice is that what is currently passed through
> > > > > > > > > Cedrus
> > > > > > > > > 
> > > > > > > > > driver:
> > > > > > > > >   [nal_type][slice_header][slice]
> > > > > > > > > 
> > > > > > > > > This matches what is being passed through VA-API. We can
> > > > > > > > > understand
> > > > > > > > > that stripping off the slice_header would be hard, since it's
> > > > > > > > > size is
> > > > > > > > > in bits. Instead we pass size and header_bit_size in
> > > > > > > > > slice_params.
> > > > > > > > 
> > > > > > > > True, there is that.
> > > > > > > > 
> > > > > > > > > About Rockchip. RK3288 is a Hantro G1 and has a bit called
> > > > > > > > > start_code_e, when you turn this off, you don't need start
> > > > > > > > > code. As a
> > > > > > > > > side effect, the bitstream becomes identical. We do now know
> > > > > > > > > that it
> > > > > > > > > works with the ffmpeg branch implement for cedrus.
> > > > > > > > 
> > > > > > > > Oh great, that makes life easier in the short term, but I guess
> > > > > > > > the
> > > > > > > > issue could arise on another decoder sooner or later.
> > > > > > > > 
> > > > > > > > > Now what's special about Hantro G1 (also found on IMX8M) is
> > > > > > > > > that it
> > > > > > > > > take care for us of reading and executing the modification
> > > > > > > > > lists found
> > > > > > > > > in the slice header. Mostly because I very disliked having to
> > > > > > > > > pass the
> > > > > > > > > p/b0/b1 parameters, is that Boris implemented in the driver
> > > > > > > > > the
> > > > > > > > > transformation from the DPB entries into this p/b0/b1 list.
> > > > > > > > > These list
> > > > > > > > > a standard, it's basically implementing 8.2.4.1 and 8.2.4.2.
> > > > > > > > > the
> > > > > > > > > following section is the execution of the modification list.
> > > > > > > > > As this
> > > > > > > > > list is not modified, it only need to be calculated per frame.
> > > > > > > > > As a
> > > > > > > > > result, we don't need these new lists, and we can work with
> > > > > > > > > the same
> > > > > > > > > H264_SLICE format as Cedrus is using.
> > > > > > > > 
> > > > > > > > Yes but I definitely think it makes more sense to pass the list
> > > > > > > > modifications rather than reconstructing those in the driver
> > > > > > > > from a
> > > > > > > > full list. IMO controls should stick to the bitstream as close
> > > > > > > > as
> > > > > > > > possible.
> > > > > > > 
> > > > > > > For Hantro and RKVDEC, the list of modification is parsed by the
> > > > > > > IP
> > > > > > > from the slice header bits. Just to make sure, because I myself
> > > > > > > was
> > > > > > > confused on this before, the slice header does not contain a list
> > > > > > > of
> > > > > > > references, instead it contains a list modification to be applied
> > > > > > > to
> > > > > > > the reference list. I need to check again, but to execute these
> > > > > > > modification, you need to filter and sort the references in a
> > > > > > > specific
> > > > > > > order. This should be what is defined in the spec as 8.2.4.1 and
> > > > > > > 8.2.4.2. Then 8.2.4.3 is the process that creates the l0/l1.
> > > > > > > 
> > > > > > > The list of references is deduced from the DPB. The DPB, which I
> > > > > > > thinks
> > > > > > > should be rename as "references", seems more useful then p/b0/b1,
> > > > > > > since
> > > > > > > this is the data that gives use the ability to implementing glue
> > > > > > > in the
> > > > > > > driver to compensate some HW differences.
> > > > > > > 
> > > > > > > In the case of Hantro / RKVDEC, we think it's natural to build the
> > > > > > > HW
> > > > > > > specific lists (p/b0/b1) from the references rather then adding HW
> > > > > > > specific list in the decode_params structure. The fact these lists
> > > > > > > are
> > > > > > > standard intermediate step of the standard is not that important.
> > > > > > 
> > > > > > Sorry I got confused (once more) about it. Boris just explained the
> > > > > > same thing to me over IRC :) Anyway my point is that we want to pass
> > > > > > what's in ffmpeg's short and long term ref lists, and name them that
> > > > > > instead of dpb.
> > > > > > 
> > > > > > > > > Now, this is just a start. For RK3399, we have a different
> > > > > > > > > CODEC
> > > > > > > > > design. This one does not have the start_code_e bit. What the
> > > > > > > > > IP does,
> > > > > > > > > is that you give it one or more slice per buffer, setup the
> > > > > > > > > params,
> > > > > > > > > start decoding, but the decoder then return the location of
> > > > > > > > > the
> > > > > > > > > following NAL. So basically you could offload the scanning of
> > > > > > > > > start
> > > > > > > > > code to the HW. That being said, with the driver layer in
> > > > > > > > > between, that
> > > > > > > > > would be amazingly inconvenient to use, and with Boyer-more
> > > > > > > > > algorithm,
> > > > > > > > > it is pretty cheap to scan this type of start-code on CPU. But
> > > > > > > > > the
> > > > > > > > > feature that this allows is to operate in frame mode. In this
> > > > > > > > > mode, you
> > > > > > > > > have 1 interrupt per frame.
> > > > > > > > 
> > > > > > > > I'm not sure there is any interest in exposing that from
> > > > > > > > userspace and
> > > > > > > > my current feeling is that we should just ditch support for
> > > > > > > > per-frame
> > > > > > > > decoding altogether. I think it mixes decoding with notions that
> > > > > > > > are
> > > > > > > > higher-level than decoding, but I agree it's a blurry line.
> > > > > > > 
> > > > > > > I'm not worried about this either. We can already support that by
> > > > > > > copying the bitstream internally to the driver, though zero-copy
> > > > > > > with
> > > > > > > this would require a new format, the one we talked about,
> > > > > > > SLICE_ANNEX_B.
> > > > > > 
> > > > > > Right, but what I'm thinking about is making that the one and only
> > > > > > format. The rationale is that it's always easier to just append a
> > > > > > start
> > > > > > code from userspace if needed. And we need a bit offset to the slice
> > > > > > data part anyway, so it doesn't hurt to require a few extra bits to
> > > > > > have the whole thing that will work in every situation.
> > > > > 
> > > > > What I'd like is to eventually allow zero-copy (aka userptr) into the
> > > > > driver. If you make the start code mandatory, any decoding from ISOMP4
> > > > > (.mp4, .mov) will require a full bitstream copy in userspace to add
> > > > > the
> > > > > start code (unless you hack your allocation in your demuxer, but it's
> > > > > a
> > > > > bit complicated since this code might come from two libraries). In
> > > > > ISOMP4, you have an AVC header, which is just the size of the NAL that
> > > > > follows.
> > > > 
> > > > Well, I think we have to do a copy from system memory to the buffer
> > > > allocated by v4l2 anyway. Our hardware pipelines can reasonably be
> > > > expected not to have any MMU unit and not allow sg import anyway.
> > > 
> > > The Rockchip has an mmu. You need one copy at least indeed,
> > 
> > Is the MMU in use currently? That can make things troublesome if we run
> > into a case where the VPU has MMU and deals with scatter-gather while
> > the display part doesn't. As far as I know, there's no way for
> > userspace to know whether a dma-buf-exported buffer is backed by CMA or
> > by scatter-gather memory. This feels like a major issue for using dma-
> > buf, since userspace can't predict whether a buffer exported on one
> > device can be imported on another when building its pipeline.
> 
> FYI, Allwinner H6 also has IOMMU, it's just that there is no mainline driver 
> for it yet. It is supported for display, both VPUs and some other devices. I 
> think no sane SoC designer would left out one or another unit without IOMMU 
> support, that just calls for troubles, as you pointed out.

Right right, I've been following that from a distance :)

Indeed I think it's realistic to expect that for now, but it may not
play out so well in the long term. For instance, maybe connecting a USB
display would require CMA when the rest of the system can do with sg.

I think it would really be useful for userspace to have a way to test
whether a buffer can be imported from one device to another. It feels
better than indicating where the memory lives, since there are
countless cases where additional restrictions apply too.

Cheers,

Paul

> Best regards,
> Jernej
> 
> > > e.g. file
> > > to mem, or udpsocket to mem. But right now, let's say with ffmpeg/mpeg-
> > > ts, first you need to copy the MPEG TS to mem, then to demux you copy
> > > that H264 stream to another buffer, you then copy in the parser,
> > > removing the start-code and finally copy in the accelerator, adding the
> > > start code. If the driver would allow userptr, it would be unusable.
> > > 
> > > GStreamer on the other side implement lazy conversion, so it would copy
> > > the mpegts to mem, copy to demux, aggregate (with lazy merging) in the
> > > parser (but stream format is negotiation, so it keeps the start-code).
> > > If you request alignment=au, you have full frame of buffers, so if your
> > > driver could do userptr, you can same that extra copy.
> > > 
> > > Now, if we demux an MP4 it's the same, the parser will need do a full
> > > copy instead of lazy aggregation in order to prepend the start code
> > > (since it had an AVC header). But userptr could save a copy.
> > > 
> > > If the driver requires no nal prefix, then we could just pass a
> > > slightly forward point to userptr and avoid ACV to ANNEX-B conversion,
> > > which is a bit slower (even know it's nothing compare to the full
> > > copies we already do.
> > > 
> > > That was my argument in favour for no NAL prefix in term of efficiency,
> > > and it does not prevent adding a control to enable start-code for cases
> > > it make sense.
> > 
> > I see, so the internal arcitecture of userspace software may not be a
> > good fit for adding these bits and it could hurt performance a bit.
> > That feels like a significant downside.
> > 
> > > > So with that in mind, asking userspace to add a startcode it already
> > > > knows doesn't seem to be asking too much.
> > > > 
> > > > > On the other end, the data_offset thing is likely just a thing for the
> > > > > RK3399 to handle, it does not affect RK3288, Cedrus or IMX8M.
> > > > 
> > > > Well, I think it's best to be fool-proof here and just require that
> > > > start code. We should also have per-slice bit offsets to the different
> > > > parts anyway, so drivers that don't need it can just ignore it.
> > > > 
> > > > In extreme cases where there is some interest in doing direct buffer
> > > > import without doing a copy in userspace, userspace could trick the
> > > > format and avoid a copy by not providing the start-code (assuming it
> > > > knows it doesn't need it) and specifying the bit offsets accordingly.
> > > > That'd be a hack for better performance, and it feels better to do
> > > > things in this order rather than having to hack around in the drivers
> > > > that need the start code in every other case.
> > > 
> > > So basically, you and Tomas are both strongly in favour of adding
> > > ANNEX-B start-code to the current uAPI. I have digged into Cedrus
> > > registers, and it seems that it does have start-code scanning support.
> > > I'm not sure it can do "full-frame" decoding, 1 interrupt per frame
> > > like the RK do. That requires the IP to deal with the modifications
> > > lists, which are per slices.
> > 
> > Actually the bitstream parser won't reconfigure the pipeline
> > configuration registers, it's only around for userspace to avoid
> > implementing bitstream parsing, but it's a standalone thing.
> > 
> > So if we want to do full-frame decoding we always need to reconfigure
> > our pipeline (or do it like we do currently and just use one of the
> > per-slice configuration and hope for the best).
> > 
> > Do we have more information on the RK3399 and what it requires exactly?
> > (Just to make sure it's not another issue altogether.)
> > 
> > > My question is, are you willing to adapt the Cedrus driver to support
> > > receiving start-code ? And will this have a performance impact or not ?
> > > On RK side, it's really just about flipping 1 bit.
> > > 
> > > On the Rockchip side, Tomas had concern about CPU wakeup and the fact
> > > that we didn't aim at supporting passing multiple slices at once to the
> > > IP (something RK supports). It's important to understand that multi-
> > > slice streams are relatively rare and mostly used for low-latency /
> > > video conferencing. So aggregating in these case defeats the purpose of
> > > using slices. So I think RK feature is not very important.
> > 
> > Agreed, let's aim for low-latency as a standard.
> > 
> > > Of course, I do believe that long term we will want to expose bot
> > > stream formats on RK (because the HW can do that), so then userspace
> > > can just pick the best when available. So that boils down to our first
> > > idea, shall we expose _SLICE_A and _SLICE_B or something like this ?
> > > Now that we have progressed on the matter, I'm quite in favour of
> > > having _SLICE in the first place, with the preferred format that
> > > everyone should support, and allow for variants later. Now, if we make
> > > one mandatory, we could also just have a menu control to allow other
> > > formats.
> > 
> > That seems fairly reasonable to me, and indeed, having one preferred
> > format at first seems to be a good move.
> > 
> > > > > > To me the breaking point was about having the slice header both in
> > > > > > raw
> > > > > > bitstream and parsed forms. Since we agree that's fine, we might as
> > > > > > well push it to its logical conclusion and include all the bits that
> > > > > > can be useful.
> > > > > 
> > > > > To take your words, the bits that contain useful information starts
> > > > > from the NAL type byte, exactly were the data was cut by VA-API and
> > > > > the
> > > > > current uAPI.
> > > > 
> > > > Agreed, but I think that the advantages of always requiring the start
> > > > code outweigh the potential (yet quite unlikely) downsides.
> > > > 
> > > > > > > > > But it also support slice mode, with an
> > > > > > > > > interrupt per slice, which is what we decided to use.
> > > > > > > > 
> > > > > > > > Easier for everyone and probably better for latency as well :)
> > > > > > > > 
> > > > > > > > > So in this case, indeed we strictly require on start-code.
> > > > > > > > > Though, to
> > > > > > > > > me this is not a great reason to make a new fourcc, so we will
> > > > > > > > > try and
> > > > > > > > > use (data_offset = 3) in order to make some space for that
> > > > > > > > > start code,
> > > > > > > > > and write it down in the driver. This is to be continued, we
> > > > > > > > > will
> > > > > > > > > report back on this later. This could have some side effect in
> > > > > > > > > the
> > > > > > > > > ability to import buffers. But most userspace don't try to do
> > > > > > > > > zero-copy
> > > > > > > > > on the encoded size and just copy anyway.
> > > > > > > > > 
> > > > > > > > > To my opinion, having a single format is a big deal, since
> > > > > > > > > userspace
> > > > > > > > > will generally be developed for one specific HW and we would
> > > > > > > > > endup with
> > > > > > > > > fragmented support. What we really want to achieve is having a
> > > > > > > > > driver
> > > > > > > > > interface which works across multiple HW, and I think this is
> > > > > > > > > quite
> > > > > > > > > possible.
> > > > > > > > 
> > > > > > > > I agree with that. The more I think about it, the more I believe
> > > > > > > > we
> > > > > > > > should just pass the whole
> > > > > > > > [nal_header][nal_type][slice_header][slice]
> > > > > > > > and the parsed list in every scenario.
> > > > > > > 
> > > > > > > What I like of the cut at nal_type, is that there is only format.
> > > > > > > If we
> > > > > > > cut at nal_header, then we need to expose 2 formats. And it makes
> > > > > > > our
> > > > > > > API similar to other accelerator API, so it's easy to "convert"
> > > > > > > existing userspace.
> > > > > > 
> > > > > > Unless we make that cut the single one and only true cut that shall
> > > > > > supersed all other cuts :)
> > > > > 
> > > > > That's basically what I've been trying to do, kill this _RAW/ANNEX_B
> > > > > thing and go back to our first idea.
> > > > 
> > > > Right, in the end I think we should go with:
> > > > V4L2_PIX_FMT_MPEG2_SLICE
> > > > V4L2_PIX_FMT_H264_SLICE
> > > > V4L2_PIX_FMT_HEVC_SLICE
> > > > 
> > > > And just require raw bitstream for the slice with emulation-prevention
> > > > bits included.
> > > 
> > > That's should be the set of format we start with indeed. The single
> > > format for which software gets written and tested, making sure software
> > > support is not fragmented, and other variants should be something to
> > > opt-in.
> > 
> > Cheers for that!
> > 
> > Paul
> > 
> > > > Cheers,
> > > > 
> > > > Paul
> > > > 
> > > > > > > > For H.265, our decoder needs some information from the NAL type
> > > > > > > > too.
> > > > > > > > We currently extract that in userspace and stick it to the
> > > > > > > > slice_header, but maybe it would make more sense to have drivers
> > > > > > > > parse
> > > > > > > > that info from the buffer if they need it. On the other hand, it
> > > > > > > > seems
> > > > > > > > quite common to pass information from the NAL type, so maybe we
> > > > > > > > should
> > > > > > > > either make a new control for it or have all the fields in the
> > > > > > > > slice_header (which would still be wrong in terms of matching
> > > > > > > > bitstream
> > > > > > > > description).
> > > > > > > 
> > > > > > > Even in userspace, it's common to just parse this in place, it's a
> > > > > > > simple mask. But yes, if we don't have it yet, we should expose
> > > > > > > the NAL
> > > > > > > type, it would be cleaner.
> > > > > > 
> > > > > > Right, works for me.
> > > > > 
> > > > > Ack.
> > > > > 
> > > > > > Cheers,
> > > > > > 
> > > > > > Paul
> > > > > > 
> > > > > > > > > > - Dropping the DPB concept in H.264/H.265
> > > > > > > > > > 
> > > > > > > > > > As far as I could understand, the decoded picture buffer
> > > > > > > > > > (DPB) is a
> > > > > > > > > > concept that only makes sense relative to a decoder
> > > > > > > > > > implementation. The
> > > > > > > > > > spec mentions how to manage it with the Hypothetical
> > > > > > > > > > reference decoder
> > > > > > > > > > (Annex C), but that's about it.
> > > > > > > > > > 
> > > > > > > > > > What's really in the bitstream is the list of modified
> > > > > > > > > > short-term and
> > > > > > > > > > long-term references, which is enough for every decoder.
> > > > > > > > > > 
> > > > > > > > > > For this reason, I strongly believe we should stop talking
> > > > > > > > > > about DPB in
> > > > > > > > > > the controls and just pass these lists agremented with
> > > > > > > > > > relevant
> > > > > > > > > > information for userspace.
> > > > > > > > > > 
> > > > > > > > > > I think it should be up to the driver to maintain a DPB and
> > > > > > > > > > we could
> > > > > > > > > > have helpers for common cases. For instance, the rockchip
> > > > > > > > > > decoder needs
> > > > > > > > > > to keep unused entries around[2] and cedrus has the same
> > > > > > > > > > requirement
> > > > > > > > > > for H.264. However for cedrus/H.265, we don't need to do any
> > > > > > > > > > book-
> > > > > > > > > > keeping in particular and can manage with the lists from the
> > > > > > > > > > bitstream
> > > > > > > > > > directly.
> > > > > > > > > 
> > > > > > > > > As discusses today, we still need to pass that list. It's
> > > > > > > > > being index
> > > > > > > > > by the HW to retrieve the extra information we have collected
> > > > > > > > > about the
> > > > > > > > > status of the reference frames. In the case of Hantro, which
> > > > > > > > > process
> > > > > > > > > the modification list from the slice header for us, we also
> > > > > > > > > need that
> > > > > > > > > list to construct the unmodified list.
> > > > > > > > > 
> > > > > > > > > So the problem here is just a naming problem. That list is not
> > > > > > > > > really a
> > > > > > > > > DPB. It is just the list of long-term/short-term references
> > > > > > > > > with the
> > > > > > > > > status of these references. So maybe we could just rename as
> > > > > > > > > references/reference_entry ?
> > > > > > > > 
> > > > > > > > What I'd like to pass is the diff to the references list, as
> > > > > > > > ffmpeg
> > > > > > > > currently provides for v4l2 request and vaapi (probably vdpau
> > > > > > > > too). No
> > > > > > > > functional change here, only that we should stop calling it a
> > > > > > > > DPB,
> > > > > > > > which confuses everyone.
> > > > > > > 
> > > > > > > Yes.
> > > > > > > 
> > > > > > > > > > - Using flags
> > > > > > > > > > 
> > > > > > > > > > The current MPEG-2 controls have lots of u8 values that can
> > > > > > > > > > be
> > > > > > > > > > represented as flags. Using flags also helps with padding.
> > > > > > > > > > It's unlikely that we'll get more than 64 flags, so using a
> > > > > > > > > > u64 by
> > > > > > > > > > default for that sounds fine (we definitely do want to keep
> > > > > > > > > > some room
> > > > > > > > > > available and I don't think using 32 bits as a default is
> > > > > > > > > > good enough).
> > > > > > > > > > 
> > > > > > > > > > I think H.264/HEVC per-control flags should also be moved to
> > > > > > > > > > u64.
> > > > > > > > > 
> > > > > > > > > Make sense, I guess bits (member : 1) are not allowed in uAPI
> > > > > > > > > right ?
> > > > > > > > 
> > > > > > > > Mhh, even if they are, it makes it much harder to verify 32/64
> > > > > > > > bit
> > > > > > > > alignment constraints (we're dealing with 64-bit platforms that
> > > > > > > > need to
> > > > > > > > have 32-bit userspace and compat_ioctl).
> > > > > > > 
> > > > > > > I see, thanks.
> > > > > > > 
> > > > > > > > > > - Clear split of controls and terminology
> > > > > > > > > > 
> > > > > > > > > > Some codecs have explicit NAL units that are good fits to
> > > > > > > > > > match as
> > > > > > > > > > controls: e.g. slice header, pps, sps. I think we should
> > > > > > > > > > stick to the
> > > > > > > > > > bitstream element names for those.
> > > > > > > > > > 
> > > > > > > > > > For H.264, that would suggest the following changes:
> > > > > > > > > > - renaming v4l2_ctrl_h264_decode_param to
> > > > > > > > > > v4l2_ctrl_h264_slice_header;
> > > > > > > > > 
> > > > > > > > > Oops, I think you meant slice_prams ? decode_params matches
> > > > > > > > > the
> > > > > > > > > information found in SPS/PPS (combined?), while slice_params
> > > > > > > > > matches
> > > > > > > > > the information extracted (and executed in case of l0/l1) from
> > > > > > > > > the
> > > > > > > > > slice headers.
> > > > > > > > 
> > > > > > > > Yes you're right, I mixed them up.
> > > > > > > > 
> > > > > > > > >  That being said, to me this name wasn't confusing, since
> > > > > > > > > 
> > > > > > > > > it's not just the slice header, and it's per slice.
> > > > > > > > 
> > > > > > > > Mhh, what exactly remains in there and where does it originate
> > > > > > > > in the
> > > > > > > > bitstream? Maybe it wouldn't be too bad to have one control per
> > > > > > > > actual
> > > > > > > > group of bitstream elements.
> > > > > > > > 
> > > > > > > > > > - killing v4l2_ctrl_h264_decode_param and having the
> > > > > > > > > > reference lists
> > > > > > > > > > where they belong, which seems to be slice_header;
> > > > > > > > > 
> > > > > > > > > There reference list is only updated by userspace (through
> > > > > > > > > it's DPB)
> > > > > > > > > base on the result of the last decoding step. I was very
> > > > > > > > > confused for a
> > > > > > > > > moment until I realize that the lists in the slice_header are
> > > > > > > > > just a
> > > > > > > > > list of modification to apply to the reference list in order
> > > > > > > > > to produce
> > > > > > > > > l0 and l1.
> > > > > > > > 
> > > > > > > > Indeed, and I'm suggesting that we pass the modifications only,
> > > > > > > > which
> > > > > > > > would fit a slice_header control.
> > > > > > > 
> > > > > > > I think I made my point why we want the dpb -> references. I'm
> > > > > > > going to
> > > > > > > validate with the VA driver now, to see if the references list
> > > > > > > there is
> > > > > > > usable with our code.
> > > > > > > 
> > > > > > > > Cheers,
> > > > > > > > 
> > > > > > > > Paul
> > > > > > > > 
> > > > > > > > > > I'm up for preparing and submitting these control changes
> > > > > > > > > > and updating
> > > > > > > > > > cedrus if they seem agreeable.
> > > > > > > > > > 
> > > > > > > > > > What do you think?
> > > > > > > > > > 
> > > > > > > > > > Cheers,
> > > > > > > > > > 
> > > > > > > > > > Paul
> > > > > > > > > > 
> > > > > > > > > > [0]: https://lkml.org/lkml/2019/3/6/82
> > > > > > > > > > [1]: https://patchwork.linuxtv.org/patch/55947/
> > > > > > > > > > [2]:
> > > > > > > > > > https://chromium.googlesource.com/chromiumos/third_party/ke
> > > > > > > > > > rnel/+/4d7cb46539a93bb6acc802f5a46acddb5aaab378
> 
> 
> 




[Index of Archives]     [Linux Input]     [Video for Linux]     [Gstreamer Embedded]     [Mplayer Users]     [Linux USB Devel]     [Linux Audio Users]     [Linux Kernel]     [Linux SCSI]     [Yosemite Backpacking]

  Powered by Linux