On Fri, 2018-10-05 at 13:49 -0300, Mauro Carvalho Chehab wrote: > Em Fri, 05 Oct 2018 12:37:43 -0300 > Ezequiel Garcia <ezequiel@xxxxxxxxxxxxx> escreveu: > > > On Fri, 2018-10-05 at 09:10 -0300, Mauro Carvalho Chehab wrote: > > > Em Thu, 04 Oct 2018 20:39:31 -0300 > > > Ezequiel Garcia <ezequiel@xxxxxxxxxxxxx> escreveu: > > > > > > > On Mon, 2018-10-01 at 14:54 -0300, Ezequiel Garcia wrote: > > > > > On Fri, 2018-09-28 at 14:33 +0200, Hans Verkuil wrote: > > > > > > On 09/17/2018 07:30 PM, Ezequiel Garcia wrote: > > > > > > > This series adds support for JPEG encoding via the VPU block > > > > > > > present in Rockchip platforms. Currently, support for RK3288 > > > > > > > and RK3399 is included. > > > > > > > > > > > > > > Please, see the previous versions of this cover letter for > > > > > > > more information. > > > > > > > > > > > > > > Compared to v5, the only change is in the V4L2_CID_JPEG_QUANTIZATION > > > > > > > control. We've decided to support only baseline profile, > > > > > > > and only add 8-bit luma and chroma tables. > > > > > > > > > > > > > > struct v4l2_ctrl_jpeg_quantization { > > > > > > > __u8 luma_quantization_matrix[64]; > > > > > > > __u8 chroma_quantization_matrix[64]; > > > > > > > }; > > > > > > > > > > > > > > By doing this, it's clear that we don't currently support anything > > > > > > > but baseline. > > > > > > > > > > > > > > This series should apply cleanly on top of > > > > > > > > > > > > > > git://linuxtv.org/hverkuil/media_tree.git br-cedrus tag. > > > > > > > > > > > > > > If everyone is happy with this series, I'd like to route the devicetree > > > > > > > changes through the rockchip tree, and the rest via the media subsystem. > > > > > > > > > > > > OK, so I have what is no doubt an annoying question: do we really need > > > > > > a JPEG_RAW format? > > > > > > > > > > > > > > > > Not annoying, as it helps clarify a few things :-) > > > > > I think we do need the JPEG_RAW format. The way I see it, using > > > > > JPEG opens a can of worms... > > > > > > > > > > > The JPEG header is really easy to parse for a decoder and really easy to > > > > > > prepend to the compressed image for the encoder. > > > > > > > > > > > > The only reason I can see for a JPEG_RAW is if the image must start at > > > > > > some specific address alignment. Although even then you can just pad the > > > > > > JPEG header that you will add according to the alignment requirements. > > > > > > > > > > > > I know I am very late with this question, but I never looked all that > > > > > > closely at what a JPEG header looks like. But this helped: > > > > > > > > > > > > https://en.wikipedia.org/wiki/JPEG_File_Interchange_Format > > > > > > > > > > > > and it doesn't seem difficult at all to parse or create the header. > > > > > > > > > > > > > > > > > > > > > > ... I think that having JPEG_RAW as the compressed format > > > > > is much more clear for userspace, as it explicitly specifies > > > > > what is expected. > > > > > > > > > > This way, for a stateless encoder, applications are required > > > > > to set quantization and/or entropy tables, and are then in > > > > > charge of using the compressed JPEG_RAW payload in whatever form > > > > > they want. Stupid simple. > > > > > > > > > > On the other side, if the stateless encoder driver supports > > > > > JPEG (creating headers in-kernel), it means that: > > > > > > > > > > *) applications should pass a quality control, if supported, > > > > > and the driver will use hardcoded tables or... > > > > > > > > > > *) applications pass quantization control and, if supported, > > > > > entropy control. The kernel uses them to create the JPEG frame. > > > > > But also, some drivers (e.g. Rockchip), use default entropy > > > > > tables, which should now be in the kernel. > > > > > > > > > > So the application would have to query controls to find out > > > > > what to do. Not exactly hard, but I think having the JPEG_RAW > > > > > is much simpler and more clear. > > > > > > > > > > Now, for stateless decoders, supporting JPEG_RAW means > > > > > the application has to pass quantization and entropy > > > > > controls, probably using the Request API. > > > > > Given the application has parsed the JPEG, > > > > > it knows the width and height and can request > > > > > buffers accordingly. > > > > > > > > > > The hardware is stateless, and so is the driver. > > > > > > > > > > On the other hand, supporting JPEG would mean that > > > > > drivers will have to parse the image, extracting > > > > > the quantization and entropy tables. > > > > > > > > > > Format negotiation is now more complex, > > > > > either we follow the stateful spec, introducing a little > > > > > state machine in the driver... or we use the Request API, > > > > > but that means parsing on both sides kernel and application. > > > > > > > > > > Either way, using JPEG_RAW is just waaay simpler and puts > > > > > things where they belong. > > > > > > > > > > > > > As discussed on IRC, I'm sending a v7 for this series, > > > > fixing only the checkpatch issues and the extra line in the > > > > binding document. > > > > > > > > We've agreed to move forward with JPEG_RAW, for the reasons > > > > stated above. > > > > > > > > Plus, we've agreed to keep this in staging until the userspace > > > > support for JPEG_RAW format is clear. > > > > > > The problem is that a new format is actually a V4L2 core change, and > > > not something for staging. > > > > > > IMHO, it is prudent to wait for an userspace code before hushing > > > merging this upstream. > > > > > > - > > > > > > I'm concerned about adding a JPEG_RAW format. We had this discussion > > > in the past. On that time, it was opted to add a code inside the > > > drivers that will be adding the jpeg headers, on the cases where > > > the hardware was providing headerless frames. > > > > > > > I suspect such discussion was held on completely different hardware. > > What we are trying to support here has specific needs, and thus > > why we need new formats. > > > > > Among the reasons, different drivers may be requiring different > > > logic to add the jpeg headers, depending on how they interpret > > > the quality parameters and on the Huffmann tables used by an > > > specific codec. > > > > > > I can't see why this couldn't be done on a stateless codec. I mean: > > > in the specific case of JPEG, each frame is independently coded. > > > It doesn't need to store anything from a previous frame in order to > > > produce a JPEG header. > > > > > > > Well, this can be done. But since it's the wrong architecture, > > it will bring a more complex API and more complex implementation. > > I'm not saying that this is a final solution. > > We're talking about a driver that it is intended to be merged > (for now) at staging. "Polluting" V4L2 core with a new almost identical > format that has a potential of bringing problems due to a staging > driver is something that doesn't sound right. > > Before changing the core, we need to be able to test your proposal on > more drivers, on both directions (capture and output). > > > In the case of a stateless decoder, it will mean a really cumbersome > > API. As I have explained above, using JPEG format on hardware > > than can only deal with JPEG_RAW format means: > > > > 1) Userspace sets JPEG format on the OUTPUT queue (sink) > > I'm assuming that you're actually talking about a M2M device whose > output stream receives a JPEG frame, and its capture stream outputs > something like RGB, right? > Right, RGB or more likely YUVish. > Also, I'm assuming the description you gave for the "raw JPEG" as a > JPEG image without JPEG headers. > > > 2) Userspace needs to know the raw format, but it has > > to wait for the driver to parse it first. > > Why would userspace need to know the raw format? It doesn't really > matter if the output is raw JPEG or full JPEG. > > For the output stream scenario: > > 1) for a headerless-JPEG (raw JPEG is really a crappy name): > > Userspace will use the headers to extract some information useful > for the decoder to work, passing them via controls. Then, it will > queue a buffer. Request API is needed, as the parameters will change > frame by frame. > > 2) for "true" JPEG (e. g. a jpeg frame that follows the standard and > has headers): > > Kernelspace will parse the jpeg headers, extracting the parameters > they need. No need to use request API, as each frame will already > contain the needed metadata. > > No matter if (1) or (2) is used, a M2M device will receive one JPEG > frame, decode it and produce one RGB frame. No need to store any > state. > > The opposite would equally work, and it is identical to what we > have already in kernel, for the "true" JPEG output. > > A M2M device whose output stream is RGB and capture is MPEG: > > 1) for "true" JPEG: > > The driver will use a Huffmann table based on a quality parameter > (V4L2_CID_JPEG_COMPRESSION_QUALITY). Such table would either > be hardcoded at the driver or at the codec's firmware. The > output stream will contain the headers. > > 1) for a headerless-JPEG: > > The driver will use a Huffmann table based on a quality parameter > (V4L2_CID_JPEG_COMPRESSION_QUALITY). Such table would either > be hardcoded at the driver or at the codec's firmware. The > output stream will not contain the headers. > > Userspace will need to figure out what it was used by the Kernel > in order to add the JPEG header. I'm not sure if we can do it > on a transparent way for all devices. > > In any case, one input produces one output. > > > > > So, 3) it will QBUF a JPEG buffer, and 4) get the V4L2_EVENT_SOURCE_CHANGE > > event in order to move forward. This implies a stateful machinery in the > > driver. Like I said, more complex API, more complex implementation... > > and only to avoid a new pixelformat? Doesn't sound a good deal :-) > > I really can't understand what you're meaning by that... I can't > see any need to generate an event to move forward. > It is needed for userspace to negotiate the format on the CAPTURE queue. See for instance: https://github.com/torvalds/linux/blob/master/drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c#L692 Which shows how the driver needs to have an internal state, and how it needs to produce an event on the .buf_queue. This sequence is specified in the stateful codec spec, and is not needed by stateless codecs such as this one. > > > > In comparison, a proper stateless decoder driver will expose > > a simpler API, provide a 1:1 output/capture buffer relationship, > > and expose the full capabilities to userspace. > > > > > So, from my side, at least for the first version of this driver, > > > please make it produce a full JPEG frame with the headers, and > > > use V4L2_PIX_FMT_JPEG format. > > > > > > > This means using some form of JPEG quantization tables and huffman > > tables in the kernel, such as in drivers/media/usb/gspca/topro.c. > > > > Really doesn't sound like the best choice, violating the > > mechanism/policy separation, as now the kernel is deciding > > the tables for the user. > > What the Huffmann tables has to do with it? No matter if the > JPEG has headers or not, the codec should be doing RLE > encoding/decoding. For that, it needs to know the Huffmann tables. > > Anyway, you can use a control to pass huffmann tables at runtime, > using the request API, no matter if JPEG has headers or not. > > > The hardware can take any quantization table, and allowing userspace > > flexible usage, but we are going to restrict that to just a specific > > usage and a specific set of tables... again, only to avoid a new > > pixelformat? > > Same here: passing quantization tables to the driver has nothing > to do with having headers or not. > > For decoding, having headers is actually better, as the table > will be at the JPEG header. For encoding, a V4L2 array control. > could be used. > > > > That should allow us to add this driver on staging without adding > > > a new JPEG_RAW format. > > > > > > > AFAIK, Rockchip and Allwinner JPEG codecs handle so-called JPEG_RAW > > format, i.e. the payload of a SOF segment. > > And so, almost all other JPEG devices we have already. > This is not true. JPEG decoder drivers currently supported, such as coda, mtk-jpeg and similar, seem to pass the JPEG frame (with all its headers) to the hardware. Rockhip and Allwinner are different, where the hardware only takes the SOF payload. We are adding a new pixelformat, because these devices handle a different format. > > I think having these > > two cases should be enough reason to add a new format. > > I'm not against adding it. Just against adding it without further > discussions, in a way that it would potentially cause troubles for > userspace to encode/decode JPEG. > > > > If you still think we'll need a V4L2_PIX_FMT_JPEG_RAW, I'd like > > > to merge it by the time we also have a decoder for libv4l. I > > > would also like to see, at the Kernel patch series, the addition > > > of such support for already existing non-staging camera > > > drivers that internally receive headless JPEG headers from the > > > hardware. This will allow us to test the userspace code with > > > different models and proof that a generic JPEG_RAW decoder on > > > userspace is doable. > > > > > > > We can add a libv4l2 plugin for this decoder if that helps moving > > forward. Although I suspect it will just fall into oblivion, > > as chromeos and android will do their own thing. > > I don't have any control what people do on forked Kernels. > > > > > JPEG is a very well known format, and it's trivial for applications > > to add the headers. > > With the current documentation you added at the V4L2 core patch, > I seriously doubt that it would be possible to do it in a way that > it would work with different vendors. > > Also, I don't know any current V4L2 generic application that would > do that. Having it at libv4l would allow those apps to work with that. > OK, perhaps we can try adding the libv4l2 plugin in a v8 submission. Regards, Eze