Re: [PATCH 0/3] usb: gadget: uvc: allocate requests based on frame interval length and buffersize

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jun 04, 2024 at 03:32:15PM -0700, Avichal Rakesh wrote:


On 5/29/24 14:24, Michael Grzeschik wrote:
On Tue, May 28, 2024 at 05:33:46PM -0700, Avichal Rakesh wrote:


On 5/28/24 15:43, Michael Grzeschik wrote:
On Tue, May 28, 2024 at 02:27:34PM -0700, Avichal Rakesh wrote:


On 5/28/24 13:22, Michael Grzeschik wrote:
On Tue, May 28, 2024 at 10:30:30AM -0700, Avichal Rakesh wrote:


On 5/22/24 10:37, Michael Grzeschik wrote:
On Wed, May 22, 2024 at 05:17:02PM +0000, Thinh Nguyen wrote:
One option to be totally sure would be to resend the sentinel request to
be properly transmitted before starting the next frame. This resend
polling would probably include some extra zero-length requests. But also
if this resend keeps failing for n times, the driver should doubt there
is anything sane going on with the USB connection and bail out somehow.

Since we try to tackle case (1) to avoid transmit errors and also avoid
creating late enqueued requests in the running isoc transfer, the over
all chance to trigger missed transfers should be minimal.

Gotcha. It seems like the UVC gadget driver implicitly assumes that EOF
flag will be used although the userspace application can technically
make it optional.

That is not all. The additional UVC_STREAM_ERR tag on the sentinel
request can be set optional by the host driver. But by spec the
userspace application has to drop the frame when the flag was set.

Looking at the UVC specs, the ERR bit doesn't seem to refer to actual
transmission error, only errors in frame generation (Section 4.3.1.7
of UVC 1.5 Class Specification). Maybe "data discontinuity" can be
used but the examples given are bad media, and encoder issues, which
suggests errors at higher level than the wire.

Oh! That is a new perspective I did not consider.

With the definition of UVC_STREAM_ERR by spec, the uvc_video driver
would in no case set this header bit for the current frame on its own?
Is that correct?

It would indeed seem so. The way gadget driver is architected makes
is impossible for the userspace application to notify the host of
any errors.


With my proposal this flag will be set, whenever we find out that
the currently transferred frame was erroneous.

Summarizing some of the discussions above:
1. UVC gadget driver should _not_ rely on the usb controller to
  enqueue 0-length requests on UVC gadget drivers behalf;
2. However keeping up the backpressure to the controller means the
  EOF request will be delayed behind all the zero-length requests.

Exactly, this is why we have to somehow finetune the timedelay between
requests that trigger interrupts. And also monitor the amount of
requests currently enqueued in the hw ringbuffer. So that our drivers
enqueue dequeue mechanism is virtually adding only the minimum amount
of necessary zero-length requests in the hardware. This should be
possible.

I am currently thinking through the remaining steps the pump worker has
to do on each wakeup to maintain the minimum threshold while waiting
with submitting requests that contain actual image payload.

Out of curiosity: What is wrong with letting the host rely on
FID alone? Decoding the jpeg payload _should_ fail if any of the
usb_requests containing the payload failed to transmit.

This is not totally true. We saw partially rendered jpeg frames on the
host stream. How the host behaves with broken data is totally undefined
if the typical uvc flags EOF/ERR are not used as specified. Then think
about uncompressed formats. So relying on the transferred image format
to solve our problems is just as wrong as relying on the gadgets
hardware behavior.

Do you know if the partially rendered frames were valid JPEGs, or
if the host was simply making a best effort at displaying a broken
JPEG? Perhaps the fix should go to the host instead?

I can fully reproduce this with linux and windows hosts. For linux
machines I saw that the host was taking the FID change as a marker
to see the previous frame as ready and just rendered what got through.
This did not lead to garbage but only to partially displayed frames
with jpeg macroblock alignment.

I was aware of linux doing so, but I only ever saw this behavior on
Windows if there were a lot of invalid frames back to back.

I am not super familiar with the guarantees of JPEG, but I suppose
it is possible to have a "valid" JPEG with some middle blocks
missing as long the EOI bits make it through? I am not sure how we
go about solving that.

It is even worse. Since we don't necessary need the EOF tag set but the
host will draw the content that it got after the FID has changed. It is
always possible that an frame that was errornous and therefor dropped
on the sendin side, will be shown on the host to the last macroblock it
received. So these partially drawn frames are more common then expected.

Following is my opinion, feel free to disagree (and correct me if
something is factually incorrect):

The fundamental issue here is that ISOC doesn't guarantee
delivery of usb_requests or even basic data consistency upon delivery.
So the gadget driver has no way to know the state of transmitted data.
The gadget driver is notified of underruns but not of any other issues,
and ideally we should never have an underrun if the zero-length
backpressure is working as intended.

So, UVC gadget driver can reduce the number of errors, but it'll never be
able to guarantee that the data transmitted to the host isn't somehow
corrupted or missing unless a more reliable mode of transmission
(bulk, for example) is used.

All of this to say: The host absolutely needs to be able to handle
all sorts of invalid and broken payloads. How the host handles it
might be undefined, but the host can never rely on perfect knowledge
about the transmission state. In cases like these, where the underlying
transport is unreliable, the burden of enforcing consistency moves up
a layer, i.e. to the encoded payload in this case. So it is perfectly
fine for the host to rely on the encoding to determine if the payload
is corrupt and handle it accordingly.

Right.

As for uncompressed format, you're correct that subtle corruptions
may not be caught, but outright missing usb_requests can be easily
checked by simply looking at the number of bytes in the payload. YUV
frames are all of the same (predetermined) size for a given resolution.

That was also my thought about five minutes after I did send you the
previous mail. So sure, this is no real issue for the host.

So my recommendation is the following:
1. Fix the bandwidth problem by splitting the encoded video frame
  into more usb_requests (as your patch already does) making sure
  there are enough free usb_request to encode the video frame in
  one burst so we don't accidentally inflate the transmission
  duration of a video frame by sneaking in zero-length requests in
  the middle.

Ack. This should already solve a lot of issues.

For this I would still suggest to move the usb_ep_queue to be done in
the pump worker again. Its a bit back and forth, but IMHO its worth the
extra mile since only this way we would respect the dwc3 interrupt
threads assumption to run *very* short.

The main reason for queuing the requests from the complete handler
was to have a single point of usb_ep_queue call, which made reasoning
through the locking simpler. But if you find a way to do so from
the video_pump thread without making the locking a nightmare, then go
for it!


2. Unless there is an unusually high rate of transmission failures
  when using the UVC gadget driver, it might be worth fixing the
  host side driver to handle broken frames better instead (assuming
  host is linux as well).

Agreed, but this needs a separate scoped undestanding of the host side
behaviour over all layers.

Agreed!


2. Tighten up the error checking in UVC gadget driver -- We drop the
  current frame whenever an EXDEV happens which is wrong. We should
  only be dropping the current frame if the EXDEV corresponds to the
  frame currently being encoded.

What do you mean by drop?

I would suggest to immediatly switch the uvc_buffer that is being
enqueued and start queueing prepared requests from the next buffers prep
list. As suggested, the idea is to have per uvc_buffer prep_list
requests which would make this task easy.

Currently, if uvc gadget driver receives an EXDEV complete callback
all it does is set the UVC_QUEUE_DROP_INCOMPLETE flag.

So let's say that we receive an EXDEV for a usb_request containing data
for video frame N. With how video_pump is currently configured, chances
are that all usb_requests containing data for video frame N has already
been queued to the controller.

When the next video frame (N+1) comes in, video_pump's encode methods
will look at the UVC_QUEUE_DROP_INCOMPLETE flag and incorrectly
determine that "current" frame needs to be dropped, and stop encoding
video frame N+1 even though the error was for video frame N. So the
encode methods incorrectly drop video frame N+1 which isn't needed.

The encode methods should only be dropping the video frame if we
received an EXDEV for a usb_request for the video frame currently
being encoded.

I hope that makes sense!

This totally makes sense. I just wanted to make sure that this does not
involve any UVC_STREAM_ERR tagging from your understanding.

I totally agree with this concept. So now we "only" have to implement
this. :)

First I will review and update my patches that will increase the amount
of requests per frame.

Regards,
Michael

--
Pengutronix e.K.                           |                             |
Steuerwalder Str. 21                       | http://www.pengutronix.de/  |
31137 Hildesheim, Germany                  | Phone: +49-5121-206917-0    |
Amtsgericht Hildesheim, HRA 2686           | Fax:   +49-5121-206917-5555 |

Attachment: signature.asc
Description: PGP signature


[Index of Archives]     [Linux Media]     [Linux Input]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [Old Linux USB Devel Archive]

  Powered by Linux