Re: [RFC] Timestamps and V4L2

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Chris,

On Monday 15 October 2012 18:25:32 Chris MacGregor wrote:
> On 10/15/2012 12:59 PM, Sakari Ailus wrote:
> > On Mon, Oct 15, 2012 at 11:53:56AM -0700, Chris MacGregor wrote:
> >> On 10/15/2012 11:45 AM, Laurent Pinchart wrote:
> >>> On Monday 15 October 2012 19:05:49 Sakari Ailus wrote:
> >>>> Hi all,
> >>>> 
> >>>> As a summar from the discussion, I think we have reached the following
> >>>> conclusion. Please say if you agree or disagree with what's below. :-)
> >>>> 
> >>>> - The drivers will be moved to use monotonic timestamps for video
> >>>> buffers.
> >>>> - The user space will learn about the type of the timestamp through
> >>>> buffer flags.
> >>>> - The timestamp source may be made selectable in the future, but buffer
> >>>> flags won't be the means for this, primarily since they're not
> >>>> available on subdevs. Possible way to do this include a new V4L2
> >>>> control or a new IOCTL.
> >>> 
> >>> That's my understanding as well. For the concept,
> >>> 
> >>> Acked-by: Laurent Pinchart <laurent.pinchart@xxxxxxxxxxxxxxxx>
> >> 
> >> I wasn't able to participate in the discussion that led to this, but
> >> I'd like to suggest and request now that an explicit requirement (of
> >> whatever scheme is selected) be that a userspace app have a
> >> reasonable and straightforward way to translate the timestamps to
> >> real wall-clock time, ideally with enough precision to allow
> >> synchronization of cameras across multiple computers.
> >> 
> >> In the systems I work on, for instance, we are recording real-world
> >> biological processes, some of which vary based on the time of day,
> >> and it is important to know when a given frame was captured so that
> >> information can be stored with the raw frame and the data derived
> >> from it. For many such purposes, an accuracy measured in multiple
> >> seconds (or even minutes) is fine.
> >> 
> >> However, when we are using multiple cameras on multiple computers
> >> (e.g., two or more BeagleBoard xM's, each with a camera connected),
> >> we would want to synchronize with an accuracy of less than 1 frame
> >> time - e.g. 10 ms or less.
> > 
> > I think you have two use cases actually: knowing when an image has been
> > taken, and synchronisation of images from multiple sources. The first one
> > is easy: you just call clock_gettime() for the realtime clock. The
> > precision will certainly be enough.
> 
> I assume you mean that I would, for instance:
> 
> clock_gettime(CLOCK_REALTIME, &realtime);
> clock_gettime(CLOCK_MONOTONIC, &monotime);
> (compute realtime-monotime and save the result)
> ...
> (later add that result to the timestamp on a frame to recover the
> approximate real-world capture time)

That would be a very simple way to do it. You might or might not need 
something more complex in practice, depending on your use cases. The point is 
that you need to synchronize images on multiple machines, so you need a common 
clock between those machines. Whether you use wall clock time, possibly 
synchronized through NTP, or another clock source is irrelevant here. The 
kernel driver has no access to the cross-machines clock, so you will need to 
convert the timestamp provided by the kernel to a timestamp relative to your 
clock.

> Agreed, for this purpose - getting a reasonable real-world timestamp on
> the frame - the precision is fine...worst case, my process gets
> rescheduled between the two clock_gettime() calls, but even that won't
> matter for this purpose.
> 
> > For the latter the realtime clock fits poorly to begin with: it jumps
> > around e.g. when the daylight saving time changes.
> > 
> > I think what I'd do is this: figure out the difference between the
> > monotonic clocks of your systems and use that as basis for
> > synchronisation. I wonder if there are existing solutions for this based
> > e.g. on the NTP.
> > 
> > The pace of the monotonic clocks on different systems is the same as the
> > real-time ones; the same NTP adjustments are done to the monotonic clock
> > as
> > well. As an added bonus you also won't be affected by daylight saving time
> > or someone setting the clock manually.
> 
> Yes, I believe that this could work. My concern is that there is some
> unpredictability in the timing of the two clock_gettime() calls, and
> thus some inaccuracy in the conversion, and while I could likely get the
> two systems sufficiently sync'd using NTP or the like at the real-world
> level, the conversion from that (on each system) to the system's
> monotonic time would contain an unpredictable (though bounded)
> inaccuracy. I suppose that as long as I sync the cameras at the hardware
> level (e.g. by tying the strobe line of one to the trigger lines of the
> rest, or tying all the trigger lines to a GPIO), the inaccuracy would be
> less than a frame time, and so I could know reliably enough which frames
> go together.

I'm pretty sure there's lots of literature on clock synchronization algorithms 
:-) The topic can quickly become pretty complex, and I'm not expert there. I 
don't expect too much problems to reach the require precision if you 
synchronize your cameras.

> However, it seems much cleaner to have a more direct way to convert the
> monotonic time to real-world time, or to get real-world time on the
> frames in the first place (for applications that want that). I don't
> know of a way to do the former.
> 
> > The conversion of the two clocks requires the knowledge of the values of
> > kernel internal variables, so performing the conversion in user space
> > later on is not an option.

We might be missing a kernel API here then.

> Sorry, you lost me on this one, unless you're talking about what I refer
> to in my paragraph just above - converting the monotonic timestamp on
> the frame to real-world time...?
> 
> > Alternatively you could just call clock_gettime() after every DQBUF call,
> > but that's indeed less precise than if the driver would get the timestamp
> > for you.
> 
> And also less efficient. The platforms I'm working on are already
> hard-pressed to keep up with all the pixels I'm trying to capture and
> process, so I don't really want to waste time trapping into kernel mode
> again if I can avoid it.
> 
> > How would this work for you?
> 
> Better than nothing, and probably I could live with it. But I think
> perhaps we can do better than that, and now seems like the right time to
> figure it out.

-- 
Regards,

Laurent Pinchart

--
To unsubscribe from this list: send the line "unsubscribe linux-media" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux Input]     [Video for Linux]     [Gstreamer Embedded]     [Mplayer Users]     [Linux USB Devel]     [Linux Audio Users]     [Linux Kernel]     [Linux SCSI]     [Yosemite Backpacking]
  Powered by Linux