Re: [st-ericsson] v4l2 vs omx for camera

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, Feb 24, 2011 at 10:17 PM, Hans Verkuil <hverkuil@xxxxxxxxx> wrote:
> On Tuesday, February 22, 2011 03:44:19 Clark, Rob wrote:
>> On Fri, Feb 18, 2011 at 10:39 AM, Robert Fekete
>> <robert.fekete@xxxxxxxxxx> wrote:
>> > Hi,
>> >
>> > In order to expand this knowledge outside of Linaro I took the Liberty of
>> > inviting both linux-media@xxxxxxxxxxxxxxx and
>> > gstreamer-devel@xxxxxxxxxxxxxxxxxxxxxx For any newcomer I really recommend
>> > to do some catch-up reading on
>> > http://lists.linaro.org/pipermail/linaro-dev/2011-February/thread.html
>> > ("v4l2 vs omx for camera" thread) before making any comments. And sign up
>> > for Linaro-dev while you are at it :-)
>> >
>> > To make a long story short:
>> > Different vendors provide custom OpenMax solutions for say Camera/ISP. In
>> > the Linux eco-system there is V4L2 doing much of this work already and is
>> > evolving with mediacontroller as well. Then there is the integration in
>> > Gstreamer...Which solution is the best way forward. Current discussions so
>> > far puts V4L2 greatly in favor of OMX.
>> > Please have in mind that OpenMAX as a concept is more like GStreamer in many
>> > senses. The question is whether Camera drivers should have OMX or V4L2 as
>> > the driver front end? This may perhaps apply to video codecs as well. Then
>> > there is how to in best of ways make use of this in GStreamer in order to
>> > achieve no copy highly efficient multimedia pipelines. Is gst-omx the way
>> > forward?
>>
>> just fwiw, there were some patches to make v4l2src work with userptr
>> buffers in case the camera has an mmu and can handle any random
>> non-physically-contiguous buffer..  so there is in theory no reason
>> why a gst capture pipeline could not be zero copy and capture directly
>> into buffers allocated from the display
>
> V4L2 also allows userspace to pass pointers to contiguous physical memory.
> On TI systems this memory is usually obtained via the out-of-tree cmem module.
>
>> Certainly a more general way to allocate buffers that any of the hw
>> blocks (display, imaging, video encoders/decoders, 3d/2d hw, etc)
>> could use, and possibly share across-process for some zero copy DRI
>> style rendering, would be nice.  Perhaps V4L2_MEMORY_GEM?
>
> There are two parts to this: first of all you need a way to allocate large
> buffers. The CMA patch series is available (but not yet merged) that does this.
> I'm not sure of the latest status of this series.
Still ARM maintainer doesn't agree these patches since it's not solve
the ARM memory different attribute mapping problem.
but try to send the CMA v9 patch soon.

We need really require the physical memory management module. Each
chip vendors use the their own implementations.
Our approach called it as CMA, others called it as cmem, carveout,
hwmon and so on.

I think Laurent's approaches is similar one.

We will try it again to merge CMA.

Thank you,
Kyungmin Park


>
> The other part is that everyone can use and share these buffers. There isn't
> anything for this yet. We have discussed this in the past and we need something
> generic for this that all subsystems can use. It's not a good idea to tie this
> to any specific framework like GEM. Instead any subsystem should be able to use
> the same subsystem-independent buffer pool API.
>
> The actual code is probably not too bad, but trying to coordinate this over all
> subsystems is not an easy task.
>
>>
>> >
>> > Let the discussion continue...
>> >
>> >
>> > On 17 February 2011 14:48, Laurent Pinchart
>> > <laurent.pinchart@xxxxxxxxxxxxxxxx> wrote:
>> >>
>> >> On Thursday 10 February 2011 08:47:15 Hans Verkuil wrote:
>> >> > On Thursday, February 10, 2011 08:17:31 Linus Walleij wrote:
>> >> > > On Wed, Feb 9, 2011 at 8:44 PM, Harald Gustafsson wrote:
>> >> > > > OMX main purpose is to handle multimedia hardware and offer an
>> >> > > > interface to that HW that looks identical indenpendent of the vendor
>> >> > > > delivering that hardware, much like the v4l2 or USB subsystems tries
>> >> > > > to
>> >> > > > do. And yes optimally it should be implemented in drivers/omx in
>> >> > > > Linux
>> >> > > > and a user space library on top of that.
>> >> > >
>> >> > > Thanks for clarifying this part, it was unclear to me. The reason
>> >> > > being
>> >> > > that it seems OMX does not imply userspace/kernelspace separation, and
>> >> > > I was thinking more of it as a userspace lib. Now my understanding is
>> >> > > that if e.g. OpenMAX defines a certain data structure, say for a PCM
>> >> > > frame or whatever, then that exact struct is supposed to be used by
>> >> > > the
>> >> > > kernelspace/userspace interface, and defined in the include file
>> >> > > exported
>> >> > > by the kernel.
>> >> > >
>> >> > > > It might be that some alignment also needs to be made between 4vl2
>> >> > > > and
>> >> > > > other OS's implementation, to ease developing drivers for many OSs
>> >> > > > (sorry I don't know these details, but you ST-E guys should know).
>> >> > >
>> >> > > The basic conflict I would say is that Linux has its own API+ABI,
>> >> > > which
>> >> > > is defined by V4L and ALSA through a community process without much
>> >> > > thought about any existing standard APIs. (In some cases also
>> >> > > predating
>> >> > > them.)
>> >> > >
>> >> > > > By the way IL is about to finalize version 1.2 of OpenMAX IL which
>> >> > > > is
>> >> > > > more than a years work of aligning all vendors and fixing unclear
>> >> > > > and
>> >> > > > buggy parts.
>> >> > >
>> >> > > I suspect that the basic problem with Khronos OpenMAX right now is
>> >> > > how to handle communities - for example the X consortium had
>> >> > > something like the same problem a while back, only member companies
>> >> > > could partake in the standard process, and they need of course to pay
>> >> > > an upfront fee for that, and the majority of these companies didn't
>> >> > > exactly send Linux community members to the meetings.
>> >> > >
>> >> > > And now all the companies who took part in OpenMAX somehow
>> >> > > end up having to do a lot of upfront community work if they want
>> >> > > to drive the API:s in a certain direction, discuss it again with the
>> >> > > V4L
>> >> > > and ALSA maintainers and so on. Which takes a lot of time and
>> >> > > patience with uncertain outcome, since this process is autonomous
>> >> > > from Khronos. Nobody seems to be doing this, I javen't seen a single
>> >> > > patch aimed at trying to unify the APIs so far. I don't know if it'd
>> >> > > be
>> >> > > welcome.
>> >> > >
>> >> > > This coupled with strict delivery deadlines and a marketing will
>> >> > > to state conformance to OpenMAX of course leads companies into
>> >> > > solutions breaking the Linux kernelspace API to be able to present
>> >> > > this.
>> >>
>> >> From my experience with OMX, one of the issues is that companies usually
>> >> extend the API to fullfill their platform's needs, without going through
>> >> any
>> >> standardization process. Coupled with the lack of open and free reference
>> >> implementation and test tools, this more or less means that OMX
>> >> implementations are not really compatible with eachother, making OMX-based
>> >> solution not better than proprietary solutions.
>> >>
>> >> > > Now I think we have a pretty clear view of the problem, I don't
>> >> > > know what could be done about it though :-/
>> >> >
>> >> > One option might be to create a OMX wrapper library around the V4L2 API.
>> >> > Something similar is already available for the old V4L1 API (now removed
>> >> > from the kernel) that allows apps that still speak V4L1 only to use the
>> >> > V4L2 API. This is done in the libv4l1 library. The various v4l libraries
>> >> > are maintained here: http://git.linuxtv.org/v4l-utils.git
>> >> >
>> >> > Adding a libomx might not be such a bad idea. Linaro might be the
>> >> > appropriate organization to look into this. Any missing pieces in V4L2
>> >> > needed to create a fully functioning omx API can be discussed and
>> >> > solved.
>> >> >
>> >> > Making this part of v4l-utils means that it is centrally maintained and
>> >> > automatically picked up by distros.
>> >> >
>> >> > It will certainly be a non-trivial exercise, but it is a one-time job
>> >> > that
>> >> > should solve a lot of problems. But someone has to do it...
>> >>
>> >> It's an option, but why would that be needed ? Again from my (probably
>> >> limited) OMX experience, platforms expose higher-level APIs to
>> >> applications,
>> >> implemented on top of OMX. If the OMX layer is itself implemented on top
>> >> of
>> >> V4L2, it would just be an extraneous useless internal layer that could
>> >> (should
>> >> ?) be removed completely.
>> >>
>> >
>> > [Robert F]
>> > This would be the case in a GStreamer driven multimedia, i.e. Implement
>> > GStreamer elements using V4L2 directly(or camerabin using v4l2 directly).
>> > Perhaps some vendors would provide a library in between as well but that
>> > could be libv4l in that case. If someone would have an OpenMAX AL/IL media
>> > framework an OMX component would make sense to have but in this case it
>> > would be a thinner OMX component which in turn is implemented using V4L2.
>> > But it might be that Khronos provides OS independent components that by
>> > vendors gets implemented as the actual HW driver forgetting that there is a
>> > big difference in the driver model of an RTOS system compared to
>> > Linux(user/kernel space) or any OS...never mind.
>> >
>>
>> Not even different vendor's omx camera implementations are
>> compatible.. there seems to be too much various in ISP architecture
>> and features for this.
>>
>> Another point, and possibly the reason that TI went the OMX camera
>> route, was that a userspace API made it possible to move the camera
>> driver all to a co-processor (with advantages of reduced interrupt
>> latency for SIMCOP processing, and a larger part of the code being OS
>> independent)..  doing this in a kernel mode driver would have required
>> even more of syslink in the kernel.
>>
>> But maybe it would be nice to have a way to have sensor driver on the
>> linux side, pipelined with hw and imaging drivers on a co-processor
>> for various algorithms and filters with configuration all exposed to
>> userspace thru MCF.. I'm not immediately sure how this would work, but
>> it sounds nice at least ;-)
>
> MCF? What does that stand for?
>
>>
>> > The question is if the Linux kernel and V4L2 is ready to incorporate several
>> > HW(DSP, CPU, ISP, xxHW) in an imaging pipeline for instance. The reason
>> > Embedded Vendors provide custom solutions is to implement low power non(or
>> > minimal) CPU intervention pipelines where dedicated HW does the work most of
>> > the time(like full screen Video Playback).
>> >
>> > A common way of managing memory would of course also be necessary as well,
>> > like hwmem(search for hwmem in Linux-mm) handles to pass buffers in between
>> > different drivers and processes all the way from sources(camera, video
>> > parser/decode) to sinks(display, hdmi, video encoders(record))
>>
>> (ahh, ok, you have some of the same thoughts as I do regarding sharing
>> buffers between various drivers)
>
> Perhaps the time is right for someone to start working on this?
>
> Regards,
>
>        Hans
>
>> > Perhaps GStreamer experts would like to comment on the future plans ahead
>> > for zero copying/IPC and low power HW use cases? Could Gstreamer adapt some
>> > ideas from OMX IL making OMX IL obsolete?
>>
>> perhaps OMX should adapt some of the ideas from GStreamer ;-)
>>
>> OpenMAX is missing some very obvious stuff to make it an API for
>> portable applications like autoplugging, discovery of
>> capabilities/formats supported, etc..  at least with gst I can drop in
>> some hw specific plugins and have apps continue to work without code
>> changes.
>>
>> Anyways, it would be an easier argument to make if GStreamer was the
>> one true framework across different OSs, or at least across linux and
>> android.
>>
>> BR,
>> -R
>>
>> > Answering these questions could be
>> > improved guidelines on what embedded device vendors in the future would
>> > provide as hw-driver front-ends. OMX is just one of these. Perhaps we could
>> > do better to fit and evolve the Linux eco-system?
>> >
>> >
>> >>
>> >> > Regarding using V4L to communicate with DSPs/other processors: that too
>> >> > could be something for Linaro to pick up: experiment with it for one
>> >> > particular board, see what (if anything) is needed to make this work. I
>> >> > expect it to be pretty easy, but again, nobody has actually done the
>> >> > initial work.
>> >>
>> >> The main issue with the V4L2 API compared with the OMX API is that V4L2 is
>> >> a
>> >> kernelspace/userspace API only, while OMX can live in userspace. When the
>> >> need
>> >> to communicate with other processors (CPUs, DSP, dedicated image
>> >> processing
>> >> hardware blocks, ...) arises, platforms usually ship with a thin kernel
>> >> layer
>> >> to handle low-level communication protocols, and a userspace OMX library
>> >> that
>> >> does the bulk of the work. We would need to be able to do something
>> >> similar
>> >> with V4L2.
>> >
>> > [Robert F]
>> > Ok, doesn.t mediacontroller/subdevices solve many of these issues?
>> >
>> >>
>> >> > Once you have an example driver, then it should be much easier for
>> >> > others
>> >> > to follow.
>> >> >
>> >> > As Linus said, companies are unlikely to start doing this by themselves,
>> >> > but it seems that this work would exactly fit the Linaro purpose. From
>> >> > the
>> >> > Linaro homepage:
>> >> >
>> >> > "Linaro™ brings together the open source community and the electronics
>> >> > industry to work on key projects, deliver great tools, reduce industry
>> >> > wide fragmentation and provide common foundations for Linux software
>> >> > distributions and stacks to land on."
>> >> >
>> >> > Spot on, I'd say :-)
>> >> >
>> >> > Just for the record, let me say again they the V4L2 community will be
>> >> > very
>> >> > happy to assist with this when it comes to extending/improving the V4L2
>> >> > API
>> >> > to make all this possible.
>> >>
>> >> The first step would probably be to decide what Linux needs. Then I'll
>> >> also be
>> >> happy to assist with the implementation phase :-)
>> >>
>> >> --
>> >> Regards,
>> >>
>> >> Laurent Pinchart
>> >>
>> >> _______________________________________________
>> >> linaro-dev mailing list
>> >> linaro-dev@xxxxxxxxxxxxxxxx
>> >> http://lists.linaro.org/mailman/listinfo/linaro-dev
>> >
>> > BR
>> > /Robert Fekete
>> >
>> >
>> > _______________________________________________
>> > linaro-dev mailing list
>> > linaro-dev@xxxxxxxxxxxxxxxx
>> > http://lists.linaro.org/mailman/listinfo/linaro-dev
>> >
>> >
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-media" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>>
>
> --
> Hans Verkuil - video4linux developer - sponsored by Cisco
> --
> To unsubscribe from this list: send the line "unsubscribe linux-media" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
--
To unsubscribe from this list: send the line "unsubscribe linux-media" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux Input]     [Video for Linux]     [Gstreamer Embedded]     [Mplayer Users]     [Linux USB Devel]     [Linux Audio Users]     [Linux Kernel]     [Linux SCSI]     [Yosemite Backpacking]
  Powered by Linux