On 02/04/2013 12:02 PM, Felipe Balbi wrote:
Hi,
On Mon, Feb 04, 2013 at 08:54:17PM +0300, Sergei Shtylyov wrote:
On Mon, Feb 04, 2013 at 08:36:38PM +0300, Sergei Shtylyov wrote:
opted out of it. From the top of my head we have CPPI 3.x, CPPI 4.1,
Inventra DMA, OMAP sDMA and ux500 DMA engines supported by the driver.
Granted, CPPI 4.1 makes some assumptions about the fact that it's
handling USB tranfers,
What CPPI 4.1 code makes this assumptions? MUSB DMA driver? Then it's just
HW makes the asumptions
Not true at all. There is a separate set of registers (at offset 0) which
copes with USB specifics, but CPPI 4.1 itself doesn't know about USB anything.
CPPI 4.1 was made for USB transfers.
I have been dealing with CPPI hardware on KeyStone platforms (CPPI 4.2).
Our experiences with this DMA hardware may help with CPPI 4.1 on
earlier generations.
CPPI 4.2 serves as a truly common interface to a number of hardware
blocks on KeyStone SoCs - including Ethernet, RapidIO, Crypto
accelerators, and a bunch of other accelerator thingies. Given the
commonality across subsystems, we've built a shared CPPI 4.2 DMA-Engine
implementation. You can take a sneak peek at this implementation at [1].
Based on our experience with fitting multiple subsystems on top of this
DMA-Engine driver, I must say that the DMA-Engine interface has proven
to be a less than ideal fit for the network driver use case.
The first problem is that the DMA-Engine interface expects to "push"
completed traffic up into the upper layer as a part of its callback.
This doesn't fit cleanly with NAPI, which expects to "pull" completed
traffic from below in the NAPI poll. We've somehow kludged together a
solution around this, but it isn't very elegant.
The second problem is one of binding fixed DMA resources to fixed users.
AFAICT, the stock DMA-Engine mechanism works best when one DMA
resource is as good as any other. To get over this problem, we've added
support for named channels, and drivers specifically request for a DMA
resource by name. Again, this is less than ideal.
We found that virtio devices offer a more elegant solution to this
problem. First, the virtqueue interface is a much better fit into NAPI
(callback --> napi schedule, napi poll --> get_buf), and this eliminates
the need for aforementioned kludges in the code. Second, the virtio
device infrastructure nicely uses the device model to solve the problem
of binding DMA users to specific DMA resources.
These patches haven't (yet) been posted on the MLs, but you can peek at
[2]. While we are on the topic, I'd certainly appreciate feedback on
the concept of using virtqueues as an interface to peripheral
independent packet oriented DMA hardware. :-)
Cheers
-- Cyril
[1] -
http://arago-project.org/git/projects/?p=linux-keystone.git;a=shortlog;h=refs/heads/rebuild/23-drivers-dmaengine
[2] -
http://arago-project.org/git/projects/?p=linux-keystone.git;a=shortlog;h=refs/heads/rebuild/21-drivers-virtio
--
To unsubscribe from this list: send the line "unsubscribe linux-doc" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html