On 02/05/2013 01:29 PM, Linus Walleij wrote:
On Tue, Feb 5, 2013 at 5:47 PM, Mark Brown
<broonie@xxxxxxxxxxxxxxxxxxxxxxxxxxx> wrote:
On Tue, Feb 05, 2013 at 05:21:48PM +0100, Linus Walleij wrote:
For IRQ mode, use the completion callback to push each cookie
to NAPI, and thus let the IRQ drive the traffic.
The whole purpose of NAPI is to avoid taking interrupts for completion
of transfers. Anything that generates interrupts when NAPI is in
polling mode is defeating the point.
So what I was trying to get across is that when you're in polling
mode you do not set DMA_PREP_INTERRUPT on your transfers,
just throw the obtained struct dma_async_tx_descriptor on some
list and then when polling use dma_async_is_tx_complete()
on the channel and the cookie inside the descriptor.
I was trying to describe that you can move from
IRQ mode to polling mode and back again by selectively
choosing to set/not set the DMA_PREP_INTERRUPT flag.
This does not work. At prep/submit time, the network driver does not
know if a particular packet buffer needs an interrupt or not. It queues
up a whole bunch of receive buffers upfront. These buffers simply sit
on the hardware queue/ring until the NIC receives traffic. The driver
throttles the receive processing rate by dynamically switching between
interrupt and poll behaviors on the completion side, not on the
submission side.
If polling is all you want you never set it.
Another point here is that it is not simply a polling vs. interrupt
problem. The system needs to dynamically switch between the two
behaviors depending on offered load conditions. This dynamic switching
is key to balancing latency (under low load) and throughput (at high
rates). It cannot be one or the other, it must be both.
Once again, I'm fairly sure that suitably reworking or extending the
dma-engine interfaces will allow network DMAs to fit in nicely.
However, I'd also appreciate inputs on the alternative approach of using
virtio devices as an interface to packet oriented DMA hardware. From my
perspective this offers the following advantages, some of which I've
already mentioned in earlier postings:
1. The virtqueue interface is nice and clean, it fits very well with
networking concepts such as NAPI. In comparison, the dma-engine API
will need extensions and/or rework to fit network DMAs.
2. Advantages from leveraging the virtio infrastructure. For example,
if a DMA pipe to a remote processor is exposed as a virtio device,
something like rpmsg could very naturally fit on top of this without any
other added glue.
3. Advantages from leveraging the driver model for binding dma clients
to dma hardware, instead of resorting to name lookups and such.
Thanks
-- Cyril.
--
To unsubscribe from this list: send the line "unsubscribe linux-doc" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html