On Wed, May 04, 2011 at 08:29:44PM +0530, Krishna Kumar2 wrote: > "Michael S. Tsirkin" <mst@xxxxxxxxxx> wrote on 05/04/2011 08:16:22 PM: > > > > A. virtio: > > > - Provide a API to get available number of slots. > > > > > > B. virtio-net: > > > - Remove stop/start txq's and associated callback. > > > - Pre-calculate the number of slots needed to transmit > > > the skb in xmit_skb and bail out early if enough space > > > is not available. My testing shows that 2.5-3% of > > > packets are benefited by using this API. > > > - Do not drop skbs but instead return TX_BUSY like other > > > drivers. > > > - When returning EBUSY, set a per-txq variable to indicate > > > to dev_queue_xmit() whether to restart xmits on this txq. > > > > > > C. net/sched/sch_generic.c: > > > Since virtio-net now returns EBUSY, the skb is requeued to > > > gso_skb. This allows adding the addional check for restart > > > xmits in just the slow-path (the first re-queued packet > > > case of dequeue_skb, where it checks for gso_skb) before > > > deciding whether to call the driver or not. > > > > > > Patch was also tested between two servers with Emulex OneConnect > > > 10G cards to confirm there is no regression. Though the patch is > > > an attempt to improve only small packet performance, there was > > > improvement for 1K, 2K and also 16K both in BW and SD. Results > > > from Guest -> Remote Host (BW in Mbps) for 1K and 16K I/O sizes: > > > > > > ________________________________________________________ > > > I/O Size: 1K > > > # BW1 BW2 (%) SD1 SD2 (%) > > > ________________________________________________________ > > > 1 1226 3313 (170.2) 6.6 1.9 (-71.2) > > > 2 3223 7705 (139.0) 18.0 7.1 (-60.5) > > > 4 7223 8716 (20.6) 36.5 29.7 (-18.6) > > > 8 8689 8693 (0) 131.5 123.0 (-6.4) > > > 16 8059 8285 (2.8) 578.3 506.2 (-12.4) > > > 32 7758 7955 (2.5) 2281.4 2244.2 (-1.6) > > > 64 7503 7895 (5.2) 9734.0 9424.4 (-3.1) > > > 96 7496 7751 (3.4) 21980.9 20169.3 (-8.2) > > > 128 7389 7741 (4.7) 40467.5 34995.5 (-13.5) > > > ________________________________________________________ > > > Summary: BW: 16.2% SD: -10.2% > > > > > > ________________________________________________________ > > > I/O Size: 16K > > > # BW1 BW2 (%) SD1 SD2 (%) > > > ________________________________________________________ > > > 1 6684 7019 (5.0) 1.1 1.1 (0) > > > 2 7674 7196 (-6.2) 5.0 4.8 (-4.0) > > > 4 7358 8032 (9.1) 21.3 20.4 (-4.2) > > > 8 7393 8015 (8.4) 82.7 82.0 (-.8) > > > 16 7958 8366 (5.1) 283.2 310.7 (9.7) > > > 32 7792 8113 (4.1) 1257.5 1363.0 (8.3) > > > 64 7673 8040 (4.7) 5723.1 5812.4 (1.5) > > > 96 7462 7883 (5.6) 12731.8 12119.8 (-4.8) > > > 128 7338 7800 (6.2) 21331.7 21094.7 (-1.1) > > > ________________________________________________________ > > > Summary: BW: 4.6% SD: -1.5% > > > > > > Signed-off-by: Krishna Kumar <krkumar2@xxxxxxxxxx> > > > --- > > > > So IIUC, we delay transmit by an arbitrary value and hope > > that the host is done with the packets by then? > > Not "hope" exactly. If the device is not ready, then > the packet is requeued. The main idea is to avoid > drops/stop/starts, etc. Yes, I see that, definitely. I guess it's a win if the interrupt takes at least a jiffy to arrive anyway, and a loss if not. Is there some reason interrupts might be delayed until the next jiffy? > > Interesting. > > > > I am currently testing an approach where > > we tell the host explicitly to interrupt us only after > > a large part of the queue is empty. > > With 256 entries in a queue, we should get 1 interrupt per > > on the order of 100 packets which does not seem like a lot. > > > > I can post it, mind testing this? > > Sure. > > - KK Just posted. Would appreciate feedback. -- MST -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html