On Wed, May 04, 2011 at 07:32:58PM +0530, Krishna Kumar wrote: > Earlier approach to improving small packet performance went > along the lines of dropping packets when the txq is full to > avoid stop/start of the txq. Though performance improved > significantly (upto 3x) for a single thread, multiple netperf > sessions showed a regression of upto -17% (starting from 4 > sessions). > > This patch proposes a different approach with the following > changes: > > A. virtio: > - Provide a API to get available number of slots. > > B. virtio-net: > - Remove stop/start txq's and associated callback. > - Pre-calculate the number of slots needed to transmit > the skb in xmit_skb and bail out early if enough space > is not available. My testing shows that 2.5-3% of > packets are benefited by using this API. > - Do not drop skbs but instead return TX_BUSY like other > drivers. > - When returning EBUSY, set a per-txq variable to indicate > to dev_queue_xmit() whether to restart xmits on this txq. > > C. net/sched/sch_generic.c: > Since virtio-net now returns EBUSY, the skb is requeued to > gso_skb. This allows adding the addional check for restart > xmits in just the slow-path (the first re-queued packet > case of dequeue_skb, where it checks for gso_skb) before > deciding whether to call the driver or not. > > Patch was also tested between two servers with Emulex OneConnect > 10G cards to confirm there is no regression. Though the patch is > an attempt to improve only small packet performance, there was > improvement for 1K, 2K and also 16K both in BW and SD. Results > from Guest -> Remote Host (BW in Mbps) for 1K and 16K I/O sizes: > > ________________________________________________________ > I/O Size: 1K > # BW1 BW2 (%) SD1 SD2 (%) > ________________________________________________________ > 1 1226 3313 (170.2) 6.6 1.9 (-71.2) > 2 3223 7705 (139.0) 18.0 7.1 (-60.5) > 4 7223 8716 (20.6) 36.5 29.7 (-18.6) > 8 8689 8693 (0) 131.5 123.0 (-6.4) > 16 8059 8285 (2.8) 578.3 506.2 (-12.4) > 32 7758 7955 (2.5) 2281.4 2244.2 (-1.6) > 64 7503 7895 (5.2) 9734.0 9424.4 (-3.1) > 96 7496 7751 (3.4) 21980.9 20169.3 (-8.2) > 128 7389 7741 (4.7) 40467.5 34995.5 (-13.5) > ________________________________________________________ > Summary: BW: 16.2% SD: -10.2% > > ________________________________________________________ > I/O Size: 16K > # BW1 BW2 (%) SD1 SD2 (%) > ________________________________________________________ > 1 6684 7019 (5.0) 1.1 1.1 (0) > 2 7674 7196 (-6.2) 5.0 4.8 (-4.0) > 4 7358 8032 (9.1) 21.3 20.4 (-4.2) > 8 7393 8015 (8.4) 82.7 82.0 (-.8) > 16 7958 8366 (5.1) 283.2 310.7 (9.7) > 32 7792 8113 (4.1) 1257.5 1363.0 (8.3) > 64 7673 8040 (4.7) 5723.1 5812.4 (1.5) > 96 7462 7883 (5.6) 12731.8 12119.8 (-4.8) > 128 7338 7800 (6.2) 21331.7 21094.7 (-1.1) > ________________________________________________________ > Summary: BW: 4.6% SD: -1.5% > > Signed-off-by: Krishna Kumar <krkumar2@xxxxxxxxxx> > --- So IIUC, we delay transmit by an arbitrary value and hope that the host is done with the packets by then? Interesting. I am currently testing an approach where we tell the host explicitly to interrupt us only after a large part of the queue is empty. With 256 entries in a queue, we should get 1 interrupt per on the order of 100 packets which does not seem like a lot. I can post it, mind testing this? -- MST -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html