On Sun, Mar 15, 2015 at 02:50:27PM +0800, Linhaifeng wrote: > Hi,Michael > > I had tested the start_xmit function by the follow code found that the tx queue's state is stopped and can't send any packets anymore. Why don't you Cc all maintainers on this email? Pls check the file MAINTAINERS for the full list. I added Cc for now. > > static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev) > { > ... ... > > > capacity = 10; //########## test code : force to call netif_stop_queue > > if (capacity < 2+MAX_SKB_FRAGS) { > netif_stop_queue(dev); So you changed code to make it think we are out of capacity, now it stops the queue. > > if (unlikely(!virtqueue_enable_cb_delayed(vi->svq))) { > /* More just got used, free them then recheck. */ > capacity += free_old_xmit_skbs(vi); > dev_warn(&dev->dev, "free_old_xmit_skbs capacity =%d MAX_SKB_FRAGS=%d", capacity, MAX_SKB_FRAGS); > > capacity = 10; //########## test code : force not to call netif_start_queue > > if (capacity >= 2+MAX_SKB_FRAGS) { > netif_start_queue(dev); > virtqueue_disable_cb(vi->svq); > } else { > //########## OTOH if often enter this branch tx queue maybe stopped. > } and changed it here so it won't restart queue if host consumed all buffers. unsurprisingly this makes driver not work. > > } > > //########## Should we start queue here? I found that sometimes skb_xmit_done run before netif_stop_queue if this occurred the queue's state is > //########## stopped and have to reload virtio-net module to restore network. With or without your changes? Is this the condition you describe? if (sq->vq->num_free < 2+MAX_SKB_FRAGS) { ---> at this point, skb_xmit_done runs. this does: /* Suppress further interrupts. */ virtqueue_disable_cb(vq); /* We were probably waiting for more output buffers. */ netif_wake_subqueue(vi->dev, vq2txq(vq)); ---> netif_stop_subqueue(dev, qnum); ---> queue is now stopped if (unlikely(!virtqueue_enable_cb_delayed(sq->vq))) { ----> this re-enables interrupts, after an interrupt skb_xmit_done will run again. /* More just got used, free them then recheck. * */ free_old_xmit_skbs(sq); if (sq->vq->num_free >= 2+MAX_SKB_FRAGS) { netif_start_subqueue(dev, qnum); virtqueue_disable_cb(sq->vq); } } } I can't see a race condition from your description above. > } > > } > > ping 9.62.1.2 -i 0.1 > 64 bytes from 9.62.1.2: icmp_seq=19 ttl=64 time=0.115 ms > 64 bytes from 9.62.1.2: icmp_seq=20 ttl=64 time=0.101 ms > 64 bytes from 9.62.1.2: icmp_seq=21 ttl=64 time=0.094 ms > 64 bytes from 9.62.1.2: icmp_seq=22 ttl=64 time=0.098 ms > 64 bytes from 9.62.1.2: icmp_seq=23 ttl=64 time=0.097 ms > 64 bytes from 9.62.1.2: icmp_seq=24 ttl=64 time=0.095 ms > 64 bytes from 9.62.1.2: icmp_seq=25 ttl=64 time=0.095 ms > .... > ping: sendmsg: No buffer space available > ping: sendmsg: No buffer space available > ping: sendmsg: No buffer space available > ping: sendmsg: No buffer space available > ping: sendmsg: No buffer space available > ping: sendmsg: No buffer space available > .... > > -- > Regards, > Haifeng I can't say what does your code-changing experiment show. It might be better to introduce delay by calling something like cpu_relax at specific points (maybe multiple times in a loop). -- MST _______________________________________________ Virtualization mailing list Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx https://lists.linuxfoundation.org/mailman/listinfo/virtualization