Hi, I have a traffic policing setup running on Linux, serving about 800 mbps of traffic. Due to the traffic growth I decided to employ network interface bonding to scale over a single GigE. The Sun X4150 server has 2xIntel E5450 QuadCore CPUs and a total of four built-in e1000e interfaces, which I grouped into two bond interfaces. With kernel 2.6.23.1, everything works fine, but the system locked up after a few days. With kernel 2.6.28.7, I get 10-20% packet loss. I get packet loss as soon as I put a classful qdisc, even prio, without even having any classes or filters. TC prio statistics report lots of drops, around 10k per sec. With exactly the same setup on 2.6.23, the number of drops is only 50 per sec. On both kernels, the system is running with approximately 70% idle CPU. The network interrupts are distributed accross the cores. I thought it was a e1000e driver issue, but tweaking e1000e ring buffers didn't help. I tried using e1000 on 2.6.28 by adding necessary PCI IDs, I tried running on a different server with bnx cards, I tried disabling NO_HZ and HRTICK, but still I have the same problem. However, if I don't utilize bond, but just apply rules on normal ethX interfaces, there is no packet loss with 2.6.28.7. So, the problem appears only when I use 2.6.28.7 + bond + tc combination. Any ideas ? -- Best Regards, Vladimir Ivashchenko Chief Technology Officer PrimeTel PLC, Cyprus - www.prime-tel.com Tel: +357 25 100100 Fax: +357 2210 2211 -- To unsubscribe from this list: send the line "unsubscribe linux-net" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html