On 2018/9/13 20:44, Eric Dumazet wrote: > On Thu, Sep 13, 2018 at 5:32 AM Greg KH <gregkh@xxxxxxxxxxxxxxxxxxxx> wrote: >> >> On Thu, Aug 16, 2018 at 05:24:09PM +0200, Greg KH wrote: >>> On Thu, Aug 16, 2018 at 02:33:56PM +0200, Michal Kubecek wrote: >>>> On Thu, Aug 16, 2018 at 08:05:50PM +0800, maowenan wrote: >>>>> On 2018/8/16 19:39, Michal Kubecek wrote: >>>>>> >>>>>> I suspect you may be doing something wrong with your tests. I checked >>>>>> the segmentsmack testcase and the CPU utilization on receiving side >>>>>> (with sending 10 times as many packets as default) went down from ~100% >>>>>> to ~3% even when comparing what is in stable 4.4 now against older 4.4 >>>>>> kernel. >>>>> >>>>> There seems no obvious problem when you send packets with default >>>>> parameter in Segmentsmack POC, Which is also very related with your >>>>> server's hardware configuration. Please try with below parameter to >>>>> form OFO packets >>>> >>>> I did and even with these (questionable, see below) changes, I did not >>>> get more than 10% (of one core) by receiving ksoftirqd. >>>> >>>>> for (i = 0; i < 1024; i++) // 128->1024 >>>> ... >>>>> usleep(10*1000); // Adjust this and packet count to match the target!, sleep 100ms->10ms >>>> >>>> The comment in the testcase source suggests to do _one_ of these two >>>> changes so that you generate 10 times as many packets as the original >>>> testcase. You did both so that you end up sending 102400 packets per >>>> second. With 55 byte long packets, this kind of attack requires at least >>>> 5.5 MB/s (44 Mb/s) of throughput. This is no longer a "low packet rate >>>> DoS", I'm afraid. >>>> >>>> Anyway, even at this rate, I only get ~10% of one core (Intel E5-2697). >>>> >>>> What I can see, though, is that with current stable 4.4 code, modified >>>> testcase which sends something like >>>> >>>> 2:3, 3:4, ..., 3001:3002, 3003:3004, 3004:3005, ... 6001:6002, ... >>>> >>>> I quickly eat 6 MB of memory for receive queue of one socket while >>>> earlier 4.4 kernels only take 200-300 KB. I didn't test latest 4.4 with >>>> Takashi's follow-up yet but I'm pretty sure it will help while >>>> preserving nice performance when using the original segmentsmack >>>> testcase (with increased packet ratio). >>> >>> Ok, for now I've applied Takashi's fix to the 4.4 stable queue and will >>> push out a new 4.4-rc later tonight. Can everyone standardize on that >>> and test and let me know if it does, or does not, fix the reported >>> issues? >>> >>> If not, we can go from there and evaluate this much larger patch series. >>> But let's try the simple thing first. >> >> So, is the issue still present on the latest 4.4 release? Has anyone >> tested it? If not, I'm more than willing to look at backported patches, >> but I want to ensure that they really are needed here. >> >> thanks, > > Honestly, TCP stack without rb-tree for the OOO queue is vulnerable, > even with non malicious sender, > but with big enough TCP receive window and a not favorable network. > > So a malicious peer can definitely send packets needed to make TCP > stack behave in O(N), which is pretty bad if N is big... > > 9f5afeae51526b3ad7b7cb21ee8b145ce6ea7a7a ("tcp: use an RB tree for ooo > receive queue") > was proven to be almost bug free [1], and should be backported if possible. > > [1] bug fixed : > 76f0dcbb5ae1a7c3dbeec13dd98233b8e6b0b32a tcp: fix a stale ooo_last_skb > after a replace Thank you for Eric's suggestion, I will do some work to backport them. > > . >