Quick and easy depends on your network infrastructure. Sometimes it is difficult or impossible to retrofit a live cluster without disruption. > On May 25, 2020, at 1:03 AM, Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx> wrote: > > > I am interested. I am always setting mtu to 9000. To be honest I cannot > imagine there is no optimization since you have less interrupt requests, > and you are able x times as much data. Every time there something > written about optimizing the first thing mention is changing to the mtu > 9000. Because it is quick and easy win. > > > > > -----Original Message----- > From: Dave Hall [mailto:kdhall@xxxxxxxxxxxxxx] > Sent: maandag 25 mei 2020 5:11 > To: Martin Verges; Suresh Rama > Cc: Amudhan P; Khodayar Doustar; ceph-users > Subject: Re: [External Email] Re: Ceph Nautius not working > after setting MTU 9000 > > All, > > Regarding Martin's observations about Jumbo Frames.... > > I have recently been gathering some notes from various internet sources > regarding Linux network performance, and Linux performance in general, > to be applied to a Ceph cluster I manage but also to the rest of the > Linux server farm I'm responsible for. > > In short, enabling Jumbo Frames without also tuning a number of other > kernel and NIC attributes will not provide the performance increases > we'd like to see. I have not yet had a chance to go through the rest of > the testing I'd like to do, but I can confirm (via iperf3) that only > enabling Jumbo Frames didn't make a significant difference. > > Some of the other attributes I'm referring to are incoming and outgoing > buffer sizes at the NIC, IP, and TCP levels, interrupt coalescing, NIC > offload functions that should or shouldn't be turned on, packet queuing > disciplines (tc), the best choice of TCP slow-start algorithms, and > other TCP features and attributes. > > The most off-beat item I saw was something about adding IPTABLES rules > to bypass CONNTRACK table lookups. > > In order to do anything meaningful to assess the effect of all of these > settings I'd like to figure out how to set them all via Ansible - so > more to learn before I can give opinions. > > --> If anybody has added this type of configuration to Ceph Ansible, > I'd be glad for some pointers. > > I have started to compile a document containing my notes. It's rough, > but I'd be glad to share if anybody is interested. > > -Dave > > Dave Hall > Binghamton University > >> On 5/24/2020 12:29 PM, Martin Verges wrote: >> >> Just save yourself the trouble. You won't have any real benefit from > MTU >> 9000. It has some smallish, but it is not worth the effort, problems, > and >> loss of reliability for most environments. >> Try it yourself and do some benchmarks, especially with your regular >> workload on the cluster (not the maximum peak performance), then drop > the >> MTU to default ;). >> >> Please if anyone has other real world benchmarks showing huge > differences >> in regular Ceph clusters, please feel free to post it here. >> >> -- >> Martin Verges >> Managing director >> >> Mobile: +49 174 9335695 >> E-Mail: martin.verges@xxxxxxxx >> Chat: https://t.me/MartinVerges >> >> croit GmbH, Freseniusstr. 31h, 81247 Munich >> CEO: Martin Verges - VAT-ID: DE310638492 >> Com. register: Amtsgericht Munich HRB 231263 >> >> Web: https://croit.io >> YouTube: https://goo.gl/PGE1Bx >> >> >>> Am So., 24. Mai 2020 um 15:54 Uhr schrieb Suresh Rama >> <sstkadu@xxxxxxxxx>: >> >>> Ping with 9000 MTU won't get response as I said and it should be > 8972. Glad >>> it is working but you should know what happened to avoid this issue > later. >>> >>>> On Sun, May 24, 2020, 3:04 AM Amudhan P <amudhan83@xxxxxxxxx> wrote: >>> >>>> No, ping with MTU size 9000 didn't work. >>>> >>>> On Sun, May 24, 2020 at 12:26 PM Khodayar Doustar > <doustar@xxxxxxxxxxxx> >>>> wrote: >>>> >>>>> Does your ping work or not? >>>>> >>>>> >>>>> On Sun, May 24, 2020 at 6:53 AM Amudhan P <amudhan83@xxxxxxxxx> > wrote: >>>>> >>>>>> Yes, I have set setting on the switch side also. >>>>>> >>>>>> On Sat 23 May, 2020, 6:47 PM Khodayar Doustar, > <doustar@xxxxxxxxxxxx> >>>>>> wrote: >>>>>> >>>>>>> Problem should be with network. When you change MTU it should be >>>> changed >>>>>>> all over the network, any single hup on your network should speak > and >>>>>>> accept 9000 MTU packets. you can check it on your hosts with >>> "ifconfig" >>>>>>> command and there is also equivalent commands for other >>>> network/security >>>>>>> devices. >>>>>>> >>>>>>> If you have just one node which it not correctly configured for > MTU >>>> 9000 >>>>>>> it wouldn't work. >>>>>>> >>>>>>> On Sat, May 23, 2020 at 2:30 PM sinan@xxxxxxxx <sinan@xxxxxxxx> >>> wrote: >>>>>>>> Can the servers/nodes ping eachother using large packet sizes? I >>> guess >>>>>>>> not. >>>>>>>> >>>>>>>> Sinan Polat >>>>>>>> >>>>>>>>> Op 23 mei 2020 om 14:21 heeft Amudhan P <amudhan83@xxxxxxxxx> > het >>>>>>>> volgende geschreven: >>>>>>>>> In OSD logs "heartbeat_check: no reply from OSD" >>>>>>>>> >>>>>>>>>> On Sat, May 23, 2020 at 5:44 PM Amudhan P > <amudhan83@xxxxxxxxx> >>>>>>>> wrote: >>>>>>>>>> Hi, >>>>>>>>>> >>>>>>>>>> I have set Network switch with MTU size 9000 and also in my >>> netplan >>>>>>>>>> configuration. >>>>>>>>>> >>>>>>>>>> What else needs to be checked? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> On Sat, May 23, 2020 at 3:39 PM Wido den Hollander < >>> wido@xxxxxxxx >>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>>> On 5/23/20 12:02 PM, Amudhan P wrote: >>>>>>>>>>>> Hi, >>>>>>>>>>>> >>>>>>>>>>>> I am using ceph Nautilus in Ubuntu 18.04 working fine wit > MTU >>>> size >>>>>>>> 1500 >>>>>>>>>>>> (default) recently i tried to update MTU size to 9000. >>>>>>>>>>>> After setting Jumbo frame running ceph -s is timing out. >>>>>>>>>>> Ceph can run just fine with an MTU of 9000. But there is >>> probably >>>>>>>>>>> something else wrong on the network which is causing this. >>>>>>>>>>> >>>>>>>>>>> Check the Jumbo Frames settings on all the switches as well > to >>>> make >>>>>>>> sure >>>>>>>>>>> they forward all the packets. >>>>>>>>>>> >>>>>>>>>>> This is definitely not a Ceph issue. >>>>>>>>>>> >>>>>>>>>>> Wido >>>>>>>>>>> >>>>>>>>>>>> regards >>>>>>>>>>>> Amudhan P >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> ceph-users mailing list -- ceph-users@xxxxxxx >>>>>>>>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> ceph-users mailing list -- ceph-users@xxxxxxx >>>>>>>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>>>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> ceph-users mailing list -- ceph-users@xxxxxxx >>>>>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>>>>>>> _______________________________________________ >>>>>>>> ceph-users mailing list -- ceph-users@xxxxxxx >>>>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>>>>>>> >>>> _______________________________________________ >>>> ceph-users mailing list -- ceph-users@xxxxxxx >>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>>> >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@xxxxxxx >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>> >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx