Pure speculation - could it be "readahead" that somehow ends up pulling data into the system that are never needed when dealing with small files? Any ways to demystify this pattern? On Thu, Apr 9, 2020 at 7:50 PM Jesper Krogh <jesper.krogh@xxxxxxxxx> wrote: > > Hi. I really dont know if this is related to the issue (page > allocation) or a separate. But It really puzzles me that I can see > dstat -ar output like this: > > Keep in mind that this is only a "network backup client" reading from > CephFS - ideally recv == send as it just "transports data through the > host. > > STARTS OUT OK. > --total-cpu-usage-- -dsk/total- -net/total- ---paging-- ---system-- --io/total- > usr sys idl wai stl| read writ| recv send| in out | int csw | read writ > 2 2 70 25 0| 0 0 | 100M 100M| 0 0 |7690 1944 | 0 0 > 4 2 74 19 0| 0 0 | 156M 154M| 0 0 | 12k 3942 | 0 0 > 4 2 70 24 0| 0 0 | 214M 127M| 0 0 | 12k 3892 | 0 0 > 4 2 65 29 0| 0 0 | 120M 163M| 0 0 |9763 2347 | 0 0 > 5 4 77 14 0| 0 0 | 216M 242M| 0 0 | 15k 4797 | 0 0 > HERE IT BALOONS > 3 14 20 63 0| 0 0 | 912M 5970k| 0 0 | 33k 16k| 0 0 > 2 14 1 83 0| 0 0 |1121M 4723k| 0 0 | 37k 14k| 0 0 > 3 16 3 78 0| 0 84k|1198M 8738k| 0 0 | 39k 15k| 0 4.00 > 3 14 14 69 0| 0 0 |1244M 5772k| 0 0 | 40k 14k| 0 0 > 2 12 15 71 0| 0 24k|1354M | 0 24k| 41k 8241 | 0 6.00 > 2 9 1 87 0| 0 0 |1271M 1540k| 0 0 | 38k 5887 | 0 0 > 2 7 0 90 0| 0 52k|1222M 1609k| 0 0 | 37k 6359 | 0 6.00 > 2 8 0 90 0| 0 96k|1260M 5676k| 0 0 | 39k 6589 | 0 20.0 > 2 6 0 92 0| 0 0 |1043M 3002k| 0 0 | 33k 6189 | 0 0 > 2 6 0 92 0| 0 0 | 946M 1223k| 0 0 | 30k 6080 | 0 0 > 2 6 0 92 0| 0 0 | 908M 5331k| 0 0 | 29k 9983 | 0 0 > 2 5 0 94 0| 0 0 | 773M 1067k| 0 0 | 26k 6691 | 0 0 > 2 4 0 94 0| 0 0 | 626M 3190k| 0 0 | 21k 5868 | 0 0 > 1 4 0 95 0| 0 0 | 505M 15M| 0 0 | 17k 4686 | 0 0 > and then it move back to normal.. > > But a pattern of 1000x more on the recieve side than send is really puzzling. > A VM on 25Gbit interconnect with all ceph nodes. > > On Mon, Apr 6, 2020 at 10:04 AM Jesper Krogh <jesper.krogh@xxxxxxxxx> wrote: > > > > This is a CephFS client - its only purpose is to run the "filedaemon" of bacula > > and transport data to the tape-library from CephFS - below 2 threads is > > essentially doing something equivalent to > > > > find /cephfs/ -type f | xargs cat | nc server > > > > 2 threads only, load exploding and the "net read vs net write" has > > more than 100x difference. > > > > Can anyone explain this as "normal" behaviour? > > Server is a VM with 16 "vCPU" and 16GB memory running libvirt/qemu > > > > jk@wombat:~$ w > > 07:50:33 up 11:25, 1 user, load average: 206.43, 76.23, 50.58 > > USER TTY FROM LOGIN@ IDLE JCPU PCPU WHAT > > jk pts/0 10.194.133.42 06:54 0.00s 0.05s 0.00s w > > jk@wombat:~$ dstat -ar > > --total-cpu-usage-- -dsk/total- -net/total- ---paging-- ---system-- --io/total- > > usr sys idl wai stl| read writ| recv send| in out | int csw | read writ > > 0 0 98 1 0| 14k 34k| 0 0 | 3B 27B| 481 294 |0.55 0.73 > > 1 1 0 98 0| 0 0 | 60M 220k| 0 0 |6402 6182 | 0 0 > > 0 1 0 98 0| 0 0 | 69M 255k| 0 0 |7305 4339 | 0 0 > > 1 2 0 98 0| 0 0 | 76M 282k| 0 0 |7914 4886 | 0 0 > > 1 1 0 99 0| 0 0 | 70M 260k| 0 0 |7293 4444 | 0 0 > > 1 1 0 98 0| 0 0 | 80M 278k| 0 0 |8018 4931 | 0 0 > > 0 1 0 98 0| 0 0 | 60M 221k| 0 0 |6435 5951 | 0 0 > > 0 1 0 99 0| 0 0 | 59M 211k| 0 0 |6163 3584 | 0 0 > > 0 1 0 98 0| 0 0 | 64M 323k| 0 0 |6653 3881 | 0 0 > > 1 0 0 99 0| 0 0 | 61M 243k| 0 0 |6822 4401 | 0 0 > > 0 1 0 99 0| 0 0 | 55M 205k| 0 0 |5975 3518 | 0 0 > > 1 1 0 98 0| 0 0 | 68M 242k| 0 0 |7094 6544 | 0 0 > > 0 1 0 99 0| 0 0 | 58M 230k| 0 0 |6639 4178 | 0 0 > > 1 2 0 98 0| 0 0 | 61M 243k| 0 0 |7117 4477 | 0 0 > > 0 1 0 99 0| 0 0 | 61M 228k| 0 0 |6500 4078 | 0 0 > > 0 1 0 99 0| 0 0 | 65M 234k| 0 0 |6595 3914 | 0 0 > > 0 1 0 98 0| 0 0 | 64M 219k| 0 0 |6507 5755 | 0 0 > > 1 1 0 99 0| 0 0 | 64M 233k| 0 0 |6869 4153 | 0 0 > > 1 2 0 98 0| 0 0 | 63M 232k| 0 0 |6632 3907 | > > 0 0 ^C > > jk@wombat:~$ w > > 07:50:56 up 11:25, 1 user, load average: 221.35, 88.07, 55.02 > > USER TTY FROM LOGIN@ IDLE JCPU PCPU WHAT > > jk pts/0 10.194.133.42 06:54 0.00s 0.05s 0.00s w > > jk@wombat:~$ > > > > Thanks.