On Mon, 2020-04-06 at 11:09 +0200, Jesper Krogh wrote: > home-directory style - median is about 3KB - but varies greatly. > > I also get this: > [41204.865818] INFO: task kworker/u16:102:21903 blocked for more than > 120 seconds. > [41204.865955] Not tainted 5.4.30 #5 > [41204.866006] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > disables this message. > [41204.866056] kworker/u16:102 D 0 21903 2 0x80004000 > [41204.866119] Workqueue: ceph-inode ceph_inode_work [ceph] > [41204.866120] Call Trace: > [41204.866156] __schedule+0x45f/0x710 > [41204.866162] ? xas_store+0x391/0x5f0 > [41204.866164] schedule+0x3e/0xa0 > [41204.866166] io_schedule+0x16/0x40 > [41204.866180] __lock_page+0x12a/0x1d0 > [41204.866182] ? file_fdatawait_range+0x30/0x30 > [41204.866187] truncate_inode_pages_range+0x52c/0x980 > [41204.866191] ? syscall_return_via_sysret+0x12/0x7f > [41204.866197] ? drop_inode_snap_realm+0x98/0xa0 [ceph] > [41204.866207] ? fsnotify_grab_connector+0x4d/0x90 > [41204.866209] truncate_inode_pages_final+0x4c/0x60 > [41204.866214] ceph_evict_inode+0x2d/0x210 [ceph] > [41204.866219] evict+0xca/0x1a0 > [41204.866221] iput+0x1ba/0x210 > [41204.866225] ceph_inode_work+0x40/0x270 [ceph] > [41204.866232] process_one_work+0x167/0x400 > [41204.866233] worker_thread+0x4d/0x460 > [41204.866236] kthread+0x105/0x140 > [41204.866237] ? rescuer_thread+0x370/0x370 > [41204.866239] ? kthread_destroy_worker+0x50/0x50 > [41204.866240] ret_from_fork+0x35/0x40 > This task is attempting to evict an inode from the cache, but one of its pagecache pages is locked (probably due to some other active user of it). Note that just above the __lock_page call is the call to io_schedule. That means that the task will be put to sleep while waiting for the page lock, and that sleep will be counted as I/O wait (which contributes to the load average). > On Mon, Apr 6, 2020 at 10:53 AM Yan, Zheng <ukernel@xxxxxxxxx> wrote: > > On Mon, Apr 6, 2020 at 4:06 PM Jesper Krogh <jesper.krogh@xxxxxxxxx> wrote: > > > This is a CephFS client - its only purpose is to run the "filedaemon" of bacula > > > and transport data to the tape-library from CephFS - below 2 threads is > > > essentially doing something equivalent to > > > > > > find /cephfs/ -type f | xargs cat | nc server > > > > > > 2 threads only, load exploding and the "net read vs net write" has > > > more than 100x difference. > > > > > > Can anyone explain this as "normal" behaviour? > > > Server is a VM with 16 "vCPU" and 16GB memory running libvirt/qemu > > > > > > jk@wombat:~$ w > > > 07:50:33 up 11:25, 1 user, load average: 206.43, 76.23, 50.58 > > > USER TTY FROM LOGIN@ IDLE JCPU PCPU WHAT > > > jk pts/0 10.194.133.42 06:54 0.00s 0.05s 0.00s w > > > jk@wombat:~$ dstat -ar > > > --total-cpu-usage-- -dsk/total- -net/total- ---paging-- ---system-- --io/total- > > > usr sys idl wai stl| read writ| recv send| in out | int csw | read writ > > > 0 0 98 1 0| 14k 34k| 0 0 | 3B 27B| 481 294 |0.55 0.73 > > > 1 1 0 98 0| 0 0 | 60M 220k| 0 0 |6402 6182 | 0 0 > > > 0 1 0 98 0| 0 0 | 69M 255k| 0 0 |7305 4339 | 0 0 > > > 1 2 0 98 0| 0 0 | 76M 282k| 0 0 |7914 4886 | 0 0 > > > 1 1 0 99 0| 0 0 | 70M 260k| 0 0 |7293 4444 | 0 0 > > > 1 1 0 98 0| 0 0 | 80M 278k| 0 0 |8018 4931 | 0 0 > > > 0 1 0 98 0| 0 0 | 60M 221k| 0 0 |6435 5951 | 0 0 > > > 0 1 0 99 0| 0 0 | 59M 211k| 0 0 |6163 3584 | 0 0 > > > 0 1 0 98 0| 0 0 | 64M 323k| 0 0 |6653 3881 | 0 0 > > > 1 0 0 99 0| 0 0 | 61M 243k| 0 0 |6822 4401 | 0 0 > > > 0 1 0 99 0| 0 0 | 55M 205k| 0 0 |5975 3518 | 0 0 > > > 1 1 0 98 0| 0 0 | 68M 242k| 0 0 |7094 6544 | 0 0 > > > 0 1 0 99 0| 0 0 | 58M 230k| 0 0 |6639 4178 | 0 0 > > > 1 2 0 98 0| 0 0 | 61M 243k| 0 0 |7117 4477 | 0 0 > > > 0 1 0 99 0| 0 0 | 61M 228k| 0 0 |6500 4078 | 0 0 > > > 0 1 0 99 0| 0 0 | 65M 234k| 0 0 |6595 3914 | 0 0 > > > 0 1 0 98 0| 0 0 | 64M 219k| 0 0 |6507 5755 | 0 0 > > > 1 1 0 99 0| 0 0 | 64M 233k| 0 0 |6869 4153 | 0 0 > > > 1 2 0 98 0| 0 0 | 63M 232k| 0 0 |6632 3907 | > > > 0 0 ^C > > > jk@wombat:~$ w > > > 07:50:56 up 11:25, 1 user, load average: 221.35, 88.07, 55.02 > > > USER TTY FROM LOGIN@ IDLE JCPU PCPU WHAT > > > jk pts/0 10.194.133.42 06:54 0.00s 0.05s 0.00s w > > > jk@wombat:~$ > > > > > > Thanks. > > > > how small these files are? -- Jeff Layton <jlayton@xxxxxxxxxx>