Yes sort of. I do have an inconsistent pg for a while, but it is on a different pool. But I take it this is related to a networking issue I currently have with rsync and broken pipe. Where exactly does it go wrong? The cephfs kernel clients is sending a request to the osd, but the osd never replies? > > >> >> >> I got one again >> >> [<ffffffff81183503>] wait_on_page_bit_killable+0x83/0xa0 >> [<ffffffff811835d2>] __lock_page_or_retry+0xb2/0xc0 >> [<ffffffff81183997>] filemap_fault+0x3b7/0x410 >> [<ffffffffa055ce9c>] ceph_filemap_fault+0x13c/0x310 [ceph] >> [<ffffffff811ac84c>] __do_fault+0x4c/0xc0 >> [<ffffffff811acce3>] do_read_fault.isra.42+0x43/0x130 >> [<ffffffff811b1471>] handle_mm_fault+0x6b1/0x1040 >> [<ffffffff81692c04>] __do_page_fault+0x154/0x450 >> [<ffffffff81692f35>] do_page_fault+0x35/0x90 >> [<ffffffff8168f148>] page_fault+0x28/0x30 >> [<ffffffffffffffff>] 0xffffffffffffffff >> >> > >This is likely caused by hang osd request, was you cluster health? > > >> >check /proc/<stuck process>/stack to find where it is stuck >> > >> >> >> >> >> >> I have a process stuck in D+ writing to cephfs kernel mount. >> Anything >> >> can be done about this? (without rebooting) >> >> >> >> >> >> CentOS Linux release 7.5.1804 (Core) >> >> Linux 3.10.0-514.21.2.el7.x86_64 >> >> >> >> _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com