On Wed, Jan 23, 2019 at 5:50 AM Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx> wrote: > > > I got one again > > [<ffffffff81183503>] wait_on_page_bit_killable+0x83/0xa0 > [<ffffffff811835d2>] __lock_page_or_retry+0xb2/0xc0 > [<ffffffff81183997>] filemap_fault+0x3b7/0x410 > [<ffffffffa055ce9c>] ceph_filemap_fault+0x13c/0x310 [ceph] > [<ffffffff811ac84c>] __do_fault+0x4c/0xc0 > [<ffffffff811acce3>] do_read_fault.isra.42+0x43/0x130 > [<ffffffff811b1471>] handle_mm_fault+0x6b1/0x1040 > [<ffffffff81692c04>] __do_page_fault+0x154/0x450 > [<ffffffff81692f35>] do_page_fault+0x35/0x90 > [<ffffffff8168f148>] page_fault+0x28/0x30 > [<ffffffffffffffff>] 0xffffffffffffffff > > This is likely caused by hang osd request, was you cluster health? > >check /proc/<stuck process>/stack to find where it is stuck > > > >> > >> > >> I have a process stuck in D+ writing to cephfs kernel mount. > Anything > >> can be done about this? (without rebooting) > >> > >> > >> CentOS Linux release 7.5.1804 (Core) > >> Linux 3.10.0-514.21.2.el7.x86_64 > >> > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com