On Tue, Jul 9, 2013 at 3:58 AM, Milosz Tanski <milosz@xxxxxxxxx> wrote: > Yan, > > Actually after playing some more today I have another one of my > clients stuck in this spot. When I look at the kernel stacks this is > what I see for all the threads: > > [<ffffffffa02d2bab>] ceph_mdsc_do_request+0xcb/0x1a0 [ceph] > [<ffffffffa02c018f>] ceph_do_getattr+0xdf/0x120 [ceph] > [<ffffffffa02c01f4>] ceph_getattr+0x24/0x100 [ceph] > [<ffffffff811775fd>] vfs_getattr+0x4d/0x80 > [<ffffffff8117784d>] vfs_fstat+0x3d/0x70 > [<ffffffff81177895>] SYSC_newfstat+0x15/0x30 > [<ffffffff8117794e>] SyS_newfstat+0xe/0x10 > [<ffffffff8155dd59>] system_call_fastpath+0x16/0x1b > [<ffffffffffffffff>] 0xffffffffffffffff > > > Anything I can do on my end to debug this issue? > find the hang request (and inode) through /sys/kernel/debug/ceph/xxx/mdsc, use 'ceph mds tell \* dumpcache' to dump mds cache. open /cachedump.xxx and check the inode's state. does your kernel include all fixes in testing branch of ceph-client ? does restarting the mds resolve the hang ? yan, zheng -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html