Hi, I'm running into a issue in combination of CephFS and Samba and I was wondering if a dev knew what is happening here. The situation: - Jewel cluster - CephFS kernel client version 4.7 - Samba re-export of CephFS - Mount options: rw,noatime,acl A copy of a 15GB file results in Samba hanging in status D: root@hlms-zaken-01:~# ps aux|grep smb|grep D jongh 8887 0.0 0.0 376656 19068 ? D 14:42 0:00 /usr/sbin/smbd -D jongh 9740 0.0 0.0 377380 19244 ? D 14:49 0:00 /usr/sbin/smbd -D root@hlms-zaken-01:~# cat /proc/8887/stack [<ffffffff8132d353>] call_rwsem_down_write_failed+0x13/0x20 [<ffffffff8121a145>] vfs_setxattr+0x55/0xb0 [<ffffffff8121a2a5>] setxattr+0x105/0x170 [<ffffffff81203aa1>] filename_lookup+0xf1/0x180 [<ffffffff8120369f>] getname_flags+0x6f/0x1e0 [<ffffffff8121a3bd>] path_setxattr+0xad/0xe0 [<ffffffff8121a4f0>] SyS_setxattr+0x10/0x20 [<ffffffff815e8b76>] entry_SYSCALL_64_fastpath+0x1e/0xa8 [<ffffffffffffffff>] 0xffffffffffffffff root@hlms-zaken-01:~# cat /proc/9740/stack [<ffffffff8132d353>] call_rwsem_down_write_failed+0x13/0x20 [<ffffffff8121a145>] vfs_setxattr+0x55/0xb0 [<ffffffff8121a2a5>] setxattr+0x105/0x170 [<ffffffff81203aa1>] filename_lookup+0xf1/0x180 [<ffffffff8120369f>] getname_flags+0x6f/0x1e0 [<ffffffff8121a3bd>] path_setxattr+0xad/0xe0 [<ffffffff8121a4f0>] SyS_setxattr+0x10/0x20 [<ffffffff815e8b76>] entry_SYSCALL_64_fastpath+0x1e/0xa8 [<ffffffffffffffff>] 0xffffffffffffffff root@hlms-zaken-01:~# Now, when I look in /sys/kernel/debug/ceph/*/osdc / mdsc there are no outstanding requests to the OSDs or MDS. Both these calls just hang for ever on these requests and they don't continue. Any pointers where to start looking for this? I tried the 4.4 kernel before, it gave me the same hang. So that's why I upgraded to 4.7 to see if it was fixed there. The Ceph cluster is currently backfilling 17 PGs, but this also happend when HEALTH_OK was around. There are no block or slow requests in the cluster. Wido -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html