Hi, we're experiencing some strange issues running ceph 0.87 in our, I think, quite large cluster (taking number of objects as a measurement). mdsmap e721086: 1/1/1 up {0=storagemds01=up:active}, 2 up:standby osdmap e143048: 92 osds: 92 up, 92 in flags noout,noscrub,nodeep-scrub pgmap v45790682: 4736 pgs, 6 pools, 109 TB data, 3841 Mobjects 255 TB used, 48892 GB / 303 TB avail Putting some higher load via cephfs on the cluster leads to messages like mds0: Client X failing to respond to capability release after some minutes. Requests from other clients start to block after a while. Rebooting the client named client resolves the issue. Clients are a mix of CentOS6 & CentOS7 running kernel 4.1.4-1.el7.elrepo.x86_64 4.1.4-1.el6.elrepo.x86_64 4.4.0-2.el6.elrepo.x86_64 but other releases show the same behavior. Currently running 3 OSD Nodes and 3 combined MDS/MON-Nodes. What information do you need to further track down this issue? Quite unsure so this is only a rough overview of the setup. We have another issue with sometimes broken files ; bad checksums after storage, but I think I will start a new thread for this ;-) Thanks! -- Kind Regards Michael _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com