I have this ceph node that will correctly recover into my ceph pool and performance looks to be normal for the rbd clients. However after a few minutes once finishing recovery the rbd clients begin to fall over and cannot write data to the pool.
I've been trying to figure this out for weeks! None of the logs contain anything relevant at all.
If I disable the node in the crushmap the rbd clients immediately begin writing to the other nodes.
Ideas?
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com