Hi Edwin, On 07/12/2013 08:03 AM, Edwin Peer wrote:
Hi there, We've been noticing nasty multi-second cluster wide latencies if an OSD drops out of an active cluster (due to power failure, or even being stopped cleanly). We've also seen this problem occur when an OSD is inserted back into the cluster.
You will probably see that Peering Groups (PGs) go into a different state then active+clean.
What does ceph -s tell you in such a case?
Obviously, this has the effect of freezing all VMs doing I/O across the cluster for several seconds when a single node fails. Is this behaviour expected? Or have I perhaps got something configured wrong?
Not really the expected behavior, but it could be CPU power limitations on the OSDs. I notice this latency with a Atom cluster as well, but that's mainly due to the fact that the Atoms aren't fast enough to figure out what's happening.
Faster AMD or Intel CPUs don't suffer from this. There will be a very short I/O stall for certain PGs when an OSD goes down, but that should be very short and not every VM should suffer.
How many OSDs do you have with how many PGs per pool? Wido
We're trying very hard to eliminate all single points of failure in our architecture, is there anything that can be done about this? Regards, Edwin Peer _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
-- Wido den Hollander 42on B.V. Phone: +31 (0)20 700 9902 Skype: contact42on _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com