Just ran into this problem: a week ago I set up a Ceph cluster on 4
systems, with one admin node and 3 mon+osd nodes, then ran a few
casual IO tests. I returned to work after a few days out of town at
a conference, and now my Ceph cluster appears to have no OSDs!
root@rts24:/var/log/ceph# ceph statuscluster 284dbfe0-e612-4732-9d26-2c5909f0fbd1health HEALTH_ERR 119 pgs degraded; 192 pgs stale; 192 pgs stuck stale; 119 pgs stuck unclean; recovery 2/4 objects degraded (50.000%); no osdsmonmap e1: 3 mons at {rts21=172.29.0.21:6789/0,rts22=172.29.0.22:6789/0,rts23=172.29.0.23:6789/0}, election epoch 32, quorum 0,1,2 rts21,rts22,rts23osdmap e33: 0 osds: 0 up, 0 inpgmap v2774: 192 pgs, 3 pools, 135 bytes data, 2 objects0 kB used, 0 kB / 0 kB avail2/4 objects degraded (50.000%)73 stale+active+clean119 stale+active+degraded
I would appreciate if anyone could explain how can something like
this happen, or where to look for any evidence that might help me
understand what happened. The log files in /var/log/ceph/ show no
activity except for the monitors' Paxos chatter.
Thx,
Dan Koren
Director of Software
DATERA | 650.210.7910 | @dateranews
dnk@datera.io
This email and any attachments thereto may contain private,
confidential, and privileged material for the sole use of the
intended recipient. Any review, copying, or distribution of
this email (or any attachments thereto) by others is strictly
prohibited. If you are not the intended recipient, please
contact the sender immediately and permanently delete the
original and any copies of this email and any attachments
thereto.
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com