Hi, due a harware failure during expanding ceph, I'm in big trouble because the cephfs doesn't mount anymore. I was adding a couple storage nodes, but a disk has failed and after a reboot the OS (ubuntu 12.04) renamed the remaining devices, so the entire node has been screwed out. Now, from the "sane new node", I'm taking some new osd up and in because the cluster is near full and I can't revert completely the situation as before *I can* afford data loss, but i need to regain access to the filesystem My setup: 3 mon + 3 mds 4 storage nodes (i was adding no. 5 and 6) Ceph 0.56.4 ceph health: HEALTH_ERR 2008 pgs backfill; 246 pgs backfill_toofull; 74 pgs backfilling; 134 pgs degraded; 790 pgs peering; 10 pgs recovering; 1116 pgs recovery_wait; 790 pgs stuck inactive; 4782 pgs stuck unclean; recovery 3049459/21926624 degraded (13.908%); recovering 6 o/s, 16316KB/s; 4 full osd(s); 30 near full osd(s); full,noup,nodown flag(s) set ceph mds dump: dumped mdsmap epoch 44 epoch 44 flags 0 created 2013-03-18 14:42:29.330548 modified 2013-04-20 17:14:32.969332 tableserver 0 root 0 session_timeout 60 session_autoclose 300 last_failure 43 last_failure_osd_epoch 18160 compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object} max_mds 1 in 0 up {0=6376} failed stopped data_pools [0] metadata_pool 1 6376: 192.168.21.11:6800/13457 'm1' mds.0.9 up:replay seq 1 5945: 192.168.21.13:6800/12999 'm3' mds.-1.0 up:standby seq 1 5963: 192.168.21.12:6800/22454 'm2' mds.-1.0 up:standby seq 1 ceph mon dump: epoch 1 fsid d634f7b3-8a8a-4893-bdfb-a95ccca7fddd last_changed 2013-03-18 14:39:42.253923 created 2013-03-18 14:39:42.253923 0: 192.168.21.11:6789/0 mon.m1 1: 192.168.21.12:6789/0 mon.m2 2: 192.168.21.13:6789/0 mon.m3 _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com