2012/11/20 Gregory Farnum <greg@xxxxxxxxxxx>: > On Mon, Nov 19, 2012 at 7:55 AM, Drunkard Zhang <gongfan193@xxxxxxxxx> wrote: >> I created a ceph cluster for test, here's mistake I made: >> Add a second mds: mds.ab, executed 'ceph mds set_max_mds 2', then >> removed the mds just added; >> Then 'ceph mds set_max_mds 1', the first mds.aa crashed, and became laggy. >> As I can't repair mds.aa, so did 'ceph mds newfs metadata data >> --yes-i-really-mean-it'; > > So this command is a mkfs sort of thing. It's deleted all the > "allocation tables" and filesystem metadata in favor of new, empty > ones. You should not run "--yes-i-really-mean-it" commands if you > don't know exactly what the command is doing and why you're using it. > >> mds.aa was back, but 1TB data was in cluster lost, but disk space >> still used, by 'ceps -s'. >> >> Is there any chance I can get my data back? If can't, how can I >> retrieve back the disk space. > > There's not currently a great way to get that data back. With > sufficient energy it could be re-constructed by looking through all > the RADOS objects and putting something together. > To retrieve the disk space, you'll want to delete the "data" and > "metadata" RADOS pools. This will of course *eliminate* the data you > have in your new filesystem, so grab that out first if there's > anything there you care about. Then create the pools and run the newfs > command again. > Also, you've got the syntax wrong on that newfs command. You should be > using pool IDs: > "ceph mds newfs 1 0 --yes-i-really-mean-it" > (Though these IDs may change after re-creating the pools.) > -Greg > Thanks, I got you. > >> Now it looks like: >> log3 ~ # ceph -s >> health HEALTH_OK >> monmap e1: 1 mons at {log3=10.205.119.2:6789/0}, election epoch 0, >> quorum 0 log3 >> osdmap e1555: 28 osds: 20 up, 20 in >> pgmap v56518: 960 pgs: 960 active+clean; 1134 GB data, 2306 GB >> used, 51353 GB / 55890 GB avail >> mdsmap e703: 1/1/1 up {0=aa=up:active}, 1 up:standby >> >> log3 ~ # df | grep osd |sort >> /dev/sdb1 2.8T 124G 2.5T 5% /ceph/osd.0 >> /dev/sdc1 2.8T 104G 2.6T 4% /ceph/osd.1 >> /dev/sdd1 2.8T 84G 2.6T 4% /ceph/osd.2 >> /dev/sde1 2.8T 117G 2.6T 5% /ceph/osd.3 >> /dev/sdf1 2.8T 105G 2.6T 4% /ceph/osd.4 >> /dev/sdg1 2.8T 84G 2.6T 4% /ceph/osd.5 >> /dev/sdh1 2.8T 140G 2.5T 6% /ceph/osd.6 >> /dev/sdi1 2.8T 134G 2.5T 5% /ceph/osd.8 >> /dev/sdj1 2.8T 112G 2.6T 5% /ceph/osd.7 >> /dev/sdk1 2.8T 159G 2.5T 6% /ceph/osd.9 >> /dev/sdl1 2.8T 126G 2.5T 5% /ceph/osd.10 >> >> osd on another host didn't show. >> -- >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> More majordomo info at http://vger.kernel.org/majordomo-info.html -- 张绍文 gongfan193@xxxxxxxxx zhangsw@xxxxxxxxxxxxx 18601633785 -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html