Recovering PGs from Dead OSD disk

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi All

Thanks in advance for any help, I was wondering if anyone can help me with a pickle I have gotten myself into!

I was in the process of adding OSD's to my small cluster (6 OSDs) and the disk died halfway through, unfort I had left the defaults from when I was bootstrapping the cluster in place which meant that size=1. I have managed to mount the old OSD disk and copy off all the files in the root and some of the files under current but when I try and recover the PG's using ceph-objectstore-tool I get the following:

ceph-objectstore-tool --op export --pgid 7.cb --data-path /mnt/temp --journal-path /mnt/temp/journal --skip-journal-replay --debug --file 7.cb.export
2017-06-03 14:03:39.906546 7f2b217cc940  0 filestore(/mnt/temp) backend xfs (magic 0x58465342)
2017-06-03 14:03:39.907153 7f2b217cc940  0 genericfilestorebackend(/mnt/temp) detect_features: FIEMAP ioctl is disabled via 'filestore fiemap' config option
2017-06-03 14:03:39.907162 7f2b217cc940  0 genericfilestorebackend(/mnt/temp) detect_features: SEEK_DATA/SEEK_HOLE is disabled via 'filestore seek data hole' config option
2017-06-03 14:03:39.907193 7f2b217cc940  0 genericfilestorebackend(/mnt/temp) detect_features: splice is supported
2017-06-03 14:03:39.999729 7f2b217cc940  0 genericfilestorebackend(/mnt/temp) detect_features: syncfs(2) syscall fully supported (by glibc and kernel)
2017-06-03 14:03:39.999865 7f2b217cc940  0 xfsfilestorebackend(/mnt/temp) detect_feature: extsize is disabled by conf
2017-06-03 14:03:40.000392 7f2b217cc940 -1 filestore(/mnt/temp) mount initial op seq is 0; something is wrong
Mount failed with '(22) Invalid argument'

I know there is data in pg 7.cb as:
$ du -sh  /mnt/temp/current/7.cb*
15G    /mnt/temp/current/7.cb_head
0    /mnt/temp/current/7.cb_TEMP

I am unsure why it is trying to mount anything as the disk is already mounted. Is one of the disabled config options stopping me from completing this or am I missing something else? This pool was used for CephFS data (pool 6 is the metadata and in a similar state).  I have only copied the pg folders for the relevant pools, there are others on there but I think they might be stored on the damaged section of the disk.

Thanks for any help
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux