Recovering some data with 2 of 2240 pg in "remapped+peering"

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all,

 

I think I know the answer to this already after reading similar queries, but I’ll ask in case times have changed.

 

After an error on my part, I have a very small number of pgs in remapped+peering. They don’t look like they’ll get out of that state. Some IO is blocked too, as you might imagine. Entirely my fault; I removed two osds when the cluster wasn’t healthy.

 

I gather the pool is now fundamentally broken because of these 2 placement groups and I’ll need to recreate another. Some VMs are throwaway, some I’ll restore from backup. Not a great loss since I’m just testing.

 

What has got me wondering is:  one VM had a ZFS filesystem across a mirror of two rbd images. The VM hangs indefinitely, which is a shame, because I figure it’s unlikely that the same bits of data are missing from each half of the mirror.

 

Is it possible to make an IO fail rather than hang? This would be helpful in the recovery process, but I’ll cut my losses now if it’s simply not possible.

 

I’m on Ceph 0.80.7, Proxmox 3.3, which I understand to be on an ‘old’ Debian kernel.

 

Thank you,

Chris

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux