Hi all, I think I know the answer to this already after reading similar queries, but I’ll ask in case times have changed. After an error on my part, I have a very small number of pgs in remapped+peering. They don’t look like they’ll get out of that state. Some IO is blocked too, as you might imagine. Entirely my fault; I removed two osds when the cluster wasn’t healthy. I gather the pool is now fundamentally broken because of these 2 placement groups and I’ll need to recreate another. Some VMs are throwaway, some I’ll restore from backup. Not a great loss since I’m just testing. What has got me wondering is: one VM had a ZFS filesystem across a mirror of two rbd images. The VM hangs indefinitely, which is a shame, because I figure it’s unlikely that the same bits of data are missing from each half of the mirror. Is it possible to make an IO fail rather than hang? This would be helpful in the recovery process, but I’ll cut my losses now if it’s simply not possible. I’m on Ceph 0.80.7, Proxmox 3.3, which I understand to be on an ‘old’ Debian kernel. Thank you, Chris |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com