Hi list,
In case of an OS disk failure on a cephadm managed storage node, is
there a way to redeploy ceph on the (reinstalled) node leaving the data
(OSDs) intact?
So instead of removing the storage node, have the cluster recover,
redeploy the storage node, and let the cluster recover, I would like to
skip both recovery steps (when all that is broken is an OS disk) and
only have a little bit of recovery while the OSDs where down.
With a package based install this is achieved pretty easily: reinstall
OS / ceph, make sure ceph keyrings are in right place / right owner
(ceph) and run: ceph-volume lvm activate --all ... and you are back in
business.
Thanks,
Stefan
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx