Thanks for the reply Sebastian. Sadly I haven't had any luck so far restoring the OSD drive's superblocks. Any other advice from this group would be welcome before I erase and start again. On Mon, 27 Sept 2021 at 01:38, Sebastian Wagner <sewagner@xxxxxxxxxx> wrote: > Hi Phil, > > > Am 27.09.21 um 10:06 schrieb Phil Merricks: > > Hey folks, > > > > A recovery scenario I'm looking at right now is this: > > > > 1: In a clean 3-node Ceph cluster (pacific, deployed with cephadm), the > OS > > Disk is lost from all nodes > > 2: Trying to be helpful, a self-healing deployment system reinstalls the > OS > > on each node, and rebuilds the ceph services > > 3: Somewhere in the deployment system are 'sensible defaults' that assume > > there are no stateful workloads, so the superblock is wiped from the > other > > block devices attached to these nodes to prevent stale metadata conflicts > > 4: The ceph rebuild has no knowledge of prior states and is expected to > > simply restore based on discovery of existing devices. > > 5: Out of 5 block devices, 3 had their superblock wiped, 1 suffered > > mechanical failure upon reboot, and 1 is completely intact, with > > 'ceph-volume lvm list' returning the correct information. > As soon as you manage to restore the OSDs, there is a command to > re-create > <https://docs.ceph.com/en/latest/cephadm/osd/#activate-existing-osds> > the OSD containers. > > > > Is there a way to restore the 3 devices with wiped superblocks? Some > basic > > attempts with fsck to find the superblocks on the disks that were > affected > > yielded nothing. > > > > Thanks for your time reading this message. > > > > Cheers > > > > Phil > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx