Hi,
there was a similar thread recently
(https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/message/WSVIAZBFJQKBXA47RKISA5U4J7BHX6DK/). Check the output of 'cepham ls' on the node where the OSD is not running and remove it with 'cephadm rm-daemon --name osd.3'. If there's an empty directory for that OSD (/var/lib/ceph/osd/ceph-3) remove it as
well.
Zitat von Stefan Kooman <stefan@xxxxxx>:
Hi,
I'm in the process of re-provisioning OSDs on a test cluster with
cephadm. One of the OSD id's that was supposedly previously living
on host3 is now alive on host2. And cephadm is not happy about that:
"Found duplicate OSDs: osd.3 in status running on host2, osd.3 in
status stopped on host3."
As far as I can see there has never been an osd 3 on host3 managed
by cephadm (there used to be an osd.3 when this cluster was
ceph-ansible managed).
How can I find what cephadm's view of the world is? And tell cephadm
it should forget about osd.3 on host3, without destroying osd.3 on
host2?
Thanks,
Gr. Stefan
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx