Hi Eugen, What a good coincidence ;-) So I ran "cephadm ceph-volume lvm list" on the OSD node which I re-instaled and it saw my osd.2 OSD. So far so good, but the following suggested command does not work as you can see below: ubuntu@ceph1f:~$ sudo cephadm deploy --name osd.2 --fsid 91a86f20-8083-40b1-8bf1-fe35fac3d677 Deploy daemon osd.2 ... Traceback (most recent call last): File "/usr/local/sbin/cephadm", line 6223, in <module> r = args.func() File "/usr/local/sbin/cephadm", line 1440, in _default_image return func() File "/usr/local/sbin/cephadm", line 3457, in command_deploy deploy_daemon(args.fsid, daemon_type, daemon_id, c, uid, gid, File "/usr/local/sbin/cephadm", line 2193, in deploy_daemon deploy_daemon_units(fsid, uid, gid, daemon_type, daemon_id, c, File "/usr/local/sbin/cephadm", line 2255, in deploy_daemon_units assert osd_fsid AssertionError Any ideas what is wrong here? Regards, Mabi ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐ On Thursday, May 27, 2021 12:13 PM, Eugen Block <eblock@xxxxxx> wrote: > Hi, > > I posted a link to the docs [1], [2] just yesterday ;-) > > You should see the respective OSD in the output of 'cephadm > ceph-volume lvm list' on that node. You should then be able to get it > back to cephadm with > > cephadm deploy --name osd.x > > But I haven't tried this yet myself, so please report back if that > works for you. > > Regards, > Eugen > > [1] https://tracker.ceph.com/issues/49159 > [2] https://tracker.ceph.com/issues/46691 > > Zitat von mabi mabi@xxxxxxxxxxxxx: > > > Hello, > > I have by mistake re-installed the OS of an OSD node of my Octopus > > cluster (managed by cephadm). Luckily the OSD data is on a separate > > disk and did not get affected by the re-install. > > Now I have the following state: > > > > health: HEALTH_WARN > > 1 stray daemon(s) not managed by cephadm > > 1 osds down > > 1 host (1 osds) down > > > > > > To fix that I tried to run: > > > > ceph orch daemon add osd ceph1f:/dev/sda > > > > ========================================= > > > > Created no osd(s) on host ceph1f; already created? > > That did not work, so I tried: > > > > ceph cephadm osd activate ceph1f > > > > ================================= > > > > no valid command found; 10 closest matches: > > ... > > Error EINVAL: invalid command > > Did not work either. So I wanted to ask how can I "adopt" back an > > OSD disk to my cluster? > > Thanks for your help. > > Regards, > > Mabi > > > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx