Re: How to add back stray OSD daemon after node re-installation

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



ubuntu@ceph1f:~$ sudo cephadm deploy --name osd.2 --fsid 91a86f20-8083-40b1-8bf1-fe35fac3d677
Deploy daemon osd.2 ...

Which fsid is it, the cluster's or the OSD's? According to the 'cephadm deploy' help page it should be the cluster fsid.


Zitat von mabi <mabi@xxxxxxxxxxxxx>:

Hi Eugen,

What a good coincidence ;-)

So I ran "cephadm ceph-volume lvm list" on the OSD node which I re-instaled and it saw my osd.2 OSD. So far so good, but the following suggested command does not work as you can see below:

ubuntu@ceph1f:~$ sudo cephadm deploy --name osd.2 --fsid 91a86f20-8083-40b1-8bf1-fe35fac3d677
Deploy daemon osd.2 ...

Traceback (most recent call last):
  File "/usr/local/sbin/cephadm", line 6223, in <module>
    r = args.func()
  File "/usr/local/sbin/cephadm", line 1440, in _default_image
    return func()
  File "/usr/local/sbin/cephadm", line 3457, in command_deploy
    deploy_daemon(args.fsid, daemon_type, daemon_id, c, uid, gid,
  File "/usr/local/sbin/cephadm", line 2193, in deploy_daemon
    deploy_daemon_units(fsid, uid, gid, daemon_type, daemon_id, c,
  File "/usr/local/sbin/cephadm", line 2255, in deploy_daemon_units
    assert osd_fsid
AssertionError

Any ideas what is wrong here?

Regards,
Mabi

‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐
On Thursday, May 27, 2021 12:13 PM, Eugen Block <eblock@xxxxxx> wrote:

Hi,

I posted a link to the docs [1], [2] just yesterday ;-)

You should see the respective OSD in the output of 'cephadm
ceph-volume lvm list' on that node. You should then be able to get it
back to cephadm with

cephadm deploy --name osd.x

But I haven't tried this yet myself, so please report back if that
works for you.

Regards,
Eugen

[1] https://tracker.ceph.com/issues/49159
[2] https://tracker.ceph.com/issues/46691

Zitat von mabi mabi@xxxxxxxxxxxxx:

> Hello,
> I have by mistake re-installed the OS of an OSD node of my Octopus
> cluster (managed by cephadm). Luckily the OSD data is on a separate
> disk and did not get affected by the re-install.
> Now I have the following state:
>
>     health: HEALTH_WARN
>             1 stray daemon(s) not managed by cephadm
>             1 osds down
>             1 host (1 osds) down
>
>
> To fix that I tried to run:
>
> ceph orch daemon add osd ceph1f:/dev/sda
>
> =========================================
>
> Created no osd(s) on host ceph1f; already created?
> That did not work, so I tried:
>
> ceph cephadm osd activate ceph1f
>
> =================================
>
> no valid command found; 10 closest matches:
> ...
> Error EINVAL: invalid command
> Did not work either. So I wanted to ask how can I "adopt" back an
> OSD disk to my cluster?
> Thanks for your help.
> Regards,
> Mabi
>
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx

ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux