Re: CEPHADM_HOST_CHECK_FAILED after reboot of nodes

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I have now opened a bug issue as this must be a bug with cephadm:

https://tracker.ceph.com/issues/51629

Hopefully someone has time to look into that.

Thank you in advance.

‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐

On Friday, July 9th, 2021 at 8:11 AM, mabi <mabi@xxxxxxxxxxxxx> wrote:

> Hello,
>
> I rebooted all the 8 nodes of my Octopus 15.2.13 cluster which runs on Ubuntu 20.04 LTS with cephadm and since then cephadm see 7 nodes as unreachable as you can see below:
>
> [WRN] CEPHADM_HOST_CHECK_FAILED: 7 hosts fail cephadm check
>
> host ceph1d failed check: Can't communicate with remote host `ceph1d`, possibly because python3 is not installed there: [Errno 32] Broken pipe
>
> host ceph1g failed check: Can't communicate with remote host `ceph1g`, possibly because python3 is not installed there: [Errno 32] Broken pipe
>
> host ceph1c failed check: Can't communicate with remote host `ceph1c`, possibly because python3 is not installed there: [Errno 32] Broken pipe
>
> host ceph1e failed check: Can't communicate with remote host `ceph1e`, possibly because python3 is not installed there: [Errno 32] Broken pipe
>
> host ceph1f failed check: Can't communicate with remote host `ceph1f`, possibly because python3 is not installed there: [Errno 32] Broken pipe
>
> host ceph1b failed check: Can't communicate with remote host `ceph1b`, possibly because python3 is not installed there: [Errno 32] Broken pipe
>
> host ceph1h failed check: Failed to connect to ceph1h (ceph1h).
>
> Please make sure that the host is reachable and accepts connections using the cephadm SSH key
>
> To add the cephadm SSH key to the host:
>
> > ceph cephadm get-pub-key > ~/ceph.pub
> >
> > ssh-copy-id -f -i ~/ceph.pub root@ceph1h
>
> To check that the host is reachable:
>
> > ceph cephadm get-ssh-config > ssh_config
> >
> > ceph config-key get mgr/cephadm/ssh_identity_key > ~/cephadm_private_key
> >
> > chmod 0600 ~/cephadm_private_key
> >
> > ssh -F ssh_config -i ~/cephadm_private_key root@ceph1h
>
> I checked and SSH is working and python3 is installed on all nodes.
>
> As you can see here "ceph orch host ls" also shows nodes as offline:
>
> ceph orch host ls
> =================
>
> HOST ADDR LABELS STATUS
>
> ceph1a ceph1a _admin mon
>
> ceph1b ceph1b _admin mon Offline
>
> ceph1c ceph1c _admin mon Offline
>
> ceph1d ceph1d Offline
>
> ceph1e ceph1e Offline
>
> ceph1f ceph1f Offline
>
> ceph1g ceph1g mds Offline
>
> ceph1h ceph1h mds Offline
>
> Does anyone have a clue how I can fix that? cephadm seems to be broken...
>
> Thank you for your help.
>
> Regards,
>
> Mabi
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux