Re: 6 hosts fail cephadm check (15.2.4)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Looks as if your cluster is still running 15.2.1.

Have a look at https://docs.ceph.com/docs/master/cephadm/upgrade/

Am 28.07.20 um 09:57 schrieb Ml Ml:
> Hello,
> 
> i get:
> 
> [WRN] CEPHADM_HOST_CHECK_FAILED: 6 hosts fail cephadm check
>     host ceph01 failed check: Failed to connect to ceph01 (ceph01).
> Check that the host is reachable and accepts connections using the
> cephadm SSH key
> you may want to run:
>> ssh -F =(ceph cephadm get-ssh-config) -i =(ceph config-key get mgr/cephadm/ssh_identity_key) root@ceph01
>     host ceph02 failed check: Failed to connect to ceph02 (10.10.1.2).
> Check that the host is reachable and accepts connections using the
> cephadm SSH key
> you may want to run:
>> ssh -F =(ceph cephadm get-ssh-config) -i =(ceph config-key get mgr/cephadm/ssh_identity_key) root@ceph02
>     host ceph03 failed check: Failed to connect to ceph03 (10.10.1.3).
> Check that the host is reachable and accepts connections using the
> cephadm SSH key
> you may want to run:
>> ssh -F =(ceph cephadm get-ssh-config) -i =(ceph config-key get mgr/cephadm/ssh_identity_key) root@ceph03
>     host ceph04 failed check: Failed to connect to ceph04 (10.10.1.4).
> Check that the host is reachable and accepts connections using the
> cephadm SSH key
> you may want to run:
>> ssh -F =(ceph cephadm get-ssh-config) -i =(ceph config-key get mgr/cephadm/ssh_identity_key) root@ceph04
>     host ceph05 failed check: Failed to connect to ceph05 (10.10.1.5).
> Check that the host is reachable and accepts connections using the
> cephadm SSH key
> you may want to run:
>> ssh -F =(ceph cephadm get-ssh-config) -i =(ceph config-key get mgr/cephadm/ssh_identity_key) root@ceph05
>     host ceph06 failed check: Failed to connect to ceph06 (10.10.1.6).
> Check that the host is reachable and accepts connections using the
> cephadm SSH key
> 
> 
> on ceph01 i run:
> ceph cephadm get-ssh-config > /tmp/ceph.conf
> ceph config-key get mgr/cephadm/ssh_identity_key > /tmp/ceph.key
> chmod 600 /tmp/ceph.key
> ssh -F /tmp/ceph.conf -i /tmp/ceph.key root@ceph01 (which works)
> 
> So i can not understand the errors above.
> 
> root@ceph01:~# ceph versions
> {
>     "mon": {
>         "ceph version 15.2.1
> (9fd2f65f91d9246fae2c841a6222d34d121680ee) octopus (stable)": 3
>     },
>     "mgr": {
>         "ceph version 15.2.1
> (9fd2f65f91d9246fae2c841a6222d34d121680ee) octopus (stable)": 3
>     },
>     "osd": {
>         "ceph version 15.2.1
> (9fd2f65f91d9246fae2c841a6222d34d121680ee) octopus (stable)": 56
>     },
>     "mds": {
>         "ceph version 15.2.1
> (9fd2f65f91d9246fae2c841a6222d34d121680ee) octopus (stable)": 1
>     },
>     "overall": {
>         "ceph version 15.2.1
> (9fd2f65f91d9246fae2c841a6222d34d121680ee) octopus (stable)": 63
>     }
> }
> 
> root@ceph01:~# dpkg -l |grep ceph
> ii  ceph-base                               15.2.4-1~bpo10+1
>    amd64        common ceph daemon libraries and management tools
> ii  ceph-common                             15.2.4-1~bpo10+1
>    amd64        common utilities to mount and interact with a ceph
> storage cluster
> ii  ceph-deploy                             2.0.1
>    all          Ceph-deploy is an easy to use configuration tool
> ii  ceph-fuse                               15.2.4-1~bpo10+1
>    amd64        FUSE-based client for the Ceph distributed file system
> ii  ceph-grafana-dashboards                 15.2.4-1~bpo10+1
>    all          grafana dashboards for the ceph dashboard
> ii  ceph-mds                                15.2.4-1~bpo10+1
>    amd64        metadata server for the ceph distributed file system
> ii  ceph-mgr                                15.2.4-1~bpo10+1
>    amd64        manager for the ceph distributed storage system
> ii  ceph-mgr-cephadm                        15.2.4-1~bpo10+1
>    all          cephadm orchestrator module for ceph-mgr
> ii  ceph-mgr-dashboard                      15.2.4-1~bpo10+1
>    all          dashboard module for ceph-mgr
> ii  ceph-mgr-diskprediction-cloud           15.2.4-1~bpo10+1
>    all          diskprediction-cloud module for ceph-mgr
> ii  ceph-mgr-diskprediction-local           15.2.4-1~bpo10+1
>    all          diskprediction-local module for ceph-mgr
> ii  ceph-mgr-k8sevents                      15.2.4-1~bpo10+1
>    all          kubernetes events module for ceph-mgr
> ii  ceph-mgr-modules-core                   15.2.4-1~bpo10+1
>    all          ceph manager modules which are always enabled
> ii  ceph-mon                                15.2.4-1~bpo10+1
>    amd64        monitor server for the ceph storage system
> ii  ceph-osd                                15.2.4-1~bpo10+1
>    amd64        OSD server for the ceph storage system
> ii  cephadm                                 15.2.4-1~bpo10+1
>    amd64        cephadm utility to bootstrap ceph daemons with systemd
> and containers
> ii  libcephfs1                              10.2.11-2
>    amd64        Ceph distributed file system client library
> ii  libcephfs2                              15.2.4-1~bpo10+1
>    amd64        Ceph distributed file system client library
> ii  python-ceph-argparse                    14.2.8-1
>    all          Python 2 utility libraries for Ceph CLI
> ii  python3-ceph-argparse                   15.2.4-1~bpo10+1
>    all          Python 3 utility libraries for Ceph CLI
> ii  python3-ceph-common                     15.2.4-1~bpo10+1
>    all          Python 3 utility libraries for Ceph
> ii  python3-cephfs                          15.2.4-1~bpo10+1
>    amd64        Python 3 libraries for the Ceph libcephfs library
> 
> root@ceph01:~# ceph -s
>   cluster:
>     id:     5436dd5d-83d4-4dc8-a93b-60ab5db145df
>     health: HEALTH_WARN
>             6 hosts fail cephadm check
>             failed to probe daemons or devices
>             7 nearfull osd(s)
>             Reduced data availability: 1 pg inactive
>             Low space hindering backfill (add storage if this doesn't
> resolve itself): 26 pgs backfill_toofull
>             Degraded data redundancy: 202495/33226941 objects degraded
> (0.609%), 26 pgs degraded, 26 pgs undersized
>             3 pool(s) nearfull
> 
>   services:
>     mon: 3 daemons, quorum ceph01,ceph02,ceph03 (age 39m)
>     mgr: ceph02(active, since 77m), standbys: ceph03, ceph01
>     mds:  2 up:standby
>     osd: 61 osds: 56 up (since 41m), 55 in (since 41m); 27 remapped pgs
> 
>   data:
>     pools:   3 pools, 2049 pgs
>     objects: 11.08M objects, 37 TiB
>     usage:   113 TiB used, 28 TiB / 141 TiB avail
>     pgs:     0.049% pgs not active
>              202495/33226941 objects degraded (0.609%)
>              9238/33226941 objects misplaced (0.028%)
>              1025 active+clean
>              887  active+clean+snaptrim_wait
>              110  active+clean+snaptrim
>              25   active+undersized+degraded+remapped+backfill_toofull
>              1    undersized+degraded+remapped+backfill_toofull+peered
>              1    active+remapped+backfilling
> 
>   io:
>     client:   1.0 KiB/s rd, 140 KiB/s wr, 0 op/s rd, 1 op/s wr
>     recovery: 30 MiB/s, 8 objects/s
> 
> I already restarted the mgr on ceph02.
> 
> Thanks,
> Michael
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> 

-- 
SUSE Software Solutions Germany GmbH, Maxfeldstr. 5, 90409 Nürnberg, Germany
(HRB 36809, AG Nürnberg). Geschäftsführer: Felix Imendörffer

Attachment: signature.asc
Description: OpenPGP digital signature

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux