Re: Ubuntu 24.02 LTS Ceph status warning

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I pulled the V19 image and cleaned up the not used images. The problem remains.

So on host hvs004 I entered the commands 'cephadm gather-facts' and 'cephadm ceph-volume lvm list' and I got sensible output without error.
Can it be a Python issue regarding coding of json input? But why this happens in a docker container on a host and not in another host...

Can anybody hint me some ideas on how to resolve this issue?

> -----Oorspronkelijk bericht-----
> Van: Dominique Ramaekers <dominique.ramaekers@xxxxxxxxxx>
> Verzonden: dinsdag 15 oktober 2024 9:28
> Aan: Eugen Block <eblock@xxxxxx>; ceph-users@xxxxxxx
> Onderwerp:  Re: Ubuntu 24.02 LTS Ceph status warning
>
> 'ceph config get mgr container_image' gives
> quay.io/ceph/ceph@sha256:200087c35811bf28e8a8073b15fa86c07cce85c575f
> 1ccd62d1d6ddbfdc6770a => OK
>
> 'ceph health detail' gives
> HEALTH_WARN failed to probe daemons or devices [WRN]
> CEPHADM_REFRESH_FAILED: failed to probe daemons or devices
>     host hvs004 `cephadm gather-facts` failed: cephadm exited with an error
> code: 1, stderr: Traceback (most recent call last):
>   File "<frozen runpy>", line 198, in _run_module_as_main
>   File "<frozen runpy>", line 88, in _run_code
>   File "/var/lib/ceph/dd4b0610-b4d2-11ec-bb58-
> d1b32ae31585/cephadm.a58127a8eed242cae13849ddbebcb9931d7a5410f406
> f2d264e3b1ed31d9605e/__main__.py", line 5579, in <module>
>   ...
>   File "/var/lib/ceph/dd4b0610-b4d2-11ec-bb58-
> d1b32ae31585/cephadm.a58127a8eed242cae13849ddbebcb9931d7a5410f406
> f2d264e3b1ed31d9605e/cephadmlib/host_facts.py", line 722, in
> _fetch_apparmor
> ValueError: too many values to unpack (expected 2)
>     host hvs004 `cephadm ceph-volume` failed: cephadm exited with an error
> code: 1, stderr: Inferring config /var/lib/ceph/dd4b0610-b4d2-11ec-bb58-
> d1b32ae31585/config/ceph.conf
> Traceback (most recent call last):
>   File "<frozen runpy>", line 198, in _run_module_as_main
>   File "<frozen runpy>", line 88, in _run_code
>   File "/var/lib/ceph/dd4b0610-b4d2-11ec-bb58-
> d1b32ae31585/cephadm.a58127a8eed242cae13849ddbebcb9931d7a5410f406
> f2d264e3b1ed31d9605e/__main__.py", line 5579, in <module>
>   ...
>   File "/var/lib/ceph/dd4b0610-b4d2-11ec-bb58-
> d1b32ae31585/cephadm.a58127a8eed242cae13849ddbebcb9931d7a5410f406
> f2d264e3b1ed31d9605e/cephadmlib/host_facts.py", line 722, in
> _fetch_apparmor
> ValueError: too many values to unpack (expected 2)
>
> I do think it's a ceph version issue, so I started to compaire the hvs004 host
> with a good behaving host hvs001. I did find this:
> 'root@hvs001~#:cephadm shell ceph -v' gives ceph version 19.2.0
> (16063ff2022298c9300e49a547a16ffda59baf13) squid (stable)
> 'root@hvs004~#:cephadm shell ceph -v' gives ceph version 19.3.0-5346-
> gcc481a63 (cc481a63bc03a534cb8e2e961293d6509ba59401) squid (dev)
>
> It seems only the shell uses a wrong docker image so I took a list from both of
> them:
> hvs0001
> ----------
> REPOSITORY                         TAG       IMAGE ID       CREATED         SIZE
> quay.io/ceph/ceph                  v19       37996728e013   2 weeks ago     1.28GB
> quay.io/ceph/ceph                  v18.2     2bc0b0f4375d   2 months ago    1.22GB
> quay.io/ceph/ceph                  v18       a27483cc3ea0   6 months ago    1.26GB
> quay.io/ceph/ceph                  v17       5a04c8b3735d   9 months ago    1.27GB
> quay.io/ceph/ceph-grafana          9.4.7     954c08fa6188   10 months ago
> 633MB
> quay.io/ceph/grafana               9.4.12    2bacad6d85d8   17 months ago   330MB
> quay.io/prometheus/prometheus      v2.43.0   a07b618ecd1d   19 months ago
> 234MB
> quay.io/prometheus/alertmanager    v0.25.0   c8568f914cd2   22 months ago
> 65.1MB
> quay.io/prometheus/node-exporter   v1.5.0    0da6a335fe13   22 months ago
> 22.5MB
> quay.io/ceph/ceph                  v17.2     0912465dcea5   2 years ago     1.34GB
> quay.io/ceph/ceph                  v17.2.3   0912465dcea5   2 years ago     1.34GB
> quay.io/ceph/ceph-grafana          8.3.5     dad864ee21e9   2 years ago     558MB
> quay.ceph.io/ceph-ci/ceph          master    c5ce177c6a5d   2 years ago     1.38GB
> quay.io/prometheus/prometheus      v2.33.4   514e6a882f6e   2 years ago
> 204MB
> quay.io/prometheus/node-exporter   v1.3.1    1dbe0e931976   2 years ago
> 20.9MB
> quay.io/prometheus/alertmanager    v0.23.0   ba2b418f427c   3 years ago
> 57.5MB
> quay.io/ceph/ceph-grafana          6.7.4     557c83e11646   3 years ago     486MB
> quay.io/prometheus/prometheus      v2.18.1   de242295e225   4 years ago
> 140MB
> quay.io/prometheus/alertmanager    v0.20.0   0881eb8f169f   4 years ago
> 52.1MB
> quay.io/prometheus/node-exporter   v0.18.1   e5a616e4b9cf   5 years ago
> 22.9MB
>
> hvs004
> ---------
> REPOSITORY                         TAG       IMAGE ID       CREATED         SIZE
> quay.ceph.io/ceph-ci/ceph          main      6e76ca06f33a   11 days ago     1.41GB
> quay.io/ceph/ceph                  <none>    37996728e013   2 weeks ago     1.28GB
> quay.io/ceph/ceph                  v17       9cea3956c04b   18 months ago   1.16GB
> quay.io/prometheus/node-exporter   v1.5.0    0da6a335fe13   22 months ago
> 22.5MB
> quay.io/prometheus/node-exporter   v1.3.1    1dbe0e931976   2 years ago
> 20.9MB
>
> I pulled on hvs004 the v19 tagged image and my cephadm shell ceph -v gave
> the correct version.
>
> It seems my docker images aren't automatically managed by ceph?
>
> Can I fix this, or do I have to pull the correct images and remove the wrong
> ones myself?
>
>
> > -----Oorspronkelijk bericht-----
> > Van: Eugen Block <eblock@xxxxxx>
> > Verzonden: vrijdag 11 oktober 2024 13:03
> > Aan: ceph-users@xxxxxxx
> > Onderwerp:  Re: Ubuntu 24.02 LTS Ceph status warning
> >
> > I don't think the warning is related to a specific ceph version. The
> > orchestrator uses the default image anyway, you can get it via:
> >
> > ceph config get mgr container_image
> >
> > 'ceph health detail' should reveal which host or daemons misbehaves. I
> > would then look into cephadm.log on that host to find more hints, what
> > exactly goes wrong. You should also look into the active MGR log, it
> > could also give you hints why that service fails.
> >
> > Zitat von Dominique Ramaekers <dominique.ramaekers@xxxxxxxxxx>:
> >
> > > I manage a 4 hosts cluster on Ubuntu 22.04 LTS with ceph installed
> > > trough cephad and containers on Docker.
> > >
> > > Last month, I've migrated to the latest Ceph 19.2. All went great.
> > >
> > > Last week I've upgraded one of my hosts to Ubuntu 24.04.1 LTS Now I
> > > get the following warning in cephadm shell -- ceph status:
> > > Failed to apply 1 service(s): osd.all-available-devices failed to
> > > probe daemons or devices
> > >
> > > Outside the ceph shell:
> > > Ceph -v results in 'ceph version 19.2.0~git20240301.4c76c50
> > > (4c76c50a73f63ba48ccdf0adccce03b00d1d80c7) squid (dev)'
> > >
> > > Inside the shell: 'ceph version 19.3.0-5346-gcc481a63
> > > (cc481a63bc03a534cb8e2e961293d6509ba59401) squid (dev)'
> > > All osd's, mon's , mgr's and mds's are on 19.2.0` (image id
> > > 37996728e013)
> > >
> > > Do I get the warning because the Ubuntu package of ceph is still on
> > > a development version?
> > > Or can I have another underling problem?
> > >
> > > Thanks for the help.
> > > _______________________________________________
> > > ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an
> > > email to ceph-users-leave@xxxxxxx
> >
> >
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an
> > email to ceph-users-leave@xxxxxxx
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email
> to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux