Re: ceph orch ps mon, mgr, osd shows <unknown> for version, image and container id

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I can see the json output for the osd you posted doesn't list any version

{

        "style": "cephadm:v1",

        "name": "osd.61",

        "fsid": "8dbfcd81-fee3-49d2-ac0c-e988c8be7178",

        "systemd_unit": ceph-8dbfcd81-fee3-49d2-ac0c-e988c8be7178@osd.61,

        "enabled": true,

        "state": "running",

        "memory_request": null,

        "memory_limit": null,

        "ports": null,

        "container_id":
"bb7d491335323689dfc6dcb8ae1b6c022f93b3721d69d46c6ed6036bbdd68255",

        "container_image_name": "docker.io/ceph/daemon:latest-pacific",

        "container_image_id":
"6e73176320aaccf3b3fb660b9945d0514222bd7a83e28b96e8440c630ba6891f",

        "container_image_digests": [


docker.io/ceph/daemon@sha256:261bbe628f4b438f5bf10de5a8ee05282f2697a5a2cb7ff7668f776b61b9d586

        ],

{


The way the version is gathered for OSDs (and ceph daemons in general) by
cephadm is to exec into the container and run "ceph -v"). I'm not sure why
that wouldn't be working for the OSD here, but is for the mon. The one
other thing I noted is the use of the
docker.io/ceph/daemon:latest-pacific image.
We haven't been putting ceph images on docker for quite some time, so
that's actually a fairly old pacific version (over 2 years old when I
checked). There are much more recent pacific images on quay. Any reason for
using that particular image? It's hard to remember if there was maybe some
bug with such an old version or something that could cause this.

On Thu, Mar 30, 2023 at 1:40 PM Adiga, Anantha <anantha.adiga@xxxxxxxxx>
wrote:

> Hi Adam,
>
>
>
>
>
> Cephadm ls lists all details:
>
>
>
> NAME                                                   HOST
> PORTS        STATUS         REFRESHED  AGE  MEM USE  MEM LIM
> VERSION         IMAGE ID      CONTAINER ID
>
> osd.61
> zp3110b001a0101               running           3m ago   8M        -
> 22.0G  <unknown>       <unknown>     <unknown>
>
> mon.zp3110b001a0101
> zp3110b001a0101               running           3m ago   8M        -
> 2048M  <unknown>       <unknown>     <unknown>
>
> {
>
>         "style": "cephadm:v1",
>
>         "name": "osd.61",
>
>         "fsid": "8dbfcd81-fee3-49d2-ac0c-e988c8be7178",
>
>         "systemd_unit": ceph-8dbfcd81-fee3-49d2-ac0c-e988c8be7178@osd.61,
>
>         "enabled": true,
>
>         "state": "running",
>
>         "memory_request": null,
>
>         "memory_limit": null,
>
>         "ports": null,
>
>         "container_id":
> "bb7d491335323689dfc6dcb8ae1b6c022f93b3721d69d46c6ed6036bbdd68255",
>
>         "container_image_name": "docker.io/ceph/daemon:latest-pacific",
>
>         "container_image_id":
> "6e73176320aaccf3b3fb660b9945d0514222bd7a83e28b96e8440c630ba6891f",
>
>         "container_image_digests": [
>
>
> docker.io/ceph/daemon@sha256:261bbe628f4b438f5bf10de5a8ee05282f2697a5a2cb7ff7668f776b61b9d586
>
>         ],
>
> {
>
>         "style": "cephadm:v1",
>
>         "name": "mon.zp3110b001a0101",
>
>         "fsid": "8dbfcd81-fee3-49d2-ac0c-e988c8be7178",
>
>         "systemd_unit":
> ceph-8dbfcd81-fee3-49d2-ac0c-e988c8be7178@mon.zp3110b001a0101,
>
>         "enabled": true,
>
>         "state": "running",
>
>         "memory_request": null,
>
>         "memory_limit": null,
>
>         "ports": null,
>
>         "container_id":
> "32ba68d042c3dd7e7cf81a12b6b753cf12dfd8ed1faa8ffc0ecf9f55f4f26fe4",
>
>         "container_image_name": "docker.io/ceph/daemon:latest-pacific",
>
>         "container_image_id":
> "6e73176320aaccf3b3fb660b9945d0514222bd7a83e28b96e8440c630ba6891f",
>
>         "container_image_digests": [
>
>
> docker.io/ceph/daemon@sha256:261bbe628f4b438f5bf10de5a8ee05282f2697a5a2cb7ff7668f776b61b9d586
>
>         ],
>
>         "memory_usage": 1104880336,
>
>         "version": "16.2.5",
>
>         "started": "2023-03-29T22:41:52.754971Z",
>
>         "created": "2022-07-13T16:31:48.766907Z",
>
>         "deployed": "2022-07-13T16:30:48.528809Z",
>
>         "configured": "2022-07-13T16:31:48.766907Z"
>
>     },
>
>
>
>
>
> The unknown is only for osd, mon and mgr  services. It is across all
> nodes.
>
> Also  some other items missing are  PORTS, STATUS (time),  MEM USE,
>
> NAME
>                                                             HOST
>         PORTS        STATUS         REFRESHED  AGE  MEM USE  MEM LIM
> VERSION         IMAGE ID      CONTAINER ID
>
> rgw.default.default.zp3110b001a0103.ftizjg       zp3110b001a0103
> *:8080       running (12h)     5m ago   8M     145M        -
>                 16.2.5          6e73176320aa  bd6c4d4262b3
>
> alertmanager.zp3110b001a0101
> zp3110b001a0101                      running               3m ago
> 8M        -                    -        <unknown>       <unknown>
> <unknown>
>
> mds.cephfs.zp3110b001a0102.sihibe
>             zp3110b001a0102               stopped           9m ago
> 4M        -        -  <unknown>       <unknown>     <unknown>
>
> mgr.zp3110b001a0101
>         zp3110b001a0101               running           3m ago   8M
> -        -  <unknown>       <unknown>     <unknown>
>
> mgr.zp3110b001a0102
>         zp3110b001a0102               running           9m ago   8M
> -        -  <unknown>       <unknown>     <unknown>
>
> mon.zp3110b001a0101
>          zp3110b001a0101               running           3m ago   8M
> -    2048M  <unknown>       <unknown>     <unknown>
>
> mon.zp3110b001a0102
>          zp3110b001a0102               running           9m ago   8M
> -    2048M  <unknown>       <unknown>     <unknown>
>
>
>
> Thank you,
>
> Anantha
>
>
>
> *From:* Adam King <adking@xxxxxxxxxx>
> *Sent:* Thursday, March 30, 2023 8:08 AM
> *To:* Adiga, Anantha <anantha.adiga@xxxxxxxxx>
> *Cc:* ceph-users@xxxxxxx
> *Subject:* Re:  ceph orch ps mon, mgr, osd shows <unknown>
> for version, image and container id
>
>
>
> if you put a copy of the cephadm binary onto one of these hosts (e.g.
> a002s002) and run "cephadm ls" what does it give for the OSDs? That's where
> the orch ps information comes from.
>
>
>
> On Thu, Mar 30, 2023 at 10:48 AM <anantha.adiga@xxxxxxxxx> wrote:
>
> Hi ,
>
> Why is ceph orch ps showing ,unknown  version, image and container id ?
>
> root@a002s002:~# cephadm shell ceph mon versions
> Inferring fsid 682863c2-812e-41c5-8d72-28fd3d228598
> Using recent ceph image
> quay.io/ceph/daemon@sha256:9889075a79f425c2f5f5a59d03c8d5bf823856ab661113fa17a8a7572b16a997
> {
>     "ceph version 16.2.5 (0883bdea7337b95e4b611c768c0279868462204a)
> pacific (stable)": 3
> }
> root@a002s002:~# cephadm shell ceph mgr versions
> Inferring fsid 682863c2-812e-41c5-8d72-28fd3d228598
> Using recent ceph image
> quay.io/ceph/daemon@sha256:9889075a79f425c2f5f5a59d03c8d5bf823856ab661113fa17a8a7572b16a997
> {
>     "ceph version 16.2.5 (0883bdea7337b95e4b611c768c0279868462204a)
> pacific (stable)": 3
> }
>
> root@a002s002:~# cephadm shell ceph orch ps --daemon-type mgr
> Inferring fsid 682863c2-812e-41c5-8d72-28fd3d228598
> Using recent ceph image
> quay.io/ceph/daemon@sha256:9889075a79f425c2f5f5a59d03c8d5bf823856ab661113fa17a8a7572b16a997
> NAME          HOST      PORTS  STATUS   REFRESHED  AGE  MEM USE  MEM LIM
> VERSION    IMAGE ID
> mgr.a002s002  a002s002         running     4m ago  11M        -        -
> <unknown>  <unknown>
> mgr.a002s003  a002s003         running    87s ago  11M        -        -
> <unknown>  <unknown>
> mgr.a002s004  a002s004         running     4m ago  11M        -        -
> <unknown>  <unknown>
> root@a002s002:~# cephadm shell ceph orch ps --daemon-type mon
> Inferring fsid 682863c2-812e-41c5-8d72-28fd3d228598
> Using recent ceph image
> quay.io/ceph/daemon@sha256:9889075a79f425c2f5f5a59d03c8d5bf823856ab661113fa17a8a7572b16a997
> NAME          HOST      PORTS  STATUS        REFRESHED  AGE  MEM USE  MEM
> LIM  VERSION    IMAGE ID      CONTAINER ID
> mon.a002s002  a002s002         running          4m ago  11M        -
> 2048M  <unknown>  <unknown>     <unknown>
> mon.a002s003  a002s003         running         95s ago  11M        -
> 2048M  <unknown>  <unknown>     <unknown>
> mon.a002s004  a002s004         running (4w)     4m ago   5M    1172M
> 2048M  16.2.5     6e73176320aa  d38b94e00d28
> root@a002s002:~# cephadm shell ceph orch ps --daemon-type osd
> Inferring fsid 682863c2-812e-41c5-8d72-28fd3d228598
> Using recent ceph image
> quay.io/ceph/daemon@sha256:9889075a79f425c2f5f5a59d03c8d5bf823856ab661113fa17a8a7572b16a997
> NAME    HOST      PORTS  STATUS   REFRESHED  AGE  MEM USE  MEM LIM
> VERSION    IMAGE ID
> osd.0   a002s002         running     8m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.1   a002s003         running     5m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.10  a002s004         running     8m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.11  a002s003         running     5m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.12  a002s002         running     8m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.13  a002s004         running     8m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.14  a002s003         running     5m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.15  a002s002         running     8m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.16  a002s004         running     8m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.17  a002s003         running     5m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.18  a002s002         running     8m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.19  a002s004         running     8m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.2   a002s004         running     8m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.20  a002s003         running     5m ago  11M        -    10.9G
> <unknown>  <unknown>
> osd.21  a002s002         running     8m ago  11M        -    10.9G
> <unknown>  <unknown>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux