Hi,
it sounds like the mds container_image is not configured properly, you
can set it via:
ceph config set mds container_image quay.io/ceph/ceph:v18.2.2
or just set it globally for all ceph daemons:
ceph config set global container_image quay.io/ceph/ceph:v18.2.2
If you bootstrap a fresh cluster, the image is set globally for you,
but it doesn't do that during an upgrade from a non-cephadm cluster,
which requires to redeploy mds daemons.
Regards,
Eugen
Zitat von opositorvlc@xxxxxxxx:
Hi All,
I migrated my CEPH 18.2.2 cluster from a non cephadm configuration.
All goes fine except MDS service was deployed in a old version: 17.0.0
I'm trying to deploy MDS daemons using ceph orch but CEPH always
download an old MDS image from docker.
How could I deploy the MDS service in the same 18.2.2 version that
the rest of services?
[root@master1 ~]# ceph orch apply mds datafs --placement="2 master1 master2"
[root@master1 ~]# ceph orch ps
NAME HOST PORTS STATUS REFRESHED
AGE MEM USE MEM LIM VERSION IMAGE ID
CONTAINER ID
mds.datafs.master1.gcpovr master1 running (36m) 6m ago
36m 37.2M - 17.0.0-7183-g54142666 75e3d7089cea
96682779c7ad
mds.datafs.master2.oqaxuy master2 running (36m) 6m ago
36m 33.1M - 17.0.0-7183-g54142666 75e3d7089cea
a9a647f87c83
mgr.master master1 running (16h) 6m ago
17h 448M - 18.2.2 3c937764e6f5
70f06fa05b70
mgr.master2 master2 running (16h) 6m ago
17h 524M - 18.2.2 3c937764e6f5
2d0d5376d8b3
mon.master master1 running (16h) 6m ago
17h 384M 2048M 18.2.2 3c937764e6f5
66a65017ce29
mon.master2 master2 running (16h) 6m ago
17h 380M 2048M 18.2.2 3c937764e6f5
51d783a9e36c
osd.0 osd00 running (16h) 3m ago
17h 432M 4096M 18.2.2 3c937764e6f5
fedff66f5ed2
osd.1 osd00 running (16h) 3m ago
17h 475M 4096M 18.2.2 3c937764e6f5
24e24a1a22e6
osd.2 osd00 running (16h) 3m ago
17h 516M 4096M 18.2.2 3c937764e6f5
ccd05451b739
osd.3 osd00 running (16h) 3m ago
17h 454M 4096M 18.2.2 3c937764e6f5
f6d8f13c8aaf
osd.4 master1 running (16h) 6m ago
17h 525M 4096M 18.2.2 3c937764e6f5
a2dcf9f1a9b7
osd.5 master2 running (16h) 6m ago
17h 331M 4096M 18.2.2 3c937764e6f5
b0011e8561a4
[root@master1 ~]# ceph orch ls
NAME PORTS RUNNING REFRESHED AGE PLACEMENT
mds.datafs 2/2 6m ago 46s master1;master2;count:2
mgr 2/0 6m ago - <unmanaged>
mon 2/0 6m ago - <unmanaged>
osd 6 6m ago - <unmanaged>
[root@master1 ~]# ceph versions
{
"mon": {
"ceph version 18.2.2
(531c0d11a1c5d39fbfe6aa8a521f023abf3bf3e2) reef (stable)": 2
},
"mgr": {
"ceph version 18.2.2
(531c0d11a1c5d39fbfe6aa8a521f023abf3bf3e2) reef (stable)": 2
},
"osd": {
"ceph version 18.2.2
(531c0d11a1c5d39fbfe6aa8a521f023abf3bf3e2) reef (stable)": 6
},
"mds": {
"ceph version 17.0.0-7183-g54142666
(54142666e5705ced88e3e2d91ddc0ff29867a362) quincy (dev)": 2
},
"overall": {
"ceph version 17.0.0-7183-g54142666
(54142666e5705ced88e3e2d91ddc0ff29867a362) quincy (dev)": 2,
"ceph version 18.2.2
(531c0d11a1c5d39fbfe6aa8a521f023abf3bf3e2) reef (stable)": 10
}
}
[root@master1 ~]# podman images
REPOSITORY TAG IMAGE ID
CREATED SIZE
quay.io/ceph/ceph v18.2.2 3c937764e6f5
7 weeks ago 1.28 GB
quay.io/ceph/ceph v18 3c937764e6f5
7 weeks ago 1.28 GB
registry.access.redhat.com/ubi8 latest c70d72aaebb4
3 months ago 212 MB
quay.io/ceph/ceph v16 0d668911f040
23 months ago 1.27 GB
quay.io/ceph/ceph-grafana 8.3.5 dad864ee21e9
2 years ago 571 MB
quay.io/prometheus/prometheus v2.33.4 514e6a882f6e
2 years ago 205 MB
quay.io/prometheus/node-exporter v1.3.1 1dbe0e931976
2 years ago 22.3 MB
quay.io/prometheus/alertmanager v0.23.0 ba2b418f427c
2 years ago 58.9 MB
docker.io/ceph/daemon-base latest-master-devel 75e3d7089cea
2 years ago 1.29 GB
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx