On 14-09-2023 17:49, Boris Behrens wrote:
I currently try to adopt our stage cluster, some hosts just pull strange
root@0cc47a6df330:/var/lib/containers/storage/overlay-images# podman ps
CONTAINER ID IMAGE COMMAND
CREATED STATUS PORTS NAMES
a532c37ebe42 docker.io/ceph/daemon-base:latest-master-devel -n
mgr.0cc47a6df3... 2 minutes ago Up 2 minutes ago
root@0cc47a6df330:~# ceph orch ps
NAME HOST PORTS STATUS
REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID
mgr.0cc47a6df14e.vqizdz 0cc47a6df14e.f00f.gridscale.dev *:9283 running
(3m) 3m ago 3m 10.8M - 16.2.11
mgr.0cc47a6df330.iijety 0cc47a6df330.f00f.gridscale.dev *:9283 running
(5s) 2s ago 4s 10.5M - 17.0.0-7183-g54142666
mgr.0cc47aad8ce8 0cc47aad8ce8.f00f.gridscale.dev running
(65m) 8m ago 60m 553M - 17.2.6
Any idea what I need to do to change that?
I want to get some things cleared up. What is the version you are
running? I see three different ceph versions active now. I see you are
running a podman ps command, but see docker images pulled. AFAIK podman
needs a different IMAGE than docker ... or do you have a mixed setup?
What does "ceph config-key get config/global/container_image" give you?
ceph config-key list |grep container_image should give you a list
(including config-history) where you can see what has been configured
cephadm logs might give a clue as well.
You can configure the IMAGE version / type that you want by setting the
key and redeploy affected containers: For example (18.1.2):
ceph config-key set config/global/container_image
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx