Hello Adam,
Just tried command ceph orch redeploy (without "--image") and it works,
the rgw image is the right version.
The command we used is
$ sudo ceph orch daemon redeploy rgw.testrgw.svtcephrgwv1.zlfzpx
quay.io/ceph/ceph:v16.2.10
The old rgw service is still alive, but it seems that it's no problem
for orchestrator.
We can now go further and fetch what is wrong.
Thanks for your helping
JM
Le 14/10/2022 à 14:09, Adam King a écrit :
For the weird image, perhaps just "ceph orch daemon redeploy
rgw.testrgw.svtcephrgwv1.invwmo --image quay.io/ceph/ceph:v16.2.10" will
resolve it. Not sure about the other things wrong with it yet but I think
the image should be fixed before looking into that.
On Fri, Oct 14, 2022 at 5:47 AM Jean-Marc FONTANA<jean-marc.fontana@xxxxxxx>
wrote:
Hello everyone !
We're operating a small cluster which contains 1 monitor-manager, 3 osds
ans 1 RGW.
Tjhe cluster was initially installed with ceph-deploy in version
Nautilus (14.2.19) then
upgraded in Octopus (15.2.16) and lastly in Pacific (16.2.9).
Ceph-deploy does not work
any more so we need to migrate the cluster in cephadm mode. We did it
following the
official Ceph Doc.
Everything is getting OK, until it concerns mon, mgr and osd. But we get
in great troubles
when migrating the RGW :
- the RGW podman image is a very exotic version (see below)
- the old service starts after a while, although having been stopped
and removed as explained in the doc,
- we never achieved to config of the new gateway with a yaml file.
Versions of the nodes after migrating :
##############
monitor / manager : cephadm inspect-image
##############
{
"ceph_version": "ceph version 16.2.10
(45fa1a083152e41a408d15505f594ec5f1b4fe17) pacific (stable)",
"image_id":
"32214388de9de06e6f5a0a6aa9591ac10c72cbe1bdd751b792946d968cd502d6",
"repo_digests": [
"
quay.io/ceph/ceph@sha256:2b68483bcd050472a18e73389c0e1f3f70d34bb7abf733f692e88c935ea0a6bd
",
"
quay.io/ceph/ceph@sha256:3cd25ee2e1589bf534c24493ab12e27caf634725b4449d50408fd5ad4796bbfa
"
]
}
##############
OSDs : cephadm inspect-image
##############
{
"ceph_version": "ceph version 16.2.10
(45fa1a083152e41a408d15505f594ec5f1b4fe17) pacific (stable)",
"image_id":
"32214388de9de06e6f5a0a6aa9591ac10c72cbe1bdd751b792946d968cd502d6",
"repo_digests": [
"
quay.io/ceph/ceph@sha256:2b68483bcd050472a18e73389c0e1f3f70d34bb7abf733f692e88c935ea0a6bd
",
"
quay.io/ceph/ceph@sha256:3cd25ee2e1589bf534c24493ab12e27caf634725b4449d50408fd5ad4796bbfa
"
]
}
##############
RGW : cephadm inspect-image
##############
{
"ceph_version": "ceph version 16.2.5-387-g7282d81d
(7282d81d2c500b5b0e929c07971b72444c6ac424) pacific (stable)",
"image_id":
"41387741ad94630f1c58b94fdba261df8d8e3dc2d4f70ad6201739764f43eb2c",
"repo_digests": [
"
docker.io/ceph/daemon-base@sha256:a038c6dc35064edff40bb7e824783f1bbd325c888e722ec5e814671406216ad5
"
]
}
=============
Orchestrator
=============
ceph orch ps :
-------------
NAME HOST PORTS STATUS
REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID
mgr.svtcephmonv1 svtcephmonv1 running (5h)
4m ago 2d 341M - 16.2.10 32214388de9d 0464e2e0c71b
mon.svtcephmonv1 svtcephmonv1 running (5h)
4m ago 2d 262M 2048M 16.2.10 32214388de9d 28aa77685767
osd.0 svtcephosdv01 running (5h)
4m ago 2d 184M 4096M 16.2.10 32214388de9d c5a4a1091cba
osd.1 svtcephosdv02 running (5h)
4m ago 2d 182M 4096M 16.2.10 32214388de9d 080c8f2b3eca
osd.2 svtcephosdv03 running (5h)
4m ago 2d 189M 4096M 16.2.10 32214388de9d b58b549a932d
osd.3 svtcephosdv01 running (5h)
4m ago 2d 245M 4096M 16.2.10 32214388de9d 9d2f781ae290
osd.4 svtcephosdv02 running (5h)
4m ago 2d 233M 4096M 16.2.10 32214388de9d 6296db28f1d4
osd.5 svtcephosdv03 running (5h)
4m ago 2d 213M 4096M 16.2.10 32214388de9d deb58248e520
rgw.testrgw.svtcephrgwv1.invwmo svtcephrgwv1 *:80 error 67s
ago 23h - - <unknown> <unknown> <unknown>
-------------
ceph orch host ls :
-------------
HOST ADDR LABELS STATUS
svtcephmonv1 192.168.90.51
svtcephosdv01 192.168.90.54
svtcephosdv02 192.168.90.55
svtcephosdv03 192.168.90.56
svtcephrgwv1 192.168.90.57 RGW
5 hosts in cluster
Every help will be welcome and we can send any information which would
be convenient to solve the problem.
_______________________________________________
ceph-users mailing list --ceph-users@xxxxxxx
To unsubscribe send an email toceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list --ceph-users@xxxxxxx
To unsubscribe send an email toceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx