Hi,
ceph.conf on the host as well as in "cephadm shell" are similar and have
the correct IPs set. Entries for osd are not present.
# minimal ceph.conf for 10489760
[global]
fsid = 10489760
mon_host = [v2:129.217.31.186:3300/0,v1:129.217.31.186:6789/0]
[v2:129.217.31.188:3300/0,v1:129.217.31.188:6789/0]
[v2:129.217.31.189:3300/0,v1:129.217.31.189:6789/0]
[v2:129.217.31.190:3300/0,v1:129.217.31.190:6789/0]
Ceph health detail does not help much either, only that there is an
"sn05" artifact still present somewhere that should be removed:
Cheers
Dominik
HEALTH_WARN 1 filesystem is degraded; 1 MDSs report slow metadata IOs;
16 osds down; 3 hosts (18 osds) down; Reduced data availability: 545
pgs inactive
[WRN] FS_DEGRADED: 1 filesystem is degraded
fs ml2r_storage is degraded
[WRN] MDS_SLOW_METADATA_IO: 1 MDSs report slow metadata IOs
mds.ml2r_storage.ml2rsn06.clizsn(mds.0): 4 slow metadata IOs are
blocked > 30 secs, oldest blocked for 4951 secs
[WRN] OSD_DOWN: 16 osds down
osd.28 (root=default,host=ml2rsn05) is down
osd.29 (root=default,host=ml2rsn05) is down
osd.30 (root=default,host=ml2rsn05) is down
osd.31 (root=default,host=ml2rsn05) is down
osd.32 (root=default,host=ml2rsn05) is down
osd.33 (root=default,host=ml2rsn05) is down
osd.34 (root=default,host=ml2rsn05) is down
osd.35 (root=default,host=ml2rsn05) is down
osd.36 (root=default,host=ml2rsn06) is down
osd.37 (root=default,host=ml2rsn06) is down
osd.38 (root=default,host=ml2rsn06) is down
osd.39 (root=default,host=ml2rsn06) is down
osd.40 (root=default,host=ml2rsn06) is down
osd.41 (root=default,host=ml2rsn06) is down
osd.42 (root=default,host=ml2rsn06) is down
osd.43 (root=default,host=ml2rsn06) is down
[WRN] OSD_HOST_DOWN: 3 hosts (18 osds) down
host ml2rsn03 (root=default) (9 osds) is down
host ml2rsn05 (root=default) (9 osds) is down
host sn05 (root=default) (0 osds) is down
[WRN] PG_AVAILABILITY: Reduced data availability: 545 pgs inactive
Am 19.12.2022 um 14:02 schrieb Eugen Block:
I re-read a thread from last year [1] but didn't find anything
extraordinary required for this to work. If your MONs already run with
the new addresses, does the ceph.conf on the OSD nodes still reflect
the old entries? The docs do mention it explicitly:
Ceph clients and other Ceph daemons use ceph.conf to discover monitors.
In that case a restart of ceph.target wouldn't be enough. I assume
cephadm should update the conf file but maybe it requires a 'ceph orch
reconfigure osd'? I haven't tried that yet, so be careful and test the
effects in a test cluster.
Regards,
Eugen
[1]
https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/thread/OWSIXB5O3Y7LXC3D2JYETEAFMRC3K7OY/
Zitat von Dominik Baack <dominik.baack@xxxxxxxxxxxxxxxxxx>:
Hi,
min_mon_release 17 (quincy)
election_strategy: 1
0: [v2:129.217.31.189:3300/0,v1:129.217.31.189:6789/0] mon.ml2rsn06
1: [v2:129.217.31.186:3300/0,v1:129.217.31.186:6789/0] mon.ml2rsn03
2: [v2:129.217.31.188:3300/0,v1:129.217.31.188:6789/0] mon.ml2rsn05
3: [v2:129.217.31.190:3300/0,v1:129.217.31.190:6789/0] mon.ml2rsn07
dumped monmap epoch 21
seems fine for me (with the exception that there is an even number of
mons).
ceph osd metadata 18
"back_addr":
"[v2:129.217.31.173:6817/1187092418,v1:129.217.31.173:6819/1187092418]",
Still shows an old address after restarting of ceph.target.
Cheers
Dominik
Am 19.12.2022 um 09:24 schrieb Eugen Block:
If you did it "the right way" then 'ceph mon dump' should reflect
the changes. Does that output show the new IP addresses?
Zitat von Dominik Baack <dominik.baack@xxxxxxxxxxxxxxxxxx>:
Hi,
I removed/added the new mons as explained in:
https://docs.ceph.com/en/latest/rados/operations/add-or-rm-mons/#changing-a-monitor-s-ip-address-the-right-way
but did not change the mon map manually.
I can retry the manual way as well and report afterwards:
https://docs.ceph.com/en/latest/rados/operations/add-or-rm-mons/#changing-a-monitor-s-ip-address-the-messy-way
Cheers
Dominik Baack
Am 17.12.2022 um 10:06 schrieb Eugen Block:
Hi,
did you also change the monmap as described in the docs [1]? There
have been multiple threads in this list with the same topic.
Simply changing the IP address of the MONs is not sufficient, but
after fixing the monmap the OSDs should connect successfully.
[1]
https://docs.ceph.com/en/quincy/rados/operations/add-or-rm-mons.html
Zitat von Dominik Baack <dominik.baack@xxxxxxxxxxxxxxxxxx>:
Hi,
we have switched our network from IP over IB to an Ethernet
configuration.
I could move the mons over to the new address range and they
connect into the cluster network. The OSD create more of a
problem, even after setting
public_network129.217.31.176/29
cluster_network 129.217.31.184/29
to their values and restarting ceph.target on all nodes but they
still report the wrong information when checked with ceph osd
metadata :
back_addr
front_addr
hb_back_addr
hb_front_addr
and are therefore not able to connect.
Where do I need to modify the address to get them connecting again?
Cheers
Dominik
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx