Hi all, I am trying to understand a issue with ceph directing clients to connect to OSDs through their cluster_network address instead of their public_network address. I have a configured a ceph cluster with a public and cluster network: >> ceph config dump|grep network global advanced cluster_network *10.114.0.0/16 <http://10.114.0.0/16>* * mon advanced public_network 10.113.0.0/16 * I upgraded the cluster from 16.2.4 to 16.2.6. After that, I am seeing that ceph is directing clients to connect to OSD's cluster_network address instead of their public_address: >> strace -f -e trace=network -s 10000 rbd ls --pool ephemeral-vms --name client.cinder .... [pid 2353692] connect(14, {sa_family=AF_INET, sin_port=htons(6806), sin_addr=inet_addr("*10.114.29.10*")}, 16) = -1 EINPROGRESS (Operation now in progress) In this case the client hangs because it is not able to access the address, since its an internal address. This appeared after upgrading to 16.2.6, but I am not sure it was due to the upgrade or it was a hidden issue that appeared after the nodes were rebooted. It can also be that I am missing something in the config, but this config was generated by the cephadm bootstrap command and not created by hand, and it worked before the upgrade/reboot so I am pretty confident with it. What do you think, can this be a bug or is more a misconfiguration on my side? Thanks, Javier _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx