Hi Phil, is the ceph public and cluster network set in your ceph config dump or is there another ceph.conf on the local servers? Joachim joachim.kraftmayer@xxxxxxxxx www.clyso.com Hohenzollernstr. 27, 80801 Munich Utting a. A. | HR: Augsburg | HRB: 25866 | USt. ID-Nr.: DE2754306 Am Sa., 21. Sept. 2024 um 23:31 Uhr schrieb Phil <infolist@xxxxxxxxxxxxxx>: > Digged further. > > ceph osd dump --format json |jq '.osds[]| select (.osd==4)' > > { > "osd": 4, > "uuid": "OSD UID", > "up": 0, > "in": 0, > "weight": 0, > "primary_affinity": 1, > "last_clean_begin": 0, > "last_clean_end": 0, > "up_from": 0, > "up_thru": 0, > "down_at": 0, > "lost_at": 0, > "public_addrs": { > "addrvec": [] > }, > "cluster_addrs": { > "addrvec": [] > }, > "heartbeat_back_addrs": { > "addrvec": [] > }, > "heartbeat_front_addrs": { > "addrvec": [] > }, > "public_addr": "(unrecognized address family 0)/0", > "cluster_addr": "(unrecognized address family 0)/0", > "heartbeat_back_addr": "(unrecognized address family 0)/0", > "heartbeat_front_addr": "(unrecognized address family 0)/0", > "state": [ > "autoout", > "exists", > "new" > ] > } > > Every other healthy osds fields are filled with proper values.. > > Clues match, but i keep on looking for the root cause. > > Best regards > > > Philippe > > > > 21 sept. 2024, 21:29 de infolist@xxxxxxxxxxxxxx: > > > Hi, on a healthy cluster, every osd creation made orphans osd. > > > > > > # ceph osd tree > > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > > -1 26.38051 root default > > -7 3.63869 host A > > 1 hdd 3.63869 osd.1 up 1.00000 1.00000 > > -11 1.81940 host F > > 2 hdd 1.81940 osd.2 up 1.00000 1.00000 > > -5 7.27737 host J > > 3 hdd 3.63869 osd.3 up 1.00000 1.00000 > > 5 hdd 3.63869 osd.5 up 1.00000 1.00000 > > -3 7.27737 host K > > 0 hdd 3.63869 osd.0 up 1.00000 1.00000 > > 8 hdd 3.63869 osd.8 up 1.00000 1.00000 > > -9 6.36768 host S > > 6 hdd 3.63869 osd.6 up 1.00000 1.00000 > > 7 hdd 2.72899 osd.7 up 1.00000 1.00000 > > 4 0 osd.4 down 0 1.00000 > > 9 0 osd.9 down 0 1.00000 > > > > neither 4 nor 9 osd appear in the crushmap (decompiled) > > Destroying and recreating osd recreate orphan osd. > > > > > > Any hint ? > > Thanks for all > > Best regards > > Philippe > > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx