Hi, We add hosts to the cluster using fqdn, manually (ceph orch host add) everything works fine. However, if we use the spec file as below, the whole thing falls apart. --- service_type: host addr: xx.xx.xx.xx hostname: ceph001.xx002.xx.xx.xx.com location: root: xx002 rack: rack01 labels: - osd - rgw --- service_type: osd service_id: object_hdd service_name: osd.object_hdd placement: host_pattern: ceph* crush_device_class: object_hdd spec: data_devices: rotational: 1 db_devices: rotational: 0 size: '3000G:' --- service_type: osd service_id: index_nvme service_name: osd.index_nvme placement: host_pattern: ceph* crush_device_class: index_nvme spec: data_devices: rotational: 0 size: ':900G' Applying this spec results in two hosts, one fqdn and the other short: root@mon001(xx002):~/cephadm# ceph osd df tree ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL %USE VAR PGS STATUS TYPE NAME -4 0 - 0 B 0 B 0 B 0 B 0 B 0 B 0 0 - root dias002 -3 0 - 0 B 0 B 0 B 0 B 0 B 0 B 0 0 - rack rack01 -2 0 - 0 B 0 B 0 B 0 B 0 B 0 B 0 0 - host ceph001.xx002.xx.xx.xx.com -1 662.71497 - 663 TiB 7.0 TiB 102 MiB 37 KiB 1.7 GiB 656 TiB 1.05 1.00 - root default -9 662.71497 - 663 TiB 7.0 TiB 102 MiB 37 KiB 1.7 GiB 656 TiB 1.05 1.00 - host ceph001 36 index_nvme 0.87329 1.00000 894 GiB 33 MiB 2.7 MiB 1 KiB 30 MiB 894 GiB 0.00 0.00 0 up osd.36 0 object_hdd 18.38449 1.00000 18 TiB 199 GiB 2.7 MiB 1 KiB 56 MiB 18 TiB 1.06 1.00 0 up osd.0 1 object_hdd 18.38449 1.00000 18 TiB 199 GiB 2.7 MiB 1 KiB 74 MiB 18 TiB 1.06 1.00 0 up osd.1 2 object_hdd 18.38449 1.00000 18 TiB 199 GiB 2.7 MiB 1 KiB 56 MiB 18 TiB 1.06 1.00 0 up osd.2 This looks like a bug, but I'm not sure, maybe someone has encountered something similar? Regards, Piotr _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx