we have a running ceph, 16.2.7, with SATA OSD and DB on nvme. and we insert some SATA to host, and the status of new host is AVAILABLE. then we apply the osd-spec.yml, it can't create the OSD automatically. # ceph orch device ls HOST PATH TYPE DEVICE ID SIZE AVAILABLE REJECT REASONS h172-18-100-100 /dev/nvme0n1 ssd INTEL SSDPF2KX038TZ_PHAC1036009Z3P8AGN 3840G LVM detected, locked h172-18-100-100 /dev/sdb hdd ST16000NM000G-2K_ZL2CB8ZR 16.0T Insufficient space (<10 extents) on vgs, LVM detected, locked h172-18-100-100 /dev/sdc hdd ST16000NM000G-2K_ZL2CB0J2 16.0T Insufficient space (<10 extents) on vgs, LVM detected, locked h172-18-100-100 /dev/sdd hdd ST16000NM000G-2K_ZL2CBFSF 16.0T Insufficient space (<10 extents) on vgs, LVM detected, locked h172-18-100-100 /dev/sde hdd ST16000NM000G-2K_ZL2CAYQB 16.0T Insufficient space (<10 extents) on vgs, LVM detected, locked h172-18-100-100 /dev/sdf hdd ST16000NM000G-2K_ZL2CBEMC 16.0T Yes h172-18-100-100 /dev/sdg hdd ST16000NM000G-2K_ZL2C427J 16.0T Yes h172-18-100-100 /dev/sdh hdd ST16000NM000G-2K_ZL2CAZCZ 16.0T Yes h172-18-100-100 /dev/sdi hdd ST16000NM000G-2K_ZL2CBM7M 16.0T Yes osd-spec.yml: service_type: osd service_id: osd-spec placement: host_pattern: '*' spec: objectstore: bluestore block_db_size: 73014444032 data_devices: rotational: 1 db_devices: rotational: 0 ceph orch apply osd -i osd-spec.yml --dry-run -- Peng Yong _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx