Hi,
this is handled by ceph-volume, do you find anything helpful in
/var/log/ceph/<CEPH_FSID>/ceph-volume.log? Also check the cephadm.log
for any hints.
Zitat von 彭勇 <ppyy@xxxxxxxxxx>:
we have a running ceph, 16.2.7, with SATA OSD and DB on nvme.
and we insert some SATA to host, and the status of new host is AVAILABLE.
then we apply the osd-spec.yml, it can't create the OSD automatically.
# ceph orch device ls
HOST PATH TYPE DEVICE ID
SIZE AVAILABLE REJECT REASONS
h172-18-100-100 /dev/nvme0n1 ssd INTEL SSDPF2KX038TZ_PHAC1036009Z3P8AGN
3840G LVM detected, locked
h172-18-100-100 /dev/sdb hdd ST16000NM000G-2K_ZL2CB8ZR
16.0T Insufficient space (<10 extents) on vgs, LVM detected,
locked
h172-18-100-100 /dev/sdc hdd ST16000NM000G-2K_ZL2CB0J2
16.0T Insufficient space (<10 extents) on vgs, LVM detected,
locked
h172-18-100-100 /dev/sdd hdd ST16000NM000G-2K_ZL2CBFSF
16.0T Insufficient space (<10 extents) on vgs, LVM detected,
locked
h172-18-100-100 /dev/sde hdd ST16000NM000G-2K_ZL2CAYQB
16.0T Insufficient space (<10 extents) on vgs, LVM detected,
locked
h172-18-100-100 /dev/sdf hdd ST16000NM000G-2K_ZL2CBEMC
16.0T Yes
h172-18-100-100 /dev/sdg hdd ST16000NM000G-2K_ZL2C427J
16.0T Yes
h172-18-100-100 /dev/sdh hdd ST16000NM000G-2K_ZL2CAZCZ
16.0T Yes
h172-18-100-100 /dev/sdi hdd ST16000NM000G-2K_ZL2CBM7M
16.0T Yes
osd-spec.yml:
service_type: osd
service_id: osd-spec
placement:
host_pattern: '*'
spec:
objectstore: bluestore
block_db_size: 73014444032
data_devices:
rotational: 1
db_devices:
rotational: 0
ceph orch apply osd -i osd-spec.yml --dry-run
--
Peng Yong
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx