Hi All I am facing some issue related to osd config in ceph-ansible I have 2 nvme disks where i have wal and db configure and i configure lvm like this in osd.yamL osd.yaml devices: - /dev/sdb - /dev/sdc - /dev/sdd - /dev/sde - /dev/sdf - /dev/sdg - /dev/sdh - /dev/sdi - /dev/sdj - /dev/sdk - /dev/sdl - /dev/sdm lvm_volumes: - data: /dev/sdb wal: wal-lv1 wal_vg: vg1 db: journal-lv1 db_vg: vg1 - data: /dev/sdb wal: wal-lv2 wal_vg: vg1 db: journal-lv2 db_vg: vg1 - data: /dev/sdb wal: wal-lv3 wal_vg: vg1 db: journal-lv3 db_vg: vg1 - data: /dev/sdb wal: wal-lv4 wal_vg: vg1 db: journal-lv4 db_vg: vg1 - data: /dev/sdb wal: wal-lv5 wal_vg: vg1 db: journal-lv5 db_vg: vg1 - data: /dev/sdc wal: wal-lv1 wal_vg: vg2 db: journal-lv1 db_vg: vg2 - data: /dev/sdc wal: wal-lv2 wal_vg: vg2 db: journal-lv2 db_vg: vg2 - data: /dev/sdc wal: wal-lv3 wal_vg: vg2 db: journal-lv3 db_vg: vg2 - data: /dev/sdc wal: wal-lv4 wal_vg: vg2 db: journal-lv4 db_vg: vg2 - data: /dev/sdc wal: wal-lv5 wal_vg: vg2 db: journal-lv5 db_vg: vg2 - data: /dev/sdd - data: /dev/sde - data: /dev/sdf - data: /dev/sdg - data: /dev/sdh - data: /dev/sdi - data: /dev/sdj - data: /dev/sdk - data: /dev/sdl - data: /dev/sdm I have also setup wal size and db size respectively in all.yaml all.yaml ntp_service_enabled: false mon_group_name: mons osd_group_name: osds mgr_group_name: mgrs ceph_origin: repository ceph_repository: community ceph_stable_release: octopus public_network: 10.18.0.0/22 cluster_network: 10.25.0.0/24 monitor_interface: ens3 configure_firewall: False ceph_test: true cephx: true ip_version: ipv4 osd_objectstore: bluestore dashboard_enabled: False grafana_server_group_name: grafana-server ceph_conf_overrides: global: bluestore block db size: 85899345920 (80GB) bluestore_block_wal_size: 1073741824 (1GB) i have also tried to to set values of db_size and journal_size in OSD options journal_size: 1024 # OSD journal size in MB block_db_size: 85899345920 But my osd partitions are not creating as I want and taking full disk space sdb 8:16 0 447.1G 0 disk └─ceph--3e81a883--86cd--4ddb--b0b0--be5892f6afe8-osd--block--b981ea8d--410d--4e2f--86f5--509112ff5ebf 253:0 0 447.1G 0 lvm sdc 8:32 0 447.1G 0 disk └─ceph--d3d82887--a7ad--40a4--9c3d--440cd9057fd9-osd--block--de6cae7d--daef--4144--a4bc--077f6044ef22 253:1 0 447.1G 0 lvm sdd 8:48 0 5.5T 0 disk └─ceph--e5aa3641--263c--45df--90d0--cd7e26772e69-osd--block--989fd7b2--d7b7--4170--8b33--cf4c8a29f5c1 253:2 0 5.5T 0 lvm sde 8:64 0 5.5T 0 disk └─ceph--286034b5--1c55--4fb6--8983--4ec6ef482dbd-osd--block--7cdc5ef4--8a88--4ff8--a08e--278985f4764c 253:3 0 5.5T 0 lvm sdf 8:80 0 5.5T 0 disk └─ceph--3bd8b866--b99e--41c4--a993--d69e85c1b658-osd--block--d32e42f4--1133--4c0d--aaf6--821b724fd220 253:4 0 5.5T 0 lvm sdg 8:96 0 5.5T 0 disk └─ceph--972ae2d1--2098--436f--a448--9015edfc4dc5-osd--block--79411ef4--2469--4243--9545--89cd266719f1 253:5 0 5.5T 0 lvm sdh 8:112 0 5.5T 0 disk └─ceph--92cd1c5f--3c67--420e--8a4d--c9ab0b75705b-osd--block--e643c7b3--cb10--49ab--86a8--31b60f4525dc 253:6 0 5.5T 0 lvm sdi 8:128 0 5.5T 0 disk └─ceph--0a27a900--0e81--4ef9--9b2c--692d0a609e7e-osd--block--30eddf48--5e12--4c2f--aed0--e06fb403ed33 253:7 0 5.5T 0 lvm sdj 8:144 0 5.5T 0 disk └─ceph--90156988--e0a3--4699--ae18--ad1f743521bb-osd--block--eb879f94--f289--4dd9--ba2e--48ff06a14708 253:8 0 5.5T 0 lvm sdk 8:160 0 5.5T 0 disk └─ceph--cdcabf84--e713--42c2--b2f0--f1bd6bd54924-osd--block--067418cf--1bd3--46a3--bdb9--b23216d6e8e3 253:9 0 5.5T 0 lvm sdl 8:176 0 5.5T 0 disk └─ceph--60829d44--ca66--44b3--b4be--05e8fe0c154e-osd--block--49a34fa3--c8ba--4fd9--ba3e--a6c4a889ecd5 253:10 0 5.5T 0 lvm sdm 8:192 0 5.5T 0 disk └─ceph--bd0a95e3--4d83--4c9e--a5fe--9f28414015fd-osd--block--abf5c032--eaf8--47b4--9966--67526009f68f 253:11 0 5.5T 0 lvm so after running playbook i get this result of osds with error failed: [10.11.2.202] (item={u'wal_vg': u'vg2', u'data': u'/dev/sdc', u'wal': u'wal-lv5', u'db': u'journal-lv5', u'db_vg': u'vg2'}) => changed=true ansible_loop_var: item cmd: - ceph-volume - --cluster - ceph - lvm - create - --bluestore - --data - /dev/sdc - --block.db - vg2/journal-lv5 - --block.wal - vg2/wal-lv5 delta: '0:00:03.260879' end: '2020-04-20 15:49:51.407784' item: data: /dev/sdc db: journal-lv5 db_vg: vg2 wal: wal-lv5 wal_vg: vg2 msg: non-zero return code rc: 1 start: '2020-04-20 15:49:48.146905' stderr: |- Running command: /usr/bin/ceph-authtool --gen-print-key Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new b26655b0-0859-4f8b-9471-0bafddfc36ef Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-b26655b0-0859-4f8b-9471-0bafddfc36ef ceph-b7e18568-6735-4a1a-9aa1-95094189e859 stderr: Calculated size of logical volume is 0 extents. Needs to be larger. --> Was unable to complete a new OSD, will rollback changes Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd purge-new osd.0 --yes-i-really-mean-it stderr: purged osd.0 Traceback (most recent call last): File "/usr/sbin/ceph-volume", line 11, in <module> load_entry_point('ceph-volume==1.0.0', 'console_scripts', 'ceph-volume')() File "/usr/lib/python3/dist-packages/ceph_volume/main.py", line 39, in __init__ self.main(self.argv) File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line 59, in newfunc return f(*a, **kw) File "/usr/lib/python3/dist-packages/ceph_volume/main.py", line 150, in main terminal.dispatch(self.mapper, subcommand_args) File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line 194, in dispatch instance.main() File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/main.py", line 42, in main terminal.dispatch(self.mapper, self.argv) File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line 194, in dispatch instance.main() File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/main.py", line 42, in main terminal.dispatch(self.mapper, self.argv) File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line 194, in dispatch instance.main() File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/create.py", line 77, in main self.create(args) File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line 16, in is_root return func(*a, **kw) File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/create.py", line 26, in create prepare_step.safe_prepare(args) File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py", line 246, in safe_prepare self.prepare() File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line 16, in is_root return func(*a, **kw) File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py", line 327, in prepare block_lv = self.prepare_data_device('block', osd_fsid) File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py", line 223, in prepare_data_device **kwargs) File "/usr/lib/python3/dist-packages/ceph_volume/api/lvm.py", line 1185, in create_lv process.run(command) File "/usr/lib/python3/dist-packages/ceph_volume/process.py", line 153, in run raise RuntimeError(msg) RuntimeError: command returned non-zero exit status: 5 So some one could please help me to figure solution for this issue . May be i am doing somewhere something wrong. If yes then please tell me. Also how sizing can be done for osds i mean how we can allow ceph-ansible to not to use default sizing? -- Thanks and Regards, Hemant Sonawane _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx