Hello all - Hi. After running this dev cluster with a single osd (/dev/sda) hdd in each node (6), I want to now put the metadata on the nvme disk which is also used as boot. There is plenty of space left on the nvme, so I re-did the logical volumes to make a 50gb LV for the metadata, thinking I'd put the metdata on thenvme/LV and use the entire /dev/sda as data. Before I really go down this rabbit hole, just want opinions if this is something that should work? I've tried both Cepth 15.2.7 & Ceph 15.2.8, each with different errors. This particular trace is ceph15.2.8. This is under Rook, so rook is doing: <snip> exec: Running command: stdbuf -oL ceph-volume --log-path /tmp/ceph-log lvm batch --prepare --bluestore --yes --osds-per-device 1 /dev/sda --db-devices /dev/cephDB/database --report provision 2021-01-28 01:46:56.186043 D | exec: --> passed data devices: 1 physical, 0 LVM provision 2021-01-28 01:46:56.186074 D | exec: --> relative data size: 1.0 provision 2021-01-28 01:46:56.186079 D | exec: --> passed block_db devices: 0 physical, 1 LVM provision 2021-01-28 01:46:56.186092 D | exec: provision 2021-01-28 01:46:56.186104 D | exec: Total OSDs: 1 provision 2021-01-28 01:46:56.186107 D | exec: provision 2021-01-28 01:46:56.186111 D | exec: Type Path LV Size % of device provision 2021-01-28 01:46:56.186114 D | exec: ---------------------------------------------------------------------------------------------------- provision 2021-01-28 01:46:56.186117 D | exec: data /dev/sda 3.64 TB 100.00% provision 2021-01-28 01:46:56.186121 D | exec: block_db /dev/cephDB/database 51.65 GB 10000.00%<snip> It fails with the stack trace below, essentially complaining it can't PARTUUID for the LV. exec: Running command: /usr/sbin/lvcreate --yes -l 953861 -n osd-block-2acd94f6-0fed-423b-8540-ae93c0621c2e ceph-b6da5679-543b-4a79-9cc2-e4e308ba61a4 exec: stderr: Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, LVM will manage logical volume symlinks in device directory. exec: stderr: Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, LVM will obtain device list by scanning device directory. exec: stderr: Udev is running and DM_DISABLE_UDEV environment variable is set. Bypassing udev, device-mapper library will manage device nodes in device directory. exec: stdout: Logical volume "osd-block-2acd94f6-0fed-423b-8540-ae93c0621c2e" created. exec: --> blkid could not detect a PARTUUID for device: /dev/cephDB/database exec: --> Was unable to complete a new OSD, will rollback changes exec: Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd purge-new osd.0 --yes-i-really-mean-it exec: stderr: purged osd.0 exec: Traceback (most recent call last): exec: File "/usr/sbin/ceph-volume", line 11, in <module> exec: load_entry_point('ceph-volume==1.0.0', 'console_scripts', 'ceph-volume')() exec: File "/usr/lib/python3.6/site-packages/ceph_volume/main.py", line 40, in __init__ exec: self.main(self.argv) exec: File "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 59, in newfunc exec: return f(*a, **kw) exec: File "/usr/lib/python3.6/site-packages/ceph_volume/main.py", line 152, in main exec: terminal.dispatch(self.mapper, subcommand_args) exec: File "/usr/lib/python3.6/site-packages/ceph_volume/terminal.py", line 194, in dispatch exec: instance.main() exec: File "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/main.py", line 42, in main exec: terminal.dispatch(self.mapper, self.argv) exec: File "/usr/lib/python3.6/site-packages/ceph_volume/terminal.py", line 194, in dispatch exec: instance.main() exec: File "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 16, in is_root exec: return func(*a, **kw) exec: File "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/batch.py", line 415, in main exec: self._execute(plan) exec: File "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/batch.py", line 431, in _execute exec: p.safe_prepare(argparse.Namespace(**args)) exec: File "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/prepare.py", line 252, in safe_prepare exec: self.prepare() exec: File "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 16, in is_root exec: return func(*a, **kw) exec: File "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/prepare.py", line 382, in prepare exec: self.args.block_db_slots) exec: File "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/prepare.py", line 189, in setup_device exec: name_uuid = self.get_ptuuid(device_name) exec: File "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/prepare.py", line 135, in get_ptuuid exec: raise RuntimeError('unable to use device') exec: RuntimeError: unable to use device provision failed to configure devices: failed to initialize devices: failed ceph-volume: exit status 1 Thanks for any ideas/help. _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx