Hello
We have a new Pacific cluster configured via Cephadm.
For the OSDs, the spec is like this, with the intention for DB and WAL to
be on NVMe:
spec:
data_devices:
rotational: true
db_devices:
model: SSDPE2KE032T8L
filter_logic: AND
objectstore: bluestore
wal_devices:
model: SSDPE2KE032T8L
This was after an initial attempt like this:
spec:
data_devices:
rotational: 1
db_devices:
rotational: 0
based on the suggestion in the docs that this would be sufficient for both
DB and WAL (
https://docs.ceph.com/en/pacific/cephadm/services/osd/#the-simple-case)
ended up with metadata on the HDD data disks, as demonstrated by quite a
lot of space being consumed even with no actual data.
With the new spec, the usage looks more normal. However, it's not clear
whether both DB and WAL are in fact on the faster devices as desired.
Here's an except of one of the new OSDs:
{
"id": 107,
"arch": "x86_64",
"back_iface": "",
"bluefs": "1",
"bluefs_dedicated_db": "0",
"bluefs_dedicated_wal": "1",
"bluefs_single_shared_device": "0",
"bluefs_wal_access_mode": "blk",
"bluefs_wal_block_size": "4096",
"bluefs_wal_dev_node": "/dev/dm-40",
"bluefs_wal_devices": "nvme0n1",
"bluefs_wal_driver": "KernelDevice",
"bluefs_wal_partition_path": "/dev/dm-40",
"bluefs_wal_rotational": "0",
"bluefs_wal_size": "355622453248",
"bluefs_wal_support_discard": "1",
"bluefs_wal_type": "ssd",
"bluestore_bdev_access_mode": "blk",
"bluestore_bdev_block_size": "4096",
"bluestore_bdev_dev_node": "/dev/dm-39",
"bluestore_bdev_devices": "sdr",
"bluestore_bdev_driver": "KernelDevice",
"bluestore_bdev_partition_path": "/dev/dm-39",
"bluestore_bdev_rotational": "1",
"bluestore_bdev_size": "8001561821184",
"bluestore_bdev_support_discard": "0",
"bluestore_bdev_type": "hdd",
"ceph_release": "pacific",
"ceph_version": "ceph version 16.2.7
(dd0603118f56ab514f133c8d2e3adfc983942503) pacific (stable)",
"ceph_version_short": "16.2.7",
8< >8
"container_image": "
quay.io/ceph/ceph@sha256:a39107f8d3daab4d756eabd6ee1630d1bc7f31eaa76fff41a77fa32d0b903061
",
"cpu": "AMD EPYC 7352 24-Core Processor",
"default_device_class": "hdd",
"device_ids":
"nvme0n1=SSDPE2KE032T8L_PHLN0195002R3P2BGN,sdr=LENOVO_ST8000NM010A_EX_WKD2CHZL0000E02930J6",
"device_paths":
"nvme0n1=/dev/disk/by-path/pci-0000:c1:00.0-nvme-1,sdr=/dev/disk/by-path/pci-0000:41:00.0-scsi-0:0:41:0",
"devices": "nvme0n1,sdr",
"distro": "centos",
"distro_description": "CentOS Stream 8",
"distro_version": "8",
8< >8
"journal_rotational": "0",
"kernel_description": "#1 SMP Thu Feb 10 16:11:23 UTC 2022",
"kernel_version": "4.18.0-365.el8.x86_64",
"mem_swap_kb": "4194300",
"mem_total_kb": "131583928",
"network_numa_unknown_ifaces": "back_iface,front_iface",
"objectstore_numa_nodes": "0",
"objectstore_numa_unknown_devices": "sdr",
"os": "Linux",
"osd_data": "/var/lib/ceph/osd/ceph-107",
"osd_objectstore": "bluestore",
"osdspec_affinity": "dashboard-admin-1645460246886",
"rotational": "1"
}
Note:
"bluefs_dedicated_db": "0",
"bluefs_dedicated_wal": "1",
"bluefs_single_shared_device": "0",
On one of our Nautilus clusters, we have:
"bluefs_single_shared_device": "1",
and the same on an Octopus cluster.
I've heard of the WAL being hosted in the DB, but not the other way
around...
Best Wishes,
Adam
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx