Hi Lukasz, RGW uses Omap objects for the index pool; Omaps are stored in Rocksdb database of each osd, not on the actual index pool, so by putting DB/WALL on an NVMe as you mentioned, you are already configuring the index pool on a non-rotational drive, you don't need to do anything else. You just need to size your DB/WALL partition accordingly. For RGW/object storage, a good starting point for the DB/Wall sizing is 4%. Example of Omap entries in the index pool using 0 bytes, as they are stored in Rocksdb: # rados -p default.rgw.buckets.index listomapkeys .dir.7fb0a3df-9553-4a76-938d-d23711e67677.34162.1.2 file1 file2 file4 file10 rados df -p default.rgw.buckets.index POOL_NAME USED OBJECTS CLONES COPIES MISSING_ON_PRIMARY UNFOUND DEGRADED RD_OPS RD WR_OPS WR USED COMPR UNDER COMPR default.rgw.buckets.index 0 B 11 0 33 0 0 0 208 207 KiB 41 20 KiB 0 B 0 B # rados -p default.rgw.buckets.index stat .dir.7fb0a3df-9553-4a76-938d-d23711e67677.34162.1.2 default.rgw.buckets.index/.dir.7fb0a3df-9553-4a76-938d-d23711e67677.34162.1.2 mtime 2022-12-20T07:32:11.000000-0500, size 0 On Sun, Apr 7, 2024 at 10:06 PM Lukasz Borek <lukasz@xxxxxxxxxxxx> wrote: > Hi! > > I'm working on a POC cluster setup dedicated to backup app writing objects > via s3 (large objects, up to 1TB transferred via multipart upload process). > > Initial setup is 18 storage nodes (12HDDs + 1 NVME card for DB/WALL) + EC > pool. Plan is to use cephadm. > > I'd like to follow good practice and put the RGW index pool on a > no-rotation drive. Question is how to do it? > > - replace a few HDDs (1 per node) with a SSD (how many? 4-6-8?) > - reserve space on NVME drive on each node, create lv based OSD and let > rgb index use the same NVME drive as DB/WALL > > Thoughts? > > -- > Lukasz > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx