Thank you guys for your replies. The "used space" there is exactly that. It is the accounting for Rocks.DB and WAL. ``` RAW USED: The sum of USED space and the space allocated the db and wal BlueStore partitions. ``` There is one detail I do not understand. We are off-loading WAL and RocksDB to an NVME device; however, Ceph still seems to think that we use our data plane disks to store those elements. We have about 375TB (5 * 5 * 15) in HDD disks, and Ceph seems to be discounting from the usable space the volume (space) dedicated to WAL and Rocks.DB, which are applied into different disks; therefore, it shows as usable space 364 TB (after removing the space dedicated to WAL and Rocks.DB, which are in another device). Is that a bug of some sort? On Tue, Apr 4, 2023 at 6:31 AM Igor Fedotov <igor.fedotov@xxxxxxxx> wrote: > Please also note that total cluster size reported below as SIZE > apparently includes DB volumes: > > # ceph df > --- RAW STORAGE --- > CLASS SIZE AVAIL USED RAW USED %RAW USED > hdd 373 TiB 364 TiB 9.3 TiB 9.3 TiB 2.50 > > On 4/4/2023 12:22 PM, Igor Fedotov wrote: > > Do you have standalone DB volumes for your OSD? > > > > If so then highly likely RAW usage is that high due to DB volumes > > space is considered as in-use one already. > > > > Could you please share "ceph osd df tree" output to prove that? > > > > > > Thanks, > > > > Igor > > > > On 4/4/2023 4:25 AM, Work Ceph wrote: > >> Hello guys! > >> > >> > >> We noticed an unexpected situation. In a recently deployed Ceph > >> cluster we > >> are seeing a raw usage, that is a bit odd. We have the following setup: > >> > >> > >> We have a new cluster with 5 nodes with the following setup: > >> > >> - 128 GB of RAM > >> - 2 cpus Intel(R) Intel Xeon Silver 4210R > >> - 1 NVME of 2 TB for the rocks DB caching > >> - 5 HDDs of 14TB > >> - 1 NIC dual port of 25GiB in BOND mode. > >> > >> > >> Right after deploying the Ceph cluster, we see a raw usage of about > >> 9TiB. > >> However, no load has been applied onto the cluster. Have you guys > >> seen such > >> a situation? Or, can you guys help understand it? > >> > >> > >> We are using Ceph Octopus, and we have set the following configurations: > >> > >> ``` > >> > >> ceph_conf_overrides: > >> > >> global: > >> > >> osd pool default size: 3 > >> > >> osd pool default min size: 1 > >> > >> osd pool default pg autoscale mode: "warn" > >> > >> perf: true > >> > >> rocksdb perf: true > >> > >> mon: > >> > >> mon osd down out interval: 120 > >> > >> osd: > >> > >> bluestore min alloc size hdd: 65536 > >> > >> > >> ``` > >> > >> > >> Any tip or help on how to explain this situation is welcome! > >> _______________________________________________ > >> ceph-users mailing list -- ceph-users@xxxxxxx > >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > -- > Igor Fedotov > Ceph Lead Developer > > Looking for help with your Ceph cluster? Contact us at https://croit.io > > croit GmbH, Freseniusstr. 31h, 81247 Munich > CEO: Martin Verges - VAT-ID: DE310638492 > Com. register: Amtsgericht Munich HRB 231263 > Web: https://croit.io | YouTube: https://goo.gl/PGE1Bx > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx