Hi Joachim, yes i assigned the stretch rule to the pool (4x replica / 2x min). The rule says that two replicas should be in every datacenter. $ ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 62.87799 root default -17 31.43900 datacenter site1 -15 31.43900 rack b7 -3 10.48000 host host01 0 ssd 1.74699 osd.0 up 1.00000 1.00000 1 ssd 1.74699 osd.1 up 1.00000 1.00000 2 ssd 1.74699 osd.2 up 1.00000 1.00000 3 ssd 1.74699 osd.3 up 1.00000 1.00000 4 ssd 1.74699 osd.4 up 1.00000 1.00000 5 ssd 1.74699 osd.5 up 1.00000 1.00000 -5 10.48000 host host02 6 ssd 1.74699 osd.6 up 1.00000 1.00000 7 ssd 1.74699 osd.7 up 1.00000 1.00000 8 ssd 1.74699 osd.8 up 1.00000 1.00000 9 ssd 1.74699 osd.9 up 1.00000 1.00000 10 ssd 1.74699 osd.10 up 1.00000 1.00000 11 ssd 1.74699 osd.11 up 1.00000 1.00000 -7 10.48000 host host03 12 ssd 1.74699 osd.12 up 1.00000 1.00000 13 ssd 1.74699 osd.13 up 1.00000 1.00000 14 ssd 1.74699 osd.14 up 1.00000 1.00000 15 ssd 1.74699 osd.15 up 1.00000 1.00000 16 ssd 1.74699 osd.16 up 1.00000 1.00000 17 ssd 1.74699 osd.17 up 1.00000 1.00000 -18 31.43900 datacenter site2 -16 31.43900 rack h2 -9 10.48000 host host04 18 ssd 1.74699 osd.18 up 1.00000 1.00000 19 ssd 1.74699 osd.19 up 1.00000 1.00000 20 ssd 1.74699 osd.20 up 1.00000 1.00000 21 ssd 1.74699 osd.21 up 1.00000 1.00000 22 ssd 1.74699 osd.22 up 1.00000 1.00000 23 ssd 1.74699 osd.23 up 1.00000 1.00000 -11 10.48000 host host05 24 ssd 1.74699 osd.24 up 1.00000 1.00000 25 ssd 1.74699 osd.25 up 1.00000 1.00000 26 ssd 1.74699 osd.26 up 1.00000 1.00000 27 ssd 1.74699 osd.27 up 1.00000 1.00000 28 ssd 1.74699 osd.28 up 1.00000 1.00000 29 ssd 1.74699 osd.29 up 1.00000 1.00000 -13 10.48000 host host06 30 ssd 1.74699 osd.30 up 1.00000 1.00000 31 ssd 1.74699 osd.31 up 1.00000 1.00000 32 ssd 1.74699 osd.32 up 1.00000 1.00000 33 ssd 1.74699 osd.33 up 1.00000 1.00000 34 ssd 1.74699 osd.34 up 1.00000 1.00000 35 ssd 1.74699 osd.35 up 1.00000 1.00000 So regarding my calculation it should be (6x Nodes * 6x SSD * 1,8TB) / 4 = 16 TB Is this maybe a bug in the stretch mode that i only get displayed half the size available? Regards, Kilian ________________________________ Von: Clyso GmbH - Ceph Foundation Member <joachim.kraftmayer@xxxxxxxxx> Gesendet: Mittwoch, 22. Juni 2022 18:20:59 An: Kilian Ries; ceph-users@xxxxxxx Betreff: Re: Ceph Stretch Cluster - df pool size (Max Avail) Hi Kilian, we do not currently use this mode of ceph clustering. but normally you need to assign the crush rule to the pool as well, otherwise it will take rule 0 as default. the following calculation for rule 0 would also work approximately: (3 Nodes *6 x SSD *1,8TB)/4 = 8,1 TB hope it helps, Joachim ___________________________________ Clyso GmbH - Ceph Foundation Member Am 22.06.22 um 18:09 schrieb Kilian Ries: > Hi, > > > i'm running a ceph stretch cluster with two datacenters. Each of the datacenters has 3x OSD nodes (in total 6x) and 2x monitors. A third monitor is deployed as arbiter node in a third datacenter. > > > Each OSD node has 6x SSDs with 1,8 TB storage - that gives me a total of about 63 TB storage (6x nodes * 6x SSD * 1,8TB = 63TB)c. > > > In stretch mode my pool is configured with replication 4x - and as far as i unterstand this should give me a max pool storage size of ~15TB (63TB / 4 = 15,75TB. But if i run "ceph df" it shows me only the half, about ~7,5TB. > > > > $ ceph df > > --- RAW STORAGE --- > > CLASS SIZE AVAIL USED RAW USED %RAW USED > > ssd 63 TiB 63 TiB 35 GiB 35 GiB 0.05 > > TOTAL 63 TiB 63 TiB 35 GiB 35 GiB 0.05 > > > > --- POOLS --- > > POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL > > device_health_metrics 1 4 4.4 MiB 36 17 MiB 0 7.5 TiB > > vm_stretch 2 64 8.2 GiB 2.19k 33 GiB 0.11 7.5 TiB > > > > My replication rule is from the ceph documentation: > > > https://docs.ceph.com/en/latest/rados/operations/stretch-mode/ > > > rule stretch_rule { > > id 1 > > min_size 1 > > max_size 10 > > type replicated > > step take site1 > > step chooseleaf firstn 2 type host > > step emit > > step take site2 > > step chooseleaf firstn 2 type host > > step emit > > } > > > > Any idea why ceph shows me only about half the size i should be able to use (with 4x replicaton on the pool) ? > > > Thanks, > > Regards > > > Kilian > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx