Hi, i'm running a ceph stretch cluster with two datacenters. Each of the datacenters has 3x OSD nodes (in total 6x) and 2x monitors. A third monitor is deployed as arbiter node in a third datacenter. Each OSD node has 6x SSDs with 1,8 TB storage - that gives me a total of about 63 TB storage (6x nodes * 6x SSD * 1,8TB = 63TB)c. In stretch mode my pool is configured with replication 4x - and as far as i unterstand this should give me a max pool storage size of ~15TB (63TB / 4 = 15,75TB. But if i run "ceph df" it shows me only the half, about ~7,5TB. $ ceph df --- RAW STORAGE --- CLASS SIZE AVAIL USED RAW USED %RAW USED ssd 63 TiB 63 TiB 35 GiB 35 GiB 0.05 TOTAL 63 TiB 63 TiB 35 GiB 35 GiB 0.05 --- POOLS --- POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL device_health_metrics 1 4 4.4 MiB 36 17 MiB 0 7.5 TiB vm_stretch 2 64 8.2 GiB 2.19k 33 GiB 0.11 7.5 TiB My replication rule is from the ceph documentation: https://docs.ceph.com/en/latest/rados/operations/stretch-mode/ rule stretch_rule { id 1 min_size 1 max_size 10 type replicated step take site1 step chooseleaf firstn 2 type host step emit step take site2 step chooseleaf firstn 2 type host step emit } Any idea why ceph shows me only about half the size i should be able to use (with 4x replicaton on the pool) ? Thanks, Regards Kilian _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx