Hello,
We have exactly the same problem. Did you find an answer or should we
open a bug report?
Sincerely,
Nicolas.
Le 23/06/2022 à 11:42, Kilian Ries a écrit :
Hi Joachim,
yes i assigned the stretch rule to the pool (4x replica / 2x min). The rule says that two replicas should be in every datacenter.
$ ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 62.87799 root default
-17 31.43900 datacenter site1
-15 31.43900 rack b7
-3 10.48000 host host01
0 ssd 1.74699 osd.0 up 1.00000 1.00000
1 ssd 1.74699 osd.1 up 1.00000 1.00000
2 ssd 1.74699 osd.2 up 1.00000 1.00000
3 ssd 1.74699 osd.3 up 1.00000 1.00000
4 ssd 1.74699 osd.4 up 1.00000 1.00000
5 ssd 1.74699 osd.5 up 1.00000 1.00000
-5 10.48000 host host02
6 ssd 1.74699 osd.6 up 1.00000 1.00000
7 ssd 1.74699 osd.7 up 1.00000 1.00000
8 ssd 1.74699 osd.8 up 1.00000 1.00000
9 ssd 1.74699 osd.9 up 1.00000 1.00000
10 ssd 1.74699 osd.10 up 1.00000 1.00000
11 ssd 1.74699 osd.11 up 1.00000 1.00000
-7 10.48000 host host03
12 ssd 1.74699 osd.12 up 1.00000 1.00000
13 ssd 1.74699 osd.13 up 1.00000 1.00000
14 ssd 1.74699 osd.14 up 1.00000 1.00000
15 ssd 1.74699 osd.15 up 1.00000 1.00000
16 ssd 1.74699 osd.16 up 1.00000 1.00000
17 ssd 1.74699 osd.17 up 1.00000 1.00000
-18 31.43900 datacenter site2
-16 31.43900 rack h2
-9 10.48000 host host04
18 ssd 1.74699 osd.18 up 1.00000 1.00000
19 ssd 1.74699 osd.19 up 1.00000 1.00000
20 ssd 1.74699 osd.20 up 1.00000 1.00000
21 ssd 1.74699 osd.21 up 1.00000 1.00000
22 ssd 1.74699 osd.22 up 1.00000 1.00000
23 ssd 1.74699 osd.23 up 1.00000 1.00000
-11 10.48000 host host05
24 ssd 1.74699 osd.24 up 1.00000 1.00000
25 ssd 1.74699 osd.25 up 1.00000 1.00000
26 ssd 1.74699 osd.26 up 1.00000 1.00000
27 ssd 1.74699 osd.27 up 1.00000 1.00000
28 ssd 1.74699 osd.28 up 1.00000 1.00000
29 ssd 1.74699 osd.29 up 1.00000 1.00000
-13 10.48000 host host06
30 ssd 1.74699 osd.30 up 1.00000 1.00000
31 ssd 1.74699 osd.31 up 1.00000 1.00000
32 ssd 1.74699 osd.32 up 1.00000 1.00000
33 ssd 1.74699 osd.33 up 1.00000 1.00000
34 ssd 1.74699 osd.34 up 1.00000 1.00000
35 ssd 1.74699 osd.35 up 1.00000 1.00000
So regarding my calculation it should be
(6x Nodes * 6x SSD * 1,8TB) / 4 = 16 TB
Is this maybe a bug in the stretch mode that i only get displayed half the size available?
Regards,
Kilian
________________________________
Von: Clyso GmbH - Ceph Foundation Member <joachim.kraftmayer(a)clyso.com>
Gesendet: Mittwoch, 22. Juni 2022 18:20:59
An: Kilian Ries; ceph-users(a)ceph.io
Betreff: Re: Ceph Stretch Cluster - df pool size (Max Avail)
Hi Kilian,
we do not currently use this mode of ceph clustering. but normally you
need to assign the crush rule to the pool as well, otherwise it will
take rule 0 as default.
the following calculation for rule 0 would also work approximately:
(3 Nodes *6 x SSD *1,8TB)/4 = 8,1 TB
hope it helps, Joachim
___________________________________
Clyso GmbH - Ceph Foundation Member
Am 22.06.22 um 18:09 schrieb Kilian Ries:
Hi,
i'm running a ceph stretch cluster with two datacenters. Each of the datacenters has 3x OSD nodes (in total 6x) and 2x monitors. A third monitor is deployed as arbiter node in a third datacenter.
Each OSD node has 6x SSDs with 1,8 TB storage - that gives me a total of about 63 TB storage (6x nodes * 6x SSD * 1,8TB = 63TB)c.
In stretch mode my pool is configured with replication 4x - and as far as i unterstand this should give me a max pool storage size of ~15TB (63TB / 4 = 15,75TB. But if i run "ceph df" it shows me only the half, about ~7,5TB.
$ ceph df
--- RAW STORAGE ---
CLASS SIZE AVAIL USED RAW USED %RAW USED
ssd 63 TiB 63 TiB 35 GiB 35 GiB 0.05
TOTAL 63 TiB 63 TiB 35 GiB 35 GiB 0.05
--- POOLS ---
POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL
device_health_metrics 1 4 4.4 MiB 36 17 MiB 0 7.5 TiB
vm_stretch 2 64 8.2 GiB 2.19k 33 GiB 0.11 7.5 TiB
My replication rule is from the ceph documentation:
https://docs.ceph.com/en/latest/rados/operations/stretch-mode/
rule stretch_rule {
id 1
min_size 1
max_size 10
type replicated
step take site1
step chooseleaf firstn 2 type host
step emit
step take site2
step chooseleaf firstn 2 type host
step emit
}
Any idea why ceph shows me only about half the size i should be able to use (with 4x replicaton on the pool) ?
Thanks,
Regards
Kilian
_______________________________________________
ceph-users mailing list -- ceph-users(a)ceph.io
To unsubscribe send an email to ceph-users-leave(a)ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx