Re: Ceph Stretch Cluster - df pool size (Max Avail)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Joachim,


yes i assigned the stretch rule to the pool (4x replica / 2x min). The rule says that two replicas should be in every datacenter.


$ ceph osd tree
ID   CLASS  WEIGHT    TYPE NAME                   STATUS  REWEIGHT  PRI-AFF
 -1         62.87799  root default
-17         31.43900      datacenter site1
-15         31.43900          rack b7
 -3         10.48000              host host01
  0    ssd   1.74699                  osd.0           up   1.00000  1.00000
  1    ssd   1.74699                  osd.1           up   1.00000  1.00000
  2    ssd   1.74699                  osd.2           up   1.00000  1.00000
  3    ssd   1.74699                  osd.3           up   1.00000  1.00000
  4    ssd   1.74699                  osd.4           up   1.00000  1.00000
  5    ssd   1.74699                  osd.5           up   1.00000  1.00000
 -5         10.48000              host host02
  6    ssd   1.74699                  osd.6           up   1.00000  1.00000
  7    ssd   1.74699                  osd.7           up   1.00000  1.00000
  8    ssd   1.74699                  osd.8           up   1.00000  1.00000
  9    ssd   1.74699                  osd.9           up   1.00000  1.00000
 10    ssd   1.74699                  osd.10          up   1.00000  1.00000
 11    ssd   1.74699                  osd.11          up   1.00000  1.00000
 -7         10.48000              host host03
 12    ssd   1.74699                  osd.12          up   1.00000  1.00000
 13    ssd   1.74699                  osd.13          up   1.00000  1.00000
 14    ssd   1.74699                  osd.14          up   1.00000  1.00000
 15    ssd   1.74699                  osd.15          up   1.00000  1.00000
 16    ssd   1.74699                  osd.16          up   1.00000  1.00000
 17    ssd   1.74699                  osd.17          up   1.00000  1.00000
-18         31.43900      datacenter site2
-16         31.43900          rack h2
 -9         10.48000              host host04
 18    ssd   1.74699                  osd.18          up   1.00000  1.00000
 19    ssd   1.74699                  osd.19          up   1.00000  1.00000
 20    ssd   1.74699                  osd.20          up   1.00000  1.00000
 21    ssd   1.74699                  osd.21          up   1.00000  1.00000
 22    ssd   1.74699                  osd.22          up   1.00000  1.00000
 23    ssd   1.74699                  osd.23          up   1.00000  1.00000
-11         10.48000              host host05
 24    ssd   1.74699                  osd.24          up   1.00000  1.00000
 25    ssd   1.74699                  osd.25          up   1.00000  1.00000
 26    ssd   1.74699                  osd.26          up   1.00000  1.00000
 27    ssd   1.74699                  osd.27          up   1.00000  1.00000
 28    ssd   1.74699                  osd.28          up   1.00000  1.00000
 29    ssd   1.74699                  osd.29          up   1.00000  1.00000
-13         10.48000              host host06
 30    ssd   1.74699                  osd.30          up   1.00000  1.00000
 31    ssd   1.74699                  osd.31          up   1.00000  1.00000
 32    ssd   1.74699                  osd.32          up   1.00000  1.00000
 33    ssd   1.74699                  osd.33          up   1.00000  1.00000
 34    ssd   1.74699                  osd.34          up   1.00000  1.00000
 35    ssd   1.74699                  osd.35          up   1.00000  1.00000


So regarding my calculation it should be


(6x Nodes * 6x SSD * 1,8TB) / 4 = 16 TB


Is this maybe a bug in the stretch mode that i only get displayed half the size available?


Regards,

Kilian


________________________________
Von: Clyso GmbH - Ceph Foundation Member <joachim.kraftmayer@xxxxxxxxx>
Gesendet: Mittwoch, 22. Juni 2022 18:20:59
An: Kilian Ries; ceph-users@xxxxxxx
Betreff: Re:  Ceph Stretch Cluster - df pool size (Max Avail)

Hi Kilian,

we do not currently use this mode of ceph clustering. but normally you
need to assign the crush rule to the pool as well, otherwise it will
take rule 0 as default.

the following calculation for rule 0 would also work approximately:

(3 Nodes *6 x SSD *1,8TB)/4 = 8,1 TB

hope it helps, Joachim


___________________________________
Clyso GmbH - Ceph Foundation Member

Am 22.06.22 um 18:09 schrieb Kilian Ries:
> Hi,
>
>
> i'm running a ceph stretch cluster with two datacenters. Each of the datacenters has 3x OSD nodes (in total 6x) and 2x monitors. A third monitor is deployed as arbiter node in a third datacenter.
>
>
> Each OSD node has 6x SSDs with 1,8 TB storage - that gives me a total of about 63 TB storage (6x nodes * 6x SSD * 1,8TB = 63TB)c.
>
>
> In stretch mode my pool is configured with replication 4x - and as far as i unterstand this should give me a max pool storage size of ~15TB (63TB / 4 = 15,75TB. But if i run "ceph df" it shows me only the half, about ~7,5TB.
>
>
>
> $ ceph df
>
> --- RAW STORAGE ---
>
> CLASS    SIZE   AVAIL    USED  RAW USED  %RAW USED
>
> ssd    63 TiB  63 TiB  35 GiB    35 GiB       0.05
>
> TOTAL  63 TiB  63 TiB  35 GiB    35 GiB       0.05
>
>
>
> --- POOLS ---
>
> POOL                   ID  PGS   STORED  OBJECTS    USED  %USED  MAX AVAIL
>
> device_health_metrics   1    4  4.4 MiB       36  17 MiB      0    7.5 TiB
>
> vm_stretch              2   64  8.2 GiB    2.19k  33 GiB   0.11    7.5 TiB
>
>
>
> My replication rule is from the ceph documentation:
>
>
> https://docs.ceph.com/en/latest/rados/operations/stretch-mode/
>
>
> rule stretch_rule {
>
>          id 1
>
>          min_size 1
>
>          max_size 10
>
>          type replicated
>
>          step take site1
>
>          step chooseleaf firstn 2 type host
>
>          step emit
>
>          step take site2
>
>          step chooseleaf firstn 2 type host
>
>          step emit
>
> }
>
>
>
> Any idea why ceph shows me only about half the size i should be able to use (with 4x replicaton on the pool) ?
>
>
> Thanks,
>
> Regards
>
>
> Kilian
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux