Multisite RGW - Secondary zone's data pool bigger than master

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello!

I'm using Nautilus 14.2.16 with Multisite RGW setup.
I have 2 zones. Working as Active-Passive(Read-Only)

On master zone "ceph df" result is:
    POOL                        ID     PGS      STORED      OBJECTS
 USED        %USED     MAX AVAIL
    prod.rgw.buckets.index      54      128     844 GiB     437.52k     844
GiB      6.43       4.0 TiB
    prod.rgw.buckets.non-ec     55       32     195 MiB       2.70k     246
MiB         0       4.0 TiB
*    prod.rgw.buckets.data       56     2048     856 TiB       1.08G
 1.3 PiB     65.75       553 TiB*

On secondary zone "ceph df" result is:
    POOL                       ID     PGS      STORED      OBJECTS
 USED        %USED     MAX AVAIL
    bck.rgw.buckets.index      20      256     137 GiB     467.21k     137
GiB      0.55       8.0 TiB
    bck.rgw.buckets.non-ec     21       32         0 B           0
 0 B         0       8.0 TiB
*    bck.rgw.buckets.data       22     1024     931 TiB     653.67M     1.3
PiB     85.85       178 TiB*

As you can see master zone stored is *856TiB* and Secondary zone *931TiB*
but used size is equal.
Both cluster are 10 nodes with 8+2 EC pools. The only differences are:
- Master pool has 2048PG
- Secondary pool has 1024PG
- On the master zone's bucket pool compression is
*off. *- On the secondary zone's bucket pool compression is *on*,

How can I dig into this?
What am I missing?
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux