Hello! I'm using Nautilus 14.2.16 with Multisite RGW setup. I have 2 zones. Working as Active-Passive(Read-Only) On master zone "ceph df" result is: POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL prod.rgw.buckets.index 54 128 844 GiB 437.52k 844 GiB 6.43 4.0 TiB prod.rgw.buckets.non-ec 55 32 195 MiB 2.70k 246 MiB 0 4.0 TiB * prod.rgw.buckets.data 56 2048 856 TiB 1.08G 1.3 PiB 65.75 553 TiB* On secondary zone "ceph df" result is: POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL bck.rgw.buckets.index 20 256 137 GiB 467.21k 137 GiB 0.55 8.0 TiB bck.rgw.buckets.non-ec 21 32 0 B 0 0 B 0 8.0 TiB * bck.rgw.buckets.data 22 1024 931 TiB 653.67M 1.3 PiB 85.85 178 TiB* As you can see master zone stored is *856TiB* and Secondary zone *931TiB* but used size is equal. Both cluster are 10 nodes with 8+2 EC pools. The only differences are: - Master pool has 2048PG - Secondary pool has 1024PG - On the master zone's bucket pool compression is *off. *- On the secondary zone's bucket pool compression is *on*, How can I dig into this? What am I missing? _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx