RGW multisite replication doesn't start

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi *,

I have 2 virtual one-node-clusters configured for multisite RGW. In the beginning the replication actually worked for some hundred MB or so, and then it stopped. In the meantime I wiped both RGWs twice to make sure the configuration is right (including wiping all pools clean). I don't see any errors in the logs but nothing happens on the secondary site. Both clusters are healthy, RGWs run with https. Uploading data directly to the secondary site also works, so the configuration seems ok to me.

These is the current rgw status:

---snip---
primary:~ # radosgw-admin sync status
          realm c7d5fd30-9c06-46a1-baf4-497f95bf3abc (hamburg)
      zonegroup 68adec15-aace-403d-bd63-f5182a6437b1 (zg-hamburg)
           zone 0fb33fa1-8110-4179-ae45-acf5f5f825c5 (z-primary)
  metadata sync no sync (zone is master)


secondary:~ # radosgw-admin sync status
2020-09-17T09:34:59.593+0200 7fdd3e706a40 1 Cannot find zone id=93ece7a6-beef-4f4e-841a-60ba0405f192 (name=z-secondary), switching to local zonegroup configuration
          realm c7d5fd30-9c06-46a1-baf4-497f95bf3abc (hamburg)
      zonegroup 68adec15-aace-403d-bd63-f5182a6437b1 (zg-hamburg)
           zone 93ece7a6-beef-4f4e-841a-60ba0405f192 (z-secondary)
  metadata sync syncing
                full sync: 64/64 shards
                full sync: 3 entries to sync
                incremental sync: 0/64 shards
                metadata is behind on 64 shards
behind shards: [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63]
      data sync source: 0fb33fa1-8110-4179-ae45-acf5f5f825c5 (z-primary)
                        syncing
                        full sync: 0/128 shards
                        incremental sync: 128/128 shards
                        data is caught up with source
---snip---


Since the data was not replicated I ran a 'radosgw-admin metadata sync run --source-zone=z-primary' but it never finishes. If I do the same with data it will show that all shards are behind on data but nothing will happen either. I also don't understand the 'Cannot find zone id=93ece7a6-beef-4f4e-841a-60ba0405f192 (name=z-secondary), switching to local zonegroup configuration' message but this didn't break the replication in the first attempt, so I ignored it. Or is this something I should fix first (if yes, how)?

Can anyone point me to what's going on here? I can provide more details if necessary, just let me know.

Thank you!
Eugen
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux