both cluster are in the same epoch and period: root@ceph-a-1:~# radosgw-admin period get-current { "current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569" } root@ceph-b-1:~# radosgw-admin period get-current { "current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569" } but the sync state is still "master is on a different period": root@ceph-b-1:~# radosgw-admin sync status realm 833e65be-268f-42c2-8f3c-9bab83ebbff2 (myrealm) zonegroup 15550dc6-a761-473f-81e8-0dc6cc5106bd (ceph) zone 082cd970-bd25-4cbc-a5fd-20f3b3f9dbd2 (b) metadata sync syncing full sync: 0/64 shards master is on a different period: master_period=b7392c41-9cbe-4d92-ad03-db607dd7d569 local_period=d306a847-77a6-4306-87c9-0bb4fa16cdc4 incremental sync: 64/64 shards metadata is caught up with master data sync source: 51019cee-86fb-4b39-b6ba-282171c459c6 (a) syncing full sync: 0/128 shards incremental sync: 128/128 shards data is caught up with source Am Dienstag, den 14.11.2017, 18:21 +0100 schrieb Kim-Norman Sahm: > both cluster are in the same epoch and period: > > root@ceph-a-1:~# radosgw-admin period get-current > { > "current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569" > } > > root@ceph-b-1:~# radosgw-admin period get-current > { > "current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569" > } > > Am Dienstag, den 14.11.2017, 17:05 +0000 schrieb David Turner: > > > > I'm assuming you've looked at the period in both places `radosgw- > > admin period get` and confirmed that the second site is behind the > > master site (based on epochs). I'm also assuming (since you linked > > the instructions) that you've done `radosgw-admin period pull` on > > the > > second site to get any period updates that have been done to the > > master site. > > > > If my assumptions are wrong. Then you should do those things. If > > my > > assumptions are correct, then running `radosgw-admin period update > > -- > > commit` on the the master site and `radosgw-admin period pull` on > > the > > second site might fix this. If you've already done that as well > > (as > > they're steps in the article you linked), then you need someone > > smarter than I am to chime in. > > > > On Tue, Nov 14, 2017 at 11:35 AM Kim-Norman Sahm <kisahm@t-online.d > > e> > > wrote: > > > > > > hi, > > > > > > i've installed a ceph multi site setup with two ceph clusters and > > > each > > > one radosgw. > > > the multi site setup was in sync, so i tried a failover. > > > cluster A is going down and i've changed the zone (b) on cluster > > > b > > > to > > > the new master zone. > > > it's working fine. > > > > > > now i start the cluster A and try to switch back the master zone > > > to > > > A. > > > cluster A believes that he is the master, cluster b is secondary. > > > but on the secondary is a different period and the bucket delta > > > is > > > not > > > synced to the new master zone: > > > > > > root@ceph-a-1:~# radosgw-admin sync status > > > realm 833e65be-268f-42c2-8f3c-9bab83ebbff2 (myrealm) > > > zonegroup 15550dc6-a761-473f-81e8-0dc6cc5106bd (ceph) > > > zone 51019cee-86fb-4b39-b6ba-282171c459c6 (a) > > > metadata sync no sync (zone is master) > > > data sync source: 082cd970-bd25-4cbc-a5fd-20f3b3f9dbd2 (b) > > > syncing > > > full sync: 0/128 shards > > > incremental sync: 128/128 shards > > > data is caught up with source > > > > > > root@ceph-b-1:~# radosgw-admin sync status > > > realm 833e65be-268f-42c2-8f3c-9bab83ebbff2 (myrealm) > > > zonegroup 15550dc6-a761-473f-81e8-0dc6cc5106bd (ceph) > > > zone 082cd970-bd25-4cbc-a5fd-20f3b3f9dbd2 (b) > > > metadata sync syncing > > > full sync: 0/64 shards > > > master is on a different period: > > > master_period=b7392c41-9cbe-4d92-ad03-db607dd7d569 > > > local_period=d306a847-77a6-4306-87c9-0bb4fa16cdc4 > > > incremental sync: 64/64 shards > > > metadata is caught up with master > > > data sync source: 51019cee-86fb-4b39-b6ba-282171c459c6 (a) > > > syncing > > > full sync: 0/128 shards > > > incremental sync: 128/128 shards > > > data is caught up with source > > > > > > how can i force sync the period and the bucket deltas? > > > i've used this howto: http://docs.ceph.com/docs/master/radosgw/mu > > > lt > > > isit > > > e/ > > > > > > br Kim > > > _______________________________________________ > > > ceph-users mailing list > > > ceph-users@xxxxxxxxxxxxxx > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com