Re: radosgw multi site different period

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



both cluster are in the same epoch and period:
 
root@ceph-a-1:~# radosgw-admin period get-current 
{
    "current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569"
}

root@ceph-b-1:~# radosgw-admin period get-current 
{
    "current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569"
}

but the sync state is still "master is on a different period":

root@ceph-b-1:~# radosgw-admin sync status
          realm 833e65be-268f-42c2-8f3c-9bab83ebbff2 (myrealm)
      zonegroup 15550dc6-a761-473f-81e8-0dc6cc5106bd (ceph)
           zone 082cd970-bd25-4cbc-a5fd-20f3b3f9dbd2 (b)
  metadata sync syncing
                full sync: 0/64 shards
                master is on a different period:
master_period=b7392c41-9cbe-4d92-ad03-db607dd7d569
local_period=d306a847-77a6-4306-87c9-0bb4fa16cdc4
                incremental sync: 64/64 shards
                metadata is caught up with master
      data sync source: 51019cee-86fb-4b39-b6ba-282171c459c6 (a)
                        syncing
                        full sync: 0/128 shards
                        incremental sync: 128/128 shards
                        data is caught up with source


Am Dienstag, den 14.11.2017, 18:21 +0100 schrieb Kim-Norman Sahm:
> both cluster are in the same epoch and period:
> 
> root@ceph-a-1:~# radosgw-admin period get-current 
> {
>     "current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569"
> }
> 
> root@ceph-b-1:~# radosgw-admin period get-current 
> {
>     "current_period": "b7392c41-9cbe-4d92-ad03-db607dd7d569"
> }
> 
> Am Dienstag, den 14.11.2017, 17:05 +0000 schrieb David Turner:
> > 
> > I'm assuming you've looked at the period in both places `radosgw-
> > admin period get` and confirmed that the second site is behind the
> > master site (based on epochs).  I'm also assuming (since you linked
> > the instructions) that you've done `radosgw-admin period pull` on
> > the
> > second site to get any period updates that have been done to the
> > master site.
> > 
> > If my assumptions are wrong.  Then you should do those things.  If
> > my
> > assumptions are correct, then running `radosgw-admin period update
> > --
> > commit` on the the master site and `radosgw-admin period pull` on
> > the
> > second site might fix this.  If you've already done that as well
> > (as
> > they're steps in the article you linked), then you need someone
> > smarter than I am to chime in.
> > 
> > On Tue, Nov 14, 2017 at 11:35 AM Kim-Norman Sahm <kisahm@t-online.d
> > e>
> > wrote:
> > > 
> > > hi,
> > > 
> > > i've installed a ceph multi site setup with two ceph clusters and
> > > each
> > > one radosgw.
> > > the multi site setup was in sync, so i tried a failover.
> > > cluster A is going down and i've changed the zone (b) on cluster
> > > b
> > > to
> > > the new master zone.
> > > it's working fine.
> > > 
> > > now i start the cluster A and try to switch back the master zone
> > > to
> > > A.
> > > cluster A believes that he is the master, cluster b is secondary.
> > > but on the secondary is a different period and the bucket delta
> > > is
> > > not
> > > synced to the new master zone:
> > > 
> > > root@ceph-a-1:~# radosgw-admin sync status
> > >           realm 833e65be-268f-42c2-8f3c-9bab83ebbff2 (myrealm)
> > >       zonegroup 15550dc6-a761-473f-81e8-0dc6cc5106bd (ceph)
> > >            zone 51019cee-86fb-4b39-b6ba-282171c459c6 (a)
> > >   metadata sync no sync (zone is master)
> > >       data sync source: 082cd970-bd25-4cbc-a5fd-20f3b3f9dbd2 (b)
> > >                         syncing
> > >                         full sync: 0/128 shards
> > >                         incremental sync: 128/128 shards
> > >                         data is caught up with source
> > > 
> > > root@ceph-b-1:~# radosgw-admin sync status
> > >           realm 833e65be-268f-42c2-8f3c-9bab83ebbff2 (myrealm)
> > >       zonegroup 15550dc6-a761-473f-81e8-0dc6cc5106bd (ceph)
> > >            zone 082cd970-bd25-4cbc-a5fd-20f3b3f9dbd2 (b)
> > >   metadata sync syncing
> > >                 full sync: 0/64 shards
> > >                 master is on a different period:
> > > master_period=b7392c41-9cbe-4d92-ad03-db607dd7d569
> > > local_period=d306a847-77a6-4306-87c9-0bb4fa16cdc4
> > >                 incremental sync: 64/64 shards
> > >                 metadata is caught up with master
> > >       data sync source: 51019cee-86fb-4b39-b6ba-282171c459c6 (a)
> > >                         syncing
> > >                         full sync: 0/128 shards
> > >                         incremental sync: 128/128 shards
> > >                         data is caught up with source
> > > 
> > > how can i force sync the period and the bucket deltas?
> > > i've used this howto: http://docs.ceph.com/docs/master/radosgw/mu
> > > lt
> > > isit
> > > e/
> > > 
> > > br Kim
> > > _______________________________________________
> > > ceph-users mailing list
> > > ceph-users@xxxxxxxxxxxxxx
> > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> > > 
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux