Resetting RGW Federated replication

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



This is embarrassing.  I ran a single radosgw-agent, verbose mode, with a
single thread... and everything works correctly.

I noticed the problem a couple days ago, and stopped radosgw-agent until I
had time to look into the issue.  Looking back through the logs, things
were working correctly a week ago.  I just didn't look far enough back in
time.

So it looks like my issues were some stale locks.  Likely caused by some
cluster stability issues I was having, due to a lot of concurrent
deep-scrubs.

On Tue, Sep 23, 2014 at 10:06 PM, Yehuda Sadeh <yehuda at redhat.com> wrote:

> On Tue, Sep 23, 2014 at 4:54 PM, Craig Lewis <clewis at centraldesktop.com>
> wrote:
> > I've had some issues in my secondary cluster.  I'd like to restart
> > replication from the beginning, without destroying the data in the
> secondary
> > cluster.
> >
> > Reading the radosgw-agent and Admin REST API code, I believe I just need
> to
> > stop replication, delete the secondary zone's log_pool, recreate the
> > log_pool, and restart replication.
> >
> > Anybody have any thoughts?  I'm still setting up some VMs to test this,
> > before I try it in production.
> >
> >
> >
> > Background:
> > I'm on Emperor (yeah, still need to upgrade).  I believe I ran into
> > http://tracker.ceph.com/issues/7595 .  My read of that patch is that it
> > prevents the problem from occurring, but doesn't correct corrupt data.  I
> > tried applying some of the suggested patches, but they only ignored the
> > error, rather than correcting it.  I finally dropped the corrupt pool.
> That
> > allowed the stock Emperor binaries to run without crashing.  The pool I
> > dropped was my secondary zone's log_pool.
> >
> > Before I dropped the pool, I copied all of the objects to local disk.
> After
> > re-creating the pool, I uploaded the objects.
> >
> > Now replication is kinda of working, but not correctly.  I have a number
> of
> > buckets that are being written to in the primary cluster, but no
> replication
> > is occurring.  radosgw-agent says a number of shards have >= 1000 log
> > entries, but then it never processes the buckets in those shards.
> >
> > Looking back at the pool's contents on local disk, all of the files are 0
> > bytes.  So I'm assuming all of the important state was stored in the
> > object's metadata.
> >
> > I'd like to completely zero out the replication state, then exploit a
> > feature in radosgw-agent 1.1 that will only replicate the first 1000
> objects
> > in buckets, if the bucket isn't being actively written to.  Then I can
> > restart radosgw-agent 1.2, and let it catch up the active buckets.
> That'll
> > save me many weeks and TB of replication.
> >
> > Obviously, I'll compare bucket listings between the two clusters when I'm
> > done.  I'll probably try to catch up the read-only bucket's state at a
> later
> > date.
> >
>
> I don't really understand what happened here. Maybe start with trying
> to understand why the sync agent isn't replicating anymore? Maybe the
> replicalog markers are off?
>
>
> Yehuda
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140924/8cc54352/attachment-0001.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux