Re: radosgw sync falling behind regularly

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hey Christian,

I'm making a while guess, but assuming this is 12.2.8. If so, it it possible that you can upgrade to 12.2.11? There's been rgw multisite bug fixes for metadata syncing and data syncing ( both separate issues ) that you could be hitting.

Thanks,

From: ceph-users <ceph-users-bounces@xxxxxxxxxxxxxx> on behalf of Christian Rice <crice@xxxxxxxxxxx>
Sent: Wednesday, February 27, 2019 7:05 PM
To: ceph-users
Subject: radosgw sync falling behind regularly
 

Debian 9; ceph 12.8.8-bpo90+1; no rbd or cephfs, just radosgw; three clusters in one zonegroup.

 

Often we find either metadata or data sync behind, and it doesn’t look to ever recover until…we restart the endpoint radosgw target service.

 

eg at 15:45:40:

 

dc11-ceph-rgw1:/var/log/ceph# radosgw-admin sync status

          realm b3e2afe7-2254-494a-9a34-ce50358779fd (savagebucket)

      zonegroup de6af748-1a2f-44a1-9d44-30799cf1313e (us)

           zone 107d29a0-b732-4bf1-a26e-1f64f820e839 (dc11-prod)

  metadata sync syncing

                full sync: 0/64 shards

                incremental sync: 64/64 shards

                metadata is behind on 2 shards

                behind shards: [19,41]

                oldest incremental change not applied: 2019-02-27 14:42:24.0.408263s

      data sync source: 1e27bf9c-3a2f-4845-85b6-33a24bbe1c04 (sv5-corp)

                        syncing

                        full sync: 0/128 shards

                        incremental sync: 128/128 shards

                        data is caught up with source

                source: 331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8 (sv3-prod)

                        syncing

                        full sync: 0/128 shards

                        incremental sync: 128/128 shards

                        data is caught up with source

 

 

so at 15:46:07:

 

dc11-ceph-rgw1:/var/log/ceph# sudo systemctl restart ceph-radosgw@rgw.dc11-ceph-rgw1.service

 

and by the time I checked at 15:48:08:

 

dc11-ceph-rgw1:/var/log/ceph# radosgw-admin sync status

          realm b3e2afe7-2254-494a-9a34-ce50358779fd (savagebucket)

      zonegroup de6af748-1a2f-44a1-9d44-30799cf1313e (us)

           zone 107d29a0-b732-4bf1-a26e-1f64f820e839 (dc11-prod)

  metadata sync syncing

                full sync: 0/64 shards

                incremental sync: 64/64 shards

                metadata is caught up with master

      data sync source: 1e27bf9c-3a2f-4845-85b6-33a24bbe1c04 (sv5-corp)

                        syncing

                        full sync: 0/128 shards

                        incremental sync: 128/128 shards

                        data is caught up with source

                source: 331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8 (sv3-prod)

                        syncing

                        full sync: 0/128 shards

                        incremental sync: 128/128 shards

                        data is caught up with source

 

 

There’s no way this is “lag.”  It’s stuck, and happens frequently, though perhaps not daily.  Any suggestions?  Our cluster isn’t heavily used yet, but it’s production.

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux