Re: radosgw sync falling behind regularly

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



(cc ceph-users)

Can you tell whether these sync errors are coming from metadata sync or data sync? Are they blocking sync from making progress according to your 'sync status'?

On 3/8/19 10:23 AM, Trey Palmer wrote:
Casey,

Having done the 'reshard stale-instances delete' earlier on the advice of another list member, we have tons of sync errors on deleted buckets, as you mention.

After 'data sync init' we're still seeing all of these errors on deleted buckets.

Since buckets are metadata, it occurred to me this morning that buckets are metadata so a 'sync init' wouldn't refresh that info.   But a 'metadata sync init' might get rid of the stale bucket sync info and stop the sync errors.   Would that be the way to go?

Thanks,

Trey



On Wed, Mar 6, 2019 at 11:47 AM Casey Bodley <cbodley@xxxxxxxxxx <mailto:cbodley@xxxxxxxxxx>> wrote:

    Hi Trey,

    I think it's more likely that these stale metadata entries are from
    deleted buckets, rather than accidental bucket reshards. When a
    bucket
    is deleted in a multisite configuration, we don't delete its bucket
    instance because other zones may still need to sync the object
    deletes -
    and they can't make progress on sync if the bucket metadata
    disappears.
    These leftover bucket instances look the same to the 'reshard
    stale-instances' commands, but I'd be cautious about using that to
    remove them in multisite, as it may cause more sync errors and
    potentially leak storage if they still contain objects.

    Regarding 'datalog trim', that alone isn't safe because it could trim
    entries that hadn't been applied on other zones yet, causing them to
    miss some updates. What you can do is run 'data sync init' on each
    zone,
    and restart gateways. This will restart with a data full sync (which
    will scan all buckets for changes), and skip past any datalog entries
    from before the full sync. I was concerned that the bug in error
    handling (ie "ERROR: init sync on...") would also affect full
    sync, but
    that doesn't appear to be the case - so I do think that's worth
    trying.

    On 3/5/19 6:24 PM, Trey Palmer wrote:
    > Casey,
    >
    > Thanks very much for the reply!
    >
    > We definitely have lots of errors on sync-disabled buckets and the
    > workaround for that is obvious (most of them are empty anyway).
    >
    > Our second form of error is stale buckets.  We had dynamic
    resharding
    > enabled but have now disabled it (having discovered it was on by
    > default, and not supported in multisite).
    >
    > We removed several hundred stale buckets via 'radosgw-admin
    sharding
    > stale-instances rm', but they are still giving us sync errors.
    >
    > I have found that these buckets do have entries in 'radosgw-admin
    > datalog list', and my guess is this could be fixed by doing a
    > 'radosgw-admin datalog trim' for each entry on the master zone.
    >
    > Does that sound right?  :-)
    >
    > Thanks again for the detailed explanation,
    >
    > Trey Palmer
    >
    > On Tue, Mar 5, 2019 at 5:55 PM Casey Bodley <cbodley@xxxxxxxxxx
    <mailto:cbodley@xxxxxxxxxx>
    > <mailto:cbodley@xxxxxxxxxx <mailto:cbodley@xxxxxxxxxx>>> wrote:
    >
    >     Hi Christian,
    >
    >     I think you've correctly intuited that the issues are related to
    >     the use
    >     of 'bucket sync disable'. There was a bug fix for that
    feature in
    > http://tracker.ceph.com/issues/26895, and I recently found that a
    >     block
    >     of code was missing from its luminous backport. That missing
    code is
    >     what handled those "ERROR: init sync on <bucket instance>
    failed,
    >     retcode=-2" errors.
    >
    >     I included a fix for that in a later backport
    >     (https://github.com/ceph/ceph/pull/26549), which I'm still
    working to
    >     get through qa. I'm afraid I can't really recommend a workaround
    >     for the
    >     issue in the meantime.
    >
    >     Looking forward though, we do plan to support something like
    s3's
    >     cross
    >     region replication so you can enable replication on a
    specific bucket
    >     without having to enable it globally.
    >
    >     Casey
    >
    >
    >     On 3/5/19 2:32 PM, Christian Rice wrote:
    >     >
    >     > Much appreciated.  We’ll continue to poke around and
    certainly will
    >     > disable the dynamic resharding.
    >     >
    >     > We started with 12.2.8 in production.  We definitely did not
    >     have it
    >     > enabled in ceph.conf
    >     >
    >     > *From: *Matthew H <matthew.heler@xxxxxxxxxxx
    <mailto:matthew.heler@xxxxxxxxxxx>
    >     <mailto:matthew.heler@xxxxxxxxxxx
    <mailto:matthew.heler@xxxxxxxxxxx>>>
    >     > *Date: *Tuesday, March 5, 2019 at 11:22 AM
    >     > *To: *Christian Rice <crice@xxxxxxxxxxx
    <mailto:crice@xxxxxxxxxxx>
    >     <mailto:crice@xxxxxxxxxxx <mailto:crice@xxxxxxxxxxx>>>,
    ceph-users
    >     > <ceph-users@xxxxxxxxxxxxxx
    <mailto:ceph-users@xxxxxxxxxxxxxx>
    <mailto:ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>>>
    >     > *Cc: *Trey Palmer <nerdmagicatl@xxxxxxxxx
    <mailto:nerdmagicatl@xxxxxxxxx>
    >     <mailto:nerdmagicatl@xxxxxxxxx <mailto:nerdmagicatl@xxxxxxxxx>>>
    >     > *Subject: *Re: radosgw sync falling behind regularly
    >     >
    >     > Hi Christian,
    >     >
    >     > To be on the safe side and future proof yourself will want
    to go
    >     ahead
    >     > and set the following in your ceph.conf file, and then issue a
    >     restart
    >     > to your RGW instances.
    >     >
    >     > rgw_dynamic_resharding = false
    >     >
    >     > There are a number of issues with dynamic resharding,
    multisite rgw
    >     > problems being just one of them. However I thought it was
    disabled
    >     > automatically when multisite rgw is used (but I will have
    to double
    >     > check the code on that). What version of Ceph did you
    initially
    >     > install the cluster with? Prior to v12.2.2 this feature was
    >     enabled by
    >     > default for all rgw use cases.
    >     >
    >     > Thanks,
    >     >
    >     >
    >
     ------------------------------------------------------------------------
    >     >
    >     > *From:*Christian Rice <crice@xxxxxxxxxxx
    <mailto:crice@xxxxxxxxxxx> <mailto:crice@xxxxxxxxxxx
    <mailto:crice@xxxxxxxxxxx>>>
    >     > *Sent:* Tuesday, March 5, 2019 2:07 PM
    >     > *To:* Matthew H; ceph-users
    >     > *Subject:* Re: radosgw sync falling behind regularly
    >     >
    >     > Matthew, first of all, let me say we very much appreciate
    your help!
    >     >
    >     > So I don’t think we turned dynamic resharding on, nor did we
    >     manually
    >     > reshard buckets. Seems like it defaults to on for luminous
    but the
    >     > mimic docs say it’s not supported in multisite. So do we
    need to
    >     > disable it manually via tell and ceph.conf?
    >     >
    >     > Also, after running the command you suggested, all the stale
    >     instances
    >     > are gone…these from my examples were in output:
    >     >
    >     >         "bucket_instance":
    >     >
    >
     "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.303",
    >     >
    >     >         "bucket_instance":
    >     >
    >
     "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299",
    >     >
    >     >         "bucket_instance":
    >     >
    >
     "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.301",
    >     >
    >     > Though we still get lots of log messages like so in rgw:
    >     >
    >     > 2019-03-05 11:01:09.526120 7f64120ae700  0 ERROR: failed
    to get
    >     bucket
    >     > instance info for
    >     >
    >
     .bucket.meta.sysad_task:sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
    >     >
    >     > 2019-03-05 11:01:09.528664 7f63e5016700  1 civetweb:
    >     0x55976f1c2000:
    >     > 172.17.136.17 - - [05/Mar/2019:10:54:06 -0800] "GET
    >     >
    >
     /admin/metadata/bucket.instance/sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299?key=sysad_task%2Fsysad-task%3A1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299&rgwx-zonegroup=de6af748-1a2f-44a1-9d44-30799cf1313e
    >
    >     > HTTP/1.1" 404 0 - -
    >     >
    >     > 2019-03-05 11:01:09.529648 7f64130b0700  0 meta sync:
    ERROR: can't
    >     > remove key:
    >     >
    >
     bucket.instance:sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
    >
    >     > ret=-2
    >     >
    >     > 2019-03-05 11:01:09.530324 7f64138b1700  0 ERROR: failed
    to get
    >     bucket
    >     > instance info for
    >     >
    >
     .bucket.meta.sysad_task:sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
    >     >
    >     > 2019-03-05 11:01:09.530345 7f6405094700  0 data sync: ERROR:
    >     failed to
    >     > retrieve bucket info for
    >     >
    >
     bucket=sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
    >     >
    >     > 2019-03-05 11:01:09.531774 7f6405094700  0 data sync: WARNING:
    >     > skipping data log entry for missing bucket
    >     >
    sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
    >     >
    >     > 2019-03-05 11:01:09.571680 7f6405094700  0 data sync: ERROR:
    >     init sync
    >     > on
    >     >
    >
     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
    >     > failed, retcode=-2
    >     >
    >     > 2019-03-05 11:01:09.573179 7f6405094700  0 data sync: WARNING:
    >     > skipping data log entry for missing bucket
    >     >
    sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
    >     >
    >     > 2019-03-05 11:01:13.504308 7f63f903e700  1 civetweb:
    >     0x55976f0f2000:
    >     > 10.105.18.20 - - [05/Mar/2019:11:00:57 -0800] "GET
    >     >
    >
     /admin/metadata/bucket.instance/sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299?key=sysad_task%2Fsysad-task%3A1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299&rgwx-zonegroup=de6af748-1a2f-44a1-9d44-30799cf1313e
    >
    >     > HTTP/1.1" 404 0 - -
    >     >
    >     > *From: *Matthew H <matthew.heler@xxxxxxxxxxx
    <mailto:matthew.heler@xxxxxxxxxxx>
    >     <mailto:matthew.heler@xxxxxxxxxxx
    <mailto:matthew.heler@xxxxxxxxxxx>>>
    >     > *Date: *Tuesday, March 5, 2019 at 10:03 AM
    >     > *To: *Christian Rice <crice@xxxxxxxxxxx
    <mailto:crice@xxxxxxxxxxx>
    >     <mailto:crice@xxxxxxxxxxx <mailto:crice@xxxxxxxxxxx>>>,
    ceph-users
    >     > <ceph-users@xxxxxxxxxxxxxx
    <mailto:ceph-users@xxxxxxxxxxxxxx>
    <mailto:ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>>>
    >     > *Subject: *Re: radosgw sync falling behind regularly
    >     >
    >     > Hi Christian,
    >     >
    >     > You have stale bucket instances that need to be clean up,
    which is
    >     > what 'radosgw-admin reshard stale-instances list' is showing
    >     you. Have
    >     > you or were you manually resharding your buckets? The
    errors you
    >     are
    >     > seeing in the logs are related to these stale instances
    being kept
    >     > around.
    >     >
    >     > In v12.2.11 this command along with 'radosgw-admin reshard
    >     > stale-instance rm' was introduced [1].
    >     >
    >     > Hopefully this helps.
    >     >
    >     > [1]
    >     >
    >     > https://ceph.com/releases/v12-2-11-luminous-released/
    >     >
    >   
     <https://urldefense.proofpoint.com/v2/url?u=https-3A__ceph.com_releases_v12-2D2-2D11-2Dluminous-2Dreleased_&d=DwMF-g&c=gFTBenQ7Vj71sUi1A4CkFnmPzqwDo07QsHw-JRepxyw&r=NE1NbWtVhgG-K7YvLdoLZigfFx8zGPwOGk6HWpYK04I&m=vdtYIn6lEKaWD9wW297aHjQLpmQdHZrOVpOhmCBqkqo&s=nGCpS4p5jnaSpPUFlziSi3Y3pFijhVDy6e3867jA9BE&e=>
    >     >
    >     > /"There have been fixes to RGW dynamic and manual resharding,
    >     which no
    >     > longer
    >     > leaves behind stale bucket instances to be removed
    manually. For
    >     > finding and
    >     > cleaning up older instances from a reshard a radosgw-admin
    >     command reshard
    >     > stale-instances list and reshard stale-instances rm should
    do the
    >     > necessary
    >     > cleanup."/
    >     >
    >     >
    >
     ------------------------------------------------------------------------
    >     >
    >     > *From:*Christian Rice <crice@xxxxxxxxxxx
    <mailto:crice@xxxxxxxxxxx> <mailto:crice@xxxxxxxxxxx
    <mailto:crice@xxxxxxxxxxx>>>
    >     > *Sent:* Tuesday, March 5, 2019 11:34 AM
    >     > *To:* Matthew H; ceph-users
    >     > *Subject:* Re: radosgw sync falling behind regularly
    >     >
    >     > The output of “radosgw-admin reshard stale-instances list”
    shows
    >     242
    >     > entries, which might embed too much proprietary info for me to
    >     list,
    >     > but here’s a tiny sample:
    >     >
    >     >
    >
     "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.303",
    >     >
    >     >
    >
     "sysad_task/sysad_task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18467.281",
    >     >
    >     >
    >
     "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299",
    >     >
    >     >
    >
     "sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.301",
    >     >
    >     > Some of appear repeatedly in the radosgw error logs like so:
    >     >
    >     > 2019-03-05 08:13:08.929206 7f6405094700  0 data sync: ERROR:
    >     init sync
    >     > on
    >     >
    >
     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
    >     > failed, retcode=-2
    >     >
    >     > 2019-03-05 08:13:08.930581 7f6405094700  0 data sync: WARNING:
    >     > skipping data log entry for missing bucket
    >     >
    sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.302
    >     >
    >     > 2019-03-05 08:13:08.972053 7f6405094700  0 data sync: ERROR:
    >     init sync
    >     > on
    >     >
    >
     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
    >     > failed, retcode=-2
    >     >
    >     > 2019-03-05 08:13:08.973442 7f6405094700  0 data sync: WARNING:
    >     > skipping data log entry for missing bucket
    >     >
    sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
    >     >
    >     > 2019-03-05 08:13:19.528295 7f6406897700  0 data sync: ERROR:
    >     init sync
    >     > on
    >     >
    >
     sysad_task/sysad-task:1e27bf9c-3a2f-4845-85b6-33a24bbe1c04.18330.299
    >     > failed, retcode=-2
    >     >
    >     > Notably, “Sync is disabled for bucket sysad-task.”  We use
    “bucket
    >     > sync disable” A LOT.  It’s the only way we’ve been able to use
    >     > multisite and a single namespace and not replicate things
    that are
    >     > unneeded to every zone.  Perhaps there’s a bug in the
    >     implementation
    >     > that’s tripping us up now, with the new sync multisite
    sync code
    >     from
    >     > 12.2.9 onward?
    >     >
    >     > What might we do with stale bucket instances, then?
    >     >
    >     > Of note, our master zone endpoint, which was timing out health
    >     checks
    >     > most of the day after the upgrade (was running but
    overworked by
    >     > cluster confusion, so we couldn’t create new buckets or do
    user
    >     ops),
    >     > has returned to availability late last night. There’s a lot of
    >     data
    >     > to look at, but in my estimation, due to lack of user
    complaints
    >     (or
    >     > their unawareness of specific issues), it seems the zones are
    >     > nominally available, even with all the errors and warnings
    being
    >     > logged.  We’ve tested simple zone replication by creating
    a few
    >     files
    >     > in one zone and seeing them show up elsewhere…
    >     >
    >     > here’s “period get” output:
    >     >
    >     > sv5-ceph-rgw1
    >     >
    >     > {
    >     >
    >     >     "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
    >     >
    >     > "epoch": 16,
    >     >
    >     > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb",
    >     >
    >     > "sync_status": [],
    >     >
    >     > "period_map": {
    >     >
    >     > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
    >     >
    >     > "zonegroups": [
    >     >
    >     > {
    >     >
    >     > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e",
    >     >
    >     > "name": "us",
    >     >
    >     > "api_name": "us",
    >     >
    >     > "is_master": "true",
    >     >
    >     > "endpoints": [
    >     >
    >     >                    "http://sv5-ceph-rgw1.savagebeast.com:8080";
    >     >
    >     > ],
    >     >
    >     > "hostnames": [],
    >     >
    >     > "hostnames_s3website": [],
    >     >
    >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "zones": [
    >     >
    >     > {
    >     >
    >     > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839",
    >     >
    >     > "name": "dc11-prod",
    >     >
    >     > "endpoints": [
    >     >
    >     > "http://dc11-ceph-rgw1:8080";
    >     >
    >     >  ],
    >     >
    >     > "log_meta": "false",
    >     >
    >     > "log_data": "true",
    >     >
    >     > "bucket_index_max_shards": 0,
    >     >
    >     > "read_only": "false",
    >     >
    >     > "tier_type": "",
    >     >
    >     >   "sync_from_all": "true",
    >     >
    >     > "sync_from": []
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "name": "sv5-corp",
    >     >
    >     > "endpoints": [
    >     >
    >     > "http://sv5-ceph-rgw1.savagebeast.com:8080";
    >     >
    >     > ],
    >     >
    >     > "log_meta": "false",
    >     >
    >     > "log_data": "true",
    >     >
    >     > "bucket_index_max_shards": 0,
    >     >
    >     >                    "read_only": "false",
    >     >
    >     > "tier_type": "",
    >     >
    >     > "sync_from_all": "true",
    >     >
    >     > "sync_from": []
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
    >     >
    >     > "name": "sv3-prod",
    >     >
    >     > "endpoints": [
    >     >
    >     > "http://sv3-ceph-rgw1:8080";
    >     >
    >     > ],
    >     >
    >     >     "log_meta": "false",
    >     >
    >     > "log_data": "true",
    >     >
    >     > "bucket_index_max_shards": 0,
    >     >
    >     > "read_only": "false",
    >     >
    >     > "tier_type": "",
    >     >
    >     > "sync_from_all": "true",
    >     >
    >     > "sync_from": []
    >     >
    >     > }
    >     >
    >     > ],
    >     >
    >     > "placement_targets": [
    >     >
    >     > {
    >     >
    >     > "name": "default-placement",
    >     >
    >     > "tags": []
    >     >
    >     > }
    >     >
    >     >                ],
    >     >
    >     > "default_placement": "default-placement",
    >     >
    >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd"
    >     >
    >     > }
    >     >
    >     >         ],
    >     >
    >     > "short_zone_ids": [
    >     >
    >     > {
    >     >
    >     > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839",
    >     >
    >     > "val": 1720993486
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "val": 2301637458
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
    >     >
    >     > "val": 1449486239
    >     >
    >     > }
    >     >
    >     >         ]
    >     >
    >     >     },
    >     >
    >     > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e",
    >     >
    >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "period_config": {
    >     >
    >     > "bucket_quota": {
    >     >
    >     > "enabled": false,
    >     >
    >     > "check_on_raw": false,
    >     >
    >     > "max_size": -1,
    >     >
    >     > "max_size_kb": 0,
    >     >
    >     > "max_objects": -1
    >     >
    >     >         },
    >     >
    >     > "user_quota": {
    >     >
    >     > "enabled": false,
    >     >
    >     > "check_on_raw": false,
    >     >
    >     > "max_size": -1,
    >     >
    >     > "max_size_kb": 0,
    >     >
    >     > "max_objects": -1
    >     >
    >     >         }
    >     >
    >     >     },
    >     >
    >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd",
    >     >
    >     > "realm_name": "savagebucket",
    >     >
    >     > "realm_epoch": 2
    >     >
    >     > }
    >     >
    >     > sv3-ceph-rgw1
    >     >
    >     > {
    >     >
    >     >     "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
    >     >
    >     > "epoch": 16,
    >     >
    >     > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb",
    >     >
    >     > "sync_status": [],
    >     >
    >     > "period_map": {
    >     >
    >     > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
    >     >
    >     > "zonegroups": [
    >     >
    >     > {
    >     >
    >     > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e",
    >     >
    >     > "name": "us",
    >     >
    >     > "api_name": "us",
    >     >
    >     > "is_master": "true",
    >     >
    >     > "endpoints": [
    >     >
    >     >         "http://sv5-ceph-rgw1.savagebeast.com:8080";
    >     >
    >     > ],
    >     >
    >     > "hostnames": [],
    >     >
    >     > "hostnames_s3website": [],
    >     >
    >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "zones": [
    >     >
    >     >          {
    >     >
    >     > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839",
    >     >
    >     > "name": "dc11-prod",
    >     >
    >     > "endpoints": [
    >     >
    >     > "http://dc11-ceph-rgw1:8080";
    >     >
    >     > ],
    >     >
    >     > "log_meta": "false",
    >     >
    >     > "log_data": "true",
    >     >
    >     > "bucket_index_max_shards": 0,
    >     >
    >     > "read_only": "false",
    >     >
    >     > "tier_type": "",
    >     >
    >     > "sync_from_all": "true",
    >     >
    >     > "sync_from": []
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "name": "sv5-corp",
    >     >
    >     > "endpoints": [
    >     >
    >     > "http://sv5-ceph-rgw1.savagebeast.com:8080";
    >     >
    >     > ],
    >     >
    >     > "log_meta": "false",
    >     >
    >     > "log_data": "true",
    >     >
    >     > "bucket_index_max_shards": 0,
    >     >
    >     >                "read_only": "false",
    >     >
    >     > "tier_type": "",
    >     >
    >     > "sync_from_all": "true",
    >     >
    >     > "sync_from": []
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
    >     >
    >     > "name": "sv3-prod",
    >     >
    >     > "endpoints": [
    >     >
    >     > "http://sv3-ceph-rgw1:8080";
    >     >
    >     > ],
    >     >
    >     > "log_meta": "false",
    >     >
    >     > "log_data": "true",
    >     >
    >     > "bucket_index_max_shards": 0,
    >     >
    >     > "read_only": "false",
    >     >
    >     > "tier_type": "",
    >     >
    >     > "sync_from_all": "true",
    >     >
    >     >   "sync_from": []
    >     >
    >     > }
    >     >
    >     > ],
    >     >
    >     > "placement_targets": [
    >     >
    >     > {
    >     >
    >     > "name": "default-placement",
    >     >
    >     > "tags": []
    >     >
    >     > }
    >     >
    >     > ],
    >     >
    >     >          "default_placement": "default-placement",
    >     >
    >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd"
    >     >
    >     > }
    >     >
    >     >         ],
    >     >
    >     > "short_zone_ids": [
    >     >
    >     > {
    >     >
    >     > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839",
    >     >
    >     >         "val": 1720993486
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "val": 2301637458
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
    >     >
    >     > "val": 1449486239
    >     >
    >     > }
    >     >
    >     >         ]
    >     >
    >     >     },
    >     >
    >     > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e",
    >     >
    >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "period_config": {
    >     >
    >     > "bucket_quota": {
    >     >
    >     > "enabled": false,
    >     >
    >     > "check_on_raw": false,
    >     >
    >     > "max_size": -1,
    >     >
    >     > "max_size_kb": 0,
    >     >
    >     > "max_objects": -1
    >     >
    >     >         },
    >     >
    >     > "user_quota": {
    >     >
    >     > "enabled": false,
    >     >
    >     > "check_on_raw": false,
    >     >
    >     > "max_size": -1,
    >     >
    >     > "max_size_kb": 0,
    >     >
    >     > "max_objects": -1
    >     >
    >     >         }
    >     >
    >     >     },
    >     >
    >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd",
    >     >
    >     > "realm_name": "savagebucket",
    >     >
    >     > "realm_epoch": 2
    >     >
    >     > }
    >     >
    >     > dc11-ceph-rgw1
    >     >
    >     > {
    >     >
    >     >     "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
    >     >
    >     > "epoch": 16,
    >     >
    >     > "predecessor_uuid": "926c74c7-c1a7-46b1-9f25-eb5c392a7fbb",
    >     >
    >     > "sync_status": [],
    >     >
    >     > "period_map": {
    >     >
    >     > "id": "3d0d40ef-90de-40ea-8c44-caa20ea8dc53",
    >     >
    >     > "zonegroups": [
    >     >
    >     >         {
    >     >
    >     > "id": "de6af748-1a2f-44a1-9d44-30799cf1313e",
    >     >
    >     > "name": "us",
    >     >
    >     > "api_name": "us",
    >     >
    >     > "is_master": "true",
    >     >
    >     > "endpoints": [
    >     >
    >     > "http://sv5-ceph-rgw1.savagebeast.com:8080";
    >     >
    >     > ],
    >     >
    >     > "hostnames": [],
    >     >
    >     > "hostnames_s3website": [],
    >     >
    >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "zones": [
    >     >
    >     > {
    >     >
    >     > "id": "107d29a0-b732-4bf1-a26e-1f64f820e839",
    >     >
    >     > "name": "dc11-prod",
    >     >
    >     > "endpoints": [
    >     >
    >     > "http://dc11-ceph-rgw1:8080";
    >     >
    >     > ],
    >     >
    >     > "log_meta": "false",
    >     >
    >     >                       "log_data": "true",
    >     >
    >     > "bucket_index_max_shards": 0,
    >     >
    >     > "read_only": "false",
    >     >
    >     > "tier_type": "",
    >     >
    >     > "sync_from_all": "true",
    >     >
    >     > "sync_from": []
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "id": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "name": "sv5-corp",
    >     >
    >     > "endpoints": [
    >     >
    >     >           "http://sv5-ceph-rgw1.savagebeast.com:8080";
    >     >
    >     > ],
    >     >
    >     > "log_meta": "false",
    >     >
    >     > "log_data": "true",
    >     >
    >     > "bucket_index_max_shards": 0,
    >     >
    >     > "read_only": "false",
    >     >
    >     > "tier_type": "",
    >     >
    >     > "sync_from_all": "true",
    >     >
    >     > "sync_from": []
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "id": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
    >     >
    >     > "name": "sv3-prod",
    >     >
    >     > "endpoints": [
    >     >
    >     > "http://sv3-ceph-rgw1:8080";
    >     >
    >     > ],
    >     >
    >     > "log_meta": "false",
    >     >
    >     > "log_data": "true",
    >     >
    >     > "bucket_index_max_shards": 0,
    >     >
    >     > "read_only": "false",
    >     >
    >     > "tier_type": "",
    >     >
    >     > "sync_from_all": "true",
    >     >
    >     > "sync_from": []
    >     >
    >     > }
    >     >
    >     > ],
    >     >
    >     > "placement_targets": [
    >     >
    >     > {
    >     >
    >     > "name": "default-placement",
    >     >
    >     > "tags": []
    >     >
    >     > }
    >     >
    >     > ],
    >     >
    >     > "default_placement": "default-placement",
    >     >
    >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd"
    >     >
    >     > }
    >     >
    >     >         ],
    >     >
    >     > "short_zone_ids": [
    >     >
    >     > {
    >     >
    >     > "key": "107d29a0-b732-4bf1-a26e-1f64f820e839",
    >     >
    >     > "val": 1720993486
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "key": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "val": 2301637458
    >     >
    >     > },
    >     >
    >     > {
    >     >
    >     > "key": "331d3f1e-1b72-4c56-bb5a-d1d0fcf6d0b8",
    >     >
    >     > "val": 1449486239
    >     >
    >     > }
    >     >
    >     >         ]
    >     >
    >     >     },
    >     >
    >     > "master_zonegroup": "de6af748-1a2f-44a1-9d44-30799cf1313e",
    >     >
    >     > "master_zone": "1e27bf9c-3a2f-4845-85b6-33a24bbe1c04",
    >     >
    >     > "period_config": {
    >     >
    >     > "bucket_quota": {
    >     >
    >     > "enabled": false,
    >     >
    >     > "check_on_raw": false,
    >     >
    >     >           "max_size": -1,
    >     >
    >     > "max_size_kb": 0,
    >     >
    >     > "max_objects": -1
    >     >
    >     >         },
    >     >
    >     > "user_quota": {
    >     >
    >     > "enabled": false,
    >     >
    >     > "check_on_raw": false,
    >     >
    >     > "max_size": -1,
    >     >
    >     > "max_size_kb": 0,
    >     >
    >     > "max_objects": -1
    >     >
    >     >         }
    >     >
    >     >     },
    >     >
    >     > "realm_id": "b3e2afe7-2254-494a-9a34-ce50358779fd",
    >     >
    >     > "realm_name": "savagebucket",
    >     >
    >     > "realm_epoch": 2
    >     >
    >     > }
    >     >
    >     > *From: *Matthew H <matthew.heler@xxxxxxxxxxx
    <mailto:matthew.heler@xxxxxxxxxxx>
    >     <mailto:matthew.heler@xxxxxxxxxxx
    <mailto:matthew.heler@xxxxxxxxxxx>>>
    >     > *Date: *Tuesday, March 5, 2019 at 4:31 AM
    >     > *To: *Christian Rice <crice@xxxxxxxxxxx
    <mailto:crice@xxxxxxxxxxx>
    >     <mailto:crice@xxxxxxxxxxx <mailto:crice@xxxxxxxxxxx>>>,
    ceph-users
    >     > <ceph-users@xxxxxxxxxxxxxx
    <mailto:ceph-users@xxxxxxxxxxxxxx>
    <mailto:ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>>>
    >     > *Subject: *Re: radosgw sync falling behind regularly
    >     >
    >     > Hi Christian,
    >     >
    >     > You haven't resharded any of your buckets have you?  You
    can run
    >     the
    >     > command below in v12.2.11 to list stale bucket instances.
    >     >
    >     > radosgw-admin reshard stale-instances list
    >     >
    >     > Can you also send the output from the following command on
    each rgw?
    >     >
    >     > radosgw-admin period get
    >     >
    >     >
    >     > _______________________________________________
    >     > ceph-users mailing list
    >     > ceph-users@xxxxxxxxxxxxxx
    <mailto:ceph-users@xxxxxxxxxxxxxx>
    <mailto:ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>>
    >     > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
    >     _______________________________________________
    >     ceph-users mailing list
    > ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
    <mailto:ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>>
    > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
    >

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux