Re: Radosgw multisite replication issues

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

what ceph versions is this? Have you also verified that any config and keyring files in /etc/ceph (in case your cluster is not cephadm yet) are in the desired state? The permission denied error suggests that one site is not allowed to sync.

Zitat von "Tarrago, Eli (RIS-BCT)" <Eli.Tarrago@xxxxxxxxxxxxxxxxxx>:

Good Afternoon,

I am experiencing an issue where east-1 is no longer able to replicate from west-1, however, after a realm pull, west-1 is now able to replicate from east-1.

In other words:
West <- Can Replicate <- East
West -> Cannot Replicate -> East

After confirming the access and secret keys are identical on both sides, I restarted all radosgw services.

Here is the current status of the cluster below.

Thank you for your help,

Eli Tarrago


root@east01:~# radosgw-admin zone get
{
    "id": "ddd66ab8-0417-46ee-a53b-043352a63f93",
    "name": "rgw-east",
    "domain_root": "rgw-east.rgw.meta:root",
    "control_pool": "rgw-east.rgw.control",
    "gc_pool": "rgw-east.rgw.log:gc",
    "lc_pool": "rgw-east.rgw.log:lc",
    "log_pool": "rgw-east.rgw.log",
    "intent_log_pool": "rgw-east.rgw.log:intent",
    "usage_log_pool": "rgw-east.rgw.log:usage",
    "roles_pool": "rgw-east.rgw.meta:roles",
    "reshard_pool": "rgw-east.rgw.log:reshard",
    "user_keys_pool": "rgw-east.rgw.meta:users.keys",
    "user_email_pool": "rgw-east.rgw.meta:users.email",
    "user_swift_pool": "rgw-east.rgw.meta:users.swift",
    "user_uid_pool": "rgw-east.rgw.meta:users.uid",
    "otp_pool": "rgw-east.rgw.otp",
    "system_key": {
        "access_key": "PxxxxxxxxxxxxxxxxW",
        "secret_key": "Hxxxxxxxxxxxxxxxx6"
    },
    "placement_pools": [
        {
            "key": "default-placement",
            "val": {
                "index_pool": "rgw-east.rgw.buckets.index",
                "storage_classes": {
                    "STANDARD": {
                        "data_pool": "rgw-east.rgw.buckets.data"
                    }
                },
                "data_extra_pool": "rgw-east.rgw.buckets.non-ec",
                "index_type": 0
            }
        }
    ],
    "realm_id": "98e0e391-16fb-48da-80a5-08437fd81789",
    "notif_pool": "rgw-east.rgw.log:notif"
}

root@west01:~# radosgw-admin zone get
{
   "id": "b2a4a31c-1505-4fdc-b2e0-ea07d9463da1",
    "name": "rgw-west",
    "domain_root": "rgw-west.rgw.meta:root",
    "control_pool": "rgw-west.rgw.control",
    "gc_pool": "rgw-west.rgw.log:gc",
    "lc_pool": "rgw-west.rgw.log:lc",
    "log_pool": "rgw-west.rgw.log",
    "intent_log_pool": "rgw-west.rgw.log:intent",
    "usage_log_pool": "rgw-west.rgw.log:usage",
    "roles_pool": "rgw-west.rgw.meta:roles",
    "reshard_pool": "rgw-west.rgw.log:reshard",
    "user_keys_pool": "rgw-west.rgw.meta:users.keys",
    "user_email_pool": "rgw-west.rgw.meta:users.email",
    "user_swift_pool": "rgw-west.rgw.meta:users.swift",
    "user_uid_pool": "rgw-west.rgw.meta:users.uid",
    "otp_pool": "rgw-west.rgw.otp",
    "system_key": {
        "access_key": "PxxxxxxxxxxxxxxW",
        "secret_key": "Hxxxxxxxxxxxxxx6"
    },
    "placement_pools": [
        {
            "key": "default-placement",
            "val": {
                "index_pool": "rgw-west.rgw.buckets.index",
                "storage_classes": {
                    "STANDARD": {
                        "data_pool": "rgw-west.rgw.buckets.data"
                    }
                },
                "data_extra_pool": "rgw-west.rgw.buckets.non-ec",
                "index_type": 0
            }
        }
    ],
    "realm_id": "98e0e391-16fb-48da-80a5-08437fd81789",
    "notif_pool": "rgw-west.rgw.log:notif"
east01:~# radosgw-admin metadata sync status
{
    "sync_status": {
        "info": {
            "status": "init",
            "num_shards": 0,
            "period": "",
            "realm_epoch": 0
        },
        "markers": []
    },
    "full_sync": {
        "total": 0,
        "complete": 0
    }
}

west01:~#  radosgw-admin metadata sync status
{
    "sync_status": {
        "info": {
            "status": "sync",
            "num_shards": 64,
            "period": "44b6b308-e2d8-4835-8518-c90447e7b55c",
            "realm_epoch": 3
        },
        "markers": [
            {
                "key": 0,
                "val": {
                    "state": 1,
                    "marker": "",
                    "next_step_marker": "",
                    "total_entries": 46,
                    "pos": 0,
                    "timestamp": "0.000000",
                    "realm_epoch": 3
                }
            },
#### goes on for a long time…
            {
                "key": 63,
                "val": {
                    "state": 1,
                    "marker": "",
                    "next_step_marker": "",
                    "total_entries": 0,
                    "pos": 0,
                    "timestamp": "0.000000",
                    "realm_epoch": 3
                }
            }
        ]
    },
    "full_sync": {
        "total": 46,
        "complete": 46
    }
}

east01:~#  radosgw-admin sync status
          realm 98e0e391-16fb-48da-80a5-08437fd81789 (rgw-blobs)
      zonegroup 0e0faf4e-39f5-402e-9dbb-4a1cdc249ddd (EastWestceph)
           zone ddd66ab8-0417-46ee-a53b-043352a63f93 (rgw-east)
  metadata sync no sync (zone is master)
2023-04-20T19:03:13.388+0000 7f25fa036c80 0 ERROR: failed to fetch datalog info
      data sync source: b2a4a31c-1505-4fdc-b2e0-ea07d9463da1 (rgw-west)
                        failed to retrieve sync info: (13) Permission denied

west01:~# radosgw-admin sync status
          realm 98e0e391-16fb-48da-80a5-08437fd81789 (rgw-blobs)
      zonegroup 0e0faf4e-39f5-402e-9dbb-4a1cdc249ddd (EastWestceph)
           zone b2a4a31c-1505-4fdc-b2e0-ea07d9463da1 (rgw-west)
  metadata sync syncing
                full sync: 0/64 shards
                incremental sync: 64/64 shards
                metadata is caught up with master
      data sync source: ddd66ab8-0417-46ee-a53b-043352a63f93 (rgw-east)
                        syncing
                        full sync: 0/128 shards
                        incremental sync: 128/128 shards
                        data is behind on 16 shards
behind shards: [5,56,62,65,66,70,76,86,87,94,104,107,111,113,120,126] oldest incremental change not applied: 2023-04-20T19:02:48.783283+0000 [5]

east01:~# radosgw-admin zonegroup get
{
    "id": "0e0faf4e-39f5-402e-9dbb-4a1cdc249ddd",
    "name": "EastWestceph",
    "api_name": "EastWestceph",
    "is_master": "true",
    "endpoints": [
        http://east01.example.net:8080,
        http://east02.example.net:8080,
        http://east03.example.net:8080,
        http://west01.example.net:8080,
        http://west02.example.net:8080,
        http://west03.example.net:8080
    ],
    "hostnames": [
        "eastvip.example.net",
        "westvip.example.net"
    ],
    "hostnames_s3website": [],
    "master_zone": "ddd66ab8-0417-46ee-a53b-043352a63f93",
    "zones": [
        {
            "id": "b2a4a31c-1505-4fdc-b2e0-ea07d9463da1",
            "name": "rgw-west",
            "endpoints": [
                http://west01.example.net:8080,
                http://west02.example.net:8080,
                http://west03.example.net:8080
            ],
            "log_meta": "false",
            "log_data": "true",
            "bucket_index_max_shards": 0,
            "read_only": "false",
            "tier_type": "",
            "sync_from_all": "true",
            "sync_from": [],
            "redirect_zone": ""
        },
        {
            "id": "ddd66ab8-0417-46ee-a53b-043352a63f93",
            "name": "rgw-east",
            "endpoints": [
                http://east01.example.net:8080,
                http://east02.example.net:8080,
                http://east03.example.net:8080
            ],
            "log_meta": "false",
            "log_data": "true",
            "bucket_index_max_shards": 0,
            "read_only": "false",
            "tier_type": "",
            "sync_from_all": "true",
            "sync_from": [],
            "redirect_zone": ""
        }
    ],
    "placement_targets": [
        {
            "name": "default-placement",
            "tags": [],
            "storage_classes": [
                "STANDARD"
            ]
        }
    ],
    "default_placement": "default-placement",
    "realm_id": "98e0e391-16fb-48da-80a5-08437fd81789",
    "sync_policy": {
        "groups": []
    }
}


________________________________
The information contained in this e-mail message is intended only for the personal and confidential use of the recipient(s) named above. This message may be an attorney-client communication and/or work product and as such is privileged and confidential. If the reader of this message is not the intended recipient or an agent responsible for delivering it to the intended recipient, you are hereby notified that you have received this document in error and that any review, dissemination, distribution, or copying of this message is strictly prohibited. If you have received this communication in error, please notify us immediately by e-mail, and delete the original message.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux