RGW multisite sync and latencies problem

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
I have problem with RGW in multisite configuration with Nautilus 14.2.11. Both zones with SSD and 10Gbps network. Master zone consist from 5x DELL R740XD servers (every 256GB RAM, 8x800GB SSD for CEPH, 24xCPU). Secondary zone (temporary for testing) consist from 3x HPE DL360 Gen10 servers (every 256GB RAM, 6x800GB SSD, 48CPU). 

We have 17 test buckets with manual sharding (101 shards). Every bucket with 10M of small objects (10kB - 15kB). Zonegroup configuration is attached bellow. Replication of 150M objects from master to secondary zone takes almost 28 hours and the replication completed with success.

After deleting objects from one bucket in master zone is not possible to sync zones properly. I tried to restart both secondary RGWs, but without success. Sync status on secondary zone is behind master. The number of objects in buckets on master zone is different than on secondary zone. 

Ceph HEALTH status is WARNING on both zones. On master zone I have 146 large objects found in pool 'prg2a-1.rgw.buckets.index' 16 large objects found in pool 'prg2a-1.rgw.log'. On secondary zone 88 large objects found in pool 'prg2a-2.rgw.log' 1584 large objects found in pool 'prg2a-2.rgw.buckets.index'. 

AVG OSD latencies on secondary zone during sync was "read 0,158ms, write 1,897ms, overwrite 1,634ms". After unsuccesfull sync (after 12h of sync fall down RGW requests, IOPS and throughput) jumps up AVG OSD latencies to "read 125ms, write 30ms, overwrite 272ms". After stopping of both RGWs on secondary zone are AVG OSD latencies almost 0ms, but when I start RGWs on secondary zone again, OSD latencies will rise again to "read 125ms, write 30ms, overwrite 272ms" with spikes up to 3 seconds.

We have seen the same behaviour of ceph multisite with large number of object in one bucket (150M+ objects), so we tried different strategy with smaller buckets, but results are same.

I will appreciate any help or advice, how tune or diagnose multisite problems.
Does anyone else have any ideas? Is there anyone else with a similar use-case? I do not know what is wrong.

Thank you and best regards,
Miroslav

radosgw-admin zonegroup get
{
    "id": "ac0005da-2e9f-4f38-835f-72b289c240d0",
    "name": "prg2a",
    "api_name": "prg2a",
    "is_master": "true",
    "endpoints": [
        "http://s3.prg1a.sys.cz:80";,
        "http://s3.prg2a.sys.cz:80";
    ],
    "hostnames": [],
    "hostnames_s3website": [],
    "master_zone": "d9ebbd1f-3312-4083-b4c2-843e1fb899ad",
    "zones": [
        {
            "id": "d9ebbd1f-3312-4083-b4c2-843e1fb899ad",
            "name": "prg2a-1",
            "endpoints": [
                "http://10.104.200.101:7480";,
                "http://10.104.200.102:7480";
            ],
            "log_meta": "false",
            "log_data": "true",
            "bucket_index_max_shards": 0,
            "read_only": "false",
            "tier_type": "",
            "sync_from_all": "true",
            "sync_from": [],
            "redirect_zone": ""
        },
        {
            "id": "fdd76c02-c679-4ec7-8e7d-c14d2ac74fb4",
            "name": "prg2a-2",
            "endpoints": [
                "http://10.104.200.221:7480";,
                "http://10.104.200.222:7480";
            ],
            "log_meta": "false",
            "log_data": "true",
            "bucket_index_max_shards": 0,
            "read_only": "false",
            "tier_type": "",
            "sync_from_all": "true",
            "sync_from": [],
            "redirect_zone": ""
        }
    ],
    "placement_targets": [
        {
            "name": "default-placement",
            "tags": [],
            "storage_classes": [
                "STANDARD"
            ]
        }
    ],
    "default_placement": "default-placement",
    "realm_id": "cb831094-e219-44b8-89f3-fe25fc288c00"

ii  radosgw                              14.2.11-pve1                 amd64        REST gateway for RADOS distributed object store
ii  ceph                                 14.2.11-pve1                 amd64        distributed storage and file system
ii  ceph-base                            14.2.11-pve1                 amd64        common ceph daemon libraries and management tools
ii  ceph-common                          14.2.11-pve1                 amd64        common utilities to mount and interact with a ceph storage cluster
ii  ceph-fuse                            14.2.11-pve1                 amd64        FUSE-based client for the Ceph distributed file system
ii  ceph-mds                             14.2.11-pve1                 amd64        metadata server for the ceph distributed file system
ii  ceph-mgr                             14.2.11-pve1                 amd64        manager for the ceph distributed storage system
ii  ceph-mon                             14.2.11-pve1                 amd64        monitor server for the ceph storage system
ii  ceph-osd                             14.2.11-pve1                 amd64        OSD server for the ceph storage system
ii  libcephfs2                           14.2.11-pve1                 amd64        Ceph distributed file system client library
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux