Re: large omap object

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



There may be a mismatch between be auto-restarting and the omap warning code. Looks like you already have 349 shards, with 13 of them warning on size!
You can increase a config value to shut that error up, but you may want to get somebody from RGW to look at how you’ve managed to exceed those default limits.
On Wed, Jun 13, 2018 at 4:36 AM stephan schultchen <stephan.schultchen@xxxxxxxxx> wrote:
Hello,

i am running a ceph 13.2.0 cluster exclusively for radosrw / s3.

i only have one big bucket. and the cluster is currently in warning state:

  cluster:
    id:     d605c463-9f1c-4d91-a390-a28eedb21650
    health: HEALTH_WARN
            13 large omap objects

i tried to google it, but i was not able to find what to do about the "large omap objects".

as far as i understand ceph should automatically re shard an s3 bucket when an omap is getting to big. or is this something i have to do?

"radosgw-admin reshard list" tells that no resharding is ongoing right now.


radosgw-admin metadata get bucket.instance:nuxeo_live:6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.4854.4
{
    "key": "bucket.instance:nuxeo_live:6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.4854.4",
    "ver": {
        "tag": "Y2epzPoujRDfxM5CNMZgKPaA",
        "ver": 6
    },
    "mtime": "2018-06-08 14:48:15.515840Z",
    "data": {
        "bucket_info": {
            "bucket": {
                "name": "nuxeo_live",
                "marker": "6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.4848.1",
                "bucket_id": "6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.4854.4",
                "tenant": "",
                "explicit_placement": {
                    "data_pool": "",
                    "data_extra_pool": "",
                    "index_pool": ""
                }
            },
            "creation_time": "2018-05-23 13:31:57.664398Z",
            "owner": "nuxeo_live",
            "flags": 0,
            "zonegroup": "506cc27c-fef5-4b89-a9f3-4c928a74b955",
            "placement_rule": "default-placement",
            "has_instance_obj": "true",
            "quota": {
                "enabled": false,
                "check_on_raw": false,
                "max_size": -1,
                "max_size_kb": 0,
                "max_objects": -1
            },
            "num_shards": 349,
            "bi_shard_hash_type": 0,
            "requester_pays": "false",
            "has_website": "false",
            "swift_versioning": "false",
            "swift_ver_location": "",
            "index_type": 0,
            "mdsearch_config": [],
            "reshard_status": 2,
            "new_bucket_instance_id": "6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.176143.1"
        },
        "attrs": [
            {
                "key": "user.rgw.acl",
                "val": "AgKpAAAAAwIhAAAACgAAAG51eGVvX2xpdmUPAAAAbnV4ZW8gbGl2ZSB1c2VyBAN8AAAAAQEAAAAKAAAAbnV4ZW9fbGl2ZQ8AAAABAAAACgAAAG51eGVvX2xpdmUFA0UAAAACAgQAAAAAAAAACgAAAG51eGVvX2xpdmUAAAAAAAAAAAICBAAAAA8AAAAPAAAAbnV4ZW8gbGl2ZSB1c2VyAAAAAAAAAAAAAAAAAAAAAA=="
            },
            {
                "key": "user.rgw.idtag",
                "val": ""
            }
        ]
 
i also tried to manually trigger a resharding. but it failed with:

  • NOTICE: operation will not remove old bucket index objects ***
  • these will need to be removed manually ***
    tenant:
    bucket name: nuxeo_live
    old bucket instance id: 6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.184670.1
    new bucket instance id: 6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.176197.1
    WARNING: RGWReshard::add failed to drop lock on bucket_name:6f85d718-fd2e-4c1b-a21d-bafb04a8cfcc.184670.1 ret=-2

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux