Failed to repair pg

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

# ceph health detail
HEALTH_ERR 3 scrub errors; Possible data damage: 1 pg inconsistent
OSD_SCRUB_ERRORS 3 scrub errors
PG_DAMAGED Possible data damage: 1 pg inconsistent
    pg 2.2bb is active+clean+inconsistent, acting [36,12,80]

# ceph pg repair 2.2bb
instructing pg 2.2bb on osd.36 to repair

But:

2019-03-07 13:23:38.636881 [ERR]  Health check update: Possible data damage: 1 pg inconsistent, 1 pg repair (PG_DAMAGED) 
2019-03-07 13:20:38.373431 [ERR]  2.2bb deep-scrub 3 errors 
2019-03-07 13:20:38.373426 [ERR]  2.2bb deep-scrub 0 missing, 1 inconsistent objects 
2019-03-07 13:20:43.486860 [ERR]  Health check update: 3 scrub errors (OSD_SCRUB_ERRORS) 
2019-03-07 13:19:17.741350 [ERR]  deep-scrub 2.2bb 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986 : is an unexpected clone 
2019-03-07 13:19:17.523042 [ERR]  2.2bb shard 36 soid 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986 : data_digest 0xffffffff != data_digest 0xfc6b9538 from shard 12, size 0 != size 4194304 from auth oi 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986(482757'14986708 client.112595650.0:344888465 dirty|omap_digest s 4194304 uv 14974021 od ffffffff alloc_hint [0 0 0]), size 0 != size 4194304 from shard 12 
2019-03-07 13:19:17.523038 [ERR]  2.2bb shard 36 soid 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986 : candidate size 0 info size 4194304 mismatch 
2019-03-07 13:16:48.542673 [ERR]  2.2bb repair 2 errors, 1 fixed 
2019-03-07 13:16:48.542656 [ERR]  2.2bb repair 1 missing, 0 inconsistent objects 
2019-03-07 13:16:53.774956 [ERR]  Health check update: Possible data damage: 1 pg inconsistent (PG_DAMAGED) 
2019-03-07 13:16:53.774916 [ERR]  Health check update: 2 scrub errors (OSD_SCRUB_ERRORS) 
2019-03-07 13:15:16.986872 [ERR]  repair 2.2bb 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986 : is an unexpected clone 
2019-03-07 13:15:16.986817 [ERR]  2.2bb shard 36 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986 : missing 
2019-03-07 13:12:18.517442 [ERR]  Health check update: Possible data damage: 1 pg inconsistent, 1 pg repair (PG_DAMAGED) 

Also tried deep-scrub and scrub, same results.

Also set noscrub,nodeep-scrub, kicked currently active scrubs one at
a time using 'ceph osd down <id>'. After the last scrub was kicked,
forced scrub ran immediately then 'ceph pg repair', no luck.

Finally tryed the manual aproach:

 - stop osd.36
 - flush-journal
 - rm rbd\udata.dfd5e2235befd0.000000000001c299__4f986_CBDE52BB__2
 - start osd.36
 - ceph pg repair 2.2bb

Also no luck...

rbd\udata.dfd5e2235befd0.000000000001c299__4f986_CBDE52BB__2 at osd.36
is empty (0 size). At osd.80 4.0M, osd.2 is bluestore (can't find it).

Ceph is 12.2.10, I'm currently migrating all my OSDs to bluestore.

Is there anything else I can do?

# rados list-inconsistent-obj 2.2bb | jq
{
  "epoch": 484655,
  "inconsistents": [
    {
      "object": {
        "name": "rbd_data.dfd5e2235befd0.000000000001c299",
        "nspace": "",
        "locator": "",
        "snap": 326022,
        "version": 14974021
      },
      "errors": [
        "data_digest_mismatch",
        "size_mismatch"
      ],
      "union_shard_errors": [
        "size_mismatch_info",
        "obj_size_info_mismatch"
      ],
      "selected_object_info": {
        "oid": {
          "oid": "rbd_data.dfd5e2235befd0.000000000001c299",
          "key": "",
          "snapid": 326022,
          "hash": 3420345019,
          "max": 0,
          "pool": 2,
          "namespace": ""
        },
        "version": "482757'14986708",
        "prior_version": "482697'14980304",
        "last_reqid": "client.112595650.0:344888465",
        "user_version": 14974021,
        "size": 4194304,
        "mtime": "2019-03-02 22:30:23.812849",
        "local_mtime": "2019-03-02 22:30:23.813281",
        "lost": 0,
        "flags": [
          "dirty",
          "omap_digest"
        ],
        "legacy_snaps": [],
        "truncate_seq": 0,
        "truncate_size": 0,
        "data_digest": "0xffffffff",
        "omap_digest": "0xffffffff",
        "expected_object_size": 0,
        "expected_write_size": 0,
        "alloc_hint_flags": 0,
        "manifest": {
          "type": 0,
          "redirect_target": {
            "oid": "",
            "key": "",
            "snapid": 0,
            "hash": 0,
            "max": 0,
            "pool": -9223372036854776000,
            "namespace": ""
          }
        },
        "watchers": {}
      },
      "shards": [
        {
          "osd": 12,
          "primary": false,
          "errors": [],
          "size": 4194304,
          "omap_digest": "0xffffffff",
          "data_digest": "0xfc6b9538"
        },
        {
          "osd": 36,
          "primary": true,
          "errors": [
            "size_mismatch_info",
            "obj_size_info_mismatch"
          ],
          "size": 0,
          "omap_digest": "0xffffffff",
          "data_digest": "0xffffffff",
          "object_info": {
            "oid": {
              "oid": "rbd_data.dfd5e2235befd0.000000000001c299",
              "key": "",
              "snapid": 326022,
              "hash": 3420345019,
              "max": 0,
              "pool": 2,
              "namespace": ""
            },
            "version": "482757'14986708",
            "prior_version": "482697'14980304",
            "last_reqid": "client.112595650.0:344888465",
            "user_version": 14974021,
            "size": 4194304,
            "mtime": "2019-03-02 22:30:23.812849",
            "local_mtime": "2019-03-02 22:30:23.813281",
            "lost": 0,
            "flags": [
              "dirty",
              "omap_digest"
            ],
            "legacy_snaps": [],
            "truncate_seq": 0,
            "truncate_size": 0,
            "data_digest": "0xffffffff",
            "omap_digest": "0xffffffff",
            "expected_object_size": 0,
            "expected_write_size": 0,
            "alloc_hint_flags": 0,
            "manifest": {
              "type": 0,
              "redirect_target": {
                "oid": "",
                "key": "",
                "snapid": 0,
                "hash": 0,
                "max": 0,
                "pool": -9223372036854776000,
                "namespace": ""
              }
            },
            "watchers": {}
          }
        },
        {
          "osd": 80,
          "primary": false,
          "errors": [],
          "size": 4194304,
          "omap_digest": "0xffffffff",
          "data_digest": "0xfc6b9538"
        }
      ]
    }
  ]
}

-- 
Herbert
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux