Re: Failed to repair pg

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



you could try reading the data from this object and write it again
using rados get then rados put.

On Fri, Mar 8, 2019 at 3:32 AM Herbert Alexander Faleiros
<herbert@xxxxxxxxxxx> wrote:
>
> On Thu, Mar 07, 2019 at 01:37:55PM -0300, Herbert Alexander Faleiros wrote:
> > Hi,
> >
> > # ceph health detail
> > HEALTH_ERR 3 scrub errors; Possible data damage: 1 pg inconsistent
> > OSD_SCRUB_ERRORS 3 scrub errors
> > PG_DAMAGED Possible data damage: 1 pg inconsistent
> >     pg 2.2bb is active+clean+inconsistent, acting [36,12,80]
> >
> > # ceph pg repair 2.2bb
> > instructing pg 2.2bb on osd.36 to repair
> >
> > But:
> >
> > 2019-03-07 13:23:38.636881 [ERR]  Health check update: Possible data damage: 1 pg inconsistent, 1 pg repair (PG_DAMAGED)
> > 2019-03-07 13:20:38.373431 [ERR]  2.2bb deep-scrub 3 errors
> > 2019-03-07 13:20:38.373426 [ERR]  2.2bb deep-scrub 0 missing, 1 inconsistent objects
> > 2019-03-07 13:20:43.486860 [ERR]  Health check update: 3 scrub errors (OSD_SCRUB_ERRORS)
> > 2019-03-07 13:19:17.741350 [ERR]  deep-scrub 2.2bb 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986 : is an unexpected clone
> > 2019-03-07 13:19:17.523042 [ERR]  2.2bb shard 36 soid 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986 : data_digest 0xffffffff != data_digest 0xfc6b9538 from shard 12, size 0 != size 4194304 from auth oi 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986(482757'14986708 client.112595650.0:344888465 dirty|omap_digest s 4194304 uv 14974021 od ffffffff alloc_hint [0 0 0]), size 0 != size 4194304 from shard 12
> > 2019-03-07 13:19:17.523038 [ERR]  2.2bb shard 36 soid 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986 : candidate size 0 info size 4194304 mismatch
> > 2019-03-07 13:16:48.542673 [ERR]  2.2bb repair 2 errors, 1 fixed
> > 2019-03-07 13:16:48.542656 [ERR]  2.2bb repair 1 missing, 0 inconsistent objects
> > 2019-03-07 13:16:53.774956 [ERR]  Health check update: Possible data damage: 1 pg inconsistent (PG_DAMAGED)
> > 2019-03-07 13:16:53.774916 [ERR]  Health check update: 2 scrub errors (OSD_SCRUB_ERRORS)
> > 2019-03-07 13:15:16.986872 [ERR]  repair 2.2bb 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986 : is an unexpected clone
> > 2019-03-07 13:15:16.986817 [ERR]  2.2bb shard 36 2:dd4a7bd3:::rbd_data.dfd5e2235befd0.000000000001c299:4f986 : missing
> > 2019-03-07 13:12:18.517442 [ERR]  Health check update: Possible data damage: 1 pg inconsistent, 1 pg repair (PG_DAMAGED)
> >
> > Also tried deep-scrub and scrub, same results.
> >
> > Also set noscrub,nodeep-scrub, kicked currently active scrubs one at
> > a time using 'ceph osd down <id>'. After the last scrub was kicked,
> > forced scrub ran immediately then 'ceph pg repair', no luck.
> >
> > Finally tryed the manual aproach:
> >
> >  - stop osd.36
> >  - flush-journal
> >  - rm rbd\udata.dfd5e2235befd0.000000000001c299__4f986_CBDE52BB__2
> >  - start osd.36
> >  - ceph pg repair 2.2bb
> >
> > Also no luck...
> >
> > rbd\udata.dfd5e2235befd0.000000000001c299__4f986_CBDE52BB__2 at osd.36
> > is empty (0 size). At osd.80 4.0M, osd.2 is bluestore (can't find it).
> >
> > Ceph is 12.2.10, I'm currently migrating all my OSDs to bluestore.
> >
> > Is there anything else I can do?
>
> Should I do something like this? (below, after stop osd.36)
>
> # ceph-objectstore-tool --data-path /var/lib/ceph/osd/ceph-36/ --journal-path /dev/sdc1 rbd_data.dfd5e2235befd0.000000000001c299 remove-clone-metadata 326022
>
> I'm no sure about rbd_data.$RBD and $CLONEID (took from rados
> list-inconsistent-obj, also below).
>
> > # rados list-inconsistent-obj 2.2bb | jq
> > {
> >   "epoch": 484655,
> >   "inconsistents": [
> >     {
> >       "object": {
> >         "name": "rbd_data.dfd5e2235befd0.000000000001c299",
> >         "nspace": "",
> >         "locator": "",
> >         "snap": 326022,
> >         "version": 14974021
> >       },
> >       "errors": [
> >         "data_digest_mismatch",
> >         "size_mismatch"
> >       ],
> >       "union_shard_errors": [
> >         "size_mismatch_info",
> >         "obj_size_info_mismatch"
> >       ],
> >       "selected_object_info": {
> >         "oid": {
> >           "oid": "rbd_data.dfd5e2235befd0.000000000001c299",
> >           "key": "",
> >           "snapid": 326022,
> >           "hash": 3420345019,
> >           "max": 0,
> >           "pool": 2,
> >           "namespace": ""
> >         },
> >         "version": "482757'14986708",
> >         "prior_version": "482697'14980304",
> >         "last_reqid": "client.112595650.0:344888465",
> >         "user_version": 14974021,
> >         "size": 4194304,
> >         "mtime": "2019-03-02 22:30:23.812849",
> >         "local_mtime": "2019-03-02 22:30:23.813281",
> >         "lost": 0,
> >         "flags": [
> >           "dirty",
> >           "omap_digest"
> >         ],
> >         "legacy_snaps": [],
> >         "truncate_seq": 0,
> >         "truncate_size": 0,
> >         "data_digest": "0xffffffff",
> >         "omap_digest": "0xffffffff",
> >         "expected_object_size": 0,
> >         "expected_write_size": 0,
> >         "alloc_hint_flags": 0,
> >         "manifest": {
> >           "type": 0,
> >           "redirect_target": {
> >             "oid": "",
> >             "key": "",
> >             "snapid": 0,
> >             "hash": 0,
> >             "max": 0,
> >             "pool": -9223372036854776000,
> >             "namespace": ""
> >           }
> >         },
> >         "watchers": {}
> >       },
> >       "shards": [
> >         {
> >           "osd": 12,
> >           "primary": false,
> >           "errors": [],
> >           "size": 4194304,
> >           "omap_digest": "0xffffffff",
> >           "data_digest": "0xfc6b9538"
> >         },
> >         {
> >           "osd": 36,
> >           "primary": true,
> >           "errors": [
> >             "size_mismatch_info",
> >             "obj_size_info_mismatch"
> >           ],
> >           "size": 0,
> >           "omap_digest": "0xffffffff",
> >           "data_digest": "0xffffffff",
> >           "object_info": {
> >             "oid": {
> >               "oid": "rbd_data.dfd5e2235befd0.000000000001c299",
> >               "key": "",
> >               "snapid": 326022,
> >               "hash": 3420345019,
> >               "max": 0,
> >               "pool": 2,
> >               "namespace": ""
> >             },
> >             "version": "482757'14986708",
> >             "prior_version": "482697'14980304",
> >             "last_reqid": "client.112595650.0:344888465",
> >             "user_version": 14974021,
> >             "size": 4194304,
> >             "mtime": "2019-03-02 22:30:23.812849",
> >             "local_mtime": "2019-03-02 22:30:23.813281",
> >             "lost": 0,
> >             "flags": [
> >               "dirty",
> >               "omap_digest"
> >             ],
> >             "legacy_snaps": [],
> >             "truncate_seq": 0,
> >             "truncate_size": 0,
> >             "data_digest": "0xffffffff",
> >             "omap_digest": "0xffffffff",
> >             "expected_object_size": 0,
> >             "expected_write_size": 0,
> >             "alloc_hint_flags": 0,
> >             "manifest": {
> >               "type": 0,
> >               "redirect_target": {
> >                 "oid": "",
> >                 "key": "",
> >                 "snapid": 0,
> >                 "hash": 0,
> >                 "max": 0,
> >                 "pool": -9223372036854776000,
> >                 "namespace": ""
> >               }
> >             },
> >             "watchers": {}
> >           }
> >         },
> >         {
> >           "osd": 80,
> >           "primary": false,
> >           "errors": [],
> >           "size": 4194304,
> >           "omap_digest": "0xffffffff",
> >           "data_digest": "0xfc6b9538"
> >         }
> >       ]
> >     }
> >   ]
> > }
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



-- 
Cheers,
Brad
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux