Is there a way to remove that object from all osds? As it is unexpected it should not harm. Greets, Stefan Am 05.08.2017 um 09:03 schrieb Stefan Priebe - Profihost AG: > Hello, > > i'm trying to fix a cluster where one pg is in > active+clean+inconsistent+snaptrim > > state. > > The log says: > 2017-08-05 08:57:43.240030 osd.20 [ERR] 3.61a repair 0 missing, 1 > inconsistent objects > 2017-08-05 08:57:43.240044 osd.20 [ERR] 3.61a repair 4 errors, 2 fixed > 2017-08-05 08:57:43.242828 osd.20 [ERR] trim_object Snap 9d455 not in clones > > find says: > > osd.20]# find . -name "*9d455*" -exec ls -la "{}" \; > -rw-r--r-- 1 ceph ceph 0 Aug 5 08:57 > ./current/3.61a_head/DIR_A/DIR_1/DIR_6/DIR_A/rbd\udata.106dd406b8b4567.000000000000018c__9d455_BCB2A61A__3 > > find . -name "*9d455*" -exec ls -la "{}" \; > osd.70]# -rw-r--r-- 1 ceph ceph 4194304 Aug 5 08:57 > ./current/3.61a_head/DIR_A/DIR_1/DIR_6/DIR_A/rbd\udata.106dd406b8b4567.000000000000018c__9d455_BCB2A61A__3 > > osd.57]# find . -name "*9d455*" -exec ls -la "{}" \; > # no file found > > I tried the following stuff to solve this issue. > > stop osd.y > copy file from osd.x to osd.y > start osd.y > > But all that does not help. It stays in this state. > > The output of a ceph pg repair is: > 2017-08-05 08:57:25.715812 osd.20 [ERR] 3.61a shard 20: soid > 3:58654d3d:::rbd_data.106dd406b8b4567.000000000000018c:9d455 data_digest > 0xffffffff != data_digest 0xbd9585ca from auth oi > 3:58654d3d:::rbd_data.106dd406b8b4567.000000000000018c:9d455(818765'26612094 > osd.20.0:562819 [9d455] dirty|data_digest|omap_digest s 4194304 uv > 26585239 dd bd9585ca od ffffffff alloc_hint [0 0]), size 0 != size > 4194304 from auth oi > 3:58654d3d:::rbd_data.106dd406b8b4567.000000000000018c:9d455(818765'26612094 > osd.20.0:562819 [9d455] dirty|data_digest|omap_digest s 4194304 uv > 26585239 dd bd9585ca od ffffffff alloc_hint [0 0]) > 2017-08-05 08:57:25.715817 osd.20 [ERR] 3.61a shard 70: soid > 3:58654d3d:::rbd_data.106dd406b8b4567.000000000000018c:9d455 data_digest > 0x43d61c5d != data_digest 0xffffffff from shard 20, data_digest > 0x43d61c5d != data_digest 0xbd9585ca from auth oi > 3:58654d3d:::rbd_data.106dd406b8b4567.000000000000018c:9d455(818765'26612094 > osd.20.0:562819 [9d455] dirty|data_digest|omap_digest s 4194304 uv > 26585239 dd bd9585ca od ffffffff alloc_hint [0 0]), size 4194304 != size > 0 from shard 20 > 2017-08-05 08:57:25.715903 osd.20 [ERR] repair 3.61a > 3:58654d3d:::rbd_data.106dd406b8b4567.000000000000018c:9d455 is an > unexpected clone > > How to get this fixed? > > Greets, > Stefan > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com