> hello
>
> I am running ceph hammer on debian jessie. using 6 old used underwhelming
> servers
>
> the cluster is a "in-migration" bastard mix of 3TB sata drives with on disk
> journal partition, beeing migrated to 5 disk raid5 MD arrays with ssd
> journals, for ram limitation reasons. There are about 18 raid5 sets atm and
> the rest is 3TB spinners.
>
> I have some challenges with scrub errors, that i am trying to sort out using
> this http://ceph.com/planet/ceph-manually-repair-object/ method. but they
> are quite stubborn/sticky
>
> i do see that osd.8 is often represented in these inconsistencies. but the
> broken objects are not allways on osd.8 itself
>
> in the instructions at http://ceph.com/planet/ceph-manually-repair-object/,
> one finds the object name by grepping in the logs.
> but some of these haven been here a while. so how can i identify the broken
> object if the log file have been rotated away ?
>
> in the end i move away the broken object with size 0 and run pg repair, but
> the error is not removed.
> does the pg need to scrub after the repair for it to clear the error. ?
>
> any advice is appreciated
>
> kind regards
> Ronny Aasen
>
>
> #ceph -s
> cluster 3c229f54-bd12-4b4e-a143-1ec73dd0f12a
> health HEALTH_ERR
> 3 pgs degraded
> 9 pgs inconsistent
> 3 pgs recovering
> 3 pgs stuck degraded
> 3 pgs stuck unclean
> recovery 88/125583766 objects degraded (0.000%)
> recovery 666778/125583766 objects misplaced (0.531%)
> recovery 88/45311043 unfound (0.000%)
> 9 scrub errors
> noout,noscrub,nodeep-scrub flag(s) set
> monmap e1: 3 mons at
> {mon1=10.24.11.11:6789/0,mon2=10.24.11.12:6789/0,mon3=10.24.11.13:6789/0}
> election epoch 60, quorum 0,1,2 mon1,mon2,mon3
> osdmap e105977: 92 osds: 92 up, 92 in; 2 remapped pgs
> flags noout,noscrub,nodeep-scrub
> pgmap v12896186: 4608 pgs, 3 pools, 117 TB data, 44249 kobjects
> 308 TB used, 107 TB / 416 TB avail
> 88/125583766 objects degraded (0.000%)
> 666778/125583766 objects misplaced (0.531%)
> 88/45311043 unfound (0.000%)
> 4593 active+clean
> 9 active+clean+inconsistent
> 3 active+clean+scrubbing
> 2 active+recovering+degraded+remapped
> 1 active+recovering+degraded
> client io 4572 kB/s rd, 1141 op/s
>
> # ceph health detail
> HEALTH_ERR 3 pgs degraded; 9 pgs inconsistent; 3 pgs recovering; 3 pgs stuck
> degraded; 3 pgs stuck unclean; recovery 88/125583766 objects degraded
> (0.000%); recovery 666778/125583766 objects misplaced (0.531%); recovery
> 88/45311043 unfound (0.000%); 9 scrub errors; noout,noscrub,nodeep-scrub
> flag(s) set
> pg 6.d4 is stuck unclean for 3770820.461291, current state
> active+recovering+degraded+remapped, last acting [62,8]
> pg 6.da is stuck unclean for 2420102.778679, current state
> active+recovering+degraded, last acting [6,110]
> pg 6.ab is stuck unclean for 3774233.330685, current state
> active+recovering+degraded+remapped, last acting [12,8]
> pg 6.d4 is stuck degraded for 304239.715211, current state
> active+recovering+degraded+remapped, last acting [62,8]
> pg 6.da is stuck degraded for 416210.309539, current state
> active+recovering+degraded, last acting [6,110]
> pg 6.ab is stuck degraded for 304239.779541, current state
> active+recovering+degraded+remapped, last acting [12,8]
> pg 1.356 is active+clean+inconsistent, acting [8,84,39]
> pg 1.1a7 is active+clean+inconsistent, acting [8,36,34]
> pg 1.11e is active+clean+inconsistent, acting [8,12,6]
> pg 6.da is active+recovering+degraded, acting [6,110], 25 unfound
> pg 6.d4 is active+recovering+degraded+remapped, acting [62,8], 25 unfound
> pg 6.ab is active+recovering+degraded+remapped, acting [12,8], 38 unfound
> pg 1.de4 is active+clean+inconsistent, acting [41,8,108]
> pg 1.c90 is active+clean+inconsistent, acting [12,71,8]
> pg 1.ae6 is active+clean+inconsistent, acting [8,36,49]
> pg 1.8bc is active+clean+inconsistent, acting [59,8,107]
> pg 1.806 is active+clean+inconsistent, acting [60,3,106]
> pg 1.675 is active+clean+inconsistent, acting [37,106,62]
> recovery 88/125583766 objects degraded (0.000%)
> recovery 666778/125583766 objects misplaced (0.531%)
> recovery 88/45311043 unfound (0.000%)
> 9 scrub errors
> noout,noscrub,nodeep-scrub flag(s) set
>
> NB: the 88 unfound objects are in a pool i experimented with size 2, so not
> important in this context.
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com