Re: Inconsistent PG's, repair ineffective

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I've checked, all the disks are fine and the cluster is healthy except for the inconsistent objects.

How would I go about manually repairing?

On May 21, 2013, at 3:26 PM, David Zafman <david.zafman@xxxxxxxxxxx> wrote:

> 
> I can't reproduce this on v0.61-2.  Could the disks for osd.13 & osd.22 be unwritable?
> 
> In your case it looks like the 3rd replica is probably the bad one, since osd.13 and osd.22 are the same.  You probably want to manually repair the 3rd replica.
> 
> David Zafman
> Senior Developer
> http://www.inktank.com
> 
> 
> 
> 
> On May 21, 2013, at 6:45 AM, John Nielsen <lists@xxxxxxxxxxxx> wrote:
> 
>> Cuttlefish on CentOS 6, ceph-0.61.2-0.el6.x86_64.
>> 
>> On May 21, 2013, at 12:13 AM, David Zafman <david.zafman@xxxxxxxxxxx> wrote:
>> 
>>> 
>>> What version of ceph are you running?
>>> 
>>> David Zafman
>>> Senior Developer
>>> http://www.inktank.com
>>> 
>>> On May 20, 2013, at 9:14 AM, John Nielsen <lists@xxxxxxxxxxxx> wrote:
>>> 
>>>> Some scrub errors showed up on our cluster last week. We had some issues with host stability a couple weeks ago; my guess is that errors were introduced at that point and a recent background scrub detected them. I was able to clear most of them via "ceph pg repair", but several remain. Based on some other posts, I'm guessing that they won't repair because it is the primary copy that has the error. All of our pools are set to size 3 so there _ought_ to be a way to verify and restore the correct data, right?
>>>> 
>>>> Below is some log output about one of the problem PG's. Can anyone suggest a way to fix the inconsistencies?
>>>> 
>>>> 2013-05-20 10:07:54.529582 osd.13 10.20.192.111:6818/20919 3451 : [ERR] 19.1b osd.13: soid 507ada1b/rb.0.6989.2ae8944a.00000000005b/5//19 digest 4289025870 != known digest 4190506501
>>>> 2013-05-20 10:07:54.529585 osd.13 10.20.192.111:6818/20919 3452 : [ERR] 19.1b osd.22: soid 507ada1b/rb.0.6989.2ae8944a.00000000005b/5//19 digest 4289025870 != known digest 4190506501
>>>> 2013-05-20 10:07:54.606034 osd.13 10.20.192.111:6818/20919 3453 : [ERR] 19.1b repair 0 missing, 1 inconsistent objects
>>>> 2013-05-20 10:07:54.606066 osd.13 10.20.192.111:6818/20919 3454 : [ERR] 19.1b repair 2 errors, 2 fixed
>>>> 2013-05-20 10:07:55.034221 osd.13 10.20.192.111:6818/20919 3455 : [ERR] 19.1b osd.13: soid 507ada1b/rb.0.6989.2ae8944a.00000000005b/5//19 digest 4289025870 != known digest 4190506501
>>>> 2013-05-20 10:07:55.034224 osd.13 10.20.192.111:6818/20919 3456 : [ERR] 19.1b osd.22: soid 507ada1b/rb.0.6989.2ae8944a.00000000005b/5//19 digest 4289025870 != known digest 4190506501
>>>> 2013-05-20 10:07:55.113230 osd.13 10.20.192.111:6818/20919 3457 : [ERR] 19.1b deep-scrub 0 missing, 1 inconsistent objects
>>>> 2013-05-20 10:07:55.113235 osd.13 10.20.192.111:6818/20919 3458 : [ERR] 19.1b deep-scrub 2 errors
>>>> 
>>>> Thanks,
>>>> 
>>>> JN
>>>> 
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@xxxxxxxxxxxxxx
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>> 
>>> 
>> 
> 
> 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux