Re: Repair inconsistent pgs..

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



>From a quick peek it looks like some of the OSDs are missing clones of
objects. I'm not sure how that could happen and I'd expect the pg
repair to handle that but if it's not there's probably something
wrong; what version of Ceph are you running? Sam, is this something
you've seen, a new bug, or some kind of config issue?
-Greg

On Tue, Aug 18, 2015 at 6:27 AM, Voloshanenko Igor
<igor.voloshanenko@xxxxxxxxx> wrote:
> Hi all, at our production cluster, due high rebalancing ((( we have 2 pgs in
> inconsistent state...
>
> root@temp:~# ceph health detail | grep inc
> HEALTH_ERR 2 pgs inconsistent; 18 scrub errors
> pg 2.490 is active+clean+inconsistent, acting [56,15,29]
> pg 2.c4 is active+clean+inconsistent, acting [56,10,42]
>
> From OSD logs, after recovery attempt:
>
> root@test:~# ceph pg dump | grep -i incons | cut -f 1 | while read i; do
> ceph pg repair ${i} ; done
> dumped all in format plain
> instructing pg 2.490 on osd.56 to repair
> instructing pg 2.c4 on osd.56 to repair
>
> /var/log/ceph/ceph-osd.56.log:51:2015-08-18 07:26:37.035910 7f94663b3700 -1
> log_channel(cluster) log [ERR] : deep-scrub 2.490
> f5759490/rbd_data.1631755377d7e.00000000000004da/head//2 expected clone
> 90c59490/rbd_data.eb486436f2beb.0000000000007a65/141//2
> /var/log/ceph/ceph-osd.56.log:52:2015-08-18 07:26:37.035960 7f94663b3700 -1
> log_channel(cluster) log [ERR] : deep-scrub 2.490
> fee49490/rbd_data.12483d3ba0794b.000000000000522f/head//2 expected clone
> f5759490/rbd_data.1631755377d7e.00000000000004da/141//2
> /var/log/ceph/ceph-osd.56.log:53:2015-08-18 07:26:37.036133 7f94663b3700 -1
> log_channel(cluster) log [ERR] : deep-scrub 2.490
> a9b39490/rbd_data.12483d3ba0794b.00000000000037b3/head//2 expected clone
> fee49490/rbd_data.12483d3ba0794b.000000000000522f/141//2
> /var/log/ceph/ceph-osd.56.log:54:2015-08-18 07:26:37.036243 7f94663b3700 -1
> log_channel(cluster) log [ERR] : deep-scrub 2.490
> bac19490/rbd_data.1238e82ae8944a.000000000000032e/head//2 expected clone
> a9b39490/rbd_data.12483d3ba0794b.00000000000037b3/141//2
> /var/log/ceph/ceph-osd.56.log:55:2015-08-18 07:26:37.036289 7f94663b3700 -1
> log_channel(cluster) log [ERR] : deep-scrub 2.490
> 98519490/rbd_data.123e9c2ae8944a.0000000000000807/head//2 expected clone
> bac19490/rbd_data.1238e82ae8944a.000000000000032e/141//2
> /var/log/ceph/ceph-osd.56.log:56:2015-08-18 07:26:37.036314 7f94663b3700 -1
> log_channel(cluster) log [ERR] : deep-scrub 2.490
> c3c09490/rbd_data.1238e82ae8944a.0000000000000c2b/head//2 expected clone
> 98519490/rbd_data.123e9c2ae8944a.0000000000000807/141//2
> /var/log/ceph/ceph-osd.56.log:57:2015-08-18 07:26:37.036363 7f94663b3700 -1
> log_channel(cluster) log [ERR] : deep-scrub 2.490
> 28809490/rbd_data.edea7460fe42b.00000000000001d9/head//2 expected clone
> c3c09490/rbd_data.1238e82ae8944a.0000000000000c2b/141//2
> /var/log/ceph/ceph-osd.56.log:58:2015-08-18 07:26:37.036432 7f94663b3700 -1
> log_channel(cluster) log [ERR] : deep-scrub 2.490
> e1509490/rbd_data.1423897545e146.00000000000009a6/head//2 expected clone
> 28809490/rbd_data.edea7460fe42b.00000000000001d9/141//2
> /var/log/ceph/ceph-osd.56.log:59:2015-08-18 07:26:38.548765 7f94663b3700 -1
> log_channel(cluster) log [ERR] : 2.490 deep-scrub 17 errors
>
> So, how i can solve "expected clone" situation by hand?
> Thank in advance!
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux