Re: pg incomplete state

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Greg,

Thanks for the insight.  I suspect things are somewhat sane given that I
did erase the primary (osd.30) and the secondary (osd.11) still contains
pg data.

If I may, could you clarify the process of backfill a little?

I understand the min_size allows I/O on the object to resume while there
are only that many replicas (ie. 1 once changed) and this would let
things move forward.

I would expect, however, that some backfill would already be on-going
for pg 3.ea on osd.30.  As far as I can tell, there isn't anything
happening.  The pg 3.ea directory is just as empty today as it was
yesterday.

Will changing the min_size actually trigger backfill to begin for an
object if has stalled or never got started?

An alternative idea I had was to take osd.30 back out of the cluster so
that pg 3.ae [30,11] would get mapped to some other osd to maintain
replication.  This seems a bit heavy handed though, given that only this
one pg is affected.

Thanks for any follow up.

~jpr 


On 10/21/2015 01:21 PM, Gregory Farnum wrote:
> On Tue, Oct 20, 2015 at 7:22 AM, John-Paul Robinson <jpr@xxxxxxx> wrote:
>> Hi folks
>>
>> I've been rebuilding drives in my cluster to add space.  This has gone
>> well so far.
>>
>> After the last batch of rebuilds, I'm left with one placement group in
>> an incomplete state.
>>
>> [sudo] password for jpr:
>> HEALTH_WARN 1 pgs incomplete; 1 pgs stuck inactive; 1 pgs stuck unclean
>> pg 3.ea is stuck inactive since forever, current state incomplete, last
>> acting [30,11]
>> pg 3.ea is stuck unclean since forever, current state incomplete, last
>> acting [30,11]
>> pg 3.ea is incomplete, acting [30,11]
>>
>> I've restarted both OSD a few times but it hasn't cleared the error.
>>
>> On the primary I see errors in the log related to slow requests:
>>
>> 2015-10-20 08:40:36.678569 7f361585c700  0 log [WRN] : 8 slow requests,
>> 3 included below; oldest blocked for > 31.922487 secs
>> 2015-10-20 08:40:36.678580 7f361585c700  0 log [WRN] : slow request
>> 31.531606 seconds old, received at 2015-10-20 08:40:05.146902:
>> osd_op(client.158903.1:343217143 rb.0.25cf8.238e1f29.00000000a044 [read
>> 1064960~262144] 3.ae9968ea RETRY) v4 currently reached pg
>> 2015-10-20 08:40:36.678592 7f361585c700  0 log [WRN] : slow request
>> 31.531591 seconds old, received at 2015-10-20 08:40:05.146917:
>> osd_op(client.158903.1:343217144 rb.0.25cf8.238e1f29.00000000a044 [read
>> 2113536~262144] 3.ae9968ea RETRY) v4 currently reached pg
>> 2015-10-20 08:40:36.678599 7f361585c700  0 log [WRN] : slow request
>> 31.531551 seconds old, received at 2015-10-20 08:40:05.146957:
>> osd_op(client.158903.1:343232634 ekessler-default.rbd [watch 35~0]
>> 3.e4bd50ea) v4 currently reached pg
>>
>> Note's online suggest this is an issue with the journal and that it may
>> be possible to export and rebuild thepg.  I don't have firefly.
>>
>> https://ceph.com/community/incomplete-pgs-oh-my/
>>
>> Interestingly, pg 3.ea appears to be complete on osd.11 (the secondary)
>> but missing entirely on osd.30 (the primary).
>>
>> on osd.33 (primary):
>>
>> crowbar@da0-36-9f-0e-2b-88:~$ du -sk
>> /var/lib/ceph/osd/ceph-30/current/3.ea_head/
>> 0       /var/lib/ceph/osd/ceph-30/current/3.ea_head/
>>
>> on osd.11 (secondary):
>>
>> crowbar@da0-36-9f-0e-2b-40:~$ du -sh
>> /var/lib/ceph/osd/ceph-11/current/3.ea_head/
>> 63G     /var/lib/ceph/osd/ceph-11/current/3.ea_head/
>>
>> This makes some sense since, my disk drive rebuilding activity
>> reformatted the primary osd.30.  It also gives me some hope that my data
>> is not lost.
>>
>> I understand incomplete means problem with journal, but is there a way
>> to dig deeper into this or possible to get the secondary's data to take
>> over?
> If you're running an older version of Ceph (Firefly or earlier,
> maybe?), "incomplete" can also mean "not enough replicas". It looks
> like that's what you're hitting here, if osd.11 is not reporting any
> issues. If so, simply setting the min_size on this pool to 1 until the
> backfilling is done should let you get going.
> -Greg

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux