Re: jewel - rgw blocked on deep-scrub of bucket index pg

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,


On 06-05-17 20:08, Wido den Hollander wrote:
>> Op 6 mei 2017 om 9:55 schreef Christian Balzer <chibi@xxxxxxx>:
>>
>>
>>
>> Hello,
>>
>> On Sat, 6 May 2017 09:25:15 +0200 (CEST) Wido den Hollander wrote:
>>
>>>> Op 5 mei 2017 om 10:33 schreef Sam Wouters <sam@xxxxxxxxx>:
>>>>
>>>>
>>>> Hi,
>>>>
>>>> we have a small cluster running on jewel 10.2.7; NL-SAS disks only, osd
>>>> data and journal co located on the disks; main purpose rgw secondary zone.
>>>>
>>>> Since the upgrade to jewel, whenever a deep scrub starts on one of the
>>>> rgw index pool pg's, slow requests start piling up and rgw requests are
>>>> blocked after some hours.
>>>> The deep-scrub doesn't seem to finish (still running after +11 hours)
>>>> and only escape I found so far is a restart of the primary osd holding
>>>> the pg.
>>>>
>>>> Maybe important to know, we have some large rgw buckets regarding
>>>> #objects (+ 3 million) with only index sharding of 8.
>>>>
>>>> scrub related settings:
>>>> osd scrub sleep = 0.1  
>>> Try removing this line, it can block threads under Jewel.
I also found the bug report (#19497) yesterday, so indeed removed the
sleep and manually started the deep-scrub. I didn't had time to check
the result until now.
After almost 26 hours the deep-scrub operation finished (2017-05-05
10:57:08 -> 2017-05-06 12:29:05), however during the scrubbing frequent
timeouts and complete rgw downtime for various periods of time occurred.

Our primary cluster is still running hammer, and on there the index
pools are on ssd's, but this still raises concerns for after the planned
upgrade of that one...

Thanks a lot for the help!

r,
Sam
>>>
>> I'd really REALLY wish that would get fixed properly, as in the original
>> functionality restored. 
> Afaik new work is being done on this. There was a recent thread on the ceph-users or devel (can't find it) that new code is out there to fix this.
>
> Wido
>
>> Because as we've learned entrusting everything into internal Ceph queues
>> with priorities isn't working as expected in all cases.
>>
>> For a second, very distant option, turn it into a NOP for the time being.
>> As it stands now, it's another self-made, Jewel introduced bug...
>>
>> Christian
>>
>>> See how that works out.
>>>
>>> Wido
>>>
>>>> osd scrub during recovery = False
>>>> osd scrub priority = 1
>>>> osd deep scrub stride = 1048576
>>>> osd scrub chunk min = 1
>>>> osd scrub chunk max = 1
>>>>
>>>> Any help on debugging / resolving would be very much appreciated...
>>>>
>>>> regards,
>>>> Sam
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@xxxxxxxxxxxxxx
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com  
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@xxxxxxxxxxxxxx
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>
>> -- 
>> Christian Balzer        Network/Systems Engineer                
>> chibi@xxxxxxx   	Global OnLine Japan/Rakuten Communications
>> http://www.gol.com/
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux