Re: ceph mds slow requests

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




Yes I was also thinking of setting this configuration. But I am just 
wondering what are the drawbacks, if there are any? Why is the ceph team 
not making this a default setting if it is so good?



-----Original Message-----
To: ceph-users@xxxxxxx
Subject:  Re: ceph mds slow requests

Hi,

although I've read many threads mentioning these two osd config options 
I didn't know what to expect of it. But since you explicitly referred 
the slow requests I decided to give it a try and changed 'osd op queue 
cut off' to "high" ('osd op queue' was already "wpq"). We've had two 
deep-scrub cycles since then, half of our PGs have been deep-scrubbed 
and I haven't seen a single slow request.

So thanks again for pointing that out!

Regards,
Eugen


Zitat von Andrej Filipcic <andrej.filipcic@xxxxxx>:

> Hi,
>
> all our slow request issues were solved with:
> [osd]
>   osd op queue = wpq
>   osd op queue cut off = high
>
> before we even had  several hours old request, since the change it 
> rarely gets above 30s even with the heaviest loads, eg >100 iops/hdd
>
> Regards,
> Andrej
>
> On 6/10/20 12:12 PM, Eugen Block wrote:
>> Hi,
>>
>> we have this message almost daily, although in our case it's almost 
>> expected. We run a nightly compile job within a cephfs subtree and 
>> the OSDs (HDD with rocksDB on SSD) are saturated during those jobs.
>> Also the deep-scrubs which also run during the night have a 
>> significant impact and the cluster reports slow requests, but since 
>> that happens outside our working hours we can live with it (for now).
>>
>> You write the OSDs are on SSDs, is that true for both data and 
>> metadata pool?
>>
>> Regards,
>> Eugen
>>
>>
>> Zitat von locallocal <locallocal@xxxxxxx>:
>>
>>> hi,guys.
>>> we have a ceph cluster which version is luminous 12.2.13. and 
>>> Recently we encountered a problem.here are some log infomations:
>>>
>>>
>>> 2020-06-08 12:33:52.706070 7f4097e2d700  0 log_channel(cluster) log 
>>> [WRN] : slow request 30.518930 seconds old, received at
>>> 2020-06-08 12:33:22.186924:  
>>> client_request(client.48978906:941633993 create 
>>> #0x100028cab8a/.filename 2020-06-08 12:33:22.197434 caller_uid=0,
>>> caller_gid=0{}) currently submit entry: journal_and_reply ...
>>> 2020-06-08 13:12:17.826727 7f4097e2d700  0 log_channel(cluster) log 
>>> [WRN] : slow request 2220.991833 seconds old, received at
>>> 2020-06-08 12:35:16.764233:  
>>> client_request(client.42390705:788369155 create 
>>> #0x1000224f999/.filename 2020-06-08 12:35:16.774553 caller_uid=0,
>>> caller_gid=0{}) currently submit entry: journal_and_reply
>>>
>>>
>>> it looks like mds can't flush journal to osd of meta pool.but the 
>>> osd type is ssd and the load is very low.this problem leads the 
>>> client can't mount and the mds can't trim log.
>>> Is there anyone have encountered this problem.Please help!
>>>
>>> _______________________________________________
>>> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an 

>>> email to ceph-users-leave@xxxxxxx
>>
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an 
>> email to ceph-users-leave@xxxxxxx
>
>
> --
> _____________________________________________________________
>    prof. dr. Andrej Filipcic,   E-mail: Andrej.Filipcic@xxxxxx
>    Department of Experimental High Energy Physics - F9
>    Jozef Stefan Institute, Jamova 39, P.o.Box 3000
>    SI-1001 Ljubljana, Slovenia
>    Tel.: +386-1-477-3674    Fax: +386-1-477-3166
> -------------------------------------------------------------
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an 
> email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an 
email to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux