Re: MDS behind on trimming every 4-5 weeks causing issue for ceph filesystem

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



@Kotresh Hiremath Ravishankar <khiremat@xxxxxxxxxx>

Can you please help on above



On Fri, 17 May, 2024, 12:26 pm Akash Warkhade, <a.warkhade98@xxxxxxxxx>
wrote:

> Hi Kotresh,
>
>
> Thanks for the reply.
> 1)There are no customer configs defined
> 2) not enabled subtree pinning
> 3) there were no warning related to rados
>
> So wanted to know In order to fix this should we increase default
> mds_cache_memory_limit from 4Gb to 6Gb or more?
> Or is there any other solution for this issue?
>
> On Fri, 17 May, 2024, 12:11 pm Kotresh Hiremath Ravishankar, <
> khiremat@xxxxxxxxxx> wrote:
>
>> Hi,
>>
>> ~6K log segments to be trimmed, that's huge.
>>
>> 1. Are there any custom configs configured on this setup ?
>> 2. Is subtree pinning enabled ?
>> 3. Are there any warnings w.r.t rados slowness ?
>> 4. Please share the mds perf dump to check for latencies and other stuff.
>>    $ceph tell mds.<id> perf dump
>>
>> Thanks and Regards,
>> Kotresh H R
>>
>> On Fri, May 17, 2024 at 11:01 AM Akash Warkhade <a.warkhade98@xxxxxxxxx>
>> wrote:
>>
>>> Hi,
>>>
>>> We are using rook-ceph with operator 1.10.8 and ceph 17.2.5.
>>> we are using ceph filesystem with 4 mds i.e 2 active & 2 standby MDS
>>> every 3-4 weeks filesystem is having issue i.e in ceph status we can see
>>> below warnings warnings :
>>>
>>> 2 MDS reports slow requests
>>> 2 MDS Behind on Trimming
>>> mds.myfs-a(mds.1) : behind on trimming (6378/128) max_segments:128,
>>> num_segments: 6378
>>> mds.myfs-c(mds.1):  behind on trimming (6560/128) max_segments:128,
>>> num_segments: 6560
>>>
>>> to fix it, we have to restart all MDS pods one by one.
>>> this is happening every 4-5 weeks.
>>>
>>> We have seen many ceph issues related to it on ceph tracker and many
>>> people
>>> are suggesting to increase mds_cache_memory_limit
>>> currently for our cluster *mds_cache_memory_limit* is set to default 4GB
>>> *mds_log_max_segments* is set to default 128
>>> Should we increase *mds_cache_memory_limit* to 8GB from default 4GB or is
>>> there any solution to fix this issue permanently?
>>> _______________________________________________
>>> ceph-users mailing list -- ceph-users@xxxxxxx
>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>>
>>>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux