Re: cephfs kernel client instability

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Ilya,

thank you for the clarification. After setting the
"osd_map_messages_max" to 10 the io errors and the MDS error
"MDS_CLIENT_LATE_RELEASE" are gone.

The messages of  "mon session lost, hunting for new new mon" didn't go
away... can it be that this is related to
https://tracker.ceph.com/issues/23537

Best,
Martin

On Thu, Jan 24, 2019 at 10:16 PM Ilya Dryomov <idryomov@xxxxxxxxx> wrote:
>
> On Thu, Jan 24, 2019 at 6:21 PM Andras Pataki
> <apataki@xxxxxxxxxxxxxxxxxxxxx> wrote:
> >
> > Hi Ilya,
> >
> > Thanks for the clarification - very helpful.
> > I've lowered osd_map_messages_max to 10, and this resolves the issue
> > about the kernel being unhappy about large messages when the OSDMap
> > changes.  One comment here though: you mentioned that Luminous uses 40
> > as the default, which is indeed the case.  The documentation for
> > Luminous (and master), however, says that the default is 100.
>
> Looks like that page hasn't been kept up to date.  I'll fix that
> section.
>
> >
> > One other follow-up question on the kernel client about something I've
> > been seeing while testing.  Does the kernel client clean up when the MDS
> > asks due to cache pressure?  On a machine I ran something that touches a
> > lot of files, so the kernel client accumulated over 4 million caps.
> > Many hours after all the activity finished (i.e. many hours after
> > anything accesses ceph on that node) the kernel client still holds
> > millions of caps, and the MDS periodically complains about clients not
> > responding to cache pressure.  How is this supposed to be handled?
> > Obviously asking the kernel to drop caches via /proc/sys/vm/drop_caches
> > does a very thorough cleanup, but something in the middle would be better.
>
> The kernel client sitting on way too many caps for way too long is
> a long standing issue.  Adding Zheng who has recently been doing some
> work to facilitate cap releases and put a limit on the overall cap
> count.
>
> Thanks,
>
>                 Ilya
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux