Re: cephfs kernel client - page cache being invaildated.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



This is a general question for the ceph list.
Should Jesper be looking at these vm tunables?
vm.dirty_ratio
vm.dirty_centisecs

What effect do they have when using Cephfs?

On Sun, 14 Oct 2018 at 14:24, John Hearns <hearnsj@xxxxxxxxxxxxxx> wrote:
Hej Jesper.
Sorry I do not have a direct answer to your question.
When looking at memory usage, I often use this command:

watch cat /rpoc/meminfo






On Sun, 14 Oct 2018 at 13:22, <jesper@xxxxxxxx> wrote:
Hi

We have a dataset of ~300 GB on CephFS which as being used for computations
over and over agian .. being refreshed daily or similar.

When hosting it on NFS after refresh, they are transferred, but from
there - they would be sitting in the kernel page cache of the client
until they are refreshed serverside.

On CephFS it look "similar" but "different". Where the "steady state"
operation over NFS would give a client/server traffic of < 1MB/s ..
CephFS contantly pulls 50-100MB/s over the network.  This has
implications for the clients that end up spending unnessary time waiting
for IO in the execution.

This is in a setting where the CephFS client mem look like this:

$ free -h
              total        used        free      shared  buff/cache 
available
Mem:           377G         17G        340G        1.2G         19G       
354G
Swap:          8.8G        430M        8.4G


If I just repeatedly run (within a few minute) something that is using the
files, then
it is fully served out of client page cache (2GB'ish / s) ..  but it looks
like
it is being evicted way faster than in the NFS setting?

This is not scientific .. but the CMD is a cat /file/on/ceph > /dev/null -
type on a total of 24GB data in 300'ish files.

$ free -h; time CMD ; sleep 1800; free -h; time CMD ; free -h; sleep 3600;
time CMD ;

              total        used        free      shared  buff/cache 
available
Mem:           377G         16G        312G        1.2G         48G       
355G
Swap:          8.8G        430M        8.4G

real    0m8.997s
user    0m2.036s
sys     0m6.915s
              total        used        free      shared  buff/cache 
available
Mem:           377G         17G        277G        1.2G         82G       
354G
Swap:          8.8G        430M        8.4G

real    3m25.904s
user    0m2.794s
sys     0m9.028s
              total        used        free      shared  buff/cache 
available
Mem:           377G         17G        283G        1.2G         76G       
353G
Swap:          8.8G        430M        8.4G

real    6m18.358s
user    0m2.847s
sys     0m10.651s


Munin graphs of the system confirms that there has been zero memory
pressure over the period.

Is there things in the CephFS case that can cause the page-cache to be
invailated?
Could less agressive "read-ahead" play a role?

Other thoughts on what root cause on the different behaviour could be?

Clients are using 4.15 kernel.. Anyone aware of newer patches in this area
that could impact ?

Jesper

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux