Exactly, I mounted the volume in a no-brick node(nodeB),
and nodeA was the server. I have set different timeout, but
when I excute "ls /mnt/glusterfs(about 3 million small files,
in other words, about 3 million dentries)", the result was the
same, memory usage in nodeB didn't change at all while nodeA's
memory usage was changed about 4GB!
On 09/02/2016 05:42 AM, Keiviw
wrote:
Even
if I set the attribute-timeout and entry-timeout to
3600s(1h), in the nodeB, it didn't cache any metadata
because the memory usage didn't change. So I was confused
that why did the client not cache dentries and inodes.
If you only want to test fuse's caching, I would try
mounting the volume on a separate machine (not on the brick
node itself), disable all gluster performance xlators, do a
find.|xargs stat on the mount 2 times in succession and see
what free(1) reports the 1st and 2nd time. You could do this
experiment with various attr/entry timeout values. Make sure
your volume has a lot of small files.
-Ravi
在 2016-09-01 16:37:00,"Ravishankar N"
<ravishankar@xxxxxxxxxx>
写道:
On 09/01/2016 01:04 PM,
Keiviw wrote:
Hi,
I have found that GlusterFS client(mounted
by FUSE) didn't cache metadata like dentries and
inodes. I have installed GlusterFS 3.6.0 in nodeA
and nodeB, and the brick1 and brick2 was in nodeA,
then in nodeB, I mounted the volume to
/mnt/glusterfs by FUSE. From my test, I excuted
'ls /mnt/glusterfs' in nodeB, and found that the
memory didn't use at all. Here are my questions:
1. In fuse kernel, the author set some
attributes to control the time-out about dentry
and inode, in other words, the fuse kernel
supports metadata cache, but in my test, dentries
and inodes were not cached. WHY?
2. Were there some options in GlusterFS
mounted to local to enable the metadata cache in
fuse kernel?
You can tweak the attribute-timeout and
entry-timeout seconds while mounting the volume.
Default is 1 second for both. `man mount.glusterfs`
lists various mount options.
-Ravi
_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel