There's another inode leak caused by an incorrect counting of
lookups on directory reads.
Here's a patch that solves the problem for
3.7:
http://review.gluster.org/13324
Hopefully with this patch the
memory leaks should disapear.
Xavi
On 29.01.2016 19:09, Oleksandr
Natalenko wrote:
Here is intermediate summary of current memory
leaks in FUSE client
investigation.
I use GlusterFS v3.7.6
release with the following patches:
===
Kaleb S KEITHLEY (1):
fuse: use-after-free fix in fuse-bridge, revisited
Pranith Kumar K
(1):
mount/fuse: Fix use-after-free crash
Soumya Koduri (3):
gfapi: Fix inode nlookup counts
inode: Retire the inodes from the lru
list in inode_table_destroy
upcall: free the xdr* allocations
===
With those patches we got API leaks fixed (I hope, brief tests show
that) and
got rid of "kernel notifier loop terminated" message.
Nevertheless, FUSE
client still leaks.
I have several test
volumes with several million of small files (100K…2M in
average). I
do 2 types of FUSE client testing:
1) find /mnt/volume -type d
2)
rsync -av -H /mnt/source_volume/* /mnt/target_volume/
And most
up-to-date results are shown below:
=== find /mnt/volume -type d
===
Memory consumption: ~4G
Statedump:
https://gist.github.com/10cde83c63f1b4f1dd7a
Valgrind:
https://gist.github.com/097afb01ebb2c5e9e78d
I guess,
fuse-bridge/fuse-resolve. related.
=== rsync -av -H
/mnt/source_volume/* /mnt/target_volume/ ===
Memory consumption:
~3.3...4G
Statedump (target volume):
https://gist.github.com/31e43110eaa4da663435
Valgrind (target volume):
https://gist.github.com/f8e0151a6878cacc9b1a
I guess,
DHT-related.
Give me more patches to test :).
_______________________________________________
Gluster-devel mailing
list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel