Thanks for sharing. Best Regards, Strahil Nikolov В вторник, 13 октомври 2020 г., 18:17:23 Гринуич+3, Benjamin Knoth <bknoth@xxxxxxx> написа: Dear all, I add the community repository, to update Gluster to 8.1. This fix my memory leak. But in my logfile I got every second many errors Oct 11 11:50:29 vm01 gluster[908]: [2020-10-11 09:50:29.642031] C [mem-pool.c:873:mem_put] (-->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(fd_close+0x6a) [0x7f92d691960a] -->/usr/lib/x86_64-linux-gnu/glusterfs/8.1/xlator/performance/open-behind.so(+0x748a) [0x7f92d0b8f48a] -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(mem_put+0xf0) [0x7f92d691c7f0] ) 0-mem-pool: invalid argument hdr->pool_list NULL [Das Argument ist ungültig] I found this fix. https://github.com/gluster/glusterfs/issues/1473 # gluster volume set <volname> open-behind off After disabling open-behind no error messages in the log. Best regards Benjamin Am 09.10.20 um 08:28 schrieb Knoth, Benjamin: > All 3 server have the same configuration with Debian Buster. I used the backports repository for GlusterFS, but I can also try to change the source to Gluster.org repositories and install the latest version at this repository. Best regards Benjamin ________________________________ Von: Strahil Nikolov <hunter86_bg@xxxxxxxxx> Gesendet: Donnerstag, 8. Oktober 2020 17:42:01 An: Gluster Users; Knoth, Benjamin Betreff: Re: Memory leak und very slow speed Do you have the option to update your cluster to 8.1 ? Are your clients in a HCI (server & client are the same system) ? Best Regards, Strahil Nikolov В четвъртък, 8 октомври 2020 г., 17:07:31 Гринуич+3, Knoth, Benjamin <bknoth@xxxxxxx> написа: Dear community, actually, I'm running a 3 Node GlusterFS. Simple Wordpress pages needs 4 -10 seconds to load. Since a month we have also problems with memory leaks. All 3 nodes got 24 GB RAM (before 12 GB RAM) but GlusterFS use all the RAM. If all the RAM is used the virtual maschine loose there mountpoint. After remount everything starts again and that 2-3 times daily. # Gluster Version: 8.0 #Affected process: This is a snapshot from top where the process starts with low memory usage and run so long RAM is available. PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 869835 root 20 0 20,9g 20,3g 4340 S 2,3 86,5 152:10.62 /usr/sbin/glusterfs --process-name fuse --volfile-server=vm01 --volfile-server=vm02 --volfile-id=/gluster /var/www # gluster volume info Volume Name: gluster Type: Replicate Volume ID: c6d3beb1-b841-45e8-aa64-bb2be1e36e39 Status: Started Snapshot Count: 0 Number of Bricks: 1 x 3 = 3 Transport-type: tcp Bricks: Brick1: vm01:/srv/glusterfs Brick2: vm02:/srv/glusterfs Brick3: vm03:/srv/glusterfs Options Reconfigured: performance.io-cache: on performance.write-behind: on performance.flush-behind: on auth.allow: 10.10.10.* performance.readdir-ahead: on performance.quick-read: off performance.cache-size: 1GB performance.cache-refresh-timeout: 10 performance.read-ahead: off performance.write-behind-window-size: 4MB network.ping-timeout: 2 performance.io-thread-count: 32 performance.cache-max-file-size: 2MB performance.md-cache-timeout: 60 features.cache-invalidation: on features.cache-invalidation-timeout: 600 performance.stat-prefetch: on network.inode-lru-limit: 90000 # Logs I can't find any critical messages on all gluster logs, but in syslog I found the oom-kill. After that, the mountpoint is history. oom-kill:constraint=CONSTRAINT_NONE,nodemask=(null),cpuset=/,mems_allowed=0,global_oom,task_memcg=/system.slice/srv-web.mount,task=glusterfs,pid=961,uid=0 [68263.478730] Out of memory: Killed process 961 (glusterfs) total-vm:21832212kB, anon-rss:21271576kB, file-rss:0kB, shmem-rss:0kB, UID:0 pgtables:41792kB oom_score_adj:0 [68264.243608] oom_reaper: reaped process 961 (glusterfs), now anon-rss:0kB, file-rss:0kB, shmem-rss:0kB And after the remount it starts again to use more and more memory. Alternatively I can also activate SWAP but this slow down the load time extremely if GlusterFS starts to use SWAP after all RAM is used. If you need more information let me know it and i will send this too. Best regards Benjamin ________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users ________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users -- Benjamin Knoth Max Planck Digital Library (MPDL) Systemadministration Amalienstrasse 33 80799 Munich, Germany http://www.mpdl.mpg.de Mail: knoth@xxxxxxxxxxx Phone: +49 89 909311 211 Fax: +49-89-38602-280 ________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users ________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users