what operations are you doing on mount point? how easy it is to reproduce memory leak? On Thu, Dec 24, 2009 at 5:08 AM, Larry Bates <larry.bates at vitalesafe.com>wrote: > I restarted the mirror and once again the glusterfs client process is > just growing. Echoing the 3 to /proc/sys/vm/drop_caches eems to shriink the > memory footprint but only by a small amount. > > BTW - it is the client process that seems to grow infinitely. Note: I have > one machine that is acting as both a client and a server (which is where the > client process is growing) and another machine that is a dedicated GFS > server. > > > > Here are my configs: > > > > Client: > > > > # > > # Add client feature and attach to remote subvolumes of gfs001 > > # > > volume gfs001brick1 > > type protocol/client > > option transport-type tcp # for TCP/IP transport > > option remote-host gfs001 # IP address of the remote volume > > option remote-subvolume vol1 # name of the remote volume > > #option transport.socket.nodelay on # undocumented option for speed > > end-volume > > > > volume gfs001brick2 > > type protocol/client > > option transport-type tcp # for TCP/IP transport > > option remote-host gfs001 # IP address of the remote volume > > option remote-subvolume vol2 # name of the remote volume > > #option transport.socket.nodelay on # undocumented option for speed > > end-volume > > > > volume coraidbrick13 > > type protocol/client > > option transport-type tcp > > option remote-host 10.0.0.71 > > option remote-subvolume coraidvol13 > > #option transport.socket.nodelay on > > end-volume > > > > volume coraidbrick14 > > type protocol/client > > option transport-type tcp > > option remote-host 10.0.0.71 > > option remote-subvolume coraidvol14 > > #option transport.socket.nodelay on > > end-volume > > > > # > > # Replicate volumes > > # > > volume afr-vol1 > > type cluster/replicate > > subvolumes gfs001brick1 coraidbrick13 > > end-volume > > > > volume afr-vol2 > > type cluster/replicate > > subvolumes gfs001brick2 coraidbrick14 > > end-volume > > > > # > > # Distribute files across bricks > > # > > volume dht-vol > > type cluster/distribute > > subvolumes afr-vol1 afr-vol2 > > option min-free-disk 2% # 2% of 1.8Tb volumes is 36Gb > > end-volume > > > > # > > # Add quick-read for small files > > # > > volume quickread > > type performance/quick-read > > option cache-timeout 1 # default 1 second > > option max-file-size 256KB # default 64Kb > > subvolumes dht-vol > > end-volume > > > > Server: > > > > volume coraidbrick13 > > type storage/posix # POSIX FS translator > > option directory /mnt/glusterfs/e101.13 # Export this directory > > option background-unlink yes # unlink in background > > end-volume > > > > volume coraidbrick14 > > type storage/posix > > option directory /mnt/glusterfs/e101.14 > > option background-unlink yes > > end-volume > > > > volume iot1 > > type performance/io-threads > > option thread-count 4 > > subvolumes coraidbrick13 > > end-volume > > > > volume iot2 > > type performance/io-threads > > option thread-count 4 > > subvolumes coraidbrick14 > > end-volume > > > > volume coraidvol13 > > type features/locks > > subvolumes iot1 > > end-volume > > > > volume coraidvol14 > > type features/locks > > subvolumes iot2 > > end-volume > > > > ## Add network serving capability to volumes > > volume server > > type protocol/server > > option transport-type tcp # For TCP/IP transport > > subvolumes coraidvol13 coraidvol14 # Expose both volumes > > option auth.addr.coraidvol13.allow 10.0.0.* > > option auth.addr.coraidvol14.allow 10.0.0.* > > end-volume > > > > Thanks, > > Larry > > > > > > > > *From:* raghavendra.hg at gmail.com [mailto:raghavendra.hg at gmail.com] *On > Behalf Of *Raghavendra G > *Sent:* Wednesday, December 23, 2009 2:43 PM > *To:* Larry Bates > *Cc:* gluster-users > *Subject:* Re: glusterFS memory leak? > > > > Hi Larry, > > What is the client and server configuration? Instead of killing glusterfs > process, can you do "echo 3 > /proc/sys/vm/drop_caches" and check whether > memory usage comes down? > > regards, > > On Wed, Dec 23, 2009 at 9:01 PM, Larry Bates <larry.bates at vitalesafe.com> > wrote: > > I've successfully set up GlusterFS 3.0 with a single server. I brought on > a 2nd server and setup AFR and have been working through the mirroring > process. I started an "ls -alR" to trigger a complete mirror of all files > between the two servers. After running for about 16 hours I started getting > kernel out of memory errors. Looking at top I see: > > PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND > 2561 root 15 0 13.0g 8.9g 860 S 7 91.6 60:56.02 glusterfs > > Seems that the client has used up all my memory (RAM and SWAP). Killing > the process returned all the memory to the OS. > > BTW - I'm working on a 2.3Tb store that contains about 2.5 million files in > 65K folders. > > Thoughts? > > Larry Bates > vitalEsafe, Inc. > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > > > > > -- > Raghavendra G > -- Raghavendra G