glusterFS memory leak?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



what operations are you doing on mount point? how easy it is to reproduce
memory leak?

On Thu, Dec 24, 2009 at 5:08 AM, Larry Bates <larry.bates at vitalesafe.com>wrote:

>  I restarted the mirror and once again the glusterfs client process is
> just growing.  Echoing the 3 to /proc/sys/vm/drop_caches eems to shriink the
> memory footprint but only by a small amount.
>
> BTW - it is the client process that seems to grow infinitely.  Note: I have
> one machine that is acting as both a client and a server (which is where the
> client process is growing) and another machine that is a dedicated GFS
> server.
>
>
>
> Here are my configs:
>
>
>
> Client:
>
>
>
> #
>
> # Add client feature and attach to remote subvolumes of gfs001
>
> #
>
> volume gfs001brick1
>
>   type protocol/client
>
>   option transport-type tcp            # for TCP/IP transport
>
>   option remote-host gfs001            # IP address of the remote volume
>
>   option remote-subvolume vol1         # name of the remote volume
>
>   #option transport.socket.nodelay on   # undocumented option for speed
>
> end-volume
>
>
>
> volume gfs001brick2
>
>   type protocol/client
>
>   option transport-type tcp            # for TCP/IP transport
>
>   option remote-host gfs001            # IP address of the remote volume
>
>   option remote-subvolume vol2         # name of the remote volume
>
>   #option transport.socket.nodelay on   # undocumented option for speed
>
> end-volume
>
>
>
> volume coraidbrick13
>
>   type protocol/client
>
>   option transport-type tcp
>
>   option remote-host 10.0.0.71
>
>   option remote-subvolume coraidvol13
>
>   #option transport.socket.nodelay on
>
> end-volume
>
>
>
> volume coraidbrick14
>
>   type protocol/client
>
>   option transport-type tcp
>
>   option remote-host 10.0.0.71
>
>   option remote-subvolume coraidvol14
>
>   #option transport.socket.nodelay on
>
> end-volume
>
>
>
> #
>
> # Replicate volumes
>
> #
>
> volume afr-vol1
>
>   type cluster/replicate
>
>   subvolumes gfs001brick1 coraidbrick13
>
> end-volume
>
>
>
> volume afr-vol2
>
>   type cluster/replicate
>
>   subvolumes gfs001brick2 coraidbrick14
>
> end-volume
>
>
>
> #
>
> # Distribute files across bricks
>
> #
>
> volume dht-vol
>
>   type cluster/distribute
>
>   subvolumes afr-vol1 afr-vol2
>
>   option min-free-disk 2%               # 2% of 1.8Tb volumes is 36Gb
>
> end-volume
>
>
>
> #
>
> # Add quick-read for small files
>
> #
>
> volume quickread
>
>   type performance/quick-read
>
>   option cache-timeout 1         # default 1 second
>
>   option max-file-size 256KB     # default 64Kb
>
>   subvolumes dht-vol
>
> end-volume
>
>
>
> Server:
>
>
>
> volume coraidbrick13
>
>   type storage/posix                          # POSIX FS translator
>
>   option directory /mnt/glusterfs/e101.13     # Export this directory
>
>   option background-unlink yes                # unlink in background
>
> end-volume
>
>
>
> volume coraidbrick14
>
>   type storage/posix
>
>   option directory /mnt/glusterfs/e101.14
>
>   option background-unlink yes
>
> end-volume
>
>
>
> volume iot1
>
>   type performance/io-threads
>
>   option thread-count 4
>
>   subvolumes coraidbrick13
>
> end-volume
>
>
>
> volume iot2
>
>   type performance/io-threads
>
>   option thread-count 4
>
>   subvolumes coraidbrick14
>
> end-volume
>
>
>
> volume coraidvol13
>
>   type features/locks
>
>   subvolumes iot1
>
> end-volume
>
>
>
> volume coraidvol14
>
>   type features/locks
>
>   subvolumes iot2
>
> end-volume
>
>
>
> ## Add network serving capability to volumes
>
> volume server
>
>   type protocol/server
>
>   option transport-type tcp                    # For TCP/IP transport
>
>   subvolumes coraidvol13 coraidvol14           # Expose both volumes
>
>   option auth.addr.coraidvol13.allow 10.0.0.*
>
>   option auth.addr.coraidvol14.allow 10.0.0.*
>
> end-volume
>
>
>
> Thanks,
>
> Larry
>
>
>
>
>
>
>
> *From:* raghavendra.hg at gmail.com [mailto:raghavendra.hg at gmail.com] *On
> Behalf Of *Raghavendra G
> *Sent:* Wednesday, December 23, 2009 2:43 PM
> *To:* Larry Bates
> *Cc:* gluster-users
> *Subject:* Re: glusterFS memory leak?
>
>
>
> Hi Larry,
>
> What is the client and server configuration? Instead of killing glusterfs
> process, can you do "echo 3 > /proc/sys/vm/drop_caches" and check whether
> memory usage comes down?
>
> regards,
>
> On Wed, Dec 23, 2009 at 9:01 PM, Larry Bates <larry.bates at vitalesafe.com>
> wrote:
>
> I've successfully set up GlusterFS 3.0 with a single server.  I brought on
> a 2nd server and setup AFR and have been working through the mirroring
> process.  I started an "ls -alR" to trigger a complete mirror of all files
> between the two servers.  After running for about 16 hours I started getting
> kernel out of memory errors.  Looking at top I see:
>
>  PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND
>  2561 root      15   0 13.0g 8.9g  860 S    7 91.6  60:56.02 glusterfs
>
> Seems that the client has used up all my memory (RAM and SWAP).  Killing
> the process returned all the memory to the OS.
>
> BTW - I'm working on a 2.3Tb store that contains about 2.5 million files in
> 65K folders.
>
> Thoughts?
>
> Larry Bates
> vitalEsafe, Inc.
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
>
>
>
> --
> Raghavendra G
>



-- 
Raghavendra G


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux