[Gluster-users] glusterFS memory leak?
Raghavendra G
raghavendra at gluster.com
Thu Dec 24 01:47:59 UTC 2009
what operations are you doing on mount point? how easy it is to reproduce
memory leak?
On Thu, Dec 24, 2009 at 5:08 AM, Larry Bates <larry.bates at vitalesafe.com>wrote:
> I restarted the mirror and once again the glusterfs client process is
> just growing. Echoing the 3 to /proc/sys/vm/drop_caches eems to shriink the
> memory footprint but only by a small amount.
>
> BTW - it is the client process that seems to grow infinitely. Note: I have
> one machine that is acting as both a client and a server (which is where the
> client process is growing) and another machine that is a dedicated GFS
> server.
>
>
>
> Here are my configs:
>
>
>
> Client:
>
>
>
> #
>
> # Add client feature and attach to remote subvolumes of gfs001
>
> #
>
> volume gfs001brick1
>
> type protocol/client
>
> option transport-type tcp # for TCP/IP transport
>
> option remote-host gfs001 # IP address of the remote volume
>
> option remote-subvolume vol1 # name of the remote volume
>
> #option transport.socket.nodelay on # undocumented option for speed
>
> end-volume
>
>
>
> volume gfs001brick2
>
> type protocol/client
>
> option transport-type tcp # for TCP/IP transport
>
> option remote-host gfs001 # IP address of the remote volume
>
> option remote-subvolume vol2 # name of the remote volume
>
> #option transport.socket.nodelay on # undocumented option for speed
>
> end-volume
>
>
>
> volume coraidbrick13
>
> type protocol/client
>
> option transport-type tcp
>
> option remote-host 10.0.0.71
>
> option remote-subvolume coraidvol13
>
> #option transport.socket.nodelay on
>
> end-volume
>
>
>
> volume coraidbrick14
>
> type protocol/client
>
> option transport-type tcp
>
> option remote-host 10.0.0.71
>
> option remote-subvolume coraidvol14
>
> #option transport.socket.nodelay on
>
> end-volume
>
>
>
> #
>
> # Replicate volumes
>
> #
>
> volume afr-vol1
>
> type cluster/replicate
>
> subvolumes gfs001brick1 coraidbrick13
>
> end-volume
>
>
>
> volume afr-vol2
>
> type cluster/replicate
>
> subvolumes gfs001brick2 coraidbrick14
>
> end-volume
>
>
>
> #
>
> # Distribute files across bricks
>
> #
>
> volume dht-vol
>
> type cluster/distribute
>
> subvolumes afr-vol1 afr-vol2
>
> option min-free-disk 2% # 2% of 1.8Tb volumes is 36Gb
>
> end-volume
>
>
>
> #
>
> # Add quick-read for small files
>
> #
>
> volume quickread
>
> type performance/quick-read
>
> option cache-timeout 1 # default 1 second
>
> option max-file-size 256KB # default 64Kb
>
> subvolumes dht-vol
>
> end-volume
>
>
>
> Server:
>
>
>
> volume coraidbrick13
>
> type storage/posix # POSIX FS translator
>
> option directory /mnt/glusterfs/e101.13 # Export this directory
>
> option background-unlink yes # unlink in background
>
> end-volume
>
>
>
> volume coraidbrick14
>
> type storage/posix
>
> option directory /mnt/glusterfs/e101.14
>
> option background-unlink yes
>
> end-volume
>
>
>
> volume iot1
>
> type performance/io-threads
>
> option thread-count 4
>
> subvolumes coraidbrick13
>
> end-volume
>
>
>
> volume iot2
>
> type performance/io-threads
>
> option thread-count 4
>
> subvolumes coraidbrick14
>
> end-volume
>
>
>
> volume coraidvol13
>
> type features/locks
>
> subvolumes iot1
>
> end-volume
>
>
>
> volume coraidvol14
>
> type features/locks
>
> subvolumes iot2
>
> end-volume
>
>
>
> ## Add network serving capability to volumes
>
> volume server
>
> type protocol/server
>
> option transport-type tcp # For TCP/IP transport
>
> subvolumes coraidvol13 coraidvol14 # Expose both volumes
>
> option auth.addr.coraidvol13.allow 10.0.0.*
>
> option auth.addr.coraidvol14.allow 10.0.0.*
>
> end-volume
>
>
>
> Thanks,
>
> Larry
>
>
>
>
>
>
>
> *From:* raghavendra.hg at gmail.com [mailto:raghavendra.hg at gmail.com] *On
> Behalf Of *Raghavendra G
> *Sent:* Wednesday, December 23, 2009 2:43 PM
> *To:* Larry Bates
> *Cc:* gluster-users
> *Subject:* Re: [Gluster-users] glusterFS memory leak?
>
>
>
> Hi Larry,
>
> What is the client and server configuration? Instead of killing glusterfs
> process, can you do "echo 3 > /proc/sys/vm/drop_caches" and check whether
> memory usage comes down?
>
> regards,
>
> On Wed, Dec 23, 2009 at 9:01 PM, Larry Bates <larry.bates at vitalesafe.com>
> wrote:
>
> I've successfully set up GlusterFS 3.0 with a single server. I brought on
> a 2nd server and setup AFR and have been working through the mirroring
> process. I started an "ls -alR" to trigger a complete mirror of all files
> between the two servers. After running for about 16 hours I started getting
> kernel out of memory errors. Looking at top I see:
>
> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
> 2561 root 15 0 13.0g 8.9g 860 S 7 91.6 60:56.02 glusterfs
>
> Seems that the client has used up all my memory (RAM and SWAP). Killing
> the process returned all the memory to the OS.
>
> BTW - I'm working on a 2.3Tb store that contains about 2.5 million files in
> 65K folders.
>
> Thoughts?
>
> Larry Bates
> vitalEsafe, Inc.
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
>
>
>
> --
> Raghavendra G
>
--
Raghavendra G
More information about the Gluster-users
mailing list