[Gluster-users] Memory leak with force merge ? gluster client takes 3gig of ram fast ?
Antoine Tremblay
hexa00 at gmail.com
Wed Dec 1 16:11:42 UTC 2010
Hi,
It was a bit big to send to the mailling list you can download it at :
http://www.openrapids.net/~hexa/glusterdump.12610.gz
You can also find my complete log at :
http://www.openrapids.net/~hexa/test-gluster.log.gz
I tryed disabling statprefetch .. it got rid of the looping log but I
stilll got 3gig of memory used..
Also you should note that my setup is composed of many directories with
small files < 50k .. if it helps..
(About 8Gigs total)
Thank you very much!
Antoine
On Wed, Dec 1, 2010 at 2:06 AM, Raghavendra G <raghavendra at gluster.com>wrote:
> Hi Antoine,
>
> Can you take statedump of glusterfs when it is consuming huge amounts of
> memory? You can get the statedump of glusterfs using,
>
> # kill -SIGUSR1 <glusterfs-pid>
>
> and the statedump can be found as /tmp/glusterdump.<glusterfs-pid>
>
> regards,
> Raghavendra.
> ----- Original Message -----
> From: "Antoine Tremblay" <hexa00 at gmail.com>
> To: gluster-users at gluster.org
> Sent: Tuesday, November 30, 2010 7:41:45 PM
> Subject: [Gluster-users] Memory leak with force merge ? gluster client
> takes 3gig of ram fast ?
>
> Hi,
>
> I'm running GlusterFS 3.0.4 built and I have this weird behavior where
> each morning it seems the gluster client has a situation which makes it
> take
> 3Gigs of RAM (close to the 32bit limit , and I run on 32bits)
>
> VmPeak: 3063696 kB
> VmSize: 3062672 kB
> VmLck: 0 kB
> VmHWM: 3000196 kB
> VmRSS: 3000196 kB
> VmData: 3059952 kB
> VmStk: 88 kB
> VmExe: 28 kB
> VmLib: 2508 kB
> VmPTE: 5972 kB
> Threads: 3
>
>
> If I check the logs I see :
>
> [2010-11-30 06:39:37] D [afr-dir-read.c:163:afr_examine_dir_readdir_cbk]
> mirror-0: checksums of directory / differ, triggering forced merge
> [2010-11-30 06:39:37] D
> [afr-self-heal-entry.c:2298:afr_sh_entry_sync_prepare] mirror-0: no active
> sources for / found. merging all entries as a conservative decision
> [2010-11-30 06:39:37] D [stat-prefetch.c:3843:sp_release] statprefetch:
> cache hits: 0, cache miss: 0
> [2010-11-30 06:39:37] D [stat-prefetch.c:3843:sp_release] statprefetch:
> cache hits: 0, cache miss: 0
> [2010-11-30 06:39:37] D [stat-prefetch.c:3843:sp_release] statprefetch:
> cache hits: 0, cache miss: 0
>
> This last message is repeated a lot like > 1000 times like if it was in a
> loop...
>
> Any ideas ?
>
> Here's my config :
>
> iven volfile:
>
> +------------------------------------------------------------------------------+
> 1: ## file auto generated by /usr/bin/glusterfs-volgen (mount.vol)
> 2: # Cmd line:
> 3: # $ /usr/bin/glusterfs-volgen --raid 1 --name data web01:/srv/data
> web02:/srv/data
> 4:
> 5: # RAID 1
> 6: # TRANSPORT-TYPE tcp
> 7: volume web02-1
> 8: type protocol/client
> 9: option transport-type tcp
> 10: option remote-host web02
> 11: option transport.socket.nodelay on
> 12: option transport.remote-port 6996
> 13: option remote-subvolume brick1
> 14: end-volume
> 15:
> 16: volume web01-1
> 17: type protocol/client
> 18: option transport-type tcp
> 19: option remote-host web01
> 20: option transport.socket.nodelay on
> 21: option transport.remote-port 6996
> 22: option remote-subvolume brick1
> 23: end-volume
> 24:
> 25: volume mirror-0
> 26: type cluster/replicate
> 27: subvolumes web01-1 web02-1
> 28: end-volume
> 29:
> 30: volume readahead
> 31: type performance/read-ahead
> 32: option page-count 4
> 33: subvolumes mirror-0
> 34: end-volume
> 35:
> 36: volume iocache
> 37: type performance/io-cache
> 38: option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
> 's/[^0-9]//g') / 5120 ))`MB
> 39: option cache-timeout 1
> 40: subvolumes readahead
> 41: end-volume
> 42:
> 43: volume quickread
> 44: type performance/quick-read
> 45: option cache-timeout 1
> 46: option max-file-size 64kB
> 47: subvolumes iocache
> 48: end-volume
> 49:
> 50: volume writebehind
> 51: type performance/write-behind
> 52: option cache-size 4MB
> 53: subvolumes quickread
> 54: end-volume
> 55:
> 56: volume statprefetch
> 57: type performance/stat-prefetch
> 58: subvolumes writebehind
> 59: end-volume
> 60:
>
>
> Thanks a lot
>
>
> Antoine Tremblay
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
More information about the Gluster-users
mailing list