[Gluster-users] Memory leak with force merge ? gluster client takes 3gig of ram fast ?
Antoine Tremblay
hexa00 at gmail.com
Tue Nov 30 15:41:45 UTC 2010
Hi,
I'm running GlusterFS 3.0.4 built and I have this weird behavior where
each morning it seems the gluster client has a situation which makes it take
3Gigs of RAM (close to the 32bit limit , and I run on 32bits)
VmPeak: 3063696 kB
VmSize: 3062672 kB
VmLck: 0 kB
VmHWM: 3000196 kB
VmRSS: 3000196 kB
VmData: 3059952 kB
VmStk: 88 kB
VmExe: 28 kB
VmLib: 2508 kB
VmPTE: 5972 kB
Threads: 3
If I check the logs I see :
[2010-11-30 06:39:37] D [afr-dir-read.c:163:afr_examine_dir_readdir_cbk]
mirror-0: checksums of directory / differ, triggering forced merge
[2010-11-30 06:39:37] D
[afr-self-heal-entry.c:2298:afr_sh_entry_sync_prepare] mirror-0: no active
sources for / found. merging all entries as a conservative decision
[2010-11-30 06:39:37] D [stat-prefetch.c:3843:sp_release] statprefetch:
cache hits: 0, cache miss: 0
[2010-11-30 06:39:37] D [stat-prefetch.c:3843:sp_release] statprefetch:
cache hits: 0, cache miss: 0
[2010-11-30 06:39:37] D [stat-prefetch.c:3843:sp_release] statprefetch:
cache hits: 0, cache miss: 0
This last message is repeated a lot like > 1000 times like if it was in a
loop...
Any ideas ?
Here's my config :
iven volfile:
+------------------------------------------------------------------------------+
1: ## file auto generated by /usr/bin/glusterfs-volgen (mount.vol)
2: # Cmd line:
3: # $ /usr/bin/glusterfs-volgen --raid 1 --name data web01:/srv/data
web02:/srv/data
4:
5: # RAID 1
6: # TRANSPORT-TYPE tcp
7: volume web02-1
8: type protocol/client
9: option transport-type tcp
10: option remote-host web02
11: option transport.socket.nodelay on
12: option transport.remote-port 6996
13: option remote-subvolume brick1
14: end-volume
15:
16: volume web01-1
17: type protocol/client
18: option transport-type tcp
19: option remote-host web01
20: option transport.socket.nodelay on
21: option transport.remote-port 6996
22: option remote-subvolume brick1
23: end-volume
24:
25: volume mirror-0
26: type cluster/replicate
27: subvolumes web01-1 web02-1
28: end-volume
29:
30: volume readahead
31: type performance/read-ahead
32: option page-count 4
33: subvolumes mirror-0
34: end-volume
35:
36: volume iocache
37: type performance/io-cache
38: option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
's/[^0-9]//g') / 5120 ))`MB
39: option cache-timeout 1
40: subvolumes readahead
41: end-volume
42:
43: volume quickread
44: type performance/quick-read
45: option cache-timeout 1
46: option max-file-size 64kB
47: subvolumes iocache
48: end-volume
49:
50: volume writebehind
51: type performance/write-behind
52: option cache-size 4MB
53: subvolumes quickread
54: end-volume
55:
56: volume statprefetch
57: type performance/stat-prefetch
58: subvolumes writebehind
59: end-volume
60:
Thanks a lot
Antoine Tremblay
More information about the Gluster-users
mailing list