[Gluster-devel] Memory leak
Brent A Nelson
brent at phys.ufl.edu
Wed Mar 7 00:02:14 UTC 2007
I've narrowed the observed memory leak down to the read-ahead translator.
I can apply stat-prefetch and write-behind without triggering the leak in
my simple test, but read-ahead will cause memory consumption in the
glusterfs process to slowly increase for a little while and then suddenly
start increasing very rapidly.
Thanks,
Brent
On Tue, 6 Mar 2007, Brent A Nelson wrote:
> I can reproduce the memory leak in the glusterfs process even with just two
> disks from two nodes unified (it doesn't just occur with mirroring or
> striping), at least when all performance translators are used except for
> io-threads (io-threads causes my dd writes to die right away).
>
> I have 2 nodes, with glusterfs unifying one disk from each node. Each node
> is also a client. I do a dd on each node, simultaneously, with no problem:
> node1: dd if=/dev/zero of=/phys/blah0 bs=10M count=1024
> node2: dd if=/dev/zero of=/phys/blah1 bs=10M count=1024
>
> When doing a read on each node simultaneously, however, things go along for
> awhile, but then glusterfs starts consuming more and more memory until it
> presumably runs out and ultimately dies or becomes useless.
>
> Can anyone else confirm? And has anyone gotten io-threads to work at all?
>
> These systems are running Ubuntu Edgy, with just the generic kernel and Fuse
> 2.6.3 applied.
>
> Thanks,
>
> Brent
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at nongnu.org
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>
More information about the Gluster-devel
mailing list