[Gluster-users] Memory leak in 3.6.*?

Mykola Ulianytskyi lystor at gmail.com
Fri Jul 22 15:47:49 UTC 2016


Hi

>  3.7 clients are not compatible with 3.6 servers

Can you provide more info?

I use some 3.7 clients with 3.6 servers and don't see issues.

Thank you

--
With best regards,
Mykola


On Fri, Jul 22, 2016 at 4:31 PM, Yannick Perret
<yannick.perret at liris.cnrs.fr> wrote:
> Note: I'm have a dev client machine so I can perform tests or recompile
> glusterfs client if it can helps getting data about that.
>
> I did not test this problem against 3.7.x version as my 2 servers are in use
> and I can't upgrade them at this time, and 3.7 clients are not compatible
> with 3.6 servers (as far as I can see from my tests).
>
> --
> Y.
>
>
> Le 22/07/2016 14:06, Yannick Perret a écrit :
>
> Hello,
> some times ago I posted about a memory leak in client process, but it was on
> a very old 32bit machine (both kernel and OS) and I don't found evidences
> about a similar problem on our recent machines.
> But I performed more tests and I have the same problem.
>
> Clients are 64bit Debian 8.2 machines. Glusterfs client on these machines is
> compiled from sources with activated stuff:
> FUSE client          : yes
> Infiniband verbs     : no
> epoll IO multiplex   : yes
> argp-standalone      : no
> fusermount           : yes
> readline             : yes
> georeplication       : yes
> Linux-AIO            : no
> Enable Debug         : no
> systemtap            : no
> Block Device xlator  : no
> glupy                : no
> Use syslog           : yes
> XML output           : yes
> QEMU Block formats   : no
> Encryption xlator    : yes
> Erasure Code xlator  : yes
>
> I tested both 3.6.7 and 3.6.9 version on client (3.6.7 is the one installed
> on our machines, even on servers, 3.6.9 is for testing with last 3.6
> version).
>
> Here are the operations on the client (also performed with similar results
> with 3.6.7 version):
> # /usr/local/sbin/glusterfs --version
> glusterfs 3.6.9 built on Jul 22 2016 13:27:42
> (…)
> # mount -t glusterfs sto1.my.domain:BACKUP-ADMIN-DATA  /zog/
> # cd /usr/
> # cp -Rp * /zog/TEMP/
> Then monitoring memory used by glusterfs process while 'cp' is running
> (resp. VSZ and RSS from 'ps'):
> 284740 70232
> 284740 70232
> 284876 71704
> 285000 72684
> 285136 74008
> 285416 75940
> (…)
> 368684 151980
> 369324 153768
> 369836 155576
> 370092 156192
> 370092 156192
> Here both sizes are stable and correspond to the end of 'cp' command.
> If I restart an other 'cp' (even on the same directories) size starts again
> to increase.
> If I perform a 'ls -lR' in the directory size also increase:
> 370756 192488
> 389964 212148
> 390948 213232
> (here I ^C the 'ls')
>
> When doing nothing the size don't increase but never decrease (calling
> 'sync' don't change the situation).
> Sending a HUP signal to glusterfs process also increases memory (390948
> 213324 → 456484 213320).
> Changing volume configuration (changing diagnostics.client-sys-log-level
> value) don't change anything.
>
> Here the actual ps:
> root     17041  4.9  5.2 456484 213320 ?       Ssl  13:29   1:21
> /usr/local/sbin/glusterfs --volfile-server=sto1.my.domain
> --volfile-id=BACKUP-ADMIN-DATA /zog
>
> Of course umouting/remounting fall back to "start" size:
> # umount /zog
> # mount -t glusterfs sto1.my.domain:BACKUP-ADMIN-DATA  /zog/
> → root     28741  0.3  0.7 273320 30484 ?        Ssl  13:57   0:00
> /usr/local/sbin/glusterfs --volfile-server=sto1.my.domain
> --volfile-id=BACKUP-ADMIN-DATA /zog
>
>
> I didn't saw this before because most of our volumes are mounted "on demand"
> for some storage activities or are permanently mounted but with very few
> activity.
> But clearly this memory usage driff is a long-term problem. On the old 32bit
> machine I had this problem ("solved" by using NFS mounts in order to wait
> for this old machine to be replaced) and it lead to glusterfs being killed
> by OS when out of free memory. It was faster than what I describe here but
> it's just a question of time.
>
>
> Thanks for any help about that.
>
> Regards,
> --
> Y.
>
>
> The corresponding volume on servers is (if it can help):
> Volume Name: BACKUP-ADMIN-DATA
> Type: Replicate
> Volume ID: 306d57f3-fb30-4bcc-8687-08bf0a3d7878
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: sto1.my.domain:/glusterfs/backup-admin/data
> Brick2: sto2.my.domain:/glusterfs/backup-admin/data
> Options Reconfigured:
> diagnostics.client-sys-log-level: WARNING
>
>
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users


More information about the Gluster-users mailing list