[Gluster-users] Memory leak und very slow speed

Benjamin Knoth bknoth at gwdg.de
Tue Oct 13 15:17:15 UTC 2020


Dear all,


I add the community repository, to update Gluster to 8.1.


This fix my memory leak. But in my logfile I got every second many errors


Oct 11 11:50:29 vm01 gluster[908]: [2020-10-11 09:50:29.642031] C
[mem-pool.c:873:mem_put]
(-->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(fd_close+0x6a)
[0x7f92d691960a]
-->/usr/lib/x86_64-linux-gnu/glusterfs/8.1/xlator/performance/open-behind.so(+0x748a)
[0x7f92d0b8f48a]
-->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(mem_put+0xf0)
[0x7f92d691c7f0] ) 0-mem-pool: invalid argument hdr->pool_list NULL [Das
Argument ist ungültig]


I found this fix.

https://github.com/gluster/glusterfs/issues/1473

|# gluster volume set <volname> open-behind off After disabling
open-behind no error messages in the log. Best regards Benjamin |



Am 09.10.20 um 08:28 schrieb Knoth, Benjamin:
>
> All 3 server have the same configuration with Debian Buster.
>
>
> I used the backports repository for GlusterFS, but I can also try to
> change the source to Gluster.org repositories and install the latest
> version at this repository.
>
>
> Best regards
>
> Benjamin
>
> ------------------------------------------------------------------------
> *Von:* Strahil Nikolov <hunter86_bg at yahoo.com>
> *Gesendet:* Donnerstag, 8. Oktober 2020 17:42:01
> *An:* Gluster Users; Knoth, Benjamin
> *Betreff:* Re: [Gluster-users] Memory leak und very slow speed
>  
> Do you have the option to update your cluster to 8.1 ?
>
> Are your clients in a HCI (server & client are the same system) ?
>
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
>
>
> В четвъртък, 8 октомври 2020 г., 17:07:31 Гринуич+3, Knoth, Benjamin
> <bknoth at gwdg.de> написа:
>
>
>
>
>
>   
>
>
> Dear community,
>
>
>
>
> actually, I'm running a 3 Node GlusterFS. Simple Wordpress pages needs
> 4 -10 seconds to load. Since a month we have also problems with memory
> leaks. All 3 nodes got 24 GB RAM (before 12 GB RAM) but GlusterFS use
> all the RAM. If all the RAM is used the virtual maschine loose there
> mountpoint. After remount everything starts again and that 2-3 times
> daily.
>
>
>
>
> # Gluster Version: 8.0
>
>
>
>
> #Affected process:  This is a snapshot from top where the process
> starts with low memory usage and run so long RAM is available.
>
>
>
>
>
>
>
>    PID USER      PR  NI    VIRT    RES    SHR S  %CPU  %MEM     TIME+
> COMMAND                                                                                                                                   
>  
> 869835 root      20   0   20,9g  20,3g   4340 S   2,3  86,5 152:10.62
> /usr/sbin/glusterfs --process-name fuse --volfile-server=vm01
> --volfile-server=vm02 --volfile-id=/gluster /var/www
>
>
>
>
>
>
>
> # gluster volume info
>
>
>
> Volume Name: gluster
> Type: Replicate
> Volume ID: c6d3beb1-b841-45e8-aa64-bb2be1e36e39
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1 x 3 = 3
> Transport-type: tcp
> Bricks:
> Brick1: vm01:/srv/glusterfs
> Brick2: vm02:/srv/glusterfs
> Brick3: vm03:/srv/glusterfs
> Options Reconfigured:
> performance.io-cache: on
> performance.write-behind: on
> performance.flush-behind: on
> auth.allow: 10.10.10.*
> performance.readdir-ahead: on
> performance.quick-read: off
> performance.cache-size: 1GB
> performance.cache-refresh-timeout: 10
> performance.read-ahead: off
> performance.write-behind-window-size: 4MB
> network.ping-timeout: 2
> performance.io-thread-count: 32
> performance.cache-max-file-size: 2MB
> performance.md-cache-timeout: 60
> features.cache-invalidation: on
> features.cache-invalidation-timeout: 600
> performance.stat-prefetch: on
> network.inode-lru-limit: 90000
>
>
>
>
>
>
> # Logs
>
> I can't find any critical messages on all gluster logs, but in syslog
> I found the oom-kill. After that, the mountpoint is history.
>
>
>
>
>
>
>  oom-kill:constraint=CONSTRAINT_NONE,nodemask=(null),cpuset=/,mems_allowed=0,global_oom,task_memcg=/system.slice/srv-web.mount,task=glusterfs,pid=961,uid=0
> [68263.478730] Out of memory: Killed process 961 (glusterfs)
> total-vm:21832212kB, anon-rss:21271576kB, file-rss:0kB, shmem-rss:0kB,
> UID:0 pgtables:41792kB oom_score_adj:0
> [68264.243608] oom_reaper: reaped process 961 (glusterfs), now
> anon-rss:0kB, file-rss:0kB, shmem-rss:0kB
>
>
> And after the remount it starts again to use more and more memory.
>
>
>
>
>
> Alternatively I can also activate SWAP but this slow down the load
> time extremely if GlusterFS starts to use SWAP after all RAM is used.
>
>
>
>
> If you need more information let me know it and i will send this too.
>
>
>
>
>
> Best regards
>
> Benjamin
>
>
>
>
>
>
>
>
>
>
>
>
>
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://bluejeans.com/441850968
>
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
>
> ________
>
>
>
> Community Meeting Calendar:
>
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://bluejeans.com/441850968
>
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users

-- 
Benjamin Knoth
Max Planck Digital Library (MPDL)
Systemadministration
Amalienstrasse 33
80799 Munich, Germany
http://www.mpdl.mpg.de

Mail: knoth at mpdl.mpg.de
Phone:  +49 89 909311 211
Fax:    +49-89-38602-280

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20201013/708b2d39/attachment.html>


More information about the Gluster-users mailing list