[Gluster-users] GlusterD uses 50% of RAM

RASTELLI Alessandro alessandro.rastelli at skytv.it
Fri Feb 20 14:53:21 UTC 2015


HTTP URL works fine.
Now shall I restart the glusterd daemon?

-----Original Message-----
From: Niels de Vos [mailto:ndevos at redhat.com] 
Sent: venerdì 20 febbraio 2015 15:47
To: RASTELLI Alessandro
Cc: Atin Mukherjee; gluster-users at gluster.org
Subject: Re: [Gluster-users] GlusterD uses 50% of RAM

On Fri, Feb 20, 2015 at 01:50:38PM +0000, RASTELLI Alessandro wrote:
> I get this:
> 
> [root at gluster03-mi glusterfs]# git fetch 
> git://review.gluster.org/glusterfs refs/changes/28/9328/4 && git 
> checkout FETCH_HEAD
> fatal: Couldn't find remote ref refs/changes/28/9328/4
> 
> What's wrong with that?

I think anonymous git does not (always) work. You could try fetching over HTTP:

  $ git fetch http://review.gluster.org/glusterfs refs/changes/28/9328/4 && git checkout FETCH_HEAD

Niels

> 
> A.
> 
> -----Original Message-----
> From: Atin Mukherjee [mailto:amukherj at redhat.com]
> Sent: venerdì 20 febbraio 2015 12:54
> To: RASTELLI Alessandro
> Cc: gluster-users at gluster.org
> Subject: Re: [Gluster-users] GlusterD uses 50% of RAM
> 
> From the cmd log history I could see lots of volume status commands were triggered parallely. This is a known issue for 3.6 and it would cause a memory leak. http://review.gluster.org/#/c/9328/ should solve it.
> 
> ~Atin
> 
> On 02/20/2015 04:36 PM, RASTELLI Alessandro wrote:
> > 10MB log
> > sorry :)
> > 
> > -----Original Message-----
> > From: Atin Mukherjee [mailto:amukherj at redhat.com]
> > Sent: venerdì 20 febbraio 2015 10:49
> > To: RASTELLI Alessandro; gluster-users at gluster.org
> > Subject: Re: [Gluster-users] GlusterD uses 50% of RAM
> > 
> > Could you please share the cmd_history.log & glusterd log file to analyze this high memory usage.
> > 
> > ~Atin
> > 
> > On 02/20/2015 03:10 PM, RASTELLI Alessandro wrote:
> >> Hi,
> >> I've noticed that one of our 6 gluster 3.6.2 nodes has "glusterd" 
> >> process using 50% of RAM, on the other nodes usage is about 5% This can be a bug?
> >> Should I restart glusterd daemon?
> >> Thank you
> >> A
> >>
> >> From: Volnei Puttini [mailto:volnei at vcplinux.com.br]
> >> Sent: lunedì 9 febbraio 2015 18:06
> >> To: RASTELLI Alessandro; gluster-users at gluster.org
> >> Subject: Re: [Gluster-users] cannot access to CIFS export
> >>
> >> Hi Alessandro,
> >>
> >> My system:
> >>
> >> CentOS 7
> >>
> >> samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64
> >> samba-winbind-4.1.1-37.el7_0.x86_64
> >> samba-libs-4.1.1-37.el7_0.x86_64
> >> samba-common-4.1.1-37.el7_0.x86_64
> >> samba-winbind-modules-4.1.1-37.el7_0.x86_64
> >> samba-winbind-clients-4.1.1-37.el7_0.x86_64
> >> samba-4.1.1-37.el7_0.x86_64
> >> samba-client-4.1.1-37.el7_0.x86_64
> >>
> >> glusterfs 3.6.2 built on Jan 22 2015 12:59:57
> >>
> >> Try this, work fine for me:
> >>
> >> [GFSVOL]
> >>     browseable = No
> >>     comment = Gluster share of volume gfsvol
> >>     path = /
> >>     read only = No
> >>     guest ok = Yes
> >>     kernel share modes = No
> >>     posix locking = No
> >>     vfs objects = glusterfs
> >>     glusterfs:loglevel = 7
> >>     glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log
> >>     glusterfs:volume = vgtest
> >>     glusterfs:volfile_server = 192.168.2.21
> >>
> >> On 09-02-2015 14:45, RASTELLI Alessandro wrote:
> >> Hi,
> >> I've created and started a new replica volume "downloadstat" with CIFS export enabled on GlusterFS 3.6.2.
> >> I can see the following piece has been added automatically to smb.conf:
> >> [gluster-downloadstat]
> >> comment = For samba share of volume downloadstat vfs objects = 
> >> glusterfs glusterfs:volume = downloadstat glusterfs:logfile = 
> >> /var/log/samba/glusterfs-downloadstat.%M.log
> >> glusterfs:loglevel = 7
> >> path = /
> >> read only = no
> >> guest ok = yes
> >>
> >> I restarted smb service, without errors.
> >> When I try to access from Win7 client to "\\gluster01-mi\gluster-downloadstat<file:///\\gluster01-mi\gluster-downloadstat>" it asks me a login (which user do I need to put?) and then gives me error "The network path was not found"
> >> and on Gluster smb.log I see:
> >> [2015/02/09 17:21:13.111639,  0] smbd/vfs.c:173(vfs_init_custom)
> >>   error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL
> >> [2015/02/09 17:21:13.111709,  0] smbd/vfs.c:315(smbd_vfs_init)
> >>   smbd_vfs_init: vfs_init_custom failed for glusterfs
> >> [2015/02/09 17:21:13.111741,  0] smbd/service.c:902(make_connection_snum)
> >>   vfs_init failed for service gluster-downloadstat
> >>
> >> Can you explain how to fix?
> >> Thanks
> >>
> >> Alessandro
> >>
> >> From: 
> >> gluster-users-bounces at gluster.org<mailto:gluster-users-bounces at glus
> >> te r .org> [mailto:gluster-users-bounces at gluster.org] On Behalf Of 
> >> David F.
> >> Robinson
> >> Sent: domenica 8 febbraio 2015 18:19
> >> To: Gluster Devel;
> >> gluster-users at gluster.org<mailto:gluster-users at gluster.org>
> >> Subject: [Gluster-users] cannot delete non-empty directory
> >>
> >> I am seeing these messsages after I delete large amounts of data using gluster 3.6.2.
> >> cannot delete non-empty directory: 
> >> old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_f
> >> in
> >> a
> >> l
> >>
> >> >From the FUSE mount (as root), the directory shows up as empty:
> >>
> >> # pwd
> >> /backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Stor
> >> ag
> >> e
> >> /Jimmy_Old/src_vj1.5_final
> >>
> >> # ls -al
> >> total 5
> >> d--------- 2 root root    4106 Feb  6 13:55 .
> >> drwxrws--- 3  601 dmiller   72 Feb  6 13:55 ..
> >>
> >> However, when you look at the bricks, the files are still there (none on brick01bkp, all files are on brick02bkp).  All of the files are 0-length and have ------T permissions.
> >> Any suggestions on how to fix this and how to prevent it from happening?
> >>
> >> #  ls -al
> >> /data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/R
> >> av e nCFD/Storage/Jimmy_Old/src_vj1.5_final
> >> /data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final:
> >> total 4
> >> d---------+ 2 root root  10 Feb  6 13:55 .
> >> drwxrws---+ 3  601 raven 36 Feb  6 13:55 ..
> >>
> >> /data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final:
> >> total 8
> >> d---------+ 3 root root  4096 Dec 31  1969 .
> >> drwxrws---+ 3  601 raven   36 Feb  6 13:55 ..
> >> ---------T  5  601 raven    0 Nov 20 00:08 read_inset.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 readbc.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 readcn.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 readinp.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 readinp_v1_2.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 readinp_v1_3.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 rotatept.f.gz
> >> d---------+ 2 root root   118 Feb  6 13:54 save1
> >> ---------T  5  601 raven    0 Nov 20 00:08 sepvec.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 shadow.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 snksrc.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 source.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 step.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 stoprog.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 summer6.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 totforc.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 tritet.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 wallrsd.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 wheat.f.gz
> >> ---------T  5  601 raven    0 Nov 20 00:08 write_inset.f.gz
> >>
> >>
> >> This is using gluster 3.6.2 on a distributed gluster volume that resides on a single machine.  Both of the bricks are on one machine consisting of 2x RAID-6 arrays.
> >>
> >> df -h | grep brick
> >> /dev/mapper/vg01-lvol1                       88T   22T   66T  25% /data/brick01bkp
> >> /dev/mapper/vg02-lvol1                       88T   22T   66T  26% /data/brick02bkp
> >>
> >> # gluster volume info homegfs_bkp
> >> Volume Name: homegfs_bkp
> >> Type: Distribute
> >> Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294
> >> Status: Started
> >> Number of Bricks: 2
> >> Transport-type: tcp
> >> Bricks:
> >> Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp
> >> Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp
> >> Options Reconfigured:
> >> storage.owner-gid: 100
> >> performance.io-thread-count: 32
> >> server.allow-insecure: on
> >> network.ping-timeout: 10
> >> performance.cache-size: 128MB
> >> performance.write-behind-window-size: 128MB
> >> server.manage-gids: on
> >> changelog.rollover-time: 15
> >> changelog.fsync-interval: 3
> >>
> >>
> >>
> >> ===============================
> >> David F. Robinson, Ph.D.
> >> President - Corvid Technologies
> >> 704.799.6944 x101 [office]
> >> 704.252.1310 [cell]
> >> 704.799.7974 [fax]
> >> David.Robinson at corvidtec.com<mailto:David.Robinson at corvidtec.com>
> >> http://www.corvidtechnologies.com
> >>
> >>
> >>
> >>
> >>
> >>
> >> _______________________________________________
> >>
> >> Gluster-users mailing list
> >>
> >> Gluster-users at gluster.org<mailto:Gluster-users at gluster.org>
> >>
> >> http://www.gluster.org/mailman/listinfo/gluster-users
> >>
> >>
> >>
> >>
> >> _______________________________________________
> >> Gluster-users mailing list
> >> Gluster-users at gluster.org
> >> http://www.gluster.org/mailman/listinfo/gluster-users
> >>
> > 
> > --
> > ~Atin
> > 
> 
> --
> ~Atin
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users


More information about the Gluster-users mailing list