[Gluster-devel] [Gluster-users] glusterfs-3.5.1beta released

Franco Broi franco.broi at iongeo.com
Thu May 29 05:21:48 UTC 2014


Also forgot to mention that it's impossible to change any volume
settings while 3.4 clients are attached but I can unmount them all,
change the setting and them mount them all again.

gluster vol set data2 nfs.disable off
volume set: failed: Staging failed on nas5-10g. Error: One or more
connected clients cannot support the feature being set. These clients
need to be upgraded or disconnected before running this command again


On Thu, 2014-05-29 at 10:34 +0800, Franco Broi wrote: 
> 
> OK. I've just unmounted the data2 volume from a machine called tape1,
> and now try to remount - it's hanging.
> 
> /bin/sh /sbin/mount.glusterfs nas5-10g:/data2 /data2 -o rw,log-level=info,log-file=/var/log/glusterfs/glusterfs_data2.log
> 
> on the server 
> # gluster vol status
> Status of volume: data2
> Gluster process						Port	Online	Pid
> ------------------------------------------------------------------------------
> Brick nas5-10g:/data17/gvol				49152	Y	6553
> Brick nas5-10g:/data18/gvol				49153	Y	6564
> Brick nas5-10g:/data19/gvol				49154	Y	6575
> Brick nas5-10g:/data20/gvol				49155	Y	6586
> Brick nas6-10g:/data21/gvol				49160	Y	20608
> Brick nas6-10g:/data22/gvol				49161	Y	20613
> Brick nas6-10g:/data23/gvol				49162	Y	20614
> Brick nas6-10g:/data24/gvol				49163	Y	20621
>  
> Task Status of Volume data2
> ------------------------------------------------------------------------------
> There are no active volume tasks
> 
> Sending the sigusr1 killed the mount processes and I don't see any state
> dumps. What path should they be in? I'm running Gluster installed via
> rpm and I don't see a /var/run/gluster directory.
> 
> On Wed, 2014-05-28 at 22:09 -0400, Krishnan Parthasarathi wrote: 
> > Franco,
> > 
> > When your clients perceive a hang, could you check the status of the bricks by running,
> > # gluster volume status VOLNAME  (run this on one of the 'server' machines in the cluster.)
> > 
> > Could you also provide the statedump of the client(s),
> > by issuing the following command.
> > 
> > # kill -SIGUSR1 pid-of-mount-process (run this on the 'client' machine.)
> > 
> > This would dump the state information of the client, like the file operations in progress,
> > memory consumed etc, onto a file under $INSTALL_PREFIX/var/run/gluster. Please attach this
> > file in your response.
> > 
> > thanks,
> > Krish
> > 
> > ----- Original Message -----
> > > Hi
> > > 
> > > My clients are running 3.4.1, when I try to mount from lots of machine
> > > simultaneously, some of the mounts hang. Stopping and starting the
> > > volume clears the hung mounts.
> > > 
> > > Errors in the client logs
> > > 
> > > [2014-05-28 01:47:15.930866] E
> > > [client-handshake.c:1741:client_query_portmap_cbk] 0-data2-client-3: failed
> > > to get the port number for remote subvolume. Please run 'gluster volume
> > > status' on server to see if brick process is running.
> > > 
> > > Let me know if you want more information.
> > > 
> > > Cheers,
> > > 
> > > On Sun, 2014-05-25 at 11:55 +0200, Niels de Vos wrote:
> > > > On Sat, 24 May, 2014 at 11:34:36PM -0700, Gluster Build System wrote:
> > > > > 
> > > > > SRC:
> > > > > http://bits.gluster.org/pub/gluster/glusterfs/src/glusterfs-3.5.1beta.tar.gz
> > > > 
> > > > This beta release is intended to verify the changes that should resolve
> > > > the bugs listed below. We appreciate tests done by anyone. Please leave
> > > > a comment in the respective bugreport with a short description of the
> > > > success or failure. Visiting one of the bugreports is as easy as opening
> > > > the bugzilla.redhat.com/$BUG URL, for the first in the list, this
> > > > results in http://bugzilla.redhat.com/765202 .
> > > > 
> > > > Bugs expected to be fixed (31 in total since 3.5.0):
> > > > 
> > > >  #765202 - lgetxattr called with invalid keys on the bricks
> > > >  #833586 - inodelk hang from marker_rename_release_newp_lock
> > > >  #859581 - self-heal process can sometimes create directories instead of
> > > >  symlinks for the root gfid file in .glusterfs
> > > >  #986429 - Backupvolfile server option should work internal to GlusterFS
> > > >  framework
> > > > #1039544 - [FEAT] "gluster volume heal info" should list the entries that
> > > > actually required to be healed.
> > > > #1046624 - Unable to heal symbolic Links
> > > > #1046853 - AFR : For every file self-heal there are warning messages
> > > > reported in glustershd.log file
> > > > #1063190 - [RHEV-RHS] Volume was not accessible after server side quorum
> > > > was met
> > > > #1064096 - The old Python Translator code (not Glupy) should be removed
> > > > #1066996 - Using sanlock on a gluster mount with replica 3 (quorum-type
> > > > auto) leads to a split-brain
> > > > #1071191 - [3.5.1] Sporadic SIGBUS with mmap() on a sparse file created
> > > > with open(), seek(), write()
> > > > #1078061 - Need ability to heal mismatching user extended attributes
> > > > without any changelogs
> > > > #1078365 - New xlators are linked as versioned .so files, creating
> > > > <xlator>.so.0.0.0
> > > > #1086748 - Add documentation for the Feature: AFR CLI enhancements
> > > > #1086750 - Add documentation for the Feature: File Snapshots in GlusterFS
> > > > #1086752 - Add documentation for the Feature: On-Wire
> > > > Compression/Decompression
> > > > #1086756 - Add documentation for the Feature: zerofill API for GlusterFS
> > > > #1086758 - Add documentation for the Feature: Changelog based parallel
> > > > geo-replication
> > > > #1086760 - Add documentation for the Feature: Write Once Read Many (WORM)
> > > > volume
> > > > #1086762 - Add documentation for the Feature: BD Xlator - Block Device
> > > > translator
> > > > #1088848 - Spelling errors in rpc/rpc-transport/rdma/src/rdma.c
> > > > #1089054 - gf-error-codes.h is missing from source tarball
> > > > #1089470 - SMB: Crash on brick process during compile kernel.
> > > > #1089934 - list dir with more than N files results in Input/output error
> > > > #1091340 - Doc: Add glfs_fini known issue to release notes 3.5
> > > > #1091392 - glusterfs.spec.in: minor/nit changes to sync with Fedora spec
> > > > #1095775 - Add support in libgfapi to fetch volume info from glusterd.
> > > > #1095971 - Stopping/Starting a Gluster volume resets ownership
> > > > #1096040 - AFR : self-heal-daemon not clearing the change-logs of all the
> > > > sources after self-heal
> > > > #1096425 - i/o error when one user tries to access RHS volume over NFS with
> > > > 100+ GIDs
> > > > #1099878 - Need support for handle based Ops to fetch/modify extended
> > > > attributes of a file
> > > > 
> > > > 
> > > > Before a final glusterfs-3.5.1 release is made, we hope to have all the
> > > > blocker bugs fixed. There are currently 13 bugs marked that still need
> > > > some work done:
> > > > 
> > > > #1081016 - glusterd needs xfsprogs and e2fsprogs packages
> > > > #1086743 - Add documentation for the Feature: RDMA-connection manager
> > > > (RDMA-CM)
> > > > #1086749 - Add documentation for the Feature: Exposing Volume Capabilities
> > > > #1086751 - Add documentation for the Feature: gfid-access
> > > > #1086754 - Add documentation for the Feature: Quota Scalability
> > > > #1086755 - Add documentation for the Feature: readdir-ahead
> > > > #1086759 - Add documentation for the Feature: Improved block device
> > > > translator
> > > > #1086766 - Add documentation for the Feature: Libgfapi
> > > > #1086774 - Add documentation for the Feature: Access Control List - Version
> > > > 3 support for Gluster NFS
> > > > #1086781 - Add documentation for the Feature: Eager locking
> > > > #1086782 - Add documentation for the Feature: glusterfs and  oVirt
> > > > integration
> > > > #1086783 - Add documentation for the Feature: qemu 1.3 - libgfapi
> > > > integration
> > > > #1095595 - Stick to IANA standard while allocating brick ports
> > > > 
> > > > A more detailed overview of the status of each of these bugs is here:
> > > > - https://bugzilla.redhat.com/showdependencytree.cgi?id=glusterfs-3.5.1
> > > > 
> > > > Cheers,
> > > > Niels
> > > > _______________________________________________
> > > > Gluster-users mailing list
> > > > Gluster-users at gluster.org
> > > > http://supercolony.gluster.org/mailman/listinfo/gluster-users
> > > 
> > > 
> > > _______________________________________________
> > > Gluster-devel mailing list
> > > Gluster-devel at gluster.org
> > > http://supercolony.gluster.org/mailman/listinfo/gluster-devel
> > > 
> 
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users




More information about the Gluster-devel mailing list