[Gluster-users] glusterfs-3.7.6 released

Raghavendra Talur rtalur at redhat.com
Mon Nov 16 10:54:54 UTC 2015

Hi all,

I'm pleased to announce the release of GlusterFS-3.7.6.
The list of fixed bugs is included below.

Tarball and RPMs can be downloaded from

Ubuntu debs are available from

Debian Unstable (sid) packages have been updated and should be
available from default repos.

NetBSD has updated ports at

Bugs Fixed:

   - 1057295 <https://bugzilla.redhat.com/1057295>: glusterfs doesn't
   include firewalld rules
   - 1219399 <https://bugzilla.redhat.com/1219399>: NFS interoperability
   problem: Gluster Striped-Replicated can't read on vmware esxi 5.x NFS client
   - 1221957 <https://bugzilla.redhat.com/1221957>: Fully support
   data-tiering in 3.7.x, remove out of 'experimental' status
   - 1258197 <https://bugzilla.redhat.com/1258197>: gNFSd: NFS mount fails
   with "Remote I/O error"
   - 1258242 <https://bugzilla.redhat.com/1258242>: Data Tiering:
   detach-tier start force command not available on a tier volume(unlike which
   is possible in force remove-brick)
   - 1258833 <https://bugzilla.redhat.com/1258833>: Data Tiering: Disallow
   attach tier on a volume where any rebalance process is in progress to avoid
   deadlock(like remove brick commit pending etc)
   - 1259167 <https://bugzilla.redhat.com/1259167>: GF_LOG_NONE logs always
   - 1261146 <https://bugzilla.redhat.com/1261146>: Legacy files
   pre-existing tier attach must be promoted
   - 1261732 <https://bugzilla.redhat.com/1261732>: Disperse volume: df -h
   on a nfs mount throws Invalid argument error
   - 1261744 <https://bugzilla.redhat.com/1261744>: Tier/shd: Tracker bug
   for tier and shd compatibility
   - 1261758 <https://bugzilla.redhat.com/1261758>: Tiering/glusted: volume
   status failed after detach tier start
   - 1262860 <https://bugzilla.redhat.com/1262860>: Data Tiering: Tiering
   deamon is seeing each part of a file in a Disperse cold volume as a
   different file
   - 1265623 <https://bugzilla.redhat.com/1265623>: Data Tiering:Promotions
   and demotions fail after quota hard limits are hit for a tier volume
   - 1266836 <https://bugzilla.redhat.com/1266836>: AFR : fuse,nfs mount
   hangs when directories with same names are created and deleted continuously
   - 1266880 <https://bugzilla.redhat.com/1266880>: Tiering: unlink failed
   with error "Invaid argument"
   - 1267816 <https://bugzilla.redhat.com/1267816>: quota/marker: marker
   code cleanup
   - 1269035 <https://bugzilla.redhat.com/1269035>: Data Tiering:Throw a
   warning when user issues a detach-tier commit command
   - 1269125 <https://bugzilla.redhat.com/1269125>: Data
   Tiering:Regression: automation blocker:vol status for tier volumes using
   xml format is not working
   - 1269344 <https://bugzilla.redhat.com/1269344>: tier/cli: number of
   bricks remains the same in v info --xml
   - 1269501 <https://bugzilla.redhat.com/1269501>: Self-heal daemon
   crashes when bricks godown at the time of data heal
   - 1269530 <https://bugzilla.redhat.com/1269530>:
   Core:Blocker:Segmentation fault when using fallocate command on a gluster
   - 1269730 <https://bugzilla.redhat.com/1269730>: Sharding - Send inode
   forgets on *all* shards if/when the protocol layer (FUSE/Gfapi) at the
   top sends a forget on the actual file
   - 1270123 <https://bugzilla.redhat.com/1270123>: Data Tiering: Database
   locks observed on tiered volumes on continous writes to a file
   - 1270527 <https://bugzilla.redhat.com/1270527>: add policy mechanism
   for promotion and demotion
   - 1270769 <https://bugzilla.redhat.com/1270769>: quota/marker: dir count
   in inode quota is not atomic
   - 1271204 <https://bugzilla.redhat.com/1271204>: Introduce priv dump in
   shard xlator for better debugging
   - 1271249 <https://bugzilla.redhat.com/1271249>: tiering:compiler
   warning with gcc v5.1.1
   - 1271490 <https://bugzilla.redhat.com/1271490>: rm -rf on
   /run/gluster/vol// is not showing quota output header for other quota limit
   applied directories
   - 1271540 <https://bugzilla.redhat.com/1271540>: RHEL7/systemd : can't
   have server in debug mode anymore
   - 1271627 <https://bugzilla.redhat.com/1271627>: Creating a already
   deleted snapshot-clone deletes the corresponding snapshot.
   - 1271967 <https://bugzilla.redhat.com/1271967>: ECVOL: glustershd log
   grows quickly and fills up the root volume
   - 1272036 <https://bugzilla.redhat.com/1272036>: Data Tiering:getting
   failed to fsync on germany-hot-dht (Structure needs cleaning) warning
   - 1272331 <https://bugzilla.redhat.com/1272331>: Tier: Do not
   promote/demote files on which POSIX locks are held
   - 1272334 <https://bugzilla.redhat.com/1272334>: Data Tiering:Promotions
   fail when brick of EC (disperse) cold layer are down
   - 1272398 <https://bugzilla.redhat.com/1272398>: Data Tiering:Lot of
   Promotions/Demotions failed error messages
   - 1273246 <https://bugzilla.redhat.com/1273246>: Tier xattr name is
   misleading (trusted.tier-gfid)
   - 1273334 <https://bugzilla.redhat.com/1273334>: Fix in afr transaction
   - 1274101 <https://bugzilla.redhat.com/1274101>: need a way to
   pause/stop tiering to take snapshot
   - 1274600 <https://bugzilla.redhat.com/1274600>: [sharding+geo-rep]: On
   existing slave mount, reading files fails to show sharded file content
   - 1275157 <https://bugzilla.redhat.com/1275157>: Reduce 'CTR disabled'
   brick log message from ERROR to INFO/DEBUG
   - 1275483 <https://bugzilla.redhat.com/1275483>: Data Tiering:heat
   counters not getting reset and also internal ops seem to be heating the
   - 1275502 <https://bugzilla.redhat.com/1275502>: [Tier]: Typo in the
   output while setting the wrong value of low/hi watermark
   - 1275921 <https://bugzilla.redhat.com/1275921>: Disk usage mismatching
   after self-heal
   - 1276029 <https://bugzilla.redhat.com/1276029>: Upgrading a subset of
   cluster to 3.7.5 leads to issues with glusterd commands
   - 1276060 <https://bugzilla.redhat.com/1276060>: dist-geo-rep: geo-rep
   status shows Active/Passive even when all the gsync processes in a node are
   - 1276208 <https://bugzilla.redhat.com/1276208>: [RFE] 'gluster volume
   help' output could be sorted alphabetically
   - 1276244 <https://bugzilla.redhat.com/1276244>: gluster-nfs : Server
   crashed due to an invalid reference
   - 1276550 <https://bugzilla.redhat.com/1276550>: FUSE clients in a
   container environment hang and do not recover post losing connections to
   all bricks
   - 1277080 <https://bugzilla.redhat.com/1277080>: quota: set quota
   version for files/directories
   - 1277394 <https://bugzilla.redhat.com/1277394>: Wrong value of
   snap-max-hard-limit observed in 'gluster volume info'.
   - 1277587 <https://bugzilla.redhat.com/1277587>: Data Tiering:tiering
   deamon crashes when trying to heat the file
   - 1277590 <https://bugzilla.redhat.com/1277590>: Tier : Move common
   functions into tier.rc
   - 1277800 <https://bugzilla.redhat.com/1277800>: [New] - Message
   displayed after attach tier is misleading
   - 1277984 <https://bugzilla.redhat.com/1277984>: Upgrading to 3.7.-5-5
   has changed volume to distributed disperse
   - 1278578 <https://bugzilla.redhat.com/1278578>: move mount-nfs-auth.t
   to failed tests lists
   - 1278603 <https://bugzilla.redhat.com/1278603>: fix lookup-unhashed for
   tiered volumes.
   - 1278640 <https://bugzilla.redhat.com/1278640>: [New] - Files in a
   tiered volume gets promoted when bitd signs them
   - 1278744 <https://bugzilla.redhat.com/1278744>: ec-readdir.t is failing
   - 1278850 <https://bugzilla.redhat.com/1278850>: Tests/tiering: Correct
   typo in bug-1214222-directories_miising_after_attach_tier.t in bad_tests

*Known Issues:*

   - Volume commands fail with "staging failed" message when few nodes in
   trusted storage pool have 3.7.6 installed and other nodes have 3.7.5
   installed. Please upgrade all nodes to recover from this error. This issue
   is not seen if upgrading from 3.7.4 or previous to 3.7.6.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20151116/bbee8c2b/attachment.html>

More information about the Gluster-users mailing list