There is one more known issues which we forgot to include in the release notes: * 1282322 - Volume start fails post add-brick on a volume which is not started - If you upgrade the cluster partially and perform an add-brick operation on a volume which is stopped, the add-brick would succeed however the volume start would fail. Workaround would be to toggle a volume tunable with volume set command and post that you could start the volume. Thanks, Atin On 11/16/2015 04:24 PM, Raghavendra Talur wrote: > Hi all, > > I'm pleased to announce the release of GlusterFS-3.7.6. > The list of fixed bugs is included below. > > Tarball and RPMs can be downloaded from > http://download.gluster.org/pub/gluster/glusterfs/3.7/3.7.6/ > > Ubuntu debs are available from > https://launchpad.net/~gluster/+archive/ubuntu/glusterfs-3.7 > > Debian Unstable (sid) packages have been updated and should be > available from default repos. > > NetBSD has updated ports at > ftp://ftp.netbsd.org/pub/pkgsrc/current/pkgsrc/filesystems/glusterfs/README.html > > > Bugs Fixed: > > * 1057295 <https://bugzilla.redhat.com/1057295>: glusterfs doesn't > include firewalld rules > * 1219399 <https://bugzilla.redhat.com/1219399>: NFS interoperability > problem: Gluster Striped-Replicated can't read on vmware esxi 5.x > NFS client > * 1221957 <https://bugzilla.redhat.com/1221957>: Fully support > data-tiering in 3.7.x, remove out of 'experimental' status > * 1258197 <https://bugzilla.redhat.com/1258197>: gNFSd: NFS mount > fails with "Remote I/O error" > * 1258242 <https://bugzilla.redhat.com/1258242>: Data Tiering: > detach-tier start force command not available on a tier > volume(unlike which is possible in force remove-brick) > * 1258833 <https://bugzilla.redhat.com/1258833>: Data Tiering: > Disallow attach tier on a volume where any rebalance process is in > progress to avoid deadlock(like remove brick commit pending etc) > * 1259167 <https://bugzilla.redhat.com/1259167>: GF_LOG_NONE logs always > * 1261146 <https://bugzilla.redhat.com/1261146>: Legacy files > pre-existing tier attach must be promoted > * 1261732 <https://bugzilla.redhat.com/1261732>: Disperse volume: df > -h on a nfs mount throws Invalid argument error > * 1261744 <https://bugzilla.redhat.com/1261744>: Tier/shd: Tracker bug > for tier and shd compatibility > * 1261758 <https://bugzilla.redhat.com/1261758>: Tiering/glusted: > volume status failed after detach tier start > * 1262860 <https://bugzilla.redhat.com/1262860>: Data Tiering: Tiering > deamon is seeing each part of a file in a Disperse cold volume as a > different file > * 1265623 <https://bugzilla.redhat.com/1265623>: Data > Tiering:Promotions and demotions fail after quota hard limits are > hit for a tier volume > * 1266836 <https://bugzilla.redhat.com/1266836>: AFR : fuse,nfs mount > hangs when directories with same names are created and deleted > continuously > * 1266880 <https://bugzilla.redhat.com/1266880>: Tiering: unlink > failed with error "Invaid argument" > * 1267816 <https://bugzilla.redhat.com/1267816>: quota/marker: marker > code cleanup > * 1269035 <https://bugzilla.redhat.com/1269035>: Data Tiering:Throw a > warning when user issues a detach-tier commit command > * 1269125 <https://bugzilla.redhat.com/1269125>: Data > Tiering:Regression: automation blocker:vol status for tier volumes > using xml format is not working > * 1269344 <https://bugzilla.redhat.com/1269344>: tier/cli: number of > bricks remains the same in v info --xml > * 1269501 <https://bugzilla.redhat.com/1269501>: Self-heal daemon > crashes when bricks godown at the time of data heal > * 1269530 <https://bugzilla.redhat.com/1269530>: > Core:Blocker:Segmentation fault when using fallocate command on a > gluster volume > * 1269730 <https://bugzilla.redhat.com/1269730>: Sharding - Send inode > forgets on /all/ shards if/when the protocol layer (FUSE/Gfapi) at > the top sends a forget on the actual file > * 1270123 <https://bugzilla.redhat.com/1270123>: Data Tiering: > Database locks observed on tiered volumes on continous writes to a file > * 1270527 <https://bugzilla.redhat.com/1270527>: add policy mechanism > for promotion and demotion > * 1270769 <https://bugzilla.redhat.com/1270769>: quota/marker: dir > count in inode quota is not atomic > * 1271204 <https://bugzilla.redhat.com/1271204>: Introduce priv dump > in shard xlator for better debugging > * 1271249 <https://bugzilla.redhat.com/1271249>: tiering:compiler > warning with gcc v5.1.1 > * 1271490 <https://bugzilla.redhat.com/1271490>: rm -rf on > /run/gluster/vol// is not showing quota output header for other > quota limit applied directories > * 1271540 <https://bugzilla.redhat.com/1271540>: RHEL7/systemd : can't > have server in debug mode anymore > * 1271627 <https://bugzilla.redhat.com/1271627>: Creating a already > deleted snapshot-clone deletes the corresponding snapshot. > * 1271967 <https://bugzilla.redhat.com/1271967>: ECVOL: glustershd log > grows quickly and fills up the root volume > * 1272036 <https://bugzilla.redhat.com/1272036>: Data Tiering:getting > failed to fsync on germany-hot-dht (Structure needs cleaning) warning > * 1272331 <https://bugzilla.redhat.com/1272331>: Tier: Do not > promote/demote files on which POSIX locks are held > * 1272334 <https://bugzilla.redhat.com/1272334>: Data > Tiering:Promotions fail when brick of EC (disperse) cold layer are down > * 1272398 <https://bugzilla.redhat.com/1272398>: Data Tiering:Lot of > Promotions/Demotions failed error messages > * 1273246 <https://bugzilla.redhat.com/1273246>: Tier xattr name is > misleading (trusted.tier-gfid) > * 1273334 <https://bugzilla.redhat.com/1273334>: Fix in afr > transaction code > * 1274101 <https://bugzilla.redhat.com/1274101>: need a way to > pause/stop tiering to take snapshot > * 1274600 <https://bugzilla.redhat.com/1274600>: [sharding+geo-rep]: > On existing slave mount, reading files fails to show sharded file > content > * 1275157 <https://bugzilla.redhat.com/1275157>: Reduce 'CTR disabled' > brick log message from ERROR to INFO/DEBUG > * 1275483 <https://bugzilla.redhat.com/1275483>: Data Tiering:heat > counters not getting reset and also internal ops seem to be heating > the files > * 1275502 <https://bugzilla.redhat.com/1275502>: [Tier]: Typo in the > output while setting the wrong value of low/hi watermark > * 1275921 <https://bugzilla.redhat.com/1275921>: Disk usage > mismatching after self-heal > * 1276029 <https://bugzilla.redhat.com/1276029>: Upgrading a subset of > cluster to 3.7.5 leads to issues with glusterd commands > * 1276060 <https://bugzilla.redhat.com/1276060>: dist-geo-rep: geo-rep > status shows Active/Passive even when all the gsync processes in a > node are killed > * 1276208 <https://bugzilla.redhat.com/1276208>: [RFE] 'gluster volume > help' output could be sorted alphabetically > * 1276244 <https://bugzilla.redhat.com/1276244>: gluster-nfs : Server > crashed due to an invalid reference > * 1276550 <https://bugzilla.redhat.com/1276550>: FUSE clients in a > container environment hang and do not recover post losing > connections to all bricks > * 1277080 <https://bugzilla.redhat.com/1277080>: quota: set quota > version for files/directories > * 1277394 <https://bugzilla.redhat.com/1277394>: Wrong value of > snap-max-hard-limit observed in 'gluster volume info'. > * 1277587 <https://bugzilla.redhat.com/1277587>: Data Tiering:tiering > deamon crashes when trying to heat the file > * 1277590 <https://bugzilla.redhat.com/1277590>: Tier : Move common > functions into tier.rc > * 1277800 <https://bugzilla.redhat.com/1277800>: [New] - Message > displayed after attach tier is misleading > * 1277984 <https://bugzilla.redhat.com/1277984>: Upgrading to 3.7.-5-5 > has changed volume to distributed disperse > * 1278578 <https://bugzilla.redhat.com/1278578>: move mount-nfs-auth.t > to failed tests lists > * 1278603 <https://bugzilla.redhat.com/1278603>: fix lookup-unhashed > for tiered volumes. > * 1278640 <https://bugzilla.redhat.com/1278640>: [New] - Files in a > tiered volume gets promoted when bitd signs them > * 1278744 <https://bugzilla.redhat.com/1278744>: ec-readdir.t is > failing consistently > * 1278850 <https://bugzilla.redhat.com/1278850>: Tests/tiering: > Correct typo in bug-1214222-directories_miising_after_attach_tier.t > in bad_tests > > *Known Issues:* > > * Volume commands fail with "staging failed" message when few nodes in > trusted storage pool have 3.7.6 installed and other nodes have 3.7.5 > installed. Please upgrade all nodes to recover from this error. This > issue is not seen if upgrading from 3.7.4 or previous to 3.7.6. > > > > _______________________________________________ > Gluster-devel mailing list > Gluster-devel@xxxxxxxxxxx > http://www.gluster.org/mailman/listinfo/gluster-devel > _______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-devel