Wednesday, August 16, 2017

GlusterFS 3.8.15 is available, likely the last 3.8 update

The next Long-Term-Maintenance release for Gluster is around the corner. Once GlusterFS-3.12 is available, the oldest maintained version (3.8) will be retired and no maintenance updates are planned. With this last update to GlusterFS-3.8 a few more bugs have been fixed.

Packages for this release will become available for the different distributions and their versions listed on the community packages page.

Release notes for Gluster 3.8.15

This is a bugfix release. The Release Notes for 3.8.0, 3.8.1, 3.8.2, 3.8.3, 3.8.4, 3.8.5, 3.8.6, 3.8.7, 3.8.8, 3.8.9, 3.8.10, 3.8.11, 3.8.12, 3.8.13 and 3.8.14 contain a listing of all the new features that were added and bugs fixed in the GlusterFS 3.8 stable release.

End Of Life Notice

This is most likely the last bugfix release for the GlusterFS 3.8 Long-Term-Support version. GlusterFS 3.12 is planned to be released at the end of August 2017 and will be the next Long-Term-Support version. It is highly recommended to upgrade any Gluster 3.8 environment to either the 3.10 or 3.12 release. More details about the different Long-Term-Support versions can be found on the release schedule.

Bugs addressed

A total of 4 patches have been merged, addressing 4 bugs:
  • #1470495: gluster volume status --xml fails when there are 100 volumes
  • #1471613: metadata heal not happening despite having an active sink
  • #1480193: Running sysbench on vm disk from plain distribute gluster volume causes disk corruption
  • #1481398: libgfapi: memory leak in glfs_h_acl_get

Friday, July 14, 2017

GlusterFS 3.8.14 is here, 3.8 even closer to End-Of-Life

The 10th of the month has passed again, that means a 3.8.x update can't be far out. So, here it is, we're announcing the availability of glusterfs-3.8.14. Note that this is one of the last updates in the 3.8 Long-Term-Maintenance release stream. This schedule on the website shows what options you have for upgrading your environment. Remember that many distributions have packages included in their standard repositories, and other versions might be available from external locations. All the details about what packages to find where are on the Community Package page in the docs.

Release notes for Gluster 3.8.14

This is a bugfix release. The Release Notes for 3.8.0, 3.8.1, 3.8.2, 3.8.3, 3.8.4, 3.8.5, 3.8.6, 3.8.7, 3.8.8, 3.8.9, 3.8.10, 3.8.11, 3.8.12 and 3.8.13 contain a listing of all the new features that were added and bugs fixed in the GlusterFS 3.8 stable release.

Bugs addressed

A total of 3 patches have been merged, addressing 2 bugs:
  • #1462447: brick maintenance - no client reconnect
  • #1467272: Heal info shows incorrect status

Wednesday, June 28, 2017

GlusterFS 3.8.13 update available, and 3.8 nearing End-Of-Life

The Gluster releases follow a 3-month cycle and, with alternating Short-Term-Maintenance and Long-Term-Maintenance versions. GlusterFS 3.8 is currently the oldest Long-Term-Maintenance release, and will become End-Of-Life with the GlusterFS 3.12 version. If all goes according to plan, 3.12 will get released in August and is the last 3.x version before Gluster 4.0 hits the disks.

There will be a few more releases in the GlusterFS 3.8 line, but users should start to plan an upgrade to a version that receives regular bugfix updates after August.

Release notes for Gluster 3.8.13

This is a bugfix release. The Release Notes for 3.8.0, 3.8.1, 3.8.2, 3.8.3, 3.8.4, 3.8.5, 3.8.6, 3.8.7, 3.8.8, 3.8.9, 3.8.10, 3.8.11 and 3.8.12 contain a listing of all the new features that were added and bugs fixed in the GlusterFS 3.8 stable release.

Bugs addressed

A total of 13 patches have been merged, addressing 8 bugs:
  • #1447523: Glusterd segmentation fault in ' _Unwind_Backtrace' while running peer probe
  • #1449782: quota: limit-usage command failed with error " Failed to start aux mount"
  • #1449941: When either killing or restarting a brick with performance.stat-prefetch on, stat sometimes returns a bad st_size value.
  • #1450055: [GANESHA] Adding a node to existing cluster failed to start pacemaker service on new node
  • #1450380: GNFS crashed while taking lock on a file from 2 different clients having same volume mounted from 2 different servers
  • #1450937: [New] - Replacing an arbiter brick while I/O happens causes vm pause
  • #1460650: posix-acl: Whitelist virtual ACL xattrs
  • #1460661: "split-brain observed [Input/output error]" error messages in samba logs during parallel rm -rf

Monday, May 22, 2017

Enjoy more bugfixes with GlusterFS 3.8.12

Like every month, there is an update for the GlusterFS 3.8 stable version. A few more bugfixes have been included in this release. Packages are already available for many distributions, some distributions might still need to promote the update from their testing repository to release, so hold tight if there is no update for your favourite OS yet.

Release notes for Gluster 3.8.12

This is a bugfix release. The Release Notes for 3.8.0, 3.8.1, 3.8.2, 3.8.3, 3.8.4, 3.8.5, 3.8.6, 3.8.7, 3.8.8, 3.8.9, 3.8.10 and 3.8.11 contain a listing of all the new features that were added and bugs fixed in the GlusterFS 3.8 stable release.

Bugs addressed

A total of 13 patches have been merged, addressing 11 bugs:
  • #1440228: NFS Sub-directory mount not working on solaris10 client
  • #1440635: Application VMs with their disk images on sharded-replica 3 volume are unable to boot after performing rebalance
  • #1440810: Update rfc.sh to check Change-Id consistency for backports
  • #1441574: [geo-rep]: rsync should not try to sync internal xattrs
  • #1441930: [geo-rep]: Worker crashes with [Errno 16] Device or resource busy: '.gfid/00000000-0000-0000-0000-000000000001/dir.166 while renaming directories
  • #1441933: [Geo-rep] If for some reason MKDIR failed to sync, it should not proceed further.
  • #1442933: Segmentation fault when creating a qcow2 with qemu-img
  • #1443012: snapshot: snapshots appear to be failing with respect to secure geo-rep slave
  • #1443319: Don't wind post-op on a brick where the fop phase failed.
  • #1445213: Unable to take snapshot on a geo-replicated volume, even after stopping the session
  • #1449314: [whql][virtio-block+glusterfs]"Disk Stress" and "Disk Verification" job always failed on win7-32/win2012/win2k8R2 guest

Tuesday, April 18, 2017

Bugfix release GlusterFS 3.8.11 has landed

An other month has passed, and more bugs have been squashed in the 3.8 release. Packages should be available or arrive soon at the usual repositories. The next 3.8 update is expected to be made available just after the 10th of May.

Release notes for Gluster 3.8.11

This is a bugfix release. The Release Notes for 3.8.0, 3.8.1, 3.8.2, 3.8.3, 3.8.4, 3.8.5, 3.8.6, 3.8.7, 3.8.8, 3.8.9 and 3.8.10 contain a listing of all the new features that were added and bugs fixed in the GlusterFS 3.8 stable release.

Bugs addressed

A total of 15 patches have been merged, addressing 13 bugs:
  • #1422788: [Replicate] "RPC call decoding failed" leading to IO hang & mount inaccessible
  • #1427390: systemic testing: seeing lot of ping time outs which would lead to splitbrains
  • #1430845: build/packaging: Debian and Ubuntu don't have /usr/libexec/; results in bad packages
  • #1431592: memory leak in features/locks xlator
  • #1434298: [Disperse] Metadata version is not healing when a brick is down
  • #1434302: Move spit-brain msg in read txn to debug
  • #1435645: Disperse: Provide description of disperse.eager-lock option.
  • #1436231: Undo pending xattrs only on the up bricks
  • #1436412: Unrecognized filesystems (i.e. btrfs, zfs) log many errors about "getinode size"
  • #1437330: Sharding: Fix a performance bug
  • #1438424: [Ganesha + EC] : Input/Output Error while creating LOTS of smallfiles
  • #1439112: File-level WORM allows ftruncate() on read-only files
  • #1440635: Application VMs with their disk images on sharded-replica 3 volume are unable to boot after performing rebalance

Friday, March 17, 2017

GlusterFS 3.8.10 is available

The 10th update for GlusterFS 3.8 is available for users of the 3.8 Long-Term-Maintenance version. Packages for this minor update are in many of the repositories for different distributions already. It is recommended to update any 3.8 installation to this latest release.

Release notes for Gluster 3.8.10

This is a bugfix release. The Release Notes for 3.8.0, 3.8.1, 3.8.2, 3.8.3, 3.8.4, 3.8.5, 3.8.6, 3.8.7, 3.8.8 and 3.8.9 contain a listing of all the new features that were added and bugs fixed in the GlusterFS 3.8 stable release.

Improved configuration with additional 'virt' options

This release includes 5 more options to group virt (for VM workloads) for optimal performance.
Updating to the glusterfs version containing this patch won't automatically set these newer options on already existing volumes that have group virt configured. The changes take effect only when post-upgrade
# gluster volume-set <VOL> group virt
is performed.
For already existing volumes the users may execute the following five commands, if not already set:
# gluster volume set <VOL> performance.low-prio-threads 32
# gluster volume set <VOL> cluster.locking-scheme granular
# gluster volume set <VOL> features.shard on
# gluster volume set <VOL> cluster.shd-max-threads 8
# gluster volume set <VOL> cluster.shd-wait-qlength 10000
# gluster volume set <VOL> user.cifs off
It is most likely that features.shard would already have been set on the volume even before the upgrade, in which case the third volume set command above may be skipped.

Bugs addressed

A total of 18 patches have been merged, addressing 16 bugs:
  • #1387878: Rebalance after add bricks corrupts files
  • #1412994: Memory leak on mount/fuse when setxattr fails
  • #1420993: Modified volume options not synced once offline nodes comes up.
  • #1422352: glustershd process crashed on systemic setup
  • #1422394: Gluster NFS server crashing in __mnt3svc_umountall
  • #1422811: [Geo-rep] Recreating geo-rep session with same slave after deleting with reset-sync-time fails to sync
  • #1424915: dht_setxattr returns EINVAL when a file is deleted during the FOP
  • #1424934: Include few more options in virt file
  • #1424974: remove-brick status shows 0 rebalanced files
  • #1425112: [Ganesha] : Unable to bring up a Ganesha HA cluster on RHEL 6.9.
  • #1425307: Fix statvfs for FreeBSD in Python
  • #1427390: systemic testing: seeing lot of ping time outs which would lead to splitbrains
  • #1427419: Warning messages throwing when EC volume offline brick comes up are difficult to understand for end user.
  • #1428743: Fix crash in dht resulting from tests/features/nuke.t
  • #1429312: Prevent reverse heal from happening
  • #1429405: Restore atime/mtime for symlinks and other non-regular files.

Thursday, February 16, 2017

GlusterFS 3.8.9 is an other Long-Term-Maintenance update

We are proud to announce the General Availability of yet the next update to the Long-Term-Stable releases for GlusterFS 3.8. Packages are being prepared to hit the mirrors expected to hit the repositories of distributions and the Gluster download server over the next few days. Details on which versions are part of which distributions can be found on the Community Packages in the documentation.

The release notes are part of the git repository, the downloadable tarball and are included in this post for easy access.

Release notes for Gluster 3.8.9

This is a bugfix release. The Release Notes for 3.8.0, 3.8.1, 3.8.2, 3.8.3, 3.8.4, 3.8.5, 3.8.6, 3.8.7 and 3.8.8 contain a listing of all the new features that were added and bugs fixed in the GlusterFS 3.8 stable release.

Bugs addressed

A total of 16 patches have been merged, addressing 14 bugs:
  • #1410852: glusterfs-server should depend on firewalld-filesystem
  • #1411899: DHT doesn't evenly balance files on FreeBSD with ZFS
  • #1412119: ganesha service crashed on all nodes of ganesha cluster on disperse volume when doing lookup while copying files remotely using scp
  • #1412888: Extra lookup/fstats are sent over the network when a brick is down.
  • #1412913: [ganesha + EC]posix compliance rename tests failed on EC volume with nfs-ganesha mount.
  • #1412915: Spurious split-brain error messages are seen in rebalance logs
  • #1412916: [ganesha+ec]: Contents of original file are not seen when hardlink is created
  • #1412922: ls and move hung on disperse volume
  • #1412941: Regression caused by enabling client-io-threads by default
  • #1414655: Upcall: Possible memleak if inode_ctx_set fails
  • #1415053: geo-rep session faulty with ChangelogException "No such file or directory"
  • #1415132: Improve output of "gluster volume status detail"
  • #1417802: debug/trace: Print iatts of individual entries in readdirp callback for better debugging experience
  • #1420184: [Remove-brick] Hardlink migration fails with "lookup failed (No such file or directory)" error messages in rebalance logs