From be560c6f41ab50f023606f0f8a3de189ee94f583 Mon Sep 17 00:00:00 2001 From: Niels de Vos Date: Fri, 31 Oct 2014 11:40:31 +0100 Subject: doc: add release notes for GlusterFS 3.5.3 Change-Id: Ic5a8be45f0397e492f56142856cf9bf9d35f91b5 BUG: 1125231 Signed-off-by: Niels de Vos Reviewed-on: http://review.gluster.org/8903 Reviewed-by: Kaleb KEITHLEY --- doc/release-notes/3.5.3.md | 84 ++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 84 insertions(+) create mode 100644 doc/release-notes/3.5.3.md diff --git a/doc/release-notes/3.5.3.md b/doc/release-notes/3.5.3.md new file mode 100644 index 00000000000..bbdac8e9570 --- /dev/null +++ b/doc/release-notes/3.5.3.md @@ -0,0 +1,84 @@ +## Release Notes for GlusterFS 3.5.3 + +This is a bugfix release. The [Release Notes for 3.5.0](3.5.0.md), +[3.5.1](3.5.1.md) and [3.5.2](3.5.2.md) contain a listing of all the new +features that were added and bugs fixed in the GlusterFS 3.5 stable release. + +### Bugs Fixed: + +- [1081016](https://bugzilla.redhat.com/1081016): glusterd needs xfsprogs and e2fsprogs packages +- [1100204](https://bugzilla.redhat.com/1100204): brick failure detection does not work for ext4 filesystems +- [1126801](https://bugzilla.redhat.com/1126801): glusterfs logrotate config file pollutes global config +- [1129527](https://bugzilla.redhat.com/1129527): DHT :- data loss - file is missing on renaming same file from multiple client at same time +- [1129541](https://bugzilla.redhat.com/1129541): [DHT:REBALANCE]: Rebalance failures are seen with error message " remote operation failed: File exists" +- [1132391](https://bugzilla.redhat.com/1132391): NFS interoperability problem: stripe-xlator removes EOF at end of READDIR +- [1133949](https://bugzilla.redhat.com/1133949): Minor typo in afr logging +- [1136221](https://bugzilla.redhat.com/1136221): The memories are exhausted quickly when handle the message which has multi fragments in a single record +- [1136835](https://bugzilla.redhat.com/1136835): crash on fsync +- [1138922](https://bugzilla.redhat.com/1138922): DHT + rebalance : rebalance process crashed + data loss + few Directories are present on sub-volumes but not visible on mount point + lookup is not healing directories +- [1139103](https://bugzilla.redhat.com/1139103): DHT + Snapshot :- If snapshot is taken when Directory is created only on hashed sub-vol; On restoring that snapshot Directory is not listed on mount point and lookup on parent is not healing +- [1139170](https://bugzilla.redhat.com/1139170): DHT :- rm -rf is not removing stale link file and because of that unable to create file having same name as stale link file +- [1139245](https://bugzilla.redhat.com/1139245): vdsm invoked oom-killer during rebalance and Killed process 4305, UID 0, (glusterfs nfs process) +- [1140338](https://bugzilla.redhat.com/1140338): rebalance is not resulting in the hash layout changes being available to nfs client +- [1140348](https://bugzilla.redhat.com/1140348): Renaming file while rebalance is in progress causes data loss +- [1140549](https://bugzilla.redhat.com/1140549): DHT: Rebalance process crash after add-brick and `rebalance start' operation +- [1140556](https://bugzilla.redhat.com/1140556): Core: client crash while doing rename operations on the mount +- [1141558](https://bugzilla.redhat.com/1141558): AFR : "gluster volume heal info" prints some random characters +- [1141733](https://bugzilla.redhat.com/1141733): data loss when rebalance + renames are in progress and bricks from replica pairs goes down and comes back +- [1142052](https://bugzilla.redhat.com/1142052): Very high memory usage during rebalance +- [1142614](https://bugzilla.redhat.com/1142614): files with open fd's getting into split-brain when bricks goes offline and comes back online +- [1144315](https://bugzilla.redhat.com/1144315): core: all brick processes crash when quota is enabled +- [1145000](https://bugzilla.redhat.com/1145000): Spec %post server does not wait for the old glusterd to exit +- [1147156](https://bugzilla.redhat.com/1147156): AFR client segmentation fault in afr_priv_destroy +- [1147243](https://bugzilla.redhat.com/1147243): nfs: volume set help says the rmtab file is in "/var/lib/glusterd/rmtab" +- [1149857](https://bugzilla.redhat.com/1149857): Option transport.socket.bind-address ignored +- [1153626](https://bugzilla.redhat.com/1153626): Sizeof bug for allocation of memory in afr_lookup +- [1153629](https://bugzilla.redhat.com/1153629): AFR : excessive logging of "Non blocking entrylks failed" in glfsheal log file. +- [1153900](https://bugzilla.redhat.com/1153900): Enabling Quota on existing data won't create pgfid xattrs +- [1153904](https://bugzilla.redhat.com/1153904): self heal info logs are filled with messages reporting ENOENT while self-heal is going on +- [1155073](https://bugzilla.redhat.com/1155073): Excessive logging in the self-heal daemon after a replace-brick +- [1157661](https://bugzilla.redhat.com/1157661): GlusterFS allows insecure SSL modes + +### Known Issues: + +- The following configuration changes are necessary for 'qemu' and 'samba vfs + plugin' integration with libgfapi to work seamlessly: + + 1. `gluster volume set server.allow-insecure on` + 2. restarting the volume is necessary + + ~~~ + gluster volume stop + gluster volume start + ~~~ + + 3. Edit `/etc/glusterfs/glusterd.vol` to contain this line: + + ~~~ + option rpc-auth-allow-insecure on + ~~~ + + 4. restarting glusterd is necessary + + ~~~ + service glusterd restart + ~~~ + + More details are also documented in the Gluster Wiki on the [Libgfapi with qemu libvirt](http://www.gluster.org/community/documentation/index.php/Libgfapi_with_qemu_libvirt) page. + +- For Block Device translator based volumes open-behind translator at the + client side needs to be disabled. + + ~~~ + gluster volume set performance.open-behind disabled + ~~~ + +- libgfapi clients calling `glfs_fini` before a successful `glfs_init` will cause the client to + hang as reported [here](http://lists.gnu.org/archive/html/gluster-devel/2014-04/msg00179.html). + The workaround is NOT to call `glfs_fini` for error cases encountered before a successful + `glfs_init`. This is being tracked in [Bug 1134050](https://bugzilla.redhat.com/1134050) for + glusterfs-3.5 and [Bug 1093594](https://bugzilla.redhat.com/1093594) for mainline. + +- If the `/var/run/gluster` directory does not exist enabling quota will likely + fail ([Bug 1117888](https://bugzilla.redhat.com/show_bug.cgi?id=1117888)). + -- cgit