summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorShyamsundarR <srangana@redhat.com>2019-03-19 10:17:17 -0400
committerShyamsundarR <srangana@redhat.com>2019-03-19 10:17:57 -0400
commit3fadf5cc41d5ea3195d2228d23e890f27fc22f87 (patch)
tree8875ba80a659bcc9099116adfcc5ad0b546d4cd7
parent7e90a3b592018f4e0d270d7102cf83aa3eb242d2 (diff)
doc: Final version of release-6 release notesv6.0
Fixes: bz#1672818 Change-Id: I6a98985a7f25bc2b85af5bd85f4be3ffac7d619d Signed-off-by: ShyamsundarR <srangana@redhat.com>
-rw-r--r--doc/release-notes/6.0.md230
1 files changed, 229 insertions, 1 deletions
diff --git a/doc/release-notes/6.0.md b/doc/release-notes/6.0.md
index cbfa8a8..7319e91 100644
--- a/doc/release-notes/6.0.md
+++ b/doc/release-notes/6.0.md
@@ -202,4 +202,232 @@ older API SDK will continue to work as before.
## Bugs addressed
Bugs addressed since release-5 are listed below.
-<TODO>
+- [#1138841](https://bugzilla.redhat.com/1138841): allow the use of the CIDR format with auth.allow
+- [#1236272](https://bugzilla.redhat.com/1236272): socket: Use newer system calls that provide better interface/performance on Linux/*BSD when available
+- [#1243991](https://bugzilla.redhat.com/1243991): "gluster volume set <VOLNAME> group <GROUP>" is not in the help text
+- [#1285126](https://bugzilla.redhat.com/1285126): RFE: GlusterFS NFS does not implement an all_squash volume setting
+- [#1343926](https://bugzilla.redhat.com/1343926): port-map: let brick choose its own port
+- [#1364707](https://bugzilla.redhat.com/1364707): Remove deprecated stripe xlator
+- [#1427397](https://bugzilla.redhat.com/1427397): script to strace processes consuming high CPU
+- [#1467614](https://bugzilla.redhat.com/1467614): Gluster read/write performance improvements on NVMe backend
+- [#1486532](https://bugzilla.redhat.com/1486532): need a script to resolve backtraces
+- [#1511339](https://bugzilla.redhat.com/1511339): In Replica volume 2*2 when quorum is set, after glusterd restart nfs server is coming up instead of self-heal daemon
+- [#1535495](https://bugzilla.redhat.com/1535495): Add option -h and --help to gluster cli
+- [#1535528](https://bugzilla.redhat.com/1535528): Gluster cli show no help message in prompt
+- [#1560561](https://bugzilla.redhat.com/1560561): systemd service file enhancements
+- [#1560969](https://bugzilla.redhat.com/1560969): Garbage collect inactive inodes in fuse-bridge
+- [#1564149](https://bugzilla.redhat.com/1564149): Agree upon a coding standard, and automate check for this in smoke
+- [#1564890](https://bugzilla.redhat.com/1564890): mount.glusterfs: can't shift that many
+- [#1575836](https://bugzilla.redhat.com/1575836): logic in S30samba-start.sh hook script needs tweaking
+- [#1579788](https://bugzilla.redhat.com/1579788): Thin-arbiter: Have the state of volume in memory
+- [#1582516](https://bugzilla.redhat.com/1582516): libgfapi: glfs init fails on afr volume with ctime feature enabled
+- [#1590385](https://bugzilla.redhat.com/1590385): Refactor dht lookup code
+- [#1593538](https://bugzilla.redhat.com/1593538): ctime: Access time is different with in same replica/EC volume
+- [#1596787](https://bugzilla.redhat.com/1596787): glusterfs rpc-clnt.c: error returned while attempting to connect to host: (null), port 0
+- [#1598345](https://bugzilla.redhat.com/1598345): gluster get-state command is crashing glusterd process when geo-replication is configured
+- [#1600145](https://bugzilla.redhat.com/1600145): [geo-rep]: Worker still ACTIVE after killing bricks
+- [#1605056](https://bugzilla.redhat.com/1605056): [RHHi] Mount hung and not accessible
+- [#1605077](https://bugzilla.redhat.com/1605077): If a node disconnects during volume delete, it assumes deleted volume as a freshly created volume when it is back online
+- [#1608512](https://bugzilla.redhat.com/1608512): cluster.server-quorum-type help text is missing possible settings
+- [#1624006](https://bugzilla.redhat.com/1624006): /var/run/gluster/metrics/ wasn't created automatically
+- [#1624332](https://bugzilla.redhat.com/1624332): [Thin-arbiter]: Add tests for thin arbiter feature
+- [#1624724](https://bugzilla.redhat.com/1624724): ctime: Enable ctime feature by default and also improve usability by providing single option to enable
+- [#1624796](https://bugzilla.redhat.com/1624796): mkdir -p fails with "No data available" when root-squash is enabled
+- [#1625850](https://bugzilla.redhat.com/1625850): tests: fixes to bug-1015990-rep.t
+- [#1625961](https://bugzilla.redhat.com/1625961): Writes taking very long time leading to system hogging
+- [#1626313](https://bugzilla.redhat.com/1626313): fix glfs_fini related problems
+- [#1626610](https://bugzilla.redhat.com/1626610): [USS]: Change gf_log to gf_msg
+- [#1626994](https://bugzilla.redhat.com/1626994): split-brain observed on parent dir
+- [#1627610](https://bugzilla.redhat.com/1627610): glusterd crash in regression build
+- [#1627620](https://bugzilla.redhat.com/1627620): SAS job aborts complaining about file doesn't exist
+- [#1628194](https://bugzilla.redhat.com/1628194): tests/dht: Additional tests for dht operations
+- [#1628605](https://bugzilla.redhat.com/1628605): One client hangs when another client loses communication with bricks during intensive write I/O
+- [#1628664](https://bugzilla.redhat.com/1628664): Update op-version from 4.2 to 5.0
+- [#1629561](https://bugzilla.redhat.com/1629561): geo-rep: geo-rep config set fails to set rsync-options
+- [#1630368](https://bugzilla.redhat.com/1630368): Low Random write IOPS in VM workloads
+- [#1630798](https://bugzilla.redhat.com/1630798): Add performance options to virt profile
+- [#1630804](https://bugzilla.redhat.com/1630804): libgfapi-python: test_listdir_with_stat and test_scandir failure on release 5 branch
+- [#1630922](https://bugzilla.redhat.com/1630922): glusterd crashed and core generated at gd_mgmt_v3_unlock_timer_cbk after huge number of volumes were created
+- [#1631128](https://bugzilla.redhat.com/1631128): rpc marks brick disconnected from glusterd & volume stop transaction gets timed out
+- [#1631357](https://bugzilla.redhat.com/1631357): glusterfsd keeping fd open in index xlator after stop the volume
+- [#1631886](https://bugzilla.redhat.com/1631886): Update database profile settings for gluster
+- [#1632161](https://bugzilla.redhat.com/1632161): [Disperse] : Set others.eager-lock on for ec-1468261.t test to pass
+- [#1632236](https://bugzilla.redhat.com/1632236): Provide indication at the console or in the logs about the progress being made with changelog processing.
+- [#1632503](https://bugzilla.redhat.com/1632503): FUSE client segfaults when performance.md-cache-statfs is enabled for a volume
+- [#1632717](https://bugzilla.redhat.com/1632717): EC crashes when running on non 64-bit architectures
+- [#1632889](https://bugzilla.redhat.com/1632889): 'df' shows half as much space on volume after upgrade to RHGS 3.4
+- [#1633926](https://bugzilla.redhat.com/1633926): Script to collect system-stats
+- [#1634102](https://bugzilla.redhat.com/1634102): MAINTAINERS: Add sunny kumar as a peer for snapshot component
+- [#1634220](https://bugzilla.redhat.com/1634220): md-cache: some problems of cache virtual glusterfs ACLs for ganesha
+- [#1635050](https://bugzilla.redhat.com/1635050): [SNAPSHOT]: with brick multiplexing, snapshot restore will make glusterd send wrong volfile
+- [#1635145](https://bugzilla.redhat.com/1635145): I/O errors observed on the application side after the creation of a 'linkto' file
+- [#1635480](https://bugzilla.redhat.com/1635480): Correction for glusterd memory leak because use "gluster volume status volume_name --detail" continuesly (cli)
+- [#1635593](https://bugzilla.redhat.com/1635593): glusterd crashed in cleanup_and_exit when glusterd comes up with upgrade mode.
+- [#1635688](https://bugzilla.redhat.com/1635688): Keep only the valid (maintained/supported) components in the build
+- [#1635820](https://bugzilla.redhat.com/1635820): Seeing defunt translator and discrepancy in volume info when issued from node which doesn't host bricks in that volume
+- [#1635863](https://bugzilla.redhat.com/1635863): Gluster peer probe doesn't work for IPv6
+- [#1636570](https://bugzilla.redhat.com/1636570): Cores due to SIGILL during multiplex regression tests
+- [#1636631](https://bugzilla.redhat.com/1636631): Issuing a "heal ... full" on a disperse volume causes permanent high CPU utilization.
+- [#1637196](https://bugzilla.redhat.com/1637196): Disperse volume 'df' usage is extremely incorrect after replace-brick.
+- [#1637249](https://bugzilla.redhat.com/1637249): gfid heal does not happen when there is no source brick
+- [#1637802](https://bugzilla.redhat.com/1637802): data-self-heal in arbiter volume results in stale locks.
+- [#1637934](https://bugzilla.redhat.com/1637934): glusterfsd is keeping fd open in index xlator
+- [#1638453](https://bugzilla.redhat.com/1638453): Gfid mismatch seen on shards when lookup and mknod are in progress at the same time
+- [#1639599](https://bugzilla.redhat.com/1639599): Improve support-ability of glusterfs
+- [#1640026](https://bugzilla.redhat.com/1640026): improper checking to avoid identical mounts
+- [#1640066](https://bugzilla.redhat.com/1640066): [Stress] : Mismatching iatt in glustershd logs during MTSH and continous IO from Ganesha mounts
+- [#1640165](https://bugzilla.redhat.com/1640165): io-stats: garbage characters in the filenames generated
+- [#1640489](https://bugzilla.redhat.com/1640489): Invalid memory read after freed in dht_rmdir_readdirp_cbk
+- [#1640495](https://bugzilla.redhat.com/1640495): [GSS] Fix log level issue with brick mux
+- [#1640581](https://bugzilla.redhat.com/1640581): [AFR] : Start crawling indices and healing only if both data bricks are UP in replica 2 (thin-arbiter)
+- [#1641344](https://bugzilla.redhat.com/1641344): Spurious failures in bug-1637802-arbiter-stale-data-heal-lock.t
+- [#1642448](https://bugzilla.redhat.com/1642448): EC volume getting created without any redundant brick
+- [#1642597](https://bugzilla.redhat.com/1642597): tests/bugs/glusterd/optimized-basic-testcases-in-cluster.t failing
+- [#1642800](https://bugzilla.redhat.com/1642800): socket: log voluntary socket close/shutdown and EOF on socket at INFO log-level
+- [#1642807](https://bugzilla.redhat.com/1642807): remove 'tier' translator from build and code
+- [#1642810](https://bugzilla.redhat.com/1642810): remove glupy from code and build
+- [#1642850](https://bugzilla.redhat.com/1642850): glusterd: raise default transport.listen-backlog to 1024
+- [#1642865](https://bugzilla.redhat.com/1642865): geo-rep: geo-replication gets stuck after file rename and gfid conflict
+- [#1643349](https://bugzilla.redhat.com/1643349): [OpenSSL] : auth.ssl-allow has no option description.
+- [#1643402](https://bugzilla.redhat.com/1643402): [Geo-Replication] Geo-rep faulty sesion because of the directories are not synced to slave.
+- [#1643519](https://bugzilla.redhat.com/1643519): Provide an option to silence glfsheal logs
+- [#1643929](https://bugzilla.redhat.com/1643929): geo-rep: gluster-mountbroker status crashes
+- [#1643932](https://bugzilla.redhat.com/1643932): geo-rep: On gluster command failure on slave, worker crashes with python3
+- [#1643935](https://bugzilla.redhat.com/1643935): cliutils: geo-rep cliutils' usage of Popen is not python3 compatible
+- [#1644129](https://bugzilla.redhat.com/1644129): Excessive logging in posix_update_utime_in_mdata
+- [#1644164](https://bugzilla.redhat.com/1644164): Use GF_ATOMIC ops to update inode->nlookup
+- [#1644629](https://bugzilla.redhat.com/1644629): [rpcsvc] Single request Queue for all event threads is a performance bottleneck
+- [#1644755](https://bugzilla.redhat.com/1644755): CVE-2018-14651 glusterfs: glusterfs server exploitable via symlinks to relative paths [fedora-all]
+- [#1644756](https://bugzilla.redhat.com/1644756): CVE-2018-14653 glusterfs: Heap-based buffer overflow via "gf_getspec_req" RPC message [fedora-all]
+- [#1644757](https://bugzilla.redhat.com/1644757): CVE-2018-14659 glusterfs: Unlimited file creation via "GF_XATTR_IOSTATS_DUMP_KEY" xattr allows for denial of service [fedora-all]
+- [#1644758](https://bugzilla.redhat.com/1644758): CVE-2018-14660 glusterfs: Repeat use of "GF_META_LOCK_KEY" xattr allows for memory exhaustion [fedora-all]
+- [#1644760](https://bugzilla.redhat.com/1644760): CVE-2018-14654 glusterfs: "features/index" translator can create arbitrary, empty files [fedora-all]
+- [#1644763](https://bugzilla.redhat.com/1644763): CVE-2018-14661 glusterfs: features/locks translator passes an user-controlled string to snprintf without a proper format string resulting in a denial of service [fedora-all]
+- [#1645986](https://bugzilla.redhat.com/1645986): tests/bugs/glusterd/optimized-basic-testcases-in-cluster.t failing in distributed regression
+- [#1646104](https://bugzilla.redhat.com/1646104): [Geo-rep]: Faulty geo-rep sessions due to link ownership on slave volume
+- [#1646728](https://bugzilla.redhat.com/1646728): [snapview-server]:forget glfs handles during inode forget
+- [#1646869](https://bugzilla.redhat.com/1646869): gNFS crashed when processing "gluster v status [vol] nfs clients"
+- [#1646892](https://bugzilla.redhat.com/1646892): Portmap entries showing stale brick entries when bricks are down
+- [#1647029](https://bugzilla.redhat.com/1647029): can't enable shared-storage
+- [#1647074](https://bugzilla.redhat.com/1647074): when peer detach is issued, throw a warning to remount volumes using other cluster IPs before proceeding
+- [#1647651](https://bugzilla.redhat.com/1647651): gfapi: fix bad dict setting of lease-id
+- [#1648237](https://bugzilla.redhat.com/1648237): Bumping up of op-version times out on a scaled system with ~1200 volumes
+- [#1648298](https://bugzilla.redhat.com/1648298): dht_revalidate may not heal attrs on the brick root
+- [#1648687](https://bugzilla.redhat.com/1648687): Incorrect usage of local->fd in afr_open_ftruncate_cbk
+- [#1648768](https://bugzilla.redhat.com/1648768): Tracker bug for all leases related issues
+- [#1649709](https://bugzilla.redhat.com/1649709): profile info doesn't work when decompounder xlator is not in graph
+- [#1650115](https://bugzilla.redhat.com/1650115): glusterd requests are timing out in a brick multiplex setup
+- [#1650389](https://bugzilla.redhat.com/1650389): rpc: log flooding with ENODATA errors
+- [#1650403](https://bugzilla.redhat.com/1650403): Memory leaks observed in brick-multiplex scenario on volume start/stop loop
+- [#1650893](https://bugzilla.redhat.com/1650893): fails to sync non-ascii (utf8) file and directory names, causes permanently faulty geo-replication state
+- [#1651059](https://bugzilla.redhat.com/1651059): [OpenSSL] : Retrieving the value of "client.ssl" option,before SSL is set up, fails .
+- [#1651165](https://bugzilla.redhat.com/1651165): Race in per-thread mem-pool when a thread is terminated
+- [#1651431](https://bugzilla.redhat.com/1651431): Resolve memory leak at the time of graph init
+- [#1651439](https://bugzilla.redhat.com/1651439): gluster-NFS crash while expanding volume
+- [#1651463](https://bugzilla.redhat.com/1651463): glusterd can't regenerate volfiles in container storage upgrade workflow
+- [#1651498](https://bugzilla.redhat.com/1651498): [geo-rep]: Failover / Failback shows fault status in a non-root setup
+- [#1651584](https://bugzilla.redhat.com/1651584): [geo-rep]: validate the config checkpoint date and fail if it is not is exact format hh:mm:ss
+- [#1652118](https://bugzilla.redhat.com/1652118): default cluster.max-bricks-per-process to 250
+- [#1652430](https://bugzilla.redhat.com/1652430): glusterd fails to start, when glusterd is restarted in a loop for every 45 seconds while volume creation is in-progress
+- [#1652852](https://bugzilla.redhat.com/1652852): "gluster volume get" doesn't show real default value for server.tcp-user-timeout
+- [#1652887](https://bugzilla.redhat.com/1652887): Geo-rep help looks to have a typo.
+- [#1652911](https://bugzilla.redhat.com/1652911): Add no-verify and ssh-port n options for create command in man page
+- [#1653277](https://bugzilla.redhat.com/1653277): bump up default value of server.event-threads
+- [#1653359](https://bugzilla.redhat.com/1653359): Self-heal:Improve heal performance
+- [#1653565](https://bugzilla.redhat.com/1653565): tests/geo-rep: Add arbiter volume test case
+- [#1654138](https://bugzilla.redhat.com/1654138): Optimize for virt store fails with distribute volume type
+- [#1654181](https://bugzilla.redhat.com/1654181): glusterd segmentation fault: glusterd_op_ac_brick_op_failed (event=0x7f44e0e63f40, ctx=0x0) at glusterd-op-sm.c:5606
+- [#1654187](https://bugzilla.redhat.com/1654187): [geo-rep]: RFE - Make slave volume read-only while setting up geo-rep (by default)
+- [#1654270](https://bugzilla.redhat.com/1654270): glusterd crashed with seg fault possibly during node reboot while volume creates and deletes were happening
+- [#1654521](https://bugzilla.redhat.com/1654521): io-stats outputs json numbers as strings
+- [#1654805](https://bugzilla.redhat.com/1654805): Bitrot: Scrub status say file is corrupted even it was just created AND 'path' in the output is broken
+- [#1654917](https://bugzilla.redhat.com/1654917): cleanup resources in server_init in case of failure
+- [#1655050](https://bugzilla.redhat.com/1655050): automatic split resolution with size as policy should not work on a directory which is in metadata splitbrain
+- [#1655052](https://bugzilla.redhat.com/1655052): Automatic Splitbrain with size as policy must not resolve splitbrains when both the copies are of same size
+- [#1655827](https://bugzilla.redhat.com/1655827): [Glusterd]: Glusterd crash while expanding volumes using heketi
+- [#1655854](https://bugzilla.redhat.com/1655854): Converting distribute to replica-3/arbiter volume fails
+- [#1656100](https://bugzilla.redhat.com/1656100): configure.ac does not enforce automake --foreign
+- [#1656264](https://bugzilla.redhat.com/1656264): Fix tests/bugs/shard/zero-flag.t
+- [#1656348](https://bugzilla.redhat.com/1656348): Commit c9bde3021202f1d5c5a2d19ac05a510fc1f788ac causes ls slowdown
+- [#1656517](https://bugzilla.redhat.com/1656517): [GSS] Gluster client logs filling with 0-glusterfs-socket: invalid port messages
+- [#1656682](https://bugzilla.redhat.com/1656682): brick memory consumed by volume is not getting released even after delete
+- [#1656771](https://bugzilla.redhat.com/1656771): [Samba-Enhancement] Need for a single group command for setting up volume options for samba
+- [#1656951](https://bugzilla.redhat.com/1656951): cluster.max-bricks-per-process 250 not working as expected
+- [#1657607](https://bugzilla.redhat.com/1657607): Convert nr_files to gf_atomic in posix_private structure
+- [#1657744](https://bugzilla.redhat.com/1657744): quorum count not updated in nfs-server vol file
+- [#1657783](https://bugzilla.redhat.com/1657783): Rename of a file leading to stale reads
+- [#1658045](https://bugzilla.redhat.com/1658045): Resolve memory leak in mgmt_pmap_signout_cbk
+- [#1658116](https://bugzilla.redhat.com/1658116): python2 to python3 compatibilty issues
+- [#1659327](https://bugzilla.redhat.com/1659327): 43% regression in small-file sequential read performance
+- [#1659432](https://bugzilla.redhat.com/1659432): Memory leak: dict_t leak in rda_opendir
+- [#1659708](https://bugzilla.redhat.com/1659708): Optimize by not stopping (restart) selfheal deamon (shd) when a volume is stopped unless it is the last volume
+- [#1659857](https://bugzilla.redhat.com/1659857): change max-port value in glusterd vol file to 60999
+- [#1659868](https://bugzilla.redhat.com/1659868): glusterd : features.selinux was missing in glusterd-volume-set file
+- [#1659869](https://bugzilla.redhat.com/1659869): improvements to io-cache
+- [#1659971](https://bugzilla.redhat.com/1659971): Setting slave volume read-only option by default results in failure
+- [#1660577](https://bugzilla.redhat.com/1660577): [Ganesha] Ganesha failed on one node while exporting volumes in loop
+- [#1660701](https://bugzilla.redhat.com/1660701): Use adaptive mutex in rpcsvc_program_register to improve performance
+- [#1661214](https://bugzilla.redhat.com/1661214): Brick is getting OOM for tests/bugs/core/bug-1432542-mpx-restart-crash.t
+- [#1662089](https://bugzilla.redhat.com/1662089): NL cache: fix typos
+- [#1662264](https://bugzilla.redhat.com/1662264): thin-arbiter: Check with thin-arbiter file before marking new entry change log
+- [#1662368](https://bugzilla.redhat.com/1662368): [ovirt-gluster] Fuse mount crashed while deleting a 1 TB image file from ovirt
+- [#1662679](https://bugzilla.redhat.com/1662679): Log connection_id in statedump for posix-locks as well for better debugging experience
+- [#1662906](https://bugzilla.redhat.com/1662906): Longevity: glusterfsd(brick process) crashed when we do volume creates and deletes
+- [#1663077](https://bugzilla.redhat.com/1663077): memory leak in mgmt handshake
+- [#1663102](https://bugzilla.redhat.com/1663102): Change default value for client side heal to off for replicate volumes
+- [#1663223](https://bugzilla.redhat.com/1663223): profile info command is not displaying information of bricks which are hosted on peers
+- [#1663243](https://bugzilla.redhat.com/1663243): rebalance status does not display localhost statistics when op-version is not bumped up
+- [#1664122](https://bugzilla.redhat.com/1664122): do not send bit-rot virtual xattrs in lookup response
+- [#1664124](https://bugzilla.redhat.com/1664124): Improve information dumped from io-threads in statedump
+- [#1664551](https://bugzilla.redhat.com/1664551): Wrong description of localtime-logging in manpages
+- [#1664647](https://bugzilla.redhat.com/1664647): dht: Add NULL check for stbuf in dht_rmdir_lookup_cbk
+- [#1664934](https://bugzilla.redhat.com/1664934): glusterfs-fuse client not benefiting from page cache on read after write
+- [#1665038](https://bugzilla.redhat.com/1665038): glusterd crashed while running "gluster get-state glusterd odir /get-state"
+- [#1665332](https://bugzilla.redhat.com/1665332): Wrong offset is used in offset for zerofill fop
+- [#1665358](https://bugzilla.redhat.com/1665358): allow regression to not run tests with nfs, if nfs is disabled.
+- [#1665363](https://bugzilla.redhat.com/1665363): Fix incorrect definition in index-mem-types.h
+- [#1665656](https://bugzilla.redhat.com/1665656): testcaes glusterd/add-brick-and-validate-replicated-volume-options.t is crash while brick_mux is enable
+- [#1665826](https://bugzilla.redhat.com/1665826): [geo-rep]: Directory renames not synced to slave in Hybrid Crawl
+- [#1666143](https://bugzilla.redhat.com/1666143): Several fixes on socket pollin and pollout return value
+- [#1666833](https://bugzilla.redhat.com/1666833): move few recurring logs to DEBUG level.
+- [#1667779](https://bugzilla.redhat.com/1667779): glusterd leaks about 1GB memory per day on single machine of storage pool
+- [#1667804](https://bugzilla.redhat.com/1667804): Unable to delete directories that contain linkto files that point to itself.
+- [#1667905](https://bugzilla.redhat.com/1667905): dict_leak in __glusterd_handle_cli_uuid_get function
+- [#1668190](https://bugzilla.redhat.com/1668190): Block hosting volume deletion via heketi-cli failed with error "target is busy" but deleted from gluster backend
+- [#1668268](https://bugzilla.redhat.com/1668268): Unable to mount gluster volume
+- [#1669077](https://bugzilla.redhat.com/1669077): [ovirt-gluster] Fuse mount crashed while creating the preallocated image
+- [#1669937](https://bugzilla.redhat.com/1669937): Rebalance : While rebalance is in progress , SGID and sticky bit which is set on the files while file migration is in progress is seen on the mount point
+- [#1670031](https://bugzilla.redhat.com/1670031): performance regression seen with smallfile workload tests
+- [#1670253](https://bugzilla.redhat.com/1670253): Writes on Gluster 5 volumes fail with EIO when "cluster.consistent-metadata" is set
+- [#1670259](https://bugzilla.redhat.com/1670259): New GFID file recreated in a replica set after a GFID mismatch resolution
+- [#1671213](https://bugzilla.redhat.com/1671213): core: move "dict is NULL" logs to DEBUG log level
+- [#1671637](https://bugzilla.redhat.com/1671637): geo-rep: Issue with configparser import
+- [#1672205](https://bugzilla.redhat.com/1672205): 'gluster get-state' command fails if volume brick doesn't exist.
+- [#1672818](https://bugzilla.redhat.com/1672818): GlusterFS 6.0 tracker
+- [#1673267](https://bugzilla.redhat.com/1673267): Fix timeouts so the tests pass on AWS
+- [#1673972](https://bugzilla.redhat.com/1673972): insufficient logging in glusterd_resolve_all_bricks
+- [#1674364](https://bugzilla.redhat.com/1674364): glusterfs-fuse client not benefiting from page cache on read after write
+- [#1676429](https://bugzilla.redhat.com/1676429): distribute: Perf regression in mkdir path
+- [#1677260](https://bugzilla.redhat.com/1677260): rm -rf fails with "Directory not empty"
+- [#1678570](https://bugzilla.redhat.com/1678570): glusterfs FUSE client crashing every few days with 'Failed to dispatch handler'
+- [#1679004](https://bugzilla.redhat.com/1679004): With parallel-readdir enabled, deleting a directory containing stale linkto files fails with "Directory not empty"
+- [#1679275](https://bugzilla.redhat.com/1679275): dht: fix double extra unref of inode at heal path
+- [#1679965](https://bugzilla.redhat.com/1679965): Upgrade from glusterfs 3.12 to gluster 4/5 broken
+- [#1679998](https://bugzilla.redhat.com/1679998): GlusterFS can be improved
+- [#1680020](https://bugzilla.redhat.com/1680020): Integer Overflow possible in md-cache.c due to data type inconsistency
+- [#1680585](https://bugzilla.redhat.com/1680585): remove glupy from code and build
+- [#1680586](https://bugzilla.redhat.com/1680586): Building RPM packages with _for_fedora_koji_builds enabled fails on el6
+- [#1683008](https://bugzilla.redhat.com/1683008): glustereventsd does not start on Ubuntu 16.04 LTS
+- [#1683506](https://bugzilla.redhat.com/1683506): remove experimental xlators informations from glusterd-volume-set.c
+- [#1683716](https://bugzilla.redhat.com/1683716): glusterfind: revert shebangs to #!/usr/bin/python3
+- [#1683880](https://bugzilla.redhat.com/1683880): Multiple shd processes are running on brick_mux environmet
+- [#1683900](https://bugzilla.redhat.com/1683900): Failed to dispatch handler
+- [#1684029](https://bugzilla.redhat.com/1684029): upgrade from 3.12, 4.1 and 5 to 6 broken
+- [#1684777](https://bugzilla.redhat.com/1684777): gNFS crashed when processing "gluster v profile [vol] info nfs"
+- [#1685771](https://bugzilla.redhat.com/1685771): glusterd memory usage grows at 98 MB/h while being monitored by RHGSWA
+- [#1686364](https://bugzilla.redhat.com/1686364): [ovirt-gluster] Rolling gluster upgrade from 3.12.5 to 5.3 led to shard on-disk xattrs disappearing
+- [#1686399](https://bugzilla.redhat.com/1686399): listing a file while writing to it causes deadlock
+- [#1686875](https://bugzilla.redhat.com/1686875): packaging: rdma on s390x, unnecessary ldconfig scriptlets
+- [#1687248](https://bugzilla.redhat.com/1687248): Error handling in /usr/sbin/gluster-eventsapi produces IndexError: tuple index out of range
+- [#1687672](https://bugzilla.redhat.com/1687672): [geo-rep]: Checksum mismatch when 2x2 vols are converted to arbiter
+- [#1688218](https://bugzilla.redhat.com/1688218): Brick process has coredumped, when starting glusterd