summaryrefslogtreecommitdiffstats
path: root/doc/release-notes/3.12.2.md
diff options
context:
space:
mode:
Diffstat (limited to 'doc/release-notes/3.12.2.md')
-rw-r--r--doc/release-notes/3.12.2.md64
1 files changed, 0 insertions, 64 deletions
diff --git a/doc/release-notes/3.12.2.md b/doc/release-notes/3.12.2.md
deleted file mode 100644
index 3f0ab9abc17..00000000000
--- a/doc/release-notes/3.12.2.md
+++ /dev/null
@@ -1,64 +0,0 @@
-# Release notes for Gluster 3.12.2
-
-This is a bugfix release. The release notes for [3.12.0](3.12.0.md), [3.12.1](3.12.1.md),
-[3.12.2](3.12.2.md) contain a listing of all the new features that were added and bugs
-fixed in the GlusterFS 3.12 stable release.
-
-## Major changes, features and limitations addressed in this release
- 1.) In a pure distribute volume there is no source to heal the replaced brick
- from and hence would cause a loss of data that was present in the replaced brick.
- The CLI has been enhanced to prevent a user from inadvertently using replace brick
- in a pure distribute volume. It is advised to use add/remove brick to migrate from
- an existing brick in a pure distribute volume.
-
-## Major issues
-1. Expanding a gluster volume that is sharded may cause file corruption
- - Sharded volumes are typically used for VM images, if such volumes are
- expanded or possibly contracted (i.e add/remove bricks and rebalance) there
- are reports of VM images getting corrupted.
- - The last known cause for corruption #1465123 is still pending, and not yet
- part of this release.
-
-2. Gluster volume restarts fail if the sub directory export feature is in use.
- Status of this issue can be tracked here, #1501315
-
-3. Mounting a gluster snapshot will fail, when attempting a FUSE based mount of
- the snapshot. So for the current users, it is recommend to only access snapshot
- via ".snaps" directory on a mounted gluster volume.
- Status of this issue can be tracked here, #1501378
-
-## Bugs addressed
-
- A total of 31 patches have been merged, addressing 28 bugs
-
-
-- [#1490493](https://bugzilla.redhat.com/1490493): Sub-directory mount details are incorrect in /proc/mounts
-- [#1491178](https://bugzilla.redhat.com/1491178): GlusterD returns a bad memory pointer in glusterd_get_args_from_dict()
-- [#1491292](https://bugzilla.redhat.com/1491292): Provide brick list as part of VOLUME_CREATE event.
-- [#1491690](https://bugzilla.redhat.com/1491690): rpc: TLSv1_2_method() is deprecated in OpenSSL-1.1
-- [#1492026](https://bugzilla.redhat.com/1492026): set the shard-block-size to 64MB in virt profile
-- [#1492061](https://bugzilla.redhat.com/1492061): CLIENT_CONNECT event not being raised
-- [#1492066](https://bugzilla.redhat.com/1492066): AFR_SUBVOL_UP and AFR_SUBVOLS_DOWN events not working
-- [#1493975](https://bugzilla.redhat.com/1493975): disallow replace brick operation on plain distribute volume
-- [#1494523](https://bugzilla.redhat.com/1494523): Spelling errors in 3.12.1
-- [#1495162](https://bugzilla.redhat.com/1495162): glusterd ends up with multiple uuids for the same node
-- [#1495397](https://bugzilla.redhat.com/1495397): Make event-history feature configurable and have it disabled by default
-- [#1495858](https://bugzilla.redhat.com/1495858): gluster volume create asks for confirmation for replica-2 volume even with force
-- [#1496238](https://bugzilla.redhat.com/1496238): [geo-rep]: Scheduler help needs correction for description of --no-color
-- [#1496317](https://bugzilla.redhat.com/1496317): [afr] split-brain observed on T files post hardlink and rename in x3 volume
-- [#1496326](https://bugzilla.redhat.com/1496326): [GNFS+EC] lock is being granted to 2 different client for the same data range at a time after performing lock acquire/release from the clients1
-- [#1497084](https://bugzilla.redhat.com/1497084): glusterfs process consume huge memory on both server and client node
-- [#1499123](https://bugzilla.redhat.com/1499123): Readdirp is considerably slower than readdir on acl clients
-- [#1499150](https://bugzilla.redhat.com/1499150): Improve performance with xattrop update.
-- [#1499158](https://bugzilla.redhat.com/1499158): client-io-threads option not working for replicated volumes
-- [#1499202](https://bugzilla.redhat.com/1499202): self-heal daemon stuck
-- [#1499392](https://bugzilla.redhat.com/1499392): [geo-rep]: Improve the output message to reflect the real failure with schedule_georep script
-- [#1500396](https://bugzilla.redhat.com/1500396): [geo-rep]: Observed "Operation not supported" error with traceback on slave log
-- [#1500472](https://bugzilla.redhat.com/1500472): Use a bitmap to store local node info instead of conf->local_nodeuuids[i].uuids
-- [#1500662](https://bugzilla.redhat.com/1500662): gluster volume heal info "healed" and "heal-failed" showing wrong information
-- [#1500835](https://bugzilla.redhat.com/1500835): [geo-rep]: Status shows ACTIVE for most workers in EC before it becomes the PASSIVE
-- [#1500841](https://bugzilla.redhat.com/1500841): [geo-rep]: Worker crashes with OSError: [Errno 61] No data available
-- [#1500845](https://bugzilla.redhat.com/1500845): [geo-rep] master worker crash with interrupted system call
-- [#1500853](https://bugzilla.redhat.com/1500853): [geo-rep]: Incorrect last sync "0" during hystory crawl after upgrade/stop-start
-- [#1501022](https://bugzilla.redhat.com/1501022): Make choose-local configurable through `volume-set` command
-- [#1501154](https://bugzilla.redhat.com/1501154): Brick Multiplexing: Gluster volume start force complains with command "Error : Request timed out" when there are multiple volumes