<feed xmlns='http://www.w3.org/2005/Atom'>
<title>glusterfs.git, branch v3.6.5</title>
<subtitle></subtitle>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/'/>
<entry>
<title>features/quota: prevent statfs frame-loss when an error happens during</title>
<updated>2015-08-20T09:04:42+00:00</updated>
<author>
<name>vmallika</name>
<email>vmallika@redhat.com</email>
</author>
<published>2015-07-29T10:49:12+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=dfa2bfb289cc73ade0e441f2e2ee88d0d819d48d'/>
<id>dfa2bfb289cc73ade0e441f2e2ee88d0d819d48d</id>
<content type='text'>
ancestry building.

This is a backport of http://review.gluster.org/#/c/9380/

We do quota_build_ancestry in function 'quota_get_limit_dir',
suppose if quota_build_ancestry fails, then we don't have a
frame saved to continue the statfs FOP and client can hang.

&gt; Change-Id: I92e25c1510d09444b9d4810afdb6b2a69dcd92c0
&gt; BUG: 1178619
&gt; Signed-off-by: Raghavendra G &lt;rgowdapp@redhat.com&gt;
&gt; Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/9380
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;

Change-Id: Ia25cf738250fdc2c766f96c26e3c31093d534aba
BUG: 1247959
Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11790
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
Reviewed-by: Raghavendra G &lt;rgowdapp@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
ancestry building.

This is a backport of http://review.gluster.org/#/c/9380/

We do quota_build_ancestry in function 'quota_get_limit_dir',
suppose if quota_build_ancestry fails, then we don't have a
frame saved to continue the statfs FOP and client can hang.

&gt; Change-Id: I92e25c1510d09444b9d4810afdb6b2a69dcd92c0
&gt; BUG: 1178619
&gt; Signed-off-by: Raghavendra G &lt;rgowdapp@redhat.com&gt;
&gt; Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/9380
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;

Change-Id: Ia25cf738250fdc2c766f96c26e3c31093d534aba
BUG: 1247959
Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11790
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
Reviewed-by: Raghavendra G &lt;rgowdapp@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>posix: fix mem-leak in posix xattrop</title>
<updated>2015-08-20T07:26:28+00:00</updated>
<author>
<name>vmallika</name>
<email>vmallika@redhat.com</email>
</author>
<published>2015-07-29T11:07:13+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=620403b48a021593883f39dec3a1d6bc16c53517'/>
<id>620403b48a021593883f39dec3a1d6bc16c53517</id>
<content type='text'>
This is a backport of http://review.gluster.org/#/c/11700/

&gt; Change-Id: I1dd70f74a98c5875eb316f3c3e560047f128685b
&gt; BUG: 1243890
&gt; Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/11700
&gt; Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt; Reviewed-by: Pranith Kumar Karampuri &lt;pkarampu@redhat.com&gt;

Change-Id: Ia192342f01330f8b2c7a520a7cc749471af7d85f
BUG: 1247970
Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11794
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
This is a backport of http://review.gluster.org/#/c/11700/

&gt; Change-Id: I1dd70f74a98c5875eb316f3c3e560047f128685b
&gt; BUG: 1243890
&gt; Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/11700
&gt; Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt; Reviewed-by: Pranith Kumar Karampuri &lt;pkarampu@redhat.com&gt;

Change-Id: Ia192342f01330f8b2c7a520a7cc749471af7d85f
BUG: 1247970
Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11794
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>nfs: make it possible to disable nfs.mount-rmtab</title>
<updated>2015-08-18T13:19:48+00:00</updated>
<author>
<name>Niels de Vos</name>
<email>ndevos@redhat.com</email>
</author>
<published>2015-06-21T10:00:52+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=21643f8427be22ab7e512acf6c6368eb8af1ec9d'/>
<id>21643f8427be22ab7e512acf6c6368eb8af1ec9d</id>
<content type='text'>
When there are many NFS-clients doing very often mount/unmount actions,
the updating of the 'rmtab' can become a bottleneck and cause delays. In
these situations, the output of 'showmount' may be less important than
the responsiveness of the (un)mounting.

By setting 'nfs.mount-rmtab' to the value "/-", the cache file is not
updated anymore, and the entries are only kept in memory.

Cherry picked from commit 40407afb529f6e5fa2f79e9778c2f527122d75eb:
&gt; Cherry picked from commit 331ef6e1a86bfc0a93f8a9dec6ad35c417873849:
&gt;&gt; BUG: 1169317
&gt;&gt; Change-Id: I40c4d8d754932f86fb2b1b2588843390464c773d
&gt;&gt; Reported-by: Cyril Peponnet &lt;cyril@peponnet.fr&gt;
&gt;&gt; Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
&gt;&gt; Reviewed-on: http://review.gluster.org/9223
&gt;&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;&gt; Reviewed-by: soumya k &lt;skoduri@redhat.com&gt;
&gt;&gt; Reviewed-by: jiffin tony Thottan &lt;jthottan@redhat.com&gt;
&gt;&gt; Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
&gt;
&gt; This change also contains the fixes to the test-case from:
&gt;&gt;
&gt;&gt; nfs: fix spurious failure in bug-1166862.t
&gt;&gt;
&gt;&gt; In some environments, "showmount" could return an NFS-client that does
&gt;&gt; not start with "1". This would cause the test-case to fail. The check is
&gt;&gt; incorrect, the number of lines should get counted instead.
&gt;&gt;
&gt;&gt; Also moving the test-case to the .../nfs/... subdirectory.
&gt;&gt;
&gt;&gt; Cherry picked from commit ee9b35a780607daddc2832b9af5ed6bf414aebc0:
&gt;&gt; BUG: 1166862
&gt;&gt; Change-Id: Ic03aa8145ca57d78aea01564466e924b03bb302a
&gt;&gt; Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
&gt;&gt; Reviewed-on: http://review.gluster.org/10419
&gt;&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;&gt; Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
&gt;&gt;
&gt;
&gt; Change-Id: I40c4d8d754932f86fb2b1b2588843390464c773d
&gt; BUG: 1215385
&gt; Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/10379
&gt; Tested-by: NetBSD Build System
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt; Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;

Change-Id: I40c4d8d754932f86fb2b1b2588843390464c773d
BUG: 1234096
Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11335
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
When there are many NFS-clients doing very often mount/unmount actions,
the updating of the 'rmtab' can become a bottleneck and cause delays. In
these situations, the output of 'showmount' may be less important than
the responsiveness of the (un)mounting.

By setting 'nfs.mount-rmtab' to the value "/-", the cache file is not
updated anymore, and the entries are only kept in memory.

Cherry picked from commit 40407afb529f6e5fa2f79e9778c2f527122d75eb:
&gt; Cherry picked from commit 331ef6e1a86bfc0a93f8a9dec6ad35c417873849:
&gt;&gt; BUG: 1169317
&gt;&gt; Change-Id: I40c4d8d754932f86fb2b1b2588843390464c773d
&gt;&gt; Reported-by: Cyril Peponnet &lt;cyril@peponnet.fr&gt;
&gt;&gt; Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
&gt;&gt; Reviewed-on: http://review.gluster.org/9223
&gt;&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;&gt; Reviewed-by: soumya k &lt;skoduri@redhat.com&gt;
&gt;&gt; Reviewed-by: jiffin tony Thottan &lt;jthottan@redhat.com&gt;
&gt;&gt; Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
&gt;
&gt; This change also contains the fixes to the test-case from:
&gt;&gt;
&gt;&gt; nfs: fix spurious failure in bug-1166862.t
&gt;&gt;
&gt;&gt; In some environments, "showmount" could return an NFS-client that does
&gt;&gt; not start with "1". This would cause the test-case to fail. The check is
&gt;&gt; incorrect, the number of lines should get counted instead.
&gt;&gt;
&gt;&gt; Also moving the test-case to the .../nfs/... subdirectory.
&gt;&gt;
&gt;&gt; Cherry picked from commit ee9b35a780607daddc2832b9af5ed6bf414aebc0:
&gt;&gt; BUG: 1166862
&gt;&gt; Change-Id: Ic03aa8145ca57d78aea01564466e924b03bb302a
&gt;&gt; Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
&gt;&gt; Reviewed-on: http://review.gluster.org/10419
&gt;&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;&gt; Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
&gt;&gt;
&gt;
&gt; Change-Id: I40c4d8d754932f86fb2b1b2588843390464c773d
&gt; BUG: 1215385
&gt; Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/10379
&gt; Tested-by: NetBSD Build System
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt; Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;

Change-Id: I40c4d8d754932f86fb2b1b2588843390464c773d
BUG: 1234096
Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11335
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: really get the inode size for a brick</title>
<updated>2015-08-18T12:55:17+00:00</updated>
<author>
<name>Niels de Vos</name>
<email>ndevos@redhat.com</email>
</author>
<published>2014-08-15T09:13:46+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=aefd42fdb74a9b276d07a51a84194805169e632e'/>
<id>aefd42fdb74a9b276d07a51a84194805169e632e</id>
<content type='text'>
backport of http://review.gluster.org/#/c/8492/

The device to get the inode size from does not get passed to the tool
(tune2fs, xfs_info or the like) that is called. This is probably just an
oversight. While correcting this, cleanup some bits of the function too.

Change-Id: Ida45852cba061631fb304bc7dd5286df1a808010
BUG: 1254421
Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Reviewed-on: http://review.gluster.org/8492
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
Tested-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11941
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
backport of http://review.gluster.org/#/c/8492/

The device to get the inode size from does not get passed to the tool
(tune2fs, xfs_info or the like) that is called. This is probably just an
oversight. While correcting this, cleanup some bits of the function too.

Change-Id: Ida45852cba061631fb304bc7dd5286df1a808010
BUG: 1254421
Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Reviewed-on: http://review.gluster.org/8492
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
Tested-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11941
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>quota: fix mem leak in quota enforcer</title>
<updated>2015-08-18T10:39:05+00:00</updated>
<author>
<name>vmallika</name>
<email>vmallika@redhat.com</email>
</author>
<published>2015-07-29T10:56:12+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=df3f1d9926bf3e7a481391182aa967f0154634a3'/>
<id>df3f1d9926bf3e7a481391182aa967f0154634a3</id>
<content type='text'>
This is a backport of http://review.gluster.org/#/c/11526/

Do inode_unref on parent

&gt; Change-Id: I21d82eb8716dd73aa2dc291b3ae8506e4fb4ea8b
&gt; BUG: 1207735
&gt; Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/11526
&gt; Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt; Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt; Reviewed-by: Raghavendra G &lt;rgowdapp@redhat.com&gt;
&gt; Tested-by: Raghavendra G &lt;rgowdapp@redhat.com&gt;

Change-Id: I8f103ea96e6853d48c934b791ff4d95b226e4453
BUG: 1247964
Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11793
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
This is a backport of http://review.gluster.org/#/c/11526/

Do inode_unref on parent

&gt; Change-Id: I21d82eb8716dd73aa2dc291b3ae8506e4fb4ea8b
&gt; BUG: 1207735
&gt; Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/11526
&gt; Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt; Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt; Reviewed-by: Raghavendra G &lt;rgowdapp@redhat.com&gt;
&gt; Tested-by: Raghavendra G &lt;rgowdapp@redhat.com&gt;

Change-Id: I8f103ea96e6853d48c934b791ff4d95b226e4453
BUG: 1247964
Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11793
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>xml output: Fix non-uniform opErrstr xml output</title>
<updated>2015-08-18T10:34:03+00:00</updated>
<author>
<name>Avra Sengupta</name>
<email>asengupt@redhat.com</email>
</author>
<published>2015-08-05T11:06:06+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=a17257727ce00b25ea88800c0dd54deec4d999ba'/>
<id>a17257727ce00b25ea88800c0dd54deec4d999ba</id>
<content type='text'>
    Backport of http://review.gluster.org/#/c/11835/

Display &lt;opErrstr/&gt; in case of no operrstr for
all xml output of gluster commands

Change-Id: Ie16f749f90b4642357c562012408c434cd38661f
BUG: 1218732
Signed-off-by: Avra Sengupta &lt;asengupt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11921
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
Reviewed-by: mohammed rafi  kc &lt;rkavunga@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
    Backport of http://review.gluster.org/#/c/11835/

Display &lt;opErrstr/&gt; in case of no operrstr for
all xml output of gluster commands

Change-Id: Ie16f749f90b4642357c562012408c434cd38661f
BUG: 1218732
Signed-off-by: Avra Sengupta &lt;asengupt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11921
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
Reviewed-by: mohammed rafi  kc &lt;rkavunga@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: fix op-version bump up flow</title>
<updated>2015-08-14T08:49:23+00:00</updated>
<author>
<name>Atin Mukherjee</name>
<email>amukherj@redhat.com</email>
</author>
<published>2015-08-06T06:48:10+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=fb601c73d366376711f5f164924a7655982e887f'/>
<id>fb601c73d366376711f5f164924a7655982e887f</id>
<content type='text'>
Backport of http://review.gluster.org/#/c/11798/

If a cluster is upgraded from 3.5 to latest version, gluster volume set all
cluster.op-version &lt;VERSION&gt; will throw an error message back to the user saying
unlocking failed. This is because of trying to release a volume wise lock in
unlock phase as the lock was taken cluster wide. The problem surfaced because
the op-version is updated in commit phase and unlocking works in the v3
framework where it should have used cluster unlock. Fix is to decide which
lock/unlock is to be followed before invoking lock phase

Change-Id: I6a82251d88cfc7af36c7deec511c634a09e55004
BUG: 1250836
Signed-off-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11849
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Backport of http://review.gluster.org/#/c/11798/

If a cluster is upgraded from 3.5 to latest version, gluster volume set all
cluster.op-version &lt;VERSION&gt; will throw an error message back to the user saying
unlocking failed. This is because of trying to release a volume wise lock in
unlock phase as the lock was taken cluster wide. The problem surfaced because
the op-version is updated in commit phase and unlocking works in the v3
framework where it should have used cluster unlock. Fix is to decide which
lock/unlock is to be followed before invoking lock phase

Change-Id: I6a82251d88cfc7af36c7deec511c634a09e55004
BUG: 1250836
Signed-off-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11849
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>nfs: a unix-domain-socket should not be created as fifo</title>
<updated>2015-07-23T10:03:02+00:00</updated>
<author>
<name>Niels de Vos</name>
<email>ndevos@redhat.com</email>
</author>
<published>2015-07-17T08:16:54+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=a28d161733e670d264999c6f486915bf738a7bfd'/>
<id>a28d161733e670d264999c6f486915bf738a7bfd</id>
<content type='text'>
Cherry picked from commit 381afd4f480533f212e1f62a0ccfd9dcf7248e17:
&gt; Change-Id: Ic6a23165df1703b330636a059967c3c674dbde57
&gt; BUG: 1235231
&gt; Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/11355
&gt; Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt; Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;

Change-Id: Ic6a23165df1703b330636a059967c3c674dbde57
BUG: 1244117
Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11709
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Cherry picked from commit 381afd4f480533f212e1f62a0ccfd9dcf7248e17:
&gt; Change-Id: Ic6a23165df1703b330636a059967c3c674dbde57
&gt; BUG: 1235231
&gt; Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/11355
&gt; Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt; Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;

Change-Id: Ic6a23165df1703b330636a059967c3c674dbde57
BUG: 1244117
Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11709
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>rpc-transport: socket_poller fixes for proper working of mgmt encryption</title>
<updated>2015-07-17T07:18:41+00:00</updated>
<author>
<name>Kaushal M</name>
<email>kaushal@redhat.com</email>
</author>
<published>2015-07-13T10:46:00+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=6263e73688ba2be49fd55f4c6a2d12609ce45f91'/>
<id>6263e73688ba2be49fd55f4c6a2d12609ce45f91</id>
<content type='text'>
  Backport of 8c39f14 from master

socket_poller, the polling function used by ssl own_thread, had two
issues which lead to GlusterD crashes when using management encryption

Issue 1
-------
socket_poller calls functions which require THIS to be set. But, THIS
was being set conditionally. Because of this, functions could sometimes
be called without THIS being set. For example, rpc_transport_notify
could be called for an accepted client socket without THIS being set, as
THIS was only set it the transport wasn't yet connected. This would
cause the process to crash when THIS was accessed by the called
functions.

To fix this, THIS is being set at the start of socket_poller
unconditionally.

Issue 2
-------
DISCONNECT notify was being sent on the listener transport instead of
the client transport. The DISCONNECT event was converted to a
LISTENER_DEAD event in rpcsvc_handle_disconnect, as it could not find
the listener socket of the listener socket. GlusterD was notified of a
LISTENER_DEAD event instead of a DISCONNECT and failed to remove the
client transport from its xprt_list. The transport would subsequently
be freed, leaving the xprt_list with a corrupted/invalid entry. Later,
when GlusterD would iterate over the xprt_list to send notifications, it
would crash when the invalid entry was accessed.

To fix this, DISCONNECT notification in socket_poller is sent on the
client socket, as it is done in the epoll handler.

Change-Id: I0370b7c6d7eb13de10ebf08d91a4a39dc7d64c7a
BUG: 1243700
Signed-off-by: Kaushal M &lt;kaushal@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11690
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra G &lt;rgowdapp@redhat.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
  Backport of 8c39f14 from master

socket_poller, the polling function used by ssl own_thread, had two
issues which lead to GlusterD crashes when using management encryption

Issue 1
-------
socket_poller calls functions which require THIS to be set. But, THIS
was being set conditionally. Because of this, functions could sometimes
be called without THIS being set. For example, rpc_transport_notify
could be called for an accepted client socket without THIS being set, as
THIS was only set it the transport wasn't yet connected. This would
cause the process to crash when THIS was accessed by the called
functions.

To fix this, THIS is being set at the start of socket_poller
unconditionally.

Issue 2
-------
DISCONNECT notify was being sent on the listener transport instead of
the client transport. The DISCONNECT event was converted to a
LISTENER_DEAD event in rpcsvc_handle_disconnect, as it could not find
the listener socket of the listener socket. GlusterD was notified of a
LISTENER_DEAD event instead of a DISCONNECT and failed to remove the
client transport from its xprt_list. The transport would subsequently
be freed, leaving the xprt_list with a corrupted/invalid entry. Later,
when GlusterD would iterate over the xprt_list to send notifications, it
would crash when the invalid entry was accessed.

To fix this, DISCONNECT notification in socket_poller is sent on the
client socket, as it is done in the epoll handler.

Change-Id: I0370b7c6d7eb13de10ebf08d91a4a39dc7d64c7a
BUG: 1243700
Signed-off-by: Kaushal M &lt;kaushal@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11690
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Raghavendra G &lt;rgowdapp@redhat.com&gt;
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cluster/afr: Pick gfid from poststat during fresh lookup for read child calculation</title>
<updated>2015-07-17T07:05:36+00:00</updated>
<author>
<name>Krutika Dhananjay</name>
<email>kdhananj@redhat.com</email>
</author>
<published>2015-06-24T02:32:51+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=f9b3ba22f105772e268765d80adfeffd8f6f2776'/>
<id>f9b3ba22f105772e268765d80adfeffd8f6f2776</id>
<content type='text'>
        Backport of : http://review.gluster.org/11373

Change-Id: I03f11af082a0decf4ea084480b67e9e156964c76
BUG: 1235601
Signed-off-by: Krutika Dhananjay &lt;kdhananj@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11408
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
Reviewed-by: Pranith Kumar Karampuri &lt;pkarampu@redhat.com&gt;
Tested-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
        Backport of : http://review.gluster.org/11373

Change-Id: I03f11af082a0decf4ea084480b67e9e156964c76
BUG: 1235601
Signed-off-by: Krutika Dhananjay &lt;kdhananj@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11408
Reviewed-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
Reviewed-by: Pranith Kumar Karampuri &lt;pkarampu@redhat.com&gt;
Tested-by: Raghavendra Bhat &lt;raghavendra@redhat.com&gt;
</pre>
</div>
</content>
</entry>
</feed>
