<feed xmlns='http://www.w3.org/2005/Atom'>
<title>glusterfs.git/xlators/mgmt/glusterd/src/glusterd-rpc-ops.c, branch v3.7.9</title>
<subtitle></subtitle>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/'/>
<entry>
<title>glusterd: fixing few memory leak in glusterd</title>
<updated>2016-03-10T15:20:04+00:00</updated>
<author>
<name>Gaurav Kumar Garg</name>
<email>garg.gaurav52@gmail.com</email>
</author>
<published>2015-12-09T14:42:17+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=af7428490253421484e61409ddb77b653c4a5e63'/>
<id>af7428490253421484e61409ddb77b653c4a5e63</id>
<content type='text'>
This patch is backport of: http://review.gluster.org/#/c/12927/

Current glusterd code base having memory leak. This is because of
memory allocate by dict_allocate_and_serialize function in
"gd_syncop_mgmt_v3_lock" and "gd_syncop_mgmt_v3_unlock"
function is not freeing up memory upon exit.

Fix is to free the memory after exit of the above function.

Thanks Carlos and Roman for finding out the issue and fix.

 &gt;&gt; Change-Id: Id67aa794c84969830ca7ea8c2374f80c64d7a639
 &gt;&gt; BUG: 1287517
 &gt;&gt; Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
 &gt;&gt; Signed-off-by: Carlos Chinea &lt;carlos.chinea@nokia.com&gt;
 &gt;&gt; Signed-off-by: Roman Tereshonkov &lt;roman.tereshonkov@nokia.com&gt;
 &gt;&gt; Reviewed-on: http://review.gluster.org/12927
 &gt;&gt; Smoke: Gluster Build System &lt;jenkins@build.gluster.com&gt;
 &gt;&gt; NetBSD-regression: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
 &gt;&gt; CentOS-regression: Gluster Build System &lt;jenkins@build.gluster.com&gt;
 &gt;&gt; Reviewed-by: Jeff Darcy &lt;jdarcy@redhat.com&gt;

Change-Id: Id67aa794c84969830ca7ea8c2374f80c64d7a639
BUG: 1311377
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
(cherry picked from commit e38bf1bdeda3c7a89be3193ad62a72b9139358dd)
Reviewed-on: http://review.gluster.org/13503
Smoke: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
Tested-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
CentOS-regression: Gluster Build System &lt;jenkins@build.gluster.com&gt;
NetBSD-regression: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
This patch is backport of: http://review.gluster.org/#/c/12927/

Current glusterd code base having memory leak. This is because of
memory allocate by dict_allocate_and_serialize function in
"gd_syncop_mgmt_v3_lock" and "gd_syncop_mgmt_v3_unlock"
function is not freeing up memory upon exit.

Fix is to free the memory after exit of the above function.

Thanks Carlos and Roman for finding out the issue and fix.

 &gt;&gt; Change-Id: Id67aa794c84969830ca7ea8c2374f80c64d7a639
 &gt;&gt; BUG: 1287517
 &gt;&gt; Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
 &gt;&gt; Signed-off-by: Carlos Chinea &lt;carlos.chinea@nokia.com&gt;
 &gt;&gt; Signed-off-by: Roman Tereshonkov &lt;roman.tereshonkov@nokia.com&gt;
 &gt;&gt; Reviewed-on: http://review.gluster.org/12927
 &gt;&gt; Smoke: Gluster Build System &lt;jenkins@build.gluster.com&gt;
 &gt;&gt; NetBSD-regression: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
 &gt;&gt; CentOS-regression: Gluster Build System &lt;jenkins@build.gluster.com&gt;
 &gt;&gt; Reviewed-by: Jeff Darcy &lt;jdarcy@redhat.com&gt;

Change-Id: Id67aa794c84969830ca7ea8c2374f80c64d7a639
BUG: 1311377
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
(cherry picked from commit e38bf1bdeda3c7a89be3193ad62a72b9139358dd)
Reviewed-on: http://review.gluster.org/13503
Smoke: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
Tested-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
CentOS-regression: Gluster Build System &lt;jenkins@build.gluster.com&gt;
NetBSD-regression: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: cli command implementation for bitrot scrub status</title>
<updated>2015-11-23T03:53:58+00:00</updated>
<author>
<name>Gaurav Kumar Garg</name>
<email>garg.gaurav52@gmail.com</email>
</author>
<published>2015-11-20T08:30:38+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=164a8dda2cbf10862483e0333ebf7e727fc87f07'/>
<id>164a8dda2cbf10862483e0333ebf7e727fc87f07</id>
<content type='text'>
This patch is backport of: http://review.gluster.org/10231

CLI command  for bitrot scrub status will be :

gluster volume bitrot &lt;volname&gt; scrub status

Above command will show the statistics of bitrot scrubber.

Upon execution of this command it will show some common
scrubber tunable value of volume &lt;VOLNAME&gt; followed by
statistics of scrubber statistics of individual nodes.

sample ouput for single node:

Volume name : &lt;VOLNAME&gt;

State of scrub: Active

Scrub frequency: biweekly

Bitrot error log location: /var/log/glusterfs/bitd.log

Scrubber error log location: /var/log/glusterfs/scrub.log

=========================================================

Node name:

Number of Scrubbed files:

Number of Unsigned files:

Last completed scrub time:

Duration of last scrub:

Error count:

=========================================================

This is just infrastructure. list of bad file, last scrub
time, error count value will be taken care by
http://review.gluster.org/#/c/12503/  and
http://review.gluster.org/#/c/12654/ patches.

    &gt;&gt; Change-Id: I3ed3c7057c9d0c894233f4079a7f185d90c202d1
    &gt;&gt; BUG: 1207627
    &gt;&gt; Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
    &gt;&gt; Reviewed-on: http://review.gluster.org/10231
    &gt;&gt; Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
    &gt;&gt; Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
    &gt;&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;

Change-Id: I45ed94e5e0e78a1e007c30eb0b252f74cf3c9187
BUG: 1283881
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
Reviewed-on: http://review.gluster.org/12704
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
This patch is backport of: http://review.gluster.org/10231

CLI command  for bitrot scrub status will be :

gluster volume bitrot &lt;volname&gt; scrub status

Above command will show the statistics of bitrot scrubber.

Upon execution of this command it will show some common
scrubber tunable value of volume &lt;VOLNAME&gt; followed by
statistics of scrubber statistics of individual nodes.

sample ouput for single node:

Volume name : &lt;VOLNAME&gt;

State of scrub: Active

Scrub frequency: biweekly

Bitrot error log location: /var/log/glusterfs/bitd.log

Scrubber error log location: /var/log/glusterfs/scrub.log

=========================================================

Node name:

Number of Scrubbed files:

Number of Unsigned files:

Last completed scrub time:

Duration of last scrub:

Error count:

=========================================================

This is just infrastructure. list of bad file, last scrub
time, error count value will be taken care by
http://review.gluster.org/#/c/12503/  and
http://review.gluster.org/#/c/12654/ patches.

    &gt;&gt; Change-Id: I3ed3c7057c9d0c894233f4079a7f185d90c202d1
    &gt;&gt; BUG: 1207627
    &gt;&gt; Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
    &gt;&gt; Reviewed-on: http://review.gluster.org/10231
    &gt;&gt; Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
    &gt;&gt; Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
    &gt;&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;

Change-Id: I45ed94e5e0e78a1e007c30eb0b252f74cf3c9187
BUG: 1283881
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
Reviewed-on: http://review.gluster.org/12704
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>Tiering: change in status for remove brick and rebalance</title>
<updated>2015-09-22T12:18:59+00:00</updated>
<author>
<name>hari gowtham</name>
<email>hgowtham@redhat.com</email>
</author>
<published>2015-09-09T13:47:17+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=5d7dbda72bd35f1b8f52543503318eb049bfe0e7'/>
<id>5d7dbda72bd35f1b8f52543503318eb049bfe0e7</id>
<content type='text'>
        backport of : http://review.gluster.org/#/c/12149/

when we trigger a detach tier start on a tier vol,
it shows in the volume status task as "remove brick" instead of "Detach tier"

Status of volume: vol1
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Hot Bricks:
Brick 10.70.42.171:/data/gluster/hbr1       49154     0          Y       25098
Cold Bricks:
Brick 10.70.42.171:/data/gluster/p1         49152     0          Y       25101
Brick 10.70.42.171:/data/gluster/p2         49153     0          Y       25112
NFS Server on localhost                     N/A       N/A        N       N/A

Task Status of Volume vol1
------------------------------------------------------------------------------
Task                 : Tier migrate
ID                   : e11d5a3d-b1ae-4c3f-8f95-b28993c60939
Status               : in progress

Status of volume: vol1
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Hot Bricks:
Brick 10.70.42.171:/data/gluster/hbr1       49154     0          Y       25098
Cold Bricks:
Brick 10.70.42.171:/data/gluster/p1         49152     0          Y       25101
Brick 10.70.42.171:/data/gluster/p2         49153     0          Y       25112
NFS Server on localhost                     N/A       N/A        N       N/A

Task Status of Volume vol1
------------------------------------------------------------------------------
Task                 : Detach tier
ID                   : 76d700b1-5bbd-43ed-95fd-1640b2b4af31
Status               : completed

&gt;Change-Id: I4bd3b340d4e700e8afed00e1478b8a8b54dfe2e2
&gt;BUG: 1261837
&gt;Signed-off-by: hari gowtham &lt;hgowtham@redhat.com&gt;
&gt;Signed-off-by: Hari Gowtham &lt;hgowtham@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/12149
&gt;Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Reviewed-by: Dan Lambright &lt;dlambrig@redhat.com&gt;
&gt;Tested-by: Dan Lambright &lt;dlambrig@redhat.com&gt;

Change-Id: Ie0e994677a9277486a546e99da334bd4660b678b
BUG: 1258340
Signed-off-by: Hari Gowtham &lt;hgowtham@redhat.com&gt;
Reviewed-on: http://review.gluster.org/12203
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Dan Lambright &lt;dlambrig@redhat.com&gt;
Tested-by: Dan Lambright &lt;dlambrig@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
        backport of : http://review.gluster.org/#/c/12149/

when we trigger a detach tier start on a tier vol,
it shows in the volume status task as "remove brick" instead of "Detach tier"

Status of volume: vol1
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Hot Bricks:
Brick 10.70.42.171:/data/gluster/hbr1       49154     0          Y       25098
Cold Bricks:
Brick 10.70.42.171:/data/gluster/p1         49152     0          Y       25101
Brick 10.70.42.171:/data/gluster/p2         49153     0          Y       25112
NFS Server on localhost                     N/A       N/A        N       N/A

Task Status of Volume vol1
------------------------------------------------------------------------------
Task                 : Tier migrate
ID                   : e11d5a3d-b1ae-4c3f-8f95-b28993c60939
Status               : in progress

Status of volume: vol1
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Hot Bricks:
Brick 10.70.42.171:/data/gluster/hbr1       49154     0          Y       25098
Cold Bricks:
Brick 10.70.42.171:/data/gluster/p1         49152     0          Y       25101
Brick 10.70.42.171:/data/gluster/p2         49153     0          Y       25112
NFS Server on localhost                     N/A       N/A        N       N/A

Task Status of Volume vol1
------------------------------------------------------------------------------
Task                 : Detach tier
ID                   : 76d700b1-5bbd-43ed-95fd-1640b2b4af31
Status               : completed

&gt;Change-Id: I4bd3b340d4e700e8afed00e1478b8a8b54dfe2e2
&gt;BUG: 1261837
&gt;Signed-off-by: hari gowtham &lt;hgowtham@redhat.com&gt;
&gt;Signed-off-by: Hari Gowtham &lt;hgowtham@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/12149
&gt;Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Reviewed-by: Dan Lambright &lt;dlambrig@redhat.com&gt;
&gt;Tested-by: Dan Lambright &lt;dlambrig@redhat.com&gt;

Change-Id: Ie0e994677a9277486a546e99da334bd4660b678b
BUG: 1258340
Signed-off-by: Hari Gowtham &lt;hgowtham@redhat.com&gt;
Reviewed-on: http://review.gluster.org/12203
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Dan Lambright &lt;dlambrig@redhat.com&gt;
Tested-by: Dan Lambright &lt;dlambrig@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: getting txn_id from frame-&gt;cookie in op_sm call back</title>
<updated>2015-08-07T06:24:22+00:00</updated>
<author>
<name>anand</name>
<email>anekkunt@redhat.com</email>
</author>
<published>2015-07-21T10:12:24+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=90beb1602cb2926c14a9bd02653cd75b40661cd4'/>
<id>90beb1602cb2926c14a9bd02653cd75b40661cd4</id>
<content type='text'>
RCA: If rebalance start is triggered from one node and one of other nodes in the cluster goes down simultaneously
we might end up in a case where callback will use the txn_id from priv-&gt;global_txn_id which is always zeros and
this means injecting an event with an incorrect txn_id will result into op-sm getting stuck.

fix: set txn_id in frame-&gt;cookie during sumbit_and_request, so that we can get txn_id in call back
functions.
Back port of :
&gt;Change-Id: I519176c259ea9d37897791a77a7c92eb96d10052
&gt;BUG: 1245142
&gt;Signed-off-by: anand &lt;anekkunt@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/11728
&gt;Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
&gt;Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;(cherry picked from commit 65e6ab1bfbbec7755f7ae2294cb83334ac65a296)

Change-Id: I376d6c791b0200a8371f590d29c3e950658a02c7
BUG: 1249925
Reviewed-on: http://review.gluster.org/11823
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
RCA: If rebalance start is triggered from one node and one of other nodes in the cluster goes down simultaneously
we might end up in a case where callback will use the txn_id from priv-&gt;global_txn_id which is always zeros and
this means injecting an event with an incorrect txn_id will result into op-sm getting stuck.

fix: set txn_id in frame-&gt;cookie during sumbit_and_request, so that we can get txn_id in call back
functions.
Back port of :
&gt;Change-Id: I519176c259ea9d37897791a77a7c92eb96d10052
&gt;BUG: 1245142
&gt;Signed-off-by: anand &lt;anekkunt@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/11728
&gt;Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
&gt;Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;(cherry picked from commit 65e6ab1bfbbec7755f7ae2294cb83334ac65a296)

Change-Id: I376d6c791b0200a8371f590d29c3e950658a02c7
BUG: 1249925
Reviewed-on: http://review.gluster.org/11823
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: Get the local txn_info based on trans_id in op_sm call backs.</title>
<updated>2015-07-13T11:33:31+00:00</updated>
<author>
<name>anand</name>
<email>anekkunt@redhat.com</email>
</author>
<published>2015-06-07T18:49:00+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=0497fb6090d1ed34ac1bdcaaff49847025d13e9e'/>
<id>0497fb6090d1ed34ac1bdcaaff49847025d13e9e</id>
<content type='text'>
Issue: when two or more transactions are running concurrently in op_sm,
global op_info might get corrupted.

Fix: Get local txn_info based on trans_id instead of using global txn_info for
commands (re-balance, profile ) which are using op_sm in originator.

TODO: Handle errors properly in call backs and completely remove the global op_info from
op_sm.
Backport of:
&gt;Change-Id: I9d61388acc125841ddc77e2bd560cb7f17ae0a5a
&gt;BUG: 1229139
&gt;Signed-off-by: anand &lt;anekkunt@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/11120
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt;Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
(cherry picked from commit c9765bcb1557ab1e921080e7de4f3ebac1e424d5)

Change-Id: I4405a4a1cf4c853b5d2ce478fd65043e45a79826
BUG: 1230523
Signed-off-by: anand &lt;anekkunt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11557
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Issue: when two or more transactions are running concurrently in op_sm,
global op_info might get corrupted.

Fix: Get local txn_info based on trans_id instead of using global txn_info for
commands (re-balance, profile ) which are using op_sm in originator.

TODO: Handle errors properly in call backs and completely remove the global op_info from
op_sm.
Backport of:
&gt;Change-Id: I9d61388acc125841ddc77e2bd560cb7f17ae0a5a
&gt;BUG: 1229139
&gt;Signed-off-by: anand &lt;anekkunt@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/11120
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt;Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
(cherry picked from commit c9765bcb1557ab1e921080e7de4f3ebac1e424d5)

Change-Id: I4405a4a1cf4c853b5d2ce478fd65043e45a79826
BUG: 1230523
Signed-off-by: anand &lt;anekkunt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11557
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: Porting left out log messages to new framework</title>
<updated>2015-06-27T06:56:09+00:00</updated>
<author>
<name>Nandaja Varma</name>
<email>nandaja.varma@gmail.com</email>
</author>
<published>2015-06-24T19:27:00+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=8708953fa3d9187997dc6d484dae663b4469c7ca'/>
<id>8708953fa3d9187997dc6d484dae663b4469c7ca</id>
<content type='text'>
This is a backport of http://review.gluster.org/11388
cherry-picked from commit 23c1e6dc0fa86c014e1a8b6aa5729675f6d69017
&gt;Change-Id: I70d40ae3b5f49a21e1b93f82885cd58fa2723647
&gt;BUG: 1235538
&gt;Signed-off-by: Nandaja Varma &lt;nandaja.varma@gmail.com&gt;

Change-Id: I70d40ae3b5f49a21e1b93f82885cd58fa2723647
BUG: 1217722
Signed-off-by: Nandaja Varma &lt;nandaja.varma@gmail.com&gt;
Reviewed-on: http://review.gluster.org/11422
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Anand Nekkunti &lt;anekkunt@redhat.com&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
This is a backport of http://review.gluster.org/11388
cherry-picked from commit 23c1e6dc0fa86c014e1a8b6aa5729675f6d69017
&gt;Change-Id: I70d40ae3b5f49a21e1b93f82885cd58fa2723647
&gt;BUG: 1235538
&gt;Signed-off-by: Nandaja Varma &lt;nandaja.varma@gmail.com&gt;

Change-Id: I70d40ae3b5f49a21e1b93f82885cd58fa2723647
BUG: 1217722
Signed-off-by: Nandaja Varma &lt;nandaja.varma@gmail.com&gt;
Reviewed-on: http://review.gluster.org/11422
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Anand Nekkunti &lt;anekkunt@redhat.com&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>ops/glusterd: Porting messages to new logging framework</title>
<updated>2015-06-21T04:27:06+00:00</updated>
<author>
<name>Nandaja Varma</name>
<email>nandaja.varma@gmail.com</email>
</author>
<published>2015-03-17T07:46:06+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=38d602e820b99b0c9e2efe788df7450e4d863196'/>
<id>38d602e820b99b0c9e2efe788df7450e4d863196</id>
<content type='text'>
cherry-picked from commit 70bb8bf65b4cefb503fa94452c200b0b71a06535
&gt;Change-Id: Iafeb07aabc1781d98f51c6c2627bf3bbdf493153
&gt;BUG: 1194640
&gt;Signed-off-by: Nandaja Varma &lt;nandaja.varma@gmail.com&gt;
&gt;Reviewed-on: http://review.gluster.org/9905
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt;Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;

Change-Id: Iafeb07aabc1781d98f51c6c2627bf3bbdf493153
BUG: 1217722
Signed-off-by: Nandaja Varma &lt;nvarma@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11230
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
cherry-picked from commit 70bb8bf65b4cefb503fa94452c200b0b71a06535
&gt;Change-Id: Iafeb07aabc1781d98f51c6c2627bf3bbdf493153
&gt;BUG: 1194640
&gt;Signed-off-by: Nandaja Varma &lt;nandaja.varma@gmail.com&gt;
&gt;Reviewed-on: http://review.gluster.org/9905
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt;Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;

Change-Id: Iafeb07aabc1781d98f51c6c2627bf3bbdf493153
BUG: 1217722
Signed-off-by: Nandaja Varma &lt;nvarma@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11230
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: Removing sync lock and unlock inside rcu read critical section</title>
<updated>2015-06-19T09:49:44+00:00</updated>
<author>
<name>anand</name>
<email>anekkunt@redhat.com</email>
</author>
<published>2015-06-17T10:41:48+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=c5f0fddc27ecc79aad1ce104d5dd088051639539'/>
<id>c5f0fddc27ecc79aad1ce104d5dd088051639539</id>
<content type='text'>
Issue : Glsuterd was crashing during peer probe.

RCA :  In glusterd, we are using big lock which is implemented based on sync
task frame work for thread synchronization, sync task frame work swap the threads
if there is no worker pool threads available. Due to this rcu lock and rcu unlock
was happening  in different threads (urcu-bp will not allow this), resulting into
glusterd crash.

fix : Removing sync lock and unlock inside rcu read critical section, which was left out by
http://review.gluster.org/#/c/10285/ patch.

Upstream link : http://review.gluster.org/#/c/11276/

Change-Id: Id358dfcc797335bcd3b491c3129017b2caa826eb
BUG: 1233041
Signed-off-by: anand &lt;anekkunt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11306
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Issue : Glsuterd was crashing during peer probe.

RCA :  In glusterd, we are using big lock which is implemented based on sync
task frame work for thread synchronization, sync task frame work swap the threads
if there is no worker pool threads available. Due to this rcu lock and rcu unlock
was happening  in different threads (urcu-bp will not allow this), resulting into
glusterd crash.

fix : Removing sync lock and unlock inside rcu read critical section, which was left out by
http://review.gluster.org/#/c/10285/ patch.

Upstream link : http://review.gluster.org/#/c/11276/

Change-Id: Id358dfcc797335bcd3b491c3129017b2caa826eb
BUG: 1233041
Signed-off-by: anand &lt;anekkunt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11306
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd/shared_storage: Provide a volume set option to create and mount the shared storage</title>
<updated>2015-06-05T09:20:26+00:00</updated>
<author>
<name>Avra Sengupta</name>
<email>asengupt@redhat.com</email>
</author>
<published>2015-05-14T09:30:59+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=ae87a7fedfcf7f6b287ef5c3860f45412363e4f6'/>
<id>ae87a7fedfcf7f6b287ef5c3860f45412363e4f6</id>
<content type='text'>
     Backport of http://review.gluster.org/#/c/10793/

Introducing a global volume set option(cluster.enable-shared-storage)
which helps create and set-up the shared storage meta volume.

gluster volume set all cluster.enable-shared-storage enable

On enabling this option, the system analyzes the number of peers
in the cluster, which are currently connected, and chooses three
such peers(including the node the command is issued from). From these
peers a volume(gluster_shared_storage) is created. Depending on the
number of peers available the volume is either a replica 3
volume(if there are 3 connected peers),  or a replica 2 volume(if there
are 2 connected peers). "/var/run/gluster/ss_brick" serves as the
brick path on each node for the shared storage volume. We also mount
the shared storage at "/var/run/gluster/shared_storage" on all the nodes
in the cluster as part of enabling this option. If there is only one node
in the cluster,  or only one node is up then the command will fail

Once the volume is created, and mounted the maintainance of the
volume like adding-bricks, removing bricks etc., is expected to
be the onus of the user.

On disabling the option, we provide the user a warning, and on
affirmation from the user we stop the shared storage volume, and unmount
it from all the nodes in the cluster.

gluster volume set all cluster.enable-shared-storage disable

Change-Id: Idd92d67b93f444244f99ede9f634ef18d2945dbc
BUG: 1228181
Signed-off-by: Avra Sengupta &lt;asengupt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11086
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Rajesh Joseph &lt;rjoseph@redhat.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
     Backport of http://review.gluster.org/#/c/10793/

Introducing a global volume set option(cluster.enable-shared-storage)
which helps create and set-up the shared storage meta volume.

gluster volume set all cluster.enable-shared-storage enable

On enabling this option, the system analyzes the number of peers
in the cluster, which are currently connected, and chooses three
such peers(including the node the command is issued from). From these
peers a volume(gluster_shared_storage) is created. Depending on the
number of peers available the volume is either a replica 3
volume(if there are 3 connected peers),  or a replica 2 volume(if there
are 2 connected peers). "/var/run/gluster/ss_brick" serves as the
brick path on each node for the shared storage volume. We also mount
the shared storage at "/var/run/gluster/shared_storage" on all the nodes
in the cluster as part of enabling this option. If there is only one node
in the cluster,  or only one node is up then the command will fail

Once the volume is created, and mounted the maintainance of the
volume like adding-bricks, removing bricks etc., is expected to
be the onus of the user.

On disabling the option, we provide the user a warning, and on
affirmation from the user we stop the shared storage volume, and unmount
it from all the nodes in the cluster.

gluster volume set all cluster.enable-shared-storage disable

Change-Id: Idd92d67b93f444244f99ede9f634ef18d2945dbc
BUG: 1228181
Signed-off-by: Avra Sengupta &lt;asengupt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11086
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Rajesh Joseph &lt;rjoseph@redhat.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd/snapshot: Return correct errno in events of failure - PATCH 1</title>
<updated>2015-06-01T07:18:15+00:00</updated>
<author>
<name>Avra Sengupta</name>
<email>asengupt@redhat.com</email>
</author>
<published>2015-05-05T05:28:10+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=e9a37d5a3471db02e1c7922cee93bf290e2f326d'/>
<id>e9a37d5a3471db02e1c7922cee93bf290e2f326d</id>
<content type='text'>
       Backport of patch http://review.gluster.org/#/c/10313/

RETCODE              ERROR
-------------------------------------------
30800                Internal Error
30801                Another Transaction In Progress

&gt;Change-Id: Ica7fd2e513b2c28717b6df73cfb2667725dbf057
&gt;BUG: 1226117
&gt;Signed-off-by: Avra Sengupta &lt;asengupt@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/10313
&gt;Reviewed-by: Rajesh Joseph &lt;rjoseph@redhat.com&gt;
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Tested-by: NetBSD Build System
&gt;Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
&gt;(cherry picked from commit 4397d7e72fdff6f01c59b72eebea421f23c1a392)

Change-Id: Iace6ac0f150919cead94e6c5d99a23d28d45046e
BUG: 1226117
Signed-off-by: Avra Sengupta &lt;asengupt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11011
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
       Backport of patch http://review.gluster.org/#/c/10313/

RETCODE              ERROR
-------------------------------------------
30800                Internal Error
30801                Another Transaction In Progress

&gt;Change-Id: Ica7fd2e513b2c28717b6df73cfb2667725dbf057
&gt;BUG: 1226117
&gt;Signed-off-by: Avra Sengupta &lt;asengupt@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/10313
&gt;Reviewed-by: Rajesh Joseph &lt;rjoseph@redhat.com&gt;
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Tested-by: NetBSD Build System
&gt;Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
&gt;(cherry picked from commit 4397d7e72fdff6f01c59b72eebea421f23c1a392)

Change-Id: Iace6ac0f150919cead94e6c5d99a23d28d45046e
BUG: 1226117
Signed-off-by: Avra Sengupta &lt;asengupt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11011
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</pre>
</div>
</content>
</entry>
</feed>
