<feed xmlns='http://www.w3.org/2005/Atom'>
<title>glusterfs.git/cli/src, branch v3.7.3</title>
<subtitle></subtitle>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/'/>
<entry>
<title>cli: Create missing logbuf_pool</title>
<updated>2015-07-24T11:19:30+00:00</updated>
<author>
<name>Anoop C S</name>
<email>anoopcs@redhat.com</email>
</author>
<published>2015-07-16T08:26:51+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=c12234ee25deafe148a1557c30a0e56688548f91'/>
<id>c12234ee25deafe148a1557c30a0e56688548f91</id>
<content type='text'>
        Backport of http://review.gluster.org/#/c/11691/

Since logbuf_pool was not created via glusterfs_ctx_defaults_init(),
the following error was present in cli logs repeateadly for each
and every execution of a gluster command.

E [mem-pool.c:417:mem_get0] (--&gt;/usr/local/lib/libglusterfs.so.0(+0x7e262)
[0x7fdbc0b1f262] --&gt;/usr/local/lib/libglusterfs.so.0(_gf_msg+0x804)
[0x7fdbc0ac7844] --&gt;/usr/local/lib/libglusterfs.so.0(mem_get0+0x78)
[0x7fdbc0af5b48] ) 0-mem-pool: invalid argument [Invalid argument]

This change creates ctx-&gt;logbuf_pool via glusterfs_ctx_defaults_init()
in cli.c so that the above error is no longer logged in cli logs.

Change-Id: I3fcd9cfefa06ddd52e1989b039ff5637372c3235
BUG: 1244103
Signed-off-by: Anoop C S &lt;anoopcs@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11705
Reviewed-by: Jeff Darcy &lt;jdarcy@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
        Backport of http://review.gluster.org/#/c/11691/

Since logbuf_pool was not created via glusterfs_ctx_defaults_init(),
the following error was present in cli logs repeateadly for each
and every execution of a gluster command.

E [mem-pool.c:417:mem_get0] (--&gt;/usr/local/lib/libglusterfs.so.0(+0x7e262)
[0x7fdbc0b1f262] --&gt;/usr/local/lib/libglusterfs.so.0(_gf_msg+0x804)
[0x7fdbc0ac7844] --&gt;/usr/local/lib/libglusterfs.so.0(mem_get0+0x78)
[0x7fdbc0af5b48] ) 0-mem-pool: invalid argument [Invalid argument]

This change creates ctx-&gt;logbuf_pool via glusterfs_ctx_defaults_init()
in cli.c so that the above error is no longer logged in cli logs.

Change-Id: I3fcd9cfefa06ddd52e1989b039ff5637372c3235
BUG: 1244103
Signed-off-by: Anoop C S &lt;anoopcs@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11705
Reviewed-by: Jeff Darcy &lt;jdarcy@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>quota/marker: fix spurious failure afr-quota-xattr-mdata-heal.t</title>
<updated>2015-07-13T06:53:27+00:00</updated>
<author>
<name>vmallika</name>
<email>vmallika@redhat.com</email>
</author>
<published>2015-07-10T07:59:13+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=1509f6f2815cb75ecf45abef6d823c8cb30ed47b'/>
<id>1509f6f2815cb75ecf45abef6d823c8cb30ed47b</id>
<content type='text'>
This is a backport of http://review.gluster.org/#/c/11583

During quota-update process if inode info is present in size-xattr and
missing in contri-xattrs, then in function '_mq_get_metadata', we set
contri-size as zero (on error -2, which means usage info present, but
inode info missing).
With this we are calculating wrong delta and updating the same.

With this patch we are ignoring errors if inode info in xattrs are
missing

&gt; Change-Id: I7940a0e299b8bb425b5b43746b1f13f775c7fb92
&gt; BUG: 1241153
&gt; Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;

Change-Id: Ie85fa84b5362ae179cc43402bd6a3a6d96a04b81
BUG: 1241831
Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11614
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
This is a backport of http://review.gluster.org/#/c/11583

During quota-update process if inode info is present in size-xattr and
missing in contri-xattrs, then in function '_mq_get_metadata', we set
contri-size as zero (on error -2, which means usage info present, but
inode info missing).
With this we are calculating wrong delta and updating the same.

With this patch we are ignoring errors if inode info in xattrs are
missing

&gt; Change-Id: I7940a0e299b8bb425b5b43746b1f13f775c7fb92
&gt; BUG: 1241153
&gt; Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;

Change-Id: Ie85fa84b5362ae179cc43402bd6a3a6d96a04b81
BUG: 1241831
Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11614
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>common-ha : Fixing add node operation</title>
<updated>2015-06-22T09:37:11+00:00</updated>
<author>
<name>Meghana M</name>
<email>mmadhusu@redhat.com</email>
</author>
<published>2015-06-18T01:09:41+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=cda237fa33c275bb82f82f6d9bfa288d30a7f4ef'/>
<id>cda237fa33c275bb82f82f6d9bfa288d30a7f4ef</id>
<content type='text'>
Resource create for the added node referenced a variable
new_node that was never passed. This led to a wrong schema
type in the cib file and hence the added node always ended
up in failed state. And also, resources were wrongly
created twice and led to more errors. I have fixed the variable
name and deleted the repetitive invocation of the recreate-resource
function.

The new node has to be added to the existing ganesha-ha config
file for correct behaviour during subsequent add-node operations.
This edited file has to be copied to all the other cluster nodes.
I have added a fix for this as well.

This is a backport of the fix that is merged on master,
http://review.gluster.org/#/c/11316/5

Change-Id: I21e3fcffe8274986be53766e12a2c049eaa2da09
BUG: 1234216
Signed-off-by: Meghana M &lt;mmadhusu@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11337
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: soumya k &lt;skoduri@redhat.com&gt;
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Resource create for the added node referenced a variable
new_node that was never passed. This led to a wrong schema
type in the cib file and hence the added node always ended
up in failed state. And also, resources were wrongly
created twice and led to more errors. I have fixed the variable
name and deleted the repetitive invocation of the recreate-resource
function.

The new node has to be added to the existing ganesha-ha config
file for correct behaviour during subsequent add-node operations.
This edited file has to be copied to all the other cluster nodes.
I have added a fix for this as well.

This is a backport of the fix that is merged on master,
http://review.gluster.org/#/c/11316/5

Change-Id: I21e3fcffe8274986be53766e12a2c049eaa2da09
BUG: 1234216
Signed-off-by: Meghana M &lt;mmadhusu@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11337
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: soumya k &lt;skoduri@redhat.com&gt;
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: Fix snapshot of a volume with geo-rep</title>
<updated>2015-06-19T06:37:31+00:00</updated>
<author>
<name>Kotresh HR</name>
<email>khiremat@redhat.com</email>
</author>
<published>2015-06-15T11:50:46+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=84479bfa28f91c993b4e4a1f0f259a3ae1576791'/>
<id>84479bfa28f91c993b4e4a1f0f259a3ae1576791</id>
<content type='text'>
Problem:
    Snapshot fails for a volume configured with geo-rep
    if geo-rep is created with root user with the following
    syntax.
    gluster vol geo-rep &lt;master_vol&gt; root@&lt;slave_host&gt;::&lt;slave_vol&gt;

    It works fine if created with following syntax.
    gluster vol geo-rep &lt;master_vol&gt; &lt;slave_host&gt;::&lt;slave_vol&gt;

Cause:
    Geo-rep maintains persistent dictionary of slave associated
    with master volume. The dictionary saves the slave info
    along with 'root@' as sent from cli.
    Snapshot while constructing the working dir path to copy
    configuration files, constructs using this dictionary.
    But the actual working dir is created with out considering
    'root@'. Hence the issue.

Fix:
    Fix is done at two layers.
    1. Parse and negelect 'root@' in cli itself.
    2. For existing geo-rep sessions and upgrade scenarios,
       parse and neglect 'root@' in snapshot code as well.

BUG: 1233056
Change-Id: Ic666e43a1c4ac2f3f0fb957caa77770e1eee4129
Signed-off-by: Kotresh HR &lt;khiremat@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11233
Reviewed-on: http://review.gluster.org/11310
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Problem:
    Snapshot fails for a volume configured with geo-rep
    if geo-rep is created with root user with the following
    syntax.
    gluster vol geo-rep &lt;master_vol&gt; root@&lt;slave_host&gt;::&lt;slave_vol&gt;

    It works fine if created with following syntax.
    gluster vol geo-rep &lt;master_vol&gt; &lt;slave_host&gt;::&lt;slave_vol&gt;

Cause:
    Geo-rep maintains persistent dictionary of slave associated
    with master volume. The dictionary saves the slave info
    along with 'root@' as sent from cli.
    Snapshot while constructing the working dir path to copy
    configuration files, constructs using this dictionary.
    But the actual working dir is created with out considering
    'root@'. Hence the issue.

Fix:
    Fix is done at two layers.
    1. Parse and negelect 'root@' in cli itself.
    2. For existing geo-rep sessions and upgrade scenarios,
       parse and neglect 'root@' in snapshot code as well.

BUG: 1233056
Change-Id: Ic666e43a1c4ac2f3f0fb957caa77770e1eee4129
Signed-off-by: Kotresh HR &lt;khiremat@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11233
Reviewed-on: http://review.gluster.org/11310
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>NFS-Ganesha: Automatically export vol that was exported before vol restart</title>
<updated>2015-06-18T17:24:42+00:00</updated>
<author>
<name>Meghana M</name>
<email>mmadhusu@redhat.com</email>
</author>
<published>2015-06-15T08:43:44+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=0e358fa5f7190f84574824b5629835abf32823a3'/>
<id>0e358fa5f7190f84574824b5629835abf32823a3</id>
<content type='text'>
Consider a volume that is exported via NFS-Ganesha. Stopping this
volume will automatically unexport the volume. Starting this volume
should automatically export it. Although the logic was already there,
there was a bug in it. Fixing the same by introducing a hook script.

Also with the new CLI options, the hook script S31ganesha-set.sh
is no longer required. Hence, removing the same.
Adding a comment to tell the user that one of the CLI
commands will take a few minutes to complete.

This is a backport of the patch merged on master,
http://review.gluster.org/#/c/11247/

Change-Id: I1c16a978b9e2093a4298ea3024d031fd1a5bb577
BUG: 1232335
Signed-off-by: Meghana Madhusudhan &lt;mmadhusu@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11259
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
Reviewed-by: soumya k &lt;skoduri@redhat.com&gt;
Reviewed-by: Niels de Vos &lt;ndevos@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Consider a volume that is exported via NFS-Ganesha. Stopping this
volume will automatically unexport the volume. Starting this volume
should automatically export it. Although the logic was already there,
there was a bug in it. Fixing the same by introducing a hook script.

Also with the new CLI options, the hook script S31ganesha-set.sh
is no longer required. Hence, removing the same.
Adding a comment to tell the user that one of the CLI
commands will take a few minutes to complete.

This is a backport of the patch merged on master,
http://review.gluster.org/#/c/11247/

Change-Id: I1c16a978b9e2093a4298ea3024d031fd1a5bb577
BUG: 1232335
Signed-off-by: Meghana Madhusudhan &lt;mmadhusu@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11259
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
Reviewed-by: soumya k &lt;skoduri@redhat.com&gt;
Reviewed-by: Niels de Vos &lt;ndevos@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>features/bitrot: tuanble object signing waiting time value for bitrot</title>
<updated>2015-06-17T16:38:59+00:00</updated>
<author>
<name>Gaurav Kumar Garg</name>
<email>ggarg@redhat.com</email>
</author>
<published>2015-06-05T08:28:28+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=8a7dfb0d4b856578e89898c2bb84a0a675ade50b'/>
<id>8a7dfb0d4b856578e89898c2bb84a0a675ade50b</id>
<content type='text'>
 Currently bitrot using 120 second waiting time for object to be signed
 after all fop's released. This signing waiting time value should be tunable.

 Command for changing the signing waiting time will be
 #gluster volume bitrot &lt;VOLNAME&gt; signing-time &lt;waiting time value in second&gt;

Change-Id: I89f3121564c1bbd0825f60aae6147413a2fbd798
BUG: 1231832
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
Signed-off-by: Venky Shankar &lt;vshankar@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11105
(cherry picked from commit 554fa0c1315d0b4b78ba35a2d332d7ac0fd07d48)
Reviewed-on: http://review.gluster.org/11235
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
 Currently bitrot using 120 second waiting time for object to be signed
 after all fop's released. This signing waiting time value should be tunable.

 Command for changing the signing waiting time will be
 #gluster volume bitrot &lt;VOLNAME&gt; signing-time &lt;waiting time value in second&gt;

Change-Id: I89f3121564c1bbd0825f60aae6147413a2fbd798
BUG: 1231832
Signed-off-by: Gaurav Kumar Garg &lt;ggarg@redhat.com&gt;
Signed-off-by: Venky Shankar &lt;vshankar@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11105
(cherry picked from commit 554fa0c1315d0b4b78ba35a2d332d7ac0fd07d48)
Reviewed-on: http://review.gluster.org/11235
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd/tier: glusterd crashed with detach-tier commit force</title>
<updated>2015-06-17T05:10:55+00:00</updated>
<author>
<name>Mohammed Rafi KC</name>
<email>rkavunga@redhat.com</email>
</author>
<published>2015-06-04T08:44:14+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=bf6a71bcb08357766027e9318edae4b8298c339c'/>
<id>bf6a71bcb08357766027e9318edae4b8298c339c</id>
<content type='text'>
glusterd crashed when doing "detach-tier commit force" on a
non-tiered volume.

&gt;Change-Id: I884771893bb80bec46ae8642c2cfd7e54ab116a6
&gt;BUG: 1228112
&gt;Signed-off-by: Mohammed Rafi KC &lt;rkavunga@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/11081
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt;Reviewed-by: Joseph Fernandes
&gt;Reviewed-by: Dan Lambright &lt;dlambrig@redhat.com&gt;
&gt;Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;

(cherry picked from commit fefc8730e2a87db41d5f79c9853a14cab86715e5)

Change-Id: Ia536da3ec4c632d7d1cc1575dd2b7b098a2c7aec
BUG: 1230563
Signed-off-by: Mohammed Rafi KC &lt;rkavunga@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11174
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Joseph Fernandes
Tested-by: Joseph Fernandes
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
glusterd crashed when doing "detach-tier commit force" on a
non-tiered volume.

&gt;Change-Id: I884771893bb80bec46ae8642c2cfd7e54ab116a6
&gt;BUG: 1228112
&gt;Signed-off-by: Mohammed Rafi KC &lt;rkavunga@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/11081
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
&gt;Reviewed-by: Joseph Fernandes
&gt;Reviewed-by: Dan Lambright &lt;dlambrig@redhat.com&gt;
&gt;Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;

(cherry picked from commit fefc8730e2a87db41d5f79c9853a14cab86715e5)

Change-Id: Ia536da3ec4c632d7d1cc1575dd2b7b098a2c7aec
BUG: 1230563
Signed-off-by: Mohammed Rafi KC &lt;rkavunga@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11174
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Joseph Fernandes
Tested-by: Joseph Fernandes
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd/shared_storage: Provide a volume set option to create and mount the shared storage</title>
<updated>2015-06-05T09:20:26+00:00</updated>
<author>
<name>Avra Sengupta</name>
<email>asengupt@redhat.com</email>
</author>
<published>2015-05-14T09:30:59+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=ae87a7fedfcf7f6b287ef5c3860f45412363e4f6'/>
<id>ae87a7fedfcf7f6b287ef5c3860f45412363e4f6</id>
<content type='text'>
     Backport of http://review.gluster.org/#/c/10793/

Introducing a global volume set option(cluster.enable-shared-storage)
which helps create and set-up the shared storage meta volume.

gluster volume set all cluster.enable-shared-storage enable

On enabling this option, the system analyzes the number of peers
in the cluster, which are currently connected, and chooses three
such peers(including the node the command is issued from). From these
peers a volume(gluster_shared_storage) is created. Depending on the
number of peers available the volume is either a replica 3
volume(if there are 3 connected peers),  or a replica 2 volume(if there
are 2 connected peers). "/var/run/gluster/ss_brick" serves as the
brick path on each node for the shared storage volume. We also mount
the shared storage at "/var/run/gluster/shared_storage" on all the nodes
in the cluster as part of enabling this option. If there is only one node
in the cluster,  or only one node is up then the command will fail

Once the volume is created, and mounted the maintainance of the
volume like adding-bricks, removing bricks etc., is expected to
be the onus of the user.

On disabling the option, we provide the user a warning, and on
affirmation from the user we stop the shared storage volume, and unmount
it from all the nodes in the cluster.

gluster volume set all cluster.enable-shared-storage disable

Change-Id: Idd92d67b93f444244f99ede9f634ef18d2945dbc
BUG: 1228181
Signed-off-by: Avra Sengupta &lt;asengupt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11086
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Rajesh Joseph &lt;rjoseph@redhat.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
     Backport of http://review.gluster.org/#/c/10793/

Introducing a global volume set option(cluster.enable-shared-storage)
which helps create and set-up the shared storage meta volume.

gluster volume set all cluster.enable-shared-storage enable

On enabling this option, the system analyzes the number of peers
in the cluster, which are currently connected, and chooses three
such peers(including the node the command is issued from). From these
peers a volume(gluster_shared_storage) is created. Depending on the
number of peers available the volume is either a replica 3
volume(if there are 3 connected peers),  or a replica 2 volume(if there
are 2 connected peers). "/var/run/gluster/ss_brick" serves as the
brick path on each node for the shared storage volume. We also mount
the shared storage at "/var/run/gluster/shared_storage" on all the nodes
in the cluster as part of enabling this option. If there is only one node
in the cluster,  or only one node is up then the command will fail

Once the volume is created, and mounted the maintainance of the
volume like adding-bricks, removing bricks etc., is expected to
be the onus of the user.

On disabling the option, we provide the user a warning, and on
affirmation from the user we stop the shared storage volume, and unmount
it from all the nodes in the cluster.

gluster volume set all cluster.enable-shared-storage disable

Change-Id: Idd92d67b93f444244f99ede9f634ef18d2945dbc
BUG: 1228181
Signed-off-by: Avra Sengupta &lt;asengupt@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11086
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Rajesh Joseph &lt;rjoseph@redhat.com&gt;
Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cli: Fix incorrect parse logic for volume heal commands</title>
<updated>2015-06-04T08:26:53+00:00</updated>
<author>
<name>Anuradha</name>
<email>atalur@redhat.com</email>
</author>
<published>2015-05-13T09:59:57+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=2f61269237170253e022fe6fc587ee7630e727e3'/>
<id>2f61269237170253e022fe6fc587ee7630e727e3</id>
<content type='text'>
heal-op was being incorrectly set to
GF_SHD_OP_SBRAIN_HEAL_FROM_BIGGER_FILE.

Change-Id: I4d4461c7737feae30102e82f7788083017485669
BUG: 1218570
Reviewed-on: http://review.gluster.org/10771
Signed-off-by: Anuradha &lt;atalur@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11043
Reviewed-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Pranith Kumar Karampuri &lt;pkarampu@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
heal-op was being incorrectly set to
GF_SHD_OP_SBRAIN_HEAL_FROM_BIGGER_FILE.

Change-Id: I4d4461c7737feae30102e82f7788083017485669
BUG: 1218570
Reviewed-on: http://review.gluster.org/10771
Signed-off-by: Anuradha &lt;atalur@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11043
Reviewed-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Pranith Kumar Karampuri &lt;pkarampu@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>tiering: Correct errors in cli and glusterd</title>
<updated>2015-06-01T04:31:53+00:00</updated>
<author>
<name>Mohammed Rafi KC</name>
<email>rkavunga@redhat.com</email>
</author>
<published>2015-05-13T11:23:22+00:00</published>
<link rel='alternate' type='text/html' href='http://git.gluster.org/cgit/glusterfs.git/commit/?id=57abe450c1e90749a4a32247836f05b1195d5313'/>
<id>57abe450c1e90749a4a32247836f05b1195d5313</id>
<content type='text'>
        Back port of http://review.gluster.org/10768

Problem 1:

volume info shows Cold Bricks instead of Tier type
  eg:
Volume Name: patchy2
Type: Tier
Volume ID: 28c25b8d-b8a1-45dc-b4b7-cbd0b344f58f
Status: Started
Number of Bricks: 3
Transport-type: tcp
Hot Tier :
Hot Tier Type : Distribute
Number of Bricks: 1
Brick1: 10.70.1.35:/home/brick43
Cold Bricks:
Cold Tier Type : Distribute
Number of Bricks: 2
Brick2: 10.70.1.35:/home/brick19
Brick3: 10.70.1.35:/home/brick16
Options Reconfigured:

Problem 2: Detach-tier sending enums of Rebalance

   detach-tier has it's own Enum to send with detach-tier command,
using that enums will make more appropriate.

Problem 3:

Wrongly sets hot_brick count during the dictionary copying for response

&gt;Change-Id: Icc054a999a679456881bc70511470d32ff8a86e4
&gt;BUG: 1211264
&gt;Signed-off-by: Mohammed Rafi KC &lt;rkavunga@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/10768
&gt;Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
&gt;Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Tested-by: NetBSD Build System

Change-Id: I889d3b83a1f197b79eb71e594a1ad1f2037ed402
BUG: 1226024
Signed-off-by: Mohammed Rafi KC &lt;rkavunga@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10978
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Joseph Fernandes
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
        Back port of http://review.gluster.org/10768

Problem 1:

volume info shows Cold Bricks instead of Tier type
  eg:
Volume Name: patchy2
Type: Tier
Volume ID: 28c25b8d-b8a1-45dc-b4b7-cbd0b344f58f
Status: Started
Number of Bricks: 3
Transport-type: tcp
Hot Tier :
Hot Tier Type : Distribute
Number of Bricks: 1
Brick1: 10.70.1.35:/home/brick43
Cold Bricks:
Cold Tier Type : Distribute
Number of Bricks: 2
Brick2: 10.70.1.35:/home/brick19
Brick3: 10.70.1.35:/home/brick16
Options Reconfigured:

Problem 2: Detach-tier sending enums of Rebalance

   detach-tier has it's own Enum to send with detach-tier command,
using that enums will make more appropriate.

Problem 3:

Wrongly sets hot_brick count during the dictionary copying for response

&gt;Change-Id: Icc054a999a679456881bc70511470d32ff8a86e4
&gt;BUG: 1211264
&gt;Signed-off-by: Mohammed Rafi KC &lt;rkavunga@redhat.com&gt;
&gt;Reviewed-on: http://review.gluster.org/10768
&gt;Reviewed-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
&gt;Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
&gt;Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt;Tested-by: NetBSD Build System

Change-Id: I889d3b83a1f197b79eb71e594a1ad1f2037ed402
BUG: 1226024
Signed-off-by: Mohammed Rafi KC &lt;rkavunga@redhat.com&gt;
Reviewed-on: http://review.gluster.org/10978
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Tested-by: NetBSD Build System &lt;jenkins@build.gluster.org&gt;
Reviewed-by: Joseph Fernandes
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</pre>
</div>
</content>
</entry>
</feed>
