summaryrefslogtreecommitdiffstats
path: root/tests
diff options
context:
space:
mode:
authorBala Konda Reddy Mekala <bmekala@redhat.com>2018-06-12 10:56:06 +0530
committerAkarsha Rai <akrai@redhat.com>2018-06-25 10:55:33 +0000
commitc7c7b425e84490300dcb2a2846526255f6eb1f5f (patch)
tree814f33c36ec048c39ce1e9fcfc9bd379c1e5ab19 /tests
parentf1e5d397c734de51dcfeab321c6f761277c3b592 (diff)
Validate snap info from detached node in the cluster
Change-Id: Ica3d1175ee5d2c6a45e7b7d6513885ee2b84d960 Signed-off-by: Bala Konda Reddy Mekala <bmekala@redhat.com>
Diffstat (limited to 'tests')
-rw-r--r--tests/functional/glusterd/test_glusterd_snap_info_on_detached_node.py143
1 files changed, 143 insertions, 0 deletions
diff --git a/tests/functional/glusterd/test_glusterd_snap_info_on_detached_node.py b/tests/functional/glusterd/test_glusterd_snap_info_on_detached_node.py
new file mode 100644
index 0000000..8c0f7a7
--- /dev/null
+++ b/tests/functional/glusterd/test_glusterd_snap_info_on_detached_node.py
@@ -0,0 +1,143 @@
+# Copyright (C) 2017-2018 Red Hat, Inc. <http://www.redhat.com>
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License as published by
+# the Free Software Foundation; either version 2 of the License, or
+# any later version.
+#
+# This program is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+# GNU General Public License for more details.
+#
+# You should have received a copy of the GNU General Public License along
+# with this program; if not, write to the Free Software Foundation, Inc.,
+# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
+import random
+from time import sleep
+from glusto.core import Glusto as g
+from glustolibs.gluster.exceptions import ExecutionError
+from glustolibs.gluster.gluster_base_class import GlusterBaseClass, runs_on
+from glustolibs.gluster.lib_utils import form_bricks_list
+from glustolibs.gluster.volume_ops import (volume_create,
+ set_volume_options, volume_start)
+from glustolibs.gluster.snap_ops import snap_create, snap_activate
+from glustolibs.gluster.peer_ops import peer_detach_servers, peer_probe
+
+
+@runs_on([['distributed'], ['glusterfs']])
+class TestSnapInfoOnPeerDetachedNode(GlusterBaseClass):
+
+ def tearDown(self):
+
+ # stopping the volume and Cleaning up the volume
+ ret = self.cleanup_volume()
+ if not ret:
+ raise ExecutionError("Failed Cleanup the Volume %s" % self.volname)
+ g.log.info("Volume deleted successfully : %s", self.volname)
+
+ # Calling GlusterBaseClass tearDown
+ GlusterBaseClass.tearDown.im_func(self)
+
+ def test_snap_info_from_detached_node(self):
+ # pylint: disable=too-many-statements
+ """
+ Create a volume with single brick
+ Create a snapshot
+ Activate the snapshot created
+ Enabled uss on the volume
+ Validated snap info on all the nodes
+ Peer detach one node
+ Validate /var/lib/glusterd/snaps on the detached node
+ Probe the detached node
+ """
+
+ # Creating volume with single brick on one node
+ servers_info_single_node = {self.servers[0]:
+ self.all_servers_info[self.servers[0]]}
+ bricks_list = form_bricks_list(self.mnode, self.volname,
+ 1, self.servers[0],
+ servers_info_single_node)
+ ret, _, _ = volume_create(self.servers[0], self.volname, bricks_list)
+ self.assertEqual(ret, 0, "Volume creation failed")
+ g.log.info("Volume %s created successfully", self.volname)
+
+ # Create a snapshot of the volume without volume start should fail
+ self.snapname = "snap1"
+ ret, _, _ = snap_create(
+ self.mnode, self.volname, self.snapname, timestamp=False)
+ self.assertNotEqual(
+ ret, 0, "Snapshot created without starting the volume")
+ g.log.info("Snapshot creation failed as expected")
+
+ # Start the volume
+ ret, _, _ = volume_start(self.mnode, self.volname)
+ self.assertEqual(
+ ret, 0, "Failed to start the volume %s" % self.volname)
+ g.log.info("Volume start succeeded")
+
+ # Create a snapshot of the volume after volume start
+ ret, _, _ = snap_create(
+ self.mnode, self.volname, self.snapname, timestamp=False)
+ self.assertEqual(
+ ret, 0, "Snapshot creation failed on the volume %s" % self.volname)
+ g.log.info("Snapshot create succeeded")
+
+ # Activate snapshot created
+ ret, _, err = snap_activate(self.mnode, self.snapname)
+ self.assertEqual(
+ ret, 0, "Snapshot activate failed with following error %s" % (err))
+ g.log.info("Snapshot activated successfully")
+
+ # Enable uss
+ self.vol_options['features.uss'] = 'enable'
+ ret = set_volume_options(self.mnode, self.volname, self.vol_options)
+ self.assertTrue(ret, "gluster volume set %s features.uss "
+ "enable failed" % self.volname)
+ g.log.info("gluster volume set %s features.uss "
+ "enable successfully", self.volname)
+
+ # Validate files /var/lib/glusterd/snaps on all the servers is same
+ self.pathname = "/var/lib/glusterd/snaps/%s" % self.snapname
+ for server in self.servers:
+ conn = g.rpyc_get_connection(server)
+ ret = conn.modules.os.path.isdir(self.pathname)
+ self.assertTrue(ret, "%s directory doesn't exist on node %s" %
+ (self.pathname, server))
+ g.log.info("%s path exists on node %s", self.pathname, server)
+ g.rpyc_close_deployed_servers()
+
+ # Peer detach one node
+ self.random_node_peer_detach = random.choice(self.servers[1:])
+ ret = peer_detach_servers(self.mnode,
+ self.random_node_peer_detach, validate=True)
+ self.assertTrue(ret, "Peer detach of node: %s failed" %
+ self.random_node_peer_detach)
+ g.log.info("Peer detach succeeded")
+
+ # /var/lib/glusterd/snaps/<snapname> directory should not present
+ conn = g.rpyc_get_connection(self.random_node_peer_detach)
+ ret = conn.modules.os.path.isdir(self.pathname)
+ self.assertFalse(ret, "%s directory should not exist on the peer"
+ "which is detached from cluster%s" % (
+ self.pathname, self.random_node_peer_detach))
+ g.log.info("Expected: %s path doesn't exist on peer detached node %s",
+ self.pathname, self.random_node_peer_detach)
+ g.rpyc_close_deployed_servers()
+
+ # Peer probe the detached node
+ ret, _, _ = peer_probe(self.mnode, self.random_node_peer_detach)
+ self.assertEqual(ret, 0, "Peer probe of node: %s failed" %
+ self.random_node_peer_detach)
+ g.log.info("Peer probe succeeded")
+
+ # Validating peers are in connected state
+ count = 0
+ while count < 10:
+ sleep(2)
+ ret = self.validate_peers_are_connected()
+ if ret:
+ break
+ count += 1
+ self.assertTrue(ret, "Peers are not in connected state")
+ g.log.info("Peer are in connected state")