summaryrefslogtreecommitdiffstats
path: root/tests/functional/glusterd/test_volume_operations.py
diff options
context:
space:
mode:
Diffstat (limited to 'tests/functional/glusterd/test_volume_operations.py')
-rw-r--r--tests/functional/glusterd/test_volume_operations.py169
1 files changed, 169 insertions, 0 deletions
diff --git a/tests/functional/glusterd/test_volume_operations.py b/tests/functional/glusterd/test_volume_operations.py
new file mode 100644
index 000000000..a8e75ad8b
--- /dev/null
+++ b/tests/functional/glusterd/test_volume_operations.py
@@ -0,0 +1,169 @@
+# Copyright (C) 2016-2017 Red Hat, Inc. <http://www.redhat.com>
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License as published by
+# the Free Software Foundation; either version 2 of the License, or
+# any later version.
+#
+# This program is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+# GNU General Public License for more details.
+#
+# You should have received a copy of the GNU General Public License along
+# with this program; if not, write to the Free Software Foundation, Inc.,
+# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
+
+from glusto.core import Glusto as g
+from glustolibs.gluster.gluster_base_class import GlusterBaseClass, runs_on
+from glustolibs.gluster.volume_ops import (volume_create, volume_start,
+ get_volume_list)
+from glustolibs.gluster.brick_libs import (are_bricks_online)
+from glustolibs.gluster.volume_libs import cleanup_volume
+from glustolibs.gluster.lib_utils import form_bricks_list
+from glustolibs.gluster.exceptions import ExecutionError
+import random
+import re
+import os
+
+
+@runs_on([['distributed'], ['glusterfs']])
+class TestVolumeCreate(GlusterBaseClass):
+
+ @classmethod
+ def setUpClass(cls):
+ # Calling GlusterBaseClass setUpClass
+ GlusterBaseClass.setUpClass.im_func(cls)
+
+ def setUp(self):
+ GlusterBaseClass.setUp.im_func(self)
+ # check whether peers are in connected state
+ ret = self.validate_peers_are_connected()
+ if not ret:
+ raise ExecutionError("Peers are not in connected state")
+
+ def tearDown(self):
+
+ # clean up all volumes
+ vol_list = get_volume_list(self.mnode)
+ if vol_list is None:
+ raise ExecutionError("Failed to get the volume list")
+
+ for volume in vol_list:
+ ret = cleanup_volume(self.mnode, volume)
+ if not ret:
+ raise ExecutionError("Unable to delete volume % s" % volume)
+ g.log.info("Volume deleted successfully : %s" % volume)
+
+ GlusterBaseClass.tearDown.im_func(self)
+
+ @classmethod
+ def tearDownClass(cls):
+
+ # Calling GlusterBaseClass tearDown
+ GlusterBaseClass.tearDownClass.im_func(cls)
+
+ def test_volume_start_force(self):
+
+ # get the brick list and create a volume
+ num_of_bricks = len(self.servers)
+ bricks_list = form_bricks_list(self.mnode, self.volname, num_of_bricks,
+ self.servers, self.all_servers_info)
+
+ ret, _, _ = volume_create(self.mnode, self.volname, bricks_list)
+ self.assertEqual(ret, 0, "Failed to create volume")
+
+ # remove brick path in one node and try to start the volume with force
+ # and without force
+ index_of_node = random.randint(0, len(bricks_list) - 1)
+ brick_node = bricks_list[index_of_node]
+ node = brick_node.split(":")[0]
+ brick_path = brick_node.split(":")[1]
+ cmd = "rm -rf %s" % brick_path
+ ret, _, _ = g.run(node, cmd)
+ self.assertEqual(ret, 0, "Failed to delete the brick")
+ g.log.info("Deleted the brick successfully")
+
+ ret, _, _ = volume_start(self.mnode, self.volname)
+ self.assertNotEqual(ret, 0, "Volume start succeeded")
+
+ ret, _, _ = volume_start(self.mnode, self.volname, force=True)
+ self.assertEqual(ret, 0, "Volume start with force failed")
+
+ # volume start force should not bring the brick online
+ ret = are_bricks_online(self.mnode, self.volname,
+ [bricks_list[index_of_node]])
+ self.assertFalse(ret, "Volume start force brought the bricks online")
+ g.log.info("Volume start force didn't bring the brick online")
+
+ def test_volume_create_on_brick_root(self):
+
+ # try to create a volume on brick root path without using force and
+ # with using force
+ self.volname = "second_volume"
+ num_of_bricks = len(self.servers)
+ bricks_list = form_bricks_list(self.mnode, self.volname, num_of_bricks,
+ self.servers, self.all_servers_info)
+
+ # save for using it later
+ same_bricks_list = bricks_list[:]
+
+ complete_brick = bricks_list[0].split(":")
+ brick_root = os.path.dirname(complete_brick[1])
+ root_brickpath = complete_brick[0] + ":" + brick_root
+ bricks_list[0] = root_brickpath
+
+ # creation of volume on root brick path should fail
+ ret, _, _ = volume_create(self.mnode, self.volname, bricks_list)
+ self.assertNotEqual(ret, 0, "Volume create on root brick path is "
+ "success")
+
+ # volume create force should succeed
+ ret, _, _ = volume_create(self.mnode, self.volname, bricks_list, True)
+ self.assertEqual(ret, 0, "Volume create on root brick path with"
+ " force is failed")
+ g.log.info("Volume create on root brick path with force is success")
+
+ # create a sub directory under root partition and create a volume
+ self.volname = "third_volume"
+
+ sub_dir_path = "%s/sub_dir" % brick_root
+ cmd = "mkdir %s" % sub_dir_path
+ ret, _, _ = g.run(self.servers[0], cmd)
+ sub_dir_brickpath_node = bricks_list[0].split(":")[0]
+ sub_dir_brickpath = sub_dir_brickpath_node + ":" + sub_dir_path
+ bricks_list[0] = sub_dir_brickpath
+
+ # volume create with previously used bricks should fail
+ ret, _, _ = volume_create(self.mnode, self.volname, bricks_list, True)
+ self.assertNotEqual(ret, 0, "Volume create with previously used bricks"
+ " is success")
+
+ # delete the volume created on root partition and clear all attributes
+ # now, creation of volume should succeed.
+ self.volname = "second_volume"
+ ret, _, _ = g.run(self.mnode, "gluster vol delete %s --mode=script"
+ % self.volname)
+ for brick in bricks_list:
+ server = brick.split(":")[0]
+ brick_root = os.path.dirname(brick.split(":")[1])
+ cmd1 = "rm -rf %s/*" % brick_root
+ cmd2 = "getfattr -d -m . %s/" % brick_root
+ cmd3 = "setfattr -x trusted.glusterfs.volume-id %s/" % brick_root
+ cmd4 = "setfattr -x trusted.gfid %s/" % brick_root
+ ret, _, _ = g.run(server, cmd1)
+ self.assertEqual(ret, 0, "Failed to delete the files")
+ g.log.info("Successfully deleted the files")
+ ret, out, err = g.run(server, cmd2)
+ if re.search("trusted.glusterfs.volume-id", out):
+ ret, _, _ = g.run(server, cmd3)
+ self.assertEqual(ret, 0, "Failed to delete the xattrs")
+ g.log.info("Deleted trusted.glusterfs.volume-id the xattrs")
+ if re.search("trusted.gfid", out):
+ ret, _, _ = g.run(server, cmd4)
+ self.assertEqual(ret, 0, "Failed to delete gfid xattrs")
+ g.log.info("Deleted trusterd.gfid xattrs")
+
+ # creation of volume should succeed
+ ret, _, _ = volume_create(self.mnode, self.volname, same_bricks_list)
+ self.assertEqual(ret, 0, "Failed to create volume")