diff options
Diffstat (limited to 'tests')
-rw-r--r-- | tests/functional/glusterd/test_volume_operations.py | 169 |
1 files changed, 169 insertions, 0 deletions
diff --git a/tests/functional/glusterd/test_volume_operations.py b/tests/functional/glusterd/test_volume_operations.py new file mode 100644 index 000000000..a8e75ad8b --- /dev/null +++ b/tests/functional/glusterd/test_volume_operations.py @@ -0,0 +1,169 @@ +# Copyright (C) 2016-2017 Red Hat, Inc. <http://www.redhat.com> +# +# This program is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation; either version 2 of the License, or +# any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License along +# with this program; if not, write to the Free Software Foundation, Inc., +# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. + +from glusto.core import Glusto as g +from glustolibs.gluster.gluster_base_class import GlusterBaseClass, runs_on +from glustolibs.gluster.volume_ops import (volume_create, volume_start, + get_volume_list) +from glustolibs.gluster.brick_libs import (are_bricks_online) +from glustolibs.gluster.volume_libs import cleanup_volume +from glustolibs.gluster.lib_utils import form_bricks_list +from glustolibs.gluster.exceptions import ExecutionError +import random +import re +import os + + +@runs_on([['distributed'], ['glusterfs']]) +class TestVolumeCreate(GlusterBaseClass): + + @classmethod + def setUpClass(cls): + # Calling GlusterBaseClass setUpClass + GlusterBaseClass.setUpClass.im_func(cls) + + def setUp(self): + GlusterBaseClass.setUp.im_func(self) + # check whether peers are in connected state + ret = self.validate_peers_are_connected() + if not ret: + raise ExecutionError("Peers are not in connected state") + + def tearDown(self): + + # clean up all volumes + vol_list = get_volume_list(self.mnode) + if vol_list is None: + raise ExecutionError("Failed to get the volume list") + + for volume in vol_list: + ret = cleanup_volume(self.mnode, volume) + if not ret: + raise ExecutionError("Unable to delete volume % s" % volume) + g.log.info("Volume deleted successfully : %s" % volume) + + GlusterBaseClass.tearDown.im_func(self) + + @classmethod + def tearDownClass(cls): + + # Calling GlusterBaseClass tearDown + GlusterBaseClass.tearDownClass.im_func(cls) + + def test_volume_start_force(self): + + # get the brick list and create a volume + num_of_bricks = len(self.servers) + bricks_list = form_bricks_list(self.mnode, self.volname, num_of_bricks, + self.servers, self.all_servers_info) + + ret, _, _ = volume_create(self.mnode, self.volname, bricks_list) + self.assertEqual(ret, 0, "Failed to create volume") + + # remove brick path in one node and try to start the volume with force + # and without force + index_of_node = random.randint(0, len(bricks_list) - 1) + brick_node = bricks_list[index_of_node] + node = brick_node.split(":")[0] + brick_path = brick_node.split(":")[1] + cmd = "rm -rf %s" % brick_path + ret, _, _ = g.run(node, cmd) + self.assertEqual(ret, 0, "Failed to delete the brick") + g.log.info("Deleted the brick successfully") + + ret, _, _ = volume_start(self.mnode, self.volname) + self.assertNotEqual(ret, 0, "Volume start succeeded") + + ret, _, _ = volume_start(self.mnode, self.volname, force=True) + self.assertEqual(ret, 0, "Volume start with force failed") + + # volume start force should not bring the brick online + ret = are_bricks_online(self.mnode, self.volname, + [bricks_list[index_of_node]]) + self.assertFalse(ret, "Volume start force brought the bricks online") + g.log.info("Volume start force didn't bring the brick online") + + def test_volume_create_on_brick_root(self): + + # try to create a volume on brick root path without using force and + # with using force + self.volname = "second_volume" + num_of_bricks = len(self.servers) + bricks_list = form_bricks_list(self.mnode, self.volname, num_of_bricks, + self.servers, self.all_servers_info) + + # save for using it later + same_bricks_list = bricks_list[:] + + complete_brick = bricks_list[0].split(":") + brick_root = os.path.dirname(complete_brick[1]) + root_brickpath = complete_brick[0] + ":" + brick_root + bricks_list[0] = root_brickpath + + # creation of volume on root brick path should fail + ret, _, _ = volume_create(self.mnode, self.volname, bricks_list) + self.assertNotEqual(ret, 0, "Volume create on root brick path is " + "success") + + # volume create force should succeed + ret, _, _ = volume_create(self.mnode, self.volname, bricks_list, True) + self.assertEqual(ret, 0, "Volume create on root brick path with" + " force is failed") + g.log.info("Volume create on root brick path with force is success") + + # create a sub directory under root partition and create a volume + self.volname = "third_volume" + + sub_dir_path = "%s/sub_dir" % brick_root + cmd = "mkdir %s" % sub_dir_path + ret, _, _ = g.run(self.servers[0], cmd) + sub_dir_brickpath_node = bricks_list[0].split(":")[0] + sub_dir_brickpath = sub_dir_brickpath_node + ":" + sub_dir_path + bricks_list[0] = sub_dir_brickpath + + # volume create with previously used bricks should fail + ret, _, _ = volume_create(self.mnode, self.volname, bricks_list, True) + self.assertNotEqual(ret, 0, "Volume create with previously used bricks" + " is success") + + # delete the volume created on root partition and clear all attributes + # now, creation of volume should succeed. + self.volname = "second_volume" + ret, _, _ = g.run(self.mnode, "gluster vol delete %s --mode=script" + % self.volname) + for brick in bricks_list: + server = brick.split(":")[0] + brick_root = os.path.dirname(brick.split(":")[1]) + cmd1 = "rm -rf %s/*" % brick_root + cmd2 = "getfattr -d -m . %s/" % brick_root + cmd3 = "setfattr -x trusted.glusterfs.volume-id %s/" % brick_root + cmd4 = "setfattr -x trusted.gfid %s/" % brick_root + ret, _, _ = g.run(server, cmd1) + self.assertEqual(ret, 0, "Failed to delete the files") + g.log.info("Successfully deleted the files") + ret, out, err = g.run(server, cmd2) + if re.search("trusted.glusterfs.volume-id", out): + ret, _, _ = g.run(server, cmd3) + self.assertEqual(ret, 0, "Failed to delete the xattrs") + g.log.info("Deleted trusted.glusterfs.volume-id the xattrs") + if re.search("trusted.gfid", out): + ret, _, _ = g.run(server, cmd4) + self.assertEqual(ret, 0, "Failed to delete gfid xattrs") + g.log.info("Deleted trusterd.gfid xattrs") + + # creation of volume should succeed + ret, _, _ = volume_create(self.mnode, self.volname, same_bricks_list) + self.assertEqual(ret, 0, "Failed to create volume") |