summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorVijay Avuthu <vavuthu@redhat.com>2018-11-20 10:45:49 +0530
committerVijay Avuthu <vavuthu@redhat.com>2018-11-22 10:43:44 +0000
commitf88abb24146f0c6a8a2179688d7639c0e32490be (patch)
tree907eed124d1caef0166f09644baa10431ee044c9
parent59fe4e38000c334541e33e8b5adbcb264d5992a8 (diff)
Fix the test case
test_client_side_quorum_auto_local_to_volume_not_cluster Change-Id: Ibf16517fe062f9335d17c0e6d88ddab44a644c7b
-rw-r--r--tests/functional/afr/test_client_side_quorum_with_multiple_volumes.py382
1 files changed, 382 insertions, 0 deletions
diff --git a/tests/functional/afr/test_client_side_quorum_with_multiple_volumes.py b/tests/functional/afr/test_client_side_quorum_with_multiple_volumes.py
new file mode 100644
index 0000000..5ad74ec
--- /dev/null
+++ b/tests/functional/afr/test_client_side_quorum_with_multiple_volumes.py
@@ -0,0 +1,382 @@
+# Copyright (C) 2016-2017 Red Hat, Inc. <http://www.redhat.com>
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License as published by
+# the Free Software Foundation; either version 2 of the License, or
+# any later version.
+#
+# This program is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+# GNU General Public License for more details.
+#
+# You should have received a copy of the GNU General Public License along
+# with this program; if not, write to the Free Software Foundation, Inc.,
+# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
+
+""" Description:
+ Test Cases in this module tests the client side quorum.
+"""
+
+import tempfile
+
+from glusto.core import Glusto as g
+from glustolibs.gluster.exceptions import ExecutionError
+from glustolibs.gluster.gluster_base_class import GlusterBaseClass, runs_on
+from glustolibs.gluster.volume_libs import (
+ set_volume_options, setup_volume, cleanup_volume)
+from glustolibs.gluster.volume_ops import get_volume_list, get_volume_options
+from glustolibs.misc.misc_libs import upload_scripts
+from glustolibs.gluster.brick_libs import (bring_bricks_offline,
+ get_all_bricks,
+ are_bricks_offline)
+from glustolibs.io.utils import (validate_io_procs,
+ is_io_procs_fail_with_rofs,
+ list_all_files_and_dirs_mounts,
+ wait_for_io_to_complete)
+from glustolibs.gluster.mount_ops import mount_volume, umount_volume
+
+
+@runs_on([['distributed-replicated'],
+ ['glusterfs']])
+class ClientSideQuorumTestsMultipleVols(GlusterBaseClass):
+
+ @classmethod
+ def setUpClass(cls):
+ # Calling GlusterBaseClass setUpClass
+ GlusterBaseClass.setUpClass.im_func(cls)
+
+ # Upload io scripts for running IO on mounts
+ g.log.info("Upload io scripts to clients %s for running IO on mounts",
+ cls.clients)
+ script_local_path = ("/usr/share/glustolibs/io/scripts/"
+ "file_dir_ops.py")
+ cls.script_upload_path = ("/usr/share/glustolibs/io/scripts/"
+ "file_dir_ops.py")
+ ret = upload_scripts(cls.clients, [script_local_path])
+ if not ret:
+ raise ExecutionError("Failed to upload IO scripts to clients %s"
+ % cls.clients)
+ g.log.info("Successfully uploaded IO scripts to clients %s",
+ cls.clients)
+
+ cls.counter = 1
+ # int: Value of counter is used for dirname-start-num argument for
+ # file_dir_ops.py create_deep_dirs_with_files.
+
+ # The --dir-length argument value for file_dir_ops.py
+ # create_deep_dirs_with_files is set to 10 (refer to the cmd in setUp
+ # method). This means every mount will create
+ # 10 top level dirs. For every mountpoint/testcase to create new set of
+ # dirs, we are incrementing the counter by --dir-length value i.e 10 in
+ # this test suite.
+
+ # If we are changing the --dir-length to new value, ensure the counter
+ # is also incremented by same value to create new set of files/dirs.
+
+ # Setup Volumes
+ if cls.volume_type == "distributed-replicated":
+ cls.volume_configs = []
+
+ # Define two 2x2 distributed-replicated volumes
+ for i in range(1, 3):
+ cls.volume['voltype'] = {
+ 'type': 'distributed-replicated',
+ 'replica_count': 2,
+ 'dist_count': 2,
+ 'transport': 'tcp',
+ }
+ cls.volume_configs.append(
+ {'name': 'testvol_%s_%d'
+ % (cls.volume['voltype']['type'], i),
+ 'servers': cls.servers,
+ 'voltype': cls.volume['voltype']})
+
+ # Define two 2x3 distributed-replicated volumes
+ for i in range(1, 3):
+ cls.volume['voltype'] = {
+ 'type': 'distributed-replicated',
+ 'replica_count': 3,
+ 'dist_count': 2,
+ 'transport': 'tcp',
+ }
+ cls.volume_configs.append(
+ {'name': 'testvol_%s_%d'
+ % (cls.volume['voltype']['type'], i+2),
+ 'servers': cls.servers,
+ 'voltype': cls.volume['voltype']})
+
+ # Define distributed volume
+ cls.volume['voltype'] = {
+ 'type': 'distributed',
+ 'dist_count': 3,
+ 'transport': 'tcp',
+ }
+ cls.volume_configs.append(
+ {'name': 'testvol_%s'
+ % cls.volume['voltype']['type'],
+ 'servers': cls.servers,
+ 'voltype': cls.volume['voltype']})
+
+ # Create and mount volumes
+ cls.mount_points = []
+ cls.mount_points_and_volnames = {}
+ cls.client = cls.clients[0]
+ for volume_config in cls.volume_configs:
+ # Setup volume
+ ret = setup_volume(mnode=cls.mnode,
+ all_servers_info=cls.all_servers_info,
+ volume_config=volume_config,
+ force=False)
+ if not ret:
+ raise ExecutionError("Failed to setup Volume"
+ " %s" % volume_config['name'])
+ g.log.info("Successful in setting volume %s",
+ volume_config['name'])
+
+ # Mount volume
+ mount_point = tempfile.mkdtemp()
+ cls.mount_points.append(mount_point)
+ cls.mount_points_and_volnames[volume_config['name']] = \
+ mount_point
+ ret, _, _ = mount_volume(volume_config['name'],
+ cls.mount_type,
+ mount_point,
+ cls.mnode,
+ cls.client)
+ if ret:
+ raise ExecutionError(
+ "Failed to do gluster mount on volume %s "
+ % cls.volname)
+ g.log.info("Successfully mounted %s on client %s",
+ cls.volname, cls.client)
+
+ def setUp(self):
+ # Calling GlusterBaseClass setUp
+ GlusterBaseClass.setUp.im_func(self)
+
+ self.all_mounts_procs = []
+ self.io_validation_complete = False
+
+ def tearDown(self):
+ """
+ If test method failed before validating IO, tearDown waits for the
+ IO's to complete and checks for the IO exit status
+ Cleanup and umount volume
+ """
+ if not self.io_validation_complete:
+ g.log.info("Wait for IO to complete as IO validation did not "
+ "succeed in test method")
+ ret = wait_for_io_to_complete(self.all_mounts_procs, self.mounts)
+ if not ret:
+ raise ExecutionError("IO failed on some of the clients")
+ g.log.info("IO is successful on all mounts")
+
+ # List all files and dirs created
+ g.log.info("List all files and directories:")
+ ret = list_all_files_and_dirs_mounts(self.mounts)
+ if not ret:
+ raise ExecutionError("Failed to list all files and dirs")
+ g.log.info("Listing all files and directories is successful")
+
+ # Calling GlusterBaseClass teardown
+ GlusterBaseClass.tearDown.im_func(self)
+
+ @classmethod
+ def tearDownClass(cls):
+ """
+ Clean up the volume and umount volume from client
+ """
+ # stopping all volumes
+ g.log.info("Starting to Cleanup all Volumes")
+ volume_list = get_volume_list(cls.mnode)
+ for volume in volume_list:
+ ret = cleanup_volume(cls.mnode, volume)
+ if not ret:
+ raise ExecutionError("Failed to cleanup Volume %s" % volume)
+ g.log.info("Volume: %s cleanup is done", volume)
+ g.log.info("Successfully Cleanedup all Volumes")
+
+ # umount all volumes
+ for mount_point in cls.mount_points:
+ ret, _, _ = umount_volume(cls.client, mount_point)
+ if ret:
+ raise ExecutionError(
+ "Failed to umount on volume %s "
+ % cls.volname)
+ g.log.info("Successfully umounted %s on client %s", cls.volname,
+ cls.client)
+
+ # calling GlusterBaseClass tearDownClass
+ GlusterBaseClass.tearDownClass.im_func(cls)
+
+ def test_client_side_quorum_auto_local_to_volume_not_cluster(self):
+ """
+ - create four volume as below
+ vol1->2x2
+ vol2->2x2
+ vol3->2x3
+ vol4->2x3
+ vol5->a pure distribute volume
+ - do IO to all vols
+ - set client side quorum to auto for vol1 and vol3
+ - get the client side quorum value for all vols and check for result
+ - bring down b0 on vol1 and b0 and b1 on vol3
+ - try to create files on all vols and check for result
+ """
+ # pylint: disable=too-many-locals,too-many-statements
+ # Creating files for all volumes
+ for mount_point in self.mount_points:
+ self.all_mounts_procs = []
+ g.log.info('Creating files...')
+ command = ("python %s create_files -f 50 "
+ "--fixed-file-size 1k %s"
+ % (self.script_upload_path, mount_point))
+
+ proc = g.run_async(self.mounts[0].client_system, command)
+ self.all_mounts_procs.append(proc)
+ self.io_validation_complete = False
+
+ # Validate IO
+ self.assertTrue(
+ validate_io_procs(self.all_mounts_procs, self.mounts),
+ "IO failed on some of the clients"
+ )
+ self.io_validation_complete = True
+
+ volumes_to_change_options = ['1', '3']
+ # set cluster.quorum-type to auto
+ for vol_number in volumes_to_change_options:
+ vol_name = ('testvol_distributed-replicated_%s'
+ % vol_number)
+ options = {"cluster.quorum-type": "auto"}
+ g.log.info("setting cluster.quorum-type to auto on "
+ "volume testvol_distributed-replicated_%s", vol_number)
+ ret = set_volume_options(self.mnode, vol_name, options)
+ self.assertTrue(ret, ("Unable to set volume option %s for "
+ "volume %s" % (options, vol_name)))
+ g.log.info("Successfully set %s for volume %s", options, vol_name)
+
+ # check is options are set correctly
+ volume_list = get_volume_list(self.mnode)
+ for volume in volume_list:
+ g.log.info('Checking for cluster.quorum-type option for %s',
+ volume)
+ volume_options_dict = get_volume_options(self.mnode,
+ volume,
+ 'cluster.quorum-type')
+ if (volume == 'testvol_distributed-replicated_1' or
+ volume == 'testvol_distributed-replicated_3' or
+ volume == 'testvol_distributed-replicated_4'):
+ self.assertEqual(volume_options_dict['cluster.quorum-type'],
+ 'auto',
+ 'Option cluster.quorum-type '
+ 'is not AUTO for %s'
+ % volume)
+ g.log.info('Option cluster.quorum-type is AUTO for %s', volume)
+ else:
+ self.assertEqual(volume_options_dict['cluster.quorum-type'],
+ 'none',
+ 'Option cluster.quorum-type '
+ 'is not NONE for %s'
+ % volume)
+ g.log.info('Option cluster.quorum-type is NONE for %s', volume)
+
+ # Get first brick server and brick path
+ # and get first file from filelist then delete it from volume
+ vols_file_list = {}
+ for volume in volume_list:
+ brick_list = get_all_bricks(self.mnode, volume)
+ brick_server, brick_path = brick_list[0].split(':')
+ ret, file_list, _ = g.run(brick_server, 'ls %s' % brick_path)
+ self.assertFalse(ret, 'Failed to ls files on %s' % brick_server)
+ file_from_vol = file_list.splitlines()[0]
+ ret, _, _ = g.run(brick_server, 'rm -rf %s/%s'
+ % (brick_path, file_from_vol))
+ self.assertFalse(ret, 'Failed to rm file on %s' % brick_server)
+ vols_file_list[volume] = file_from_vol
+
+ # bring bricks offline
+ # bring first brick for testvol_distributed-replicated_1
+ volname = 'testvol_distributed-replicated_1'
+ brick_list = get_all_bricks(self.mnode, volname)
+ bricks_to_bring_offline = brick_list[0:1]
+ g.log.info('Bringing bricks %s offline...', bricks_to_bring_offline)
+ ret = bring_bricks_offline(volname, bricks_to_bring_offline)
+ self.assertTrue(ret, 'Failed to bring bricks %s offline' %
+ bricks_to_bring_offline)
+
+ ret = are_bricks_offline(self.mnode, volname,
+ bricks_to_bring_offline)
+ self.assertTrue(ret, 'Bricks %s are not offline'
+ % bricks_to_bring_offline)
+ g.log.info('Bringing bricks %s offline is successful',
+ bricks_to_bring_offline)
+
+ # bring first two bricks for testvol_distributed-replicated_3
+ volname = 'testvol_distributed-replicated_3'
+ brick_list = get_all_bricks(self.mnode, volname)
+ bricks_to_bring_offline = brick_list[0:2]
+ g.log.info('Bringing bricks %s offline...', bricks_to_bring_offline)
+ ret = bring_bricks_offline(volname, bricks_to_bring_offline)
+ self.assertTrue(ret, 'Failed to bring bricks %s offline' %
+ bricks_to_bring_offline)
+
+ ret = are_bricks_offline(self.mnode, volname,
+ bricks_to_bring_offline)
+ self.assertTrue(ret, 'Bricks %s are not offline'
+ % bricks_to_bring_offline)
+ g.log.info('Bringing bricks %s offline is successful',
+ bricks_to_bring_offline)
+
+ # merge two dicts (volname: file_to_delete) and (volname: mountpoint)
+ temp_dict = [vols_file_list, self.mount_points_and_volnames]
+ file_to_delete_to_mountpoint_dict = {}
+ for k in vols_file_list.iterkeys():
+ file_to_delete_to_mountpoint_dict[k] = (
+ tuple(file_to_delete_to_mountpoint_dict[k]
+ for file_to_delete_to_mountpoint_dict in
+ temp_dict))
+
+ # create files on all volumes and check for result
+ for volname, file_and_mountpoint in \
+ file_to_delete_to_mountpoint_dict.iteritems():
+ filename, mountpoint = file_and_mountpoint
+
+ # check for ROFS error for read-only file system for
+ # testvol_distributed-replicated_1 and
+ # testvol_distributed-replicated_3
+ if (volname == 'testvol_distributed-replicated_1' or
+ volname == 'testvol_distributed-replicated_3'):
+ # create new file taken from vols_file_list
+ g.log.info("Start creating new file on all mounts...")
+ all_mounts_procs = []
+ cmd = ("touch %s/%s" % (mountpoint, filename))
+
+ proc = g.run_async(self.client, cmd)
+ all_mounts_procs.append(proc)
+
+ # Validate IO
+ g.log.info("Validating if IO failed with read-only filesystem")
+ ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
+ self.mounts)
+ self.assertTrue(ret, ("Unexpected error and IO successful"
+ " on read-only filesystem"))
+ g.log.info("EXPECTED: "
+ "Read-only file system in IO while creating file")
+
+ # check for no errors for all the rest volumes
+ else:
+ # create new file taken from vols_file_list
+ g.log.info("Start creating new file on all mounts...")
+ all_mounts_procs = []
+ cmd = ("touch %s/%s" % (mountpoint, filename))
+
+ proc = g.run_async(self.client, cmd)
+ all_mounts_procs.append(proc)
+
+ # Validate IO
+ self.assertTrue(
+ validate_io_procs(all_mounts_procs, self.mounts),
+ "IO failed on some of the clients"
+ )