From ec70a07bf0d5a612024b12e490c075e9f80bf7fb Mon Sep 17 00:00:00 2001 From: Anees Patel Date: Sun, 9 Dec 2018 19:38:47 +0530 Subject: Fixing the test-case to delete brick directories of replaced bricks, As this was not handled by teardown class Change-Id: I789adbf0909c5edd0a2eb19ed4ccebcb654700fd Signed-off-by: Anees Patel --- .../afr/heal/test_self_heal_daemon_process.py | 158 ---------------- .../test_impact_of_replace_brick_for_glustershd.py | 206 +++++++++++++++++++++ 2 files changed, 206 insertions(+), 158 deletions(-) mode change 100755 => 100644 tests/functional/afr/heal/test_self_heal_daemon_process.py create mode 100644 tests/functional/arbiter/brick_cases/test_impact_of_replace_brick_for_glustershd.py diff --git a/tests/functional/afr/heal/test_self_heal_daemon_process.py b/tests/functional/afr/heal/test_self_heal_daemon_process.py old mode 100755 new mode 100644 index b66ae25d2..06093effb --- a/tests/functional/afr/heal/test_self_heal_daemon_process.py +++ b/tests/functional/afr/heal/test_self_heal_daemon_process.py @@ -33,7 +33,6 @@ from glustolibs.gluster.rebalance_ops import (rebalance_start, from glustolibs.gluster.brick_libs import ( get_all_bricks, bring_bricks_offline, bring_bricks_online, are_bricks_online, select_bricks_to_bring_offline, are_bricks_offline) -from glustolibs.gluster.brick_ops import replace_brick from glustolibs.gluster.heal_libs import (get_self_heal_daemon_pid, do_bricks_exist_in_shd_volfile, is_shd_daemonized, @@ -657,163 +656,6 @@ class SelfHealDaemonProcessTests(GlusterBaseClass): self.io_validation_complete = True -@runs_on([['replicated', 'distributed-replicated'], - ['glusterfs', 'nfs', 'cifs']]) -class ImpactOfReplaceBrickForGlustershdTests(GlusterBaseClass): - """ - ClientSideQuorumTests contains tests which verifies the - client side quorum Test Cases - """ - - @classmethod - def setUpClass(cls): - # Calling GlusterBaseClass setUpClass - GlusterBaseClass.setUpClass.im_func(cls) - - # Override Volumes - if cls.volume_type == "distributed-replicated": - # Define distributed-replicated volume - cls.volume['voltype'] = { - 'type': 'distributed-replicated', - 'dist_count': 2, - 'replica_count': 3, - 'arbiter_count': 1, - 'transport': 'tcp'} - - cls.glustershd = "/var/lib/glusterd/glustershd/glustershd-server.vol" - - def setUp(self): - """ - setUp method for every test - """ - - # calling GlusterBaseClass setUp - GlusterBaseClass.setUp.im_func(self) - - self.all_mounts_procs = [] - self.io_validation_complete = False - - # Setup Volume and Mount Volume - g.log.info("Starting to Setup Volume %s", self.volname) - ret = self.setup_volume_and_mount_volume(self.mounts) - if not ret: - raise ExecutionError("Failed to Setup_Volume and Mount_Volume") - g.log.info("Successful in Setup Volume and Mount Volume") - - def tearDown(self): - """ - If test method failed before validating IO, tearDown waits for the - IO's to complete and checks for the IO exit status - - Cleanup and umount volume - """ - # Cleanup and umount volume - g.log.info("Starting to Unmount Volume and Cleanup Volume") - ret = self.unmount_volume_and_cleanup_volume(mounts=self.mounts) - if not ret: - raise ExecutionError("Failed to umount the vol & cleanup Volume") - g.log.info("Successful in umounting the volume and Cleanup") - - # Calling GlusterBaseClass teardown - GlusterBaseClass.tearDown.im_func(self) - - def test_impact_of_replace_brick_for_glustershd(self): - nodes = self.volume['servers'] - - # check the self-heal daemon process - g.log.info("Starting to get self-heal daemon process on " - "nodes %s", nodes) - ret, pids = get_self_heal_daemon_pid(nodes) - self.assertTrue(ret, ("Either No self heal daemon process found or " - "more than One self heal daemon process " - "found : %s" % pids)) - g.log.info("Successful in getting Single self heal daemon process" - " on all nodes %s", nodes) - glustershd_pids = pids - - # get the bricks for the volume - g.log.info("Fetching bricks for the volume : %s", self.volname) - bricks_list = get_all_bricks(self.mnode, self.volname) - g.log.info("Brick List : %s", bricks_list) - - # validate the bricks present in volume info with - # glustershd server volume file - g.log.info("Starting parsing file %s on " - "node %s", self.glustershd, self.mnode) - ret = do_bricks_exist_in_shd_volfile(self.mnode, self.volname, - bricks_list) - self.assertTrue(ret, ("Brick List from volume info is different " - "from glustershd server volume file. " - "Please check log file for details")) - g.log.info("Successfully parsed %s file", self.glustershd) - - # replace brick - brick_to_replace = bricks_list[-1] - new_brick = brick_to_replace + 'new' - g.log.info("Replacing the brick %s for the volume : %s", - brick_to_replace, self.volname) - ret, _, err = replace_brick(self.mnode, self.volname, - brick_to_replace, new_brick) - self.assertFalse(ret, err) - g.log.info('Replaced brick %s to %s successfully', - brick_to_replace, new_brick) - - # check bricks - bricks_list = get_all_bricks(self.mnode, self.volname) - self.assertEqual(bricks_list[-1], new_brick, 'Replaced brick and ' - 'new brick are not equal') - - # Verify volume's all process are online for 60 sec - g.log.info("Verifying volume's all process are online") - ret = wait_for_volume_process_to_be_online(self.mnode, self.volname, - timeout=60) - self.assertTrue(ret, ("Volume %s : All process are not " - "online", self.volname)) - g.log.info("Successfully Verified volume %s processes are online", - self.volname) - - # Verify glustershd process releases its parent process - ret = is_shd_daemonized(nodes) - self.assertTrue(ret, ("Either No self heal daemon process found or " - "more than One self heal daemon process found")) - - # check the self-heal daemon process - g.log.info("Starting to get self-heal daemon process on " - "nodes %s", nodes) - ret, pids = get_self_heal_daemon_pid(nodes) - self.assertTrue(ret, ("Either No self heal daemon process found or " - "more than One self heal daemon process " - "found : %s" % pids)) - g.log.info("Successful in getting Single self heal daemon process" - " on all nodes %s", nodes) - glustershd_pids_after_replacement = pids - - # Compare pids before and after replacing - self.assertNotEqual(glustershd_pids, - glustershd_pids_after_replacement, - "Self Daemon process is same before and" - " after replacing bricks") - g.log.info("Self Heal Daemon Process is different before and " - "after replacing bricks") - - # get the bricks for the volume after replacing - bricks_list_after_replacing = get_all_bricks(self.mnode, self.volname) - g.log.info("Brick List after expanding " - "volume: %s", bricks_list_after_replacing) - - # validate the bricks present in volume info - # with glustershd server volume file after replacing bricks - g.log.info("Starting parsing file %s", self.glustershd) - ret = do_bricks_exist_in_shd_volfile(self.mnode, self.volname, - bricks_list_after_replacing) - - self.assertTrue(ret, ("Brick List from volume info is different " - "from glustershd server volume file after " - "replacing bricks. Please check log file " - "for details")) - g.log.info("Successfully parsed %s file", self.glustershd) - - class SelfHealDaemonProcessTestsWithMultipleVolumes(GlusterBaseClass): """ SelfHealDaemonProcessTestsWithMultipleVolumes contains tests which diff --git a/tests/functional/arbiter/brick_cases/test_impact_of_replace_brick_for_glustershd.py b/tests/functional/arbiter/brick_cases/test_impact_of_replace_brick_for_glustershd.py new file mode 100644 index 000000000..a8a222c24 --- /dev/null +++ b/tests/functional/arbiter/brick_cases/test_impact_of_replace_brick_for_glustershd.py @@ -0,0 +1,206 @@ +# Copyright (C) 2016-2017 Red Hat, Inc. +# +# This program is free software; you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation; either version 2 of the License, or +# any later version. +# +# This program is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License along +# with this program; if not, write to the Free Software Foundation, Inc., +# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. + +""" Description: + Test Cases in this module tests the self heal daemon process. +""" + +from glusto.core import Glusto as g +from glustolibs.gluster.exceptions import ExecutionError +from glustolibs.gluster.gluster_base_class import GlusterBaseClass, runs_on +from glustolibs.gluster.volume_libs import ( + wait_for_volume_process_to_be_online, get_subvols) +from glustolibs.gluster.brick_libs import get_all_bricks +from glustolibs.gluster.brick_ops import replace_brick +from glustolibs.gluster.heal_libs import (get_self_heal_daemon_pid, + do_bricks_exist_in_shd_volfile, + is_shd_daemonized) + + +@runs_on([['replicated', 'distributed-replicated'], + ['glusterfs', 'nfs']]) +class ImpactOfReplaceBrickForGlustershdTests(GlusterBaseClass): + """ + ClientSideQuorumTests contains tests which verifies the + client side quorum Test Cases + """ + + @classmethod + def setUpClass(cls): + # Calling GlusterBaseClass setUpClass + GlusterBaseClass.setUpClass.im_func(cls) + + # Override Volumes + if cls.volume_type == "distributed-replicated": + # Define distributed-replicated volume + cls.volume['voltype'] = { + 'type': 'distributed-replicated', + 'dist_count': 2, + 'replica_count': 3, + 'arbiter_count': 1, + 'transport': 'tcp'} + + cls.glustershd = "/var/lib/glusterd/glustershd/glustershd-server.vol" + + def setUp(self): + """ + setUp method for every test + """ + + # calling GlusterBaseClass setUp + GlusterBaseClass.setUp.im_func(self) + + self.all_mounts_procs = [] + self.io_validation_complete = False + + # Setup Volume and Mount Volume + g.log.info("Starting to Setup Volume %s", self.volname) + ret = self.setup_volume_and_mount_volume(self.mounts) + if not ret: + raise ExecutionError("Failed to Setup_Volume and Mount_Volume") + g.log.info("Successful in Setup Volume and Mount Volume") + + def tearDown(self): + """ + If test method failed before validating IO, tearDown waits for the + IO's to complete and checks for the IO exit status + + Cleanup and umount volume + """ + # Cleanup and umount volume + g.log.info("Starting to Unmount Volume and Cleanup Volume") + ret = self.unmount_volume_and_cleanup_volume(mounts=self.mounts) + if not ret: + raise ExecutionError("Failed to umount the vol & cleanup Volume") + g.log.info("Successful in umounting the volume and Cleanup") + + # Calling GlusterBaseClass teardown + GlusterBaseClass.tearDown.im_func(self) + + def test_impact_of_replace_brick_for_glustershd(self): + # pylint: disable=too-many-statements,too-many-branches,too-many-locals + nodes = self.volume['servers'] + replaced_bricks = [] + + # check the self-heal daemon process + g.log.info("Starting to get self-heal daemon process on " + "nodes %s", nodes) + ret, pids = get_self_heal_daemon_pid(nodes) + self.assertTrue(ret, ("Either No self heal daemon process found or " + "more than One self heal daemon process " + "found : %s" % pids)) + g.log.info("Successful in getting Single self heal daemon process" + " on all nodes %s", nodes) + glustershd_pids = pids + + # get the bricks for the volume + g.log.info("Fetching bricks for the volume : %s", self.volname) + bricks_list = get_all_bricks(self.mnode, self.volname) + g.log.info("Brick List : %s", bricks_list) + + # validate the bricks present in volume info with + # glustershd server volume file + g.log.info("Starting parsing file %s on " + "node %s", self.glustershd, self.mnode) + ret = do_bricks_exist_in_shd_volfile(self.mnode, self.volname, + bricks_list) + self.assertTrue(ret, ("Brick List from volume info is different " + "from glustershd server volume file. " + "Please check log file for details")) + g.log.info("Successfully parsed %s file", self.glustershd) + + # get the subvolumes + g.log.info("Starting to get sub-volumes for volume %s", self.volname) + subvols_dict = get_subvols(self.mnode, self.volname) + num_subvols = len(subvols_dict['volume_subvols']) + g.log.info("Number of subvolumes in volume %s:", num_subvols) + + # replace brick from each sub-vol + for i in range(0, num_subvols): + subvol_brick_list = subvols_dict['volume_subvols'][i] + g.log.info("sub-volume %s brick list : %s", i, subvol_brick_list) + brick_to_replace = subvol_brick_list[-1] + new_brick = brick_to_replace + 'new' + g.log.info("Replacing the brick %s for the volume : %s", + brick_to_replace, self.volname) + ret, _, err = replace_brick(self.mnode, self.volname, + brick_to_replace, new_brick) + self.assertFalse(ret, err) + g.log.info('Replaced brick %s to %s successfully', + brick_to_replace, new_brick) + replaced_bricks.append(brick_to_replace) + + # Verify volume's all process are online for 60 sec + g.log.info("Verifying volume's all process are online") + ret = wait_for_volume_process_to_be_online(self.mnode, self.volname, + timeout=60) + self.assertTrue(ret, ("Volume %s : All process are not " + "online", self.volname)) + g.log.info("Successfully Verified volume %s processes are online", + self.volname) + + # Verify glustershd process releases its parent process + ret = is_shd_daemonized(nodes) + self.assertTrue(ret, ("Either No self heal daemon process found or " + "more than One self heal daemon process found")) + + # check the self-heal daemon process + g.log.info("Starting to get self-heal daemon process on nodes " + "%s", nodes) + ret, pids = get_self_heal_daemon_pid(nodes) + self.assertTrue(ret, ("Either No self heal daemon process found or" + " more than One self heal daemon process" + " found : %s" % pids)) + g.log.info("Successful in getting Single self heal daemon process" + " on all nodes %s", nodes) + glustershd_pids_after_replacement = pids + + # Compare pids before and after replacing + self.assertNotEqual(glustershd_pids, + glustershd_pids_after_replacement, + "Self Daemon process is same before and" + " after replacing bricks") + g.log.info("Self Heal Daemon Process is different before and " + "after replacing bricks") + + # get the bricks for the volume after replacing + bricks_list_after_replacing = get_all_bricks(self.mnode, self.volname) + g.log.info("Brick List after expanding " + "volume: %s", bricks_list_after_replacing) + + # validate the bricks present in volume info + # with glustershd server volume file after replacing bricks + g.log.info("Starting parsing file %s", self.glustershd) + ret = do_bricks_exist_in_shd_volfile(self.mnode, self.volname, + bricks_list_after_replacing) + + self.assertTrue(ret, ("Brick List from volume info is different " + "from glustershd server volume file after " + "replacing bricks. Please check log file " + "for details")) + g.log.info("Successfully parsed %s file", self.glustershd) + g.log.info("Starting to delete replaced brick dir's") + + # Remove brick directories of the replaced bricks as this is not + # handled by tearDown class + for bricks in replaced_bricks: + node, brick_path = bricks.split(r':') + cmd = "rm -rf " + brick_path + ret, _, _ = g.run(node, cmd) + if ret: + raise ExecutionError("Failed to delete the brick dir's for" + " %s and brick %s" % (node, brick_path)) + g.log.info("Successfully deleted brick dir's for replaced bricks") -- cgit