summaryrefslogtreecommitdiffstats
path: root/tests
diff options
context:
space:
mode:
authorRajesh Madaka <rmadaka@redhat.com>2018-01-23 11:39:16 +0530
committerShwethaHPanduranga <shwetha-h-p@review.gluster.org>2018-01-28 09:53:24 +0000
commit7d4bf9511ba4b5b685717fc05177d50e14b0a6b1 (patch)
tree46097dc0adc41354f0b83a525ead892823610496 /tests
parentfe5ac28d9f422eb14c168dd0bda68c25e2167aa0 (diff)
Test case for bitd, scrub, snapd checking after volume reset and volume
reset force Description: Create Distribute volume, then enable bitrot and uss on that volume, then check the bitd, scrub, snapd daemons running or not. Then peform the volume reset, after volume reset only snap daemon will get kill but bitd scrub daemons will be remain running. Then perform volume reset with force, after volume reset with force all three(bitd, scrub, snapd) daemons will get kill, these daemons will not be running. below are the steps performed for developing this test case: -> Create Distributed volume -> Enable BitD, Scrub and Uss on volume -> Verify the BitD, Scrub and Uss daemons are running on every node -> Reset the volume -> Verify the Daemons (BitD, Scrub & Uss ) are running or not -> Eanble Uss on same volume -> Reset the volume with force -> Verify all the daemons(BitD, Scrub & Uss) are running or not Change-Id: I15d71d1434ec84d80293fda2ab6a8d02a3af5fd6 Signed-off-by: Rajesh Madaka <rmadaka@redhat.com>
Diffstat (limited to 'tests')
-rw-r--r--tests/functional/glusterd/test_volume_reset.py200
1 files changed, 200 insertions, 0 deletions
diff --git a/tests/functional/glusterd/test_volume_reset.py b/tests/functional/glusterd/test_volume_reset.py
new file mode 100644
index 000000000..2bb8c4c24
--- /dev/null
+++ b/tests/functional/glusterd/test_volume_reset.py
@@ -0,0 +1,200 @@
+# Copyright (C) 2017-2018 Red Hat, Inc. <http://www.redhat.com>
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License as published by
+# the Free Software Foundation; either version 2 of the License, or
+# any later version.
+#
+# This program is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+# GNU General Public License for more details.
+#
+# You should have received a copy of the GNU General Public License along
+# with this program; if not, write to the Free Software Foundation, Inc.,
+# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
+
+""" Description:
+ Test Cases in this module related to Glusterd volume reset validation
+ with bitd, scrub and snapd daemons running or not
+"""
+from glusto.core import Glusto as g
+from glustolibs.gluster.exceptions import ExecutionError
+from glustolibs.gluster.gluster_base_class import GlusterBaseClass, runs_on
+from glustolibs.gluster.peer_ops import peer_probe_servers
+from glustolibs.gluster.volume_libs import cleanup_volume
+from glustolibs.gluster.bitrot_ops import (enable_bitrot, is_bitd_running,
+ is_scrub_process_running)
+from glustolibs.gluster.uss_ops import enable_uss, is_snapd_running
+
+
+@runs_on([['distributed', 'replicated', 'distributed-replicated',
+ 'dispersed', 'distributed-dispersed'], ['glusterfs']])
+class GlusterdVolumeReset(GlusterBaseClass):
+ @classmethod
+ def setUpClass(cls):
+ GlusterBaseClass.setUpClass.im_func(cls)
+ g.log.info("Starting %s " % cls.__name__)
+ '''
+ checking for peer status from every node, if peers are in not
+ connected state, performing peer probe.
+ '''
+ ret = cls.validate_peers_are_connected()
+ if not ret:
+ ret = peer_probe_servers(cls.mnode, cls.servers)
+ if ret:
+ g.log.info("peers are connected successfully from %s to other \
+ servers in severlist %s:" % (cls.mnode, cls.servers))
+ else:
+ g.log.error("Peer probe failed from %s to other \
+ servers in severlist %s:" % (cls.mnode, cls.servers))
+ raise ExecutionError("Peer probe failed ")
+ else:
+ g.log.info("All server peers are already in connected state\
+ %s:" % cls.servers)
+
+ # Creating Volume
+ g.log.info("Started creating volume")
+ ret = cls.setup_volume()
+ if ret:
+ g.log.info("Volme created successfully : %s" % cls.volname)
+ else:
+ raise ExecutionError("Volume creation failed: %s" % cls.volname)
+
+ def setUp(self):
+ """
+ setUp method for every test
+ """
+ # calling GlusterBaseClass setUp
+ GlusterBaseClass.setUp.im_func(self)
+
+ # command for volume reset
+ g.log.info("started resetting volume")
+ cmd = "gluster volume reset " + self.volname
+ ret, out, _ = g.run(self.mnode, cmd)
+ if (ret == 0):
+ g.log.info("volume restted successfully :%s" % self.volname)
+ else:
+ raise ExecutionError("Volume reset Failed :%s" % self.volname)
+
+ def tearDown(self):
+ """
+ tearDown for every test
+ """
+ # Calling GlusterBaseClass tearDown
+ GlusterBaseClass.tearDown.im_func(self)
+
+ @classmethod
+ def tearDownClass(cls):
+ # stopping the volume and Cleaning up the volume
+ ret = cleanup_volume(cls.mnode, cls.volname)
+ if ret:
+ g.log.info("Volume deleted successfully : %s" % cls.volname)
+ else:
+ raise ExecutionError("Failed Cleanup the Volume %s" % cls.volname)
+
+ def test_bitd_scrubd_snapd_after_volume_reset(self):
+ '''
+ -> Create volume
+ -> Enable BitD, Scrub and Uss on volume
+ -> Verify the BitD, Scrub and Uss daemons are running on every node
+ -> Reset the volume
+ -> Verify the Daemons (BitD, Scrub & Uss ) are running or not
+ -> Eanble Uss on same volume
+ -> Reset the volume with force
+ -> Verify all the daemons(BitD, Scrub & Uss) are running or not
+ :return:
+ '''
+
+ # enable bitrot and scrub on volume
+ g.log.info("Enabling bitrot")
+ ret, out, _ = enable_bitrot(self.mnode, self.volname)
+ self.assertEqual(ret, 0, "Failed to enable bitrot on\
+ volume: %s" % self.volname)
+ g.log.info("Bitd and scrub daemons enabled\
+ successfully on volume :%s" % self.volname)
+
+ # enable uss on volume
+ g.log.info("Enabling snaphot(uss)")
+ ret, out, _ = enable_uss(self.mnode, self.volname)
+ self.assertEqual(ret, 0, "Failed to enable uss on\
+ volume: %s" % self.volname)
+ g.log.info("uss enabled successfully on volume :%s" % self.volname)
+
+ # Checks bitd, snapd, scrub daemons running or not
+ g.log.info("checking snapshot, scrub and bitrot\
+ daemons running or not")
+ for mnode in self.servers:
+ ret = is_bitd_running(mnode, self.volname)
+ self.assertTrue(ret, "Bitrot Daemon\
+ not running on %s server:" % mnode)
+ ret = is_scrub_process_running(mnode, self.volname)
+ self.assertTrue(ret, "Scrub Daemon\
+ not running on %s server:" % mnode)
+ ret = is_snapd_running(mnode, self.volname)
+ self.assertTrue(ret, "Snap Daemon\
+ not running %s server:" % mnode)
+ g.log.info("bitd, scrub and snapd running\
+ successflly on volume :%s" % self.volname)
+
+ # command for volume reset
+ g.log.info("started resetting volume")
+ cmd = "gluster volume reset " + self.volname
+ ret, out, _ = g.run(self.mnode, cmd)
+ self.assertEqual(ret, 0, "volume reset failed\
+ for : %s" % self.volname)
+ g.log.info("volume resetted succefully :%s" % self.volname)
+
+ '''
+ After volume reset snap daemon will not be running,
+ bitd and scrub deamons will be in running state.
+ '''
+ g.log.info("checking snapshot, scrub and bitrot daemons\
+ running or not after volume reset")
+ for mnode in self.servers:
+ ret = is_bitd_running(mnode, self.volname)
+ self.assertTrue(ret, "Bitrot Daemon\
+ not running on %s server:" % mnode)
+ ret = is_scrub_process_running(mnode, self.volname)
+ self.assertTrue(ret, "Scrub Daemon\
+ not running on %s server:" % mnode)
+ ret = is_snapd_running(mnode, self.volname)
+ self.assertFalse(ret, "Snap Daemon should not be\
+ running on %s server after volume reset:" % mnode)
+ g.log.info("bitd and scrub daemons are running after volume reset\
+ snapd is not running as expected on volume :%s" % self.volname)
+
+ # enable uss on volume
+ g.log.info("Enabling snaphot(uss)")
+ ret, out, _ = enable_uss(self.mnode, self.volname)
+ self.assertEqual(ret, 0, "Failed to enable\
+ uss on volume: %s" % self.volname)
+ g.log.info("uss enabled successfully on volume :%s" % self.volname)
+
+ # command for volume reset with force
+ g.log.info("started resetting volume with force option")
+ cmd = "gluster volume reset " + self.volname + " force"
+ ret, out, _ = g.run(self.mnode, cmd)
+ self.assertEqual(ret, 0, "volume reset fail\
+ for : %s" % self.volname)
+ g.log.info("Volume resetted sucessfully with\
+ force option :%s" % self.volname)
+
+ '''
+ After volume reset bitd, snapd, scrub daemons will not be running,
+ all three daemons will get die
+ '''
+ g.log.info("checking snapshot, scrub and bitrot daemons\
+ running or not after volume reset with force")
+ for mnode in self.servers:
+ ret = is_bitd_running(mnode, self.volname)
+ self.assertFalse(ret, "Bitrot Daemon should not be\
+ running on %s server after volume reset with force:" % mnode)
+ ret = is_scrub_process_running(mnode, self.volname)
+ self.assertFalse(ret, "Scrub Daemon shiuld not be running\
+ on %s server after volume reset with force:" % mnode)
+ ret = is_snapd_running(mnode, self.volname)
+ self.assertFalse(ret, "Snap Daemon should not be\
+ running on %s server after volume reset force:" % mnode)
+ g.log.info("After volume reset bitd, scrub and snapd are not running after\
+ volume reset with force on volume :%s" % self.volname)