summaryrefslogtreecommitdiffstats
path: root/tests/bvt/test_bvt_lite_and_plus.py
diff options
context:
space:
mode:
authorShwetha Panduranga <spandura@redhat.com>2016-10-03 15:38:38 +0530
committerShwetha Panduranga <spandura@redhat.com>2016-10-26 00:11:27 +0530
commit01a5f8306b9f772fa1f5f46c07355b747a2c0af6 (patch)
treeacbddfde982c2eb9893e9d62264eb4671df0aa70 /tests/bvt/test_bvt_lite_and_plus.py
parent0e0f958776091fb155822323467efd114d7326b5 (diff)
Adding a bvt testcase
Change-Id: Ide7e3bac46fbaf354a2a5c8baef8510b4aefec78 Signed-off-by: Shwetha Panduranga <spandura@redhat.com>
Diffstat (limited to 'tests/bvt/test_bvt_lite_and_plus.py')
-rw-r--r--tests/bvt/test_bvt_lite_and_plus.py213
1 files changed, 213 insertions, 0 deletions
diff --git a/tests/bvt/test_bvt_lite_and_plus.py b/tests/bvt/test_bvt_lite_and_plus.py
new file mode 100644
index 000000000..e47c13477
--- /dev/null
+++ b/tests/bvt/test_bvt_lite_and_plus.py
@@ -0,0 +1,213 @@
+#!/usr/bin/env python
+# Copyright (C) 2015-2016 Red Hat, Inc. <http://www.redhat.com>
+#
+# This program is free software; you can redistribute it and/or modify
+# it under the terms of the GNU General Public License as published by
+# the Free Software Foundation; either version 2 of the License, or
+# any later version.
+#
+# This program is distributed in the hope that it will be useful,
+# but WITHOUT ANY WARRANTY; without even the implied warranty of
+# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+# GNU General Public License for more details.
+#
+# You should have received a copy of the GNU General Public License along
+# with this program; if not, write to the Free Software Foundation, Inc.,
+# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
+
+import pytest
+import os
+from glusto.core import Glusto as g
+from glustolibs.gluster.gluster_base_class import GlusterBaseClass, runs_on
+from glustolibs.gluster.gluster_init import start_glusterd
+from glustolibs.gluster.peer_ops import (peer_probe_servers, is_peer_connected,
+ peer_status)
+from glustolibs.gluster.volume_libs import setup_volume, cleanup_volume
+from glustolibs.gluster.volume_ops import volume_info, volume_status
+import time
+
+
+@runs_on([['replicated', 'distributed', 'distributed-replicated',
+ 'dispersed', 'distributed-dispersed'],
+ ['glusterfs', 'nfs', 'cifs']])
+class BvtTestsClass(GlusterBaseClass):
+ """Class containing case for : BVT Lite and BVT Plus.
+
+ BVT Lite: Run the case on dis-rep volume with glusterfs, nfs, cifs
+ protocols
+
+ BVT Plus: Run the case on all volume types and all protocol types
+ combinations
+ """
+ @classmethod
+ def setUpClass(cls):
+ """Following are the setps in setupclass
+ - Start glusterd on all servers
+ - Peer Probe
+ - Setup the volume
+ - Mount the volume
+ """
+ GlusterBaseClass.setUpClass.im_func(cls)
+ g.log.info("Starting %s:" % cls.__name__)
+
+ # Start Glusterd
+ ret = start_glusterd(servers=cls.servers)
+ assert (ret == True), "glusterd did not start on at least one server"
+
+ # PeerProbe servers
+ ret = peer_probe_servers(mnode=cls.servers[0], servers=cls.servers[1:])
+ assert (ret == True), "Unable to peer probe one or more servers"
+
+ # Validate if peer is connected from all the servers
+ for server in cls.servers:
+ ret = is_peer_connected(server, cls.servers)
+ assert (ret == True), "Validating Peers to be in Cluster Failed"
+
+ # Print Peer Status from mnode
+ _, _, _ = peer_status(cls.mnode)
+
+ # Setup Volume
+ ret = setup_volume(mnode=cls.mnode,
+ all_servers_info=cls.all_servers_info,
+ volume_config=cls.volume, force=True)
+ assert (ret == True), "Setup volume %s failed" % cls.volname
+ time.sleep(10)
+
+ # Print Volume Info and Status
+ _, _, _ = volume_info(cls.mnode, cls.volname)
+
+ _, _, _ = volume_status(cls.mnode, cls.volname)
+
+ # Validate if volume is exported or not
+ if 'nfs' in cls.mount_type:
+ cmd = "showmount -e localhost"
+ _, _, _ = g.run(cls.mnode, cmd)
+
+ cmd = "showmount -e localhost | grep %s" % cls.volname
+ ret, _, _ = g.run(cls.mnode, cmd)
+ assert (ret == 0), "Volume %s not exported" % cls.volname
+
+ if 'cifs' in cls.mount_type:
+ cmd = "smbclient -L localhost"
+ _, _, _ = g.run(cls.mnode, cmd)
+
+ cmd = ("smbclient -L localhost -U | grep -i -Fw gluster-%s " %
+ cls.volname)
+ ret, _, _ = g.run(cls.mnode, cmd)
+ assert (ret == 0), ("Volume %s not accessable via SMB/CIFS share" %
+ cls.volname)
+
+ # Create Mounts
+ rc = True
+ for mount_obj in cls.mounts:
+ ret = mount_obj.mount()
+ if not ret:
+ g.log.error("Unable to mount volume '%s:%s' on '%s:%s'" %
+ (mount_obj.server_system, mount_obj.volname,
+ mount_obj.client_system, mount_obj.mountpoint))
+ rc = False
+ assert (rc == True), ("Mounting volume %s on few clients failed" %
+ cls.volname)
+
+ # Upload io scripts
+ cls.script_local_path = ("/usr/share/glustolibs/io/"
+ "scripts/file_dir_ops.py")
+ cls.script_upload_path = "/tmp/file_dir_ops.py"
+ ret = os.path.exists(cls.script_local_path)
+ assert (ret == True), ("Unable to find the io scripts")
+
+ for client in cls.clients:
+ g.upload(client, cls.script_local_path, cls.script_upload_path)
+ g.run(client, "ls -l %s" % cls.script_upload_path)
+ g.run(client, "chmod +x %s" % cls.script_upload_path)
+ g.run(client, "ls -l %s" % cls.script_upload_path)
+
+ def setUp(self):
+ pass
+
+ def test_bvt(self):
+ """Test IO from the mounts.
+ """
+ g.log.info("Starting Test: %s on %s %s" %
+ (self.id(), self.volume_type, self.mount_type))
+
+ # Get stat of mount before the IO
+ for mount_obj in self.mounts:
+ cmd = "mount | grep %s" % mount_obj.mountpoint
+ ret, out, err = g.run(mount_obj.client_system, cmd)
+ cmd = "df -h %s" % mount_obj.mountpoint
+ ret, out, err = g.run(mount_obj.client_system, cmd)
+ cmd = "ls -ld %s" % mount_obj.mountpoint
+ ret, out, err = g.run(mount_obj.client_system, cmd)
+ cmd = "stat %s" % mount_obj.mountpoint
+ ret, out, err = g.run(mount_obj.client_system, cmd)
+
+ # Start IO on all mounts.
+ all_mounts_procs = []
+ count = 1
+ for mount_obj in self.mounts:
+ cmd = ("python %s create_deep_dirs_with_files "
+ "--dirname-start-num %d "
+ "--dir-depth 2 "
+ "--dir-length 10 "
+ "--max-num-of-dirs 5 "
+ "--num-of-files 5 %s" % (self.script_upload_path,
+ count, mount_obj.mountpoint))
+ proc = g.run_async(mount_obj.client_system, cmd,
+ user=mount_obj.user)
+ all_mounts_procs.append(proc)
+ count = count + 10
+
+ # Get IO status
+ rc = True
+ for i, proc in enumerate(all_mounts_procs):
+ ret, _, _ = proc.async_communicate()
+ if ret != 0:
+ g.log.error("IO Failed on %s:%s" %
+ (self.mounts[i].client_system,
+ self.mounts[i].mountpoint))
+ rc = False
+ assert (rc == True), "IO failed on some of the clients"
+
+ # Get stat of all the files/dirs created.
+ all_mounts_procs = []
+ for mount_obj in self.mounts:
+ cmd = ("python %s stat "
+ "-R %s" % (self.script_upload_path, mount_obj.mountpoint))
+ proc = g.run_async(mount_obj.client_system, cmd,
+ user=mount_obj.user)
+ all_mounts_procs.append(proc)
+ rc = True
+ for i, proc in enumerate(all_mounts_procs):
+ ret, _, _ = proc.async_communicate()
+ if ret != 0:
+ g.log.error("Stat of files and dirs under %s:%s Failed" %
+ (self.mounts[i].client_system,
+ self.mounts[i].mountpoint))
+ rc = False
+ assert (rc == True), "Stat failed on some of the clients"
+
+ def tearDown(self):
+ pass
+
+ @classmethod
+ def tearDownClass(cls):
+ """Cleanup mount and Cleanup the volume
+ """
+ GlusterBaseClass.tearDownClass.im_func(cls)
+
+ # Unmount mounts
+ rc = True
+ for mount_obj in cls.mounts:
+ ret = mount_obj.unmount()
+ if not ret:
+ g.log.error("Unable to unmount volume '%s:%s' on '%s:%s'" %
+ (mount_obj.server_system, mount_obj.volname,
+ mount_obj.client_system, mount_obj.mountpoint))
+ rc = False
+ assert (rc == True), ("UnMounting volume %s on few clients failed" %
+ cls.volname)
+
+ # Cleanup Volume
+ ret = cleanup_volume(mnode=cls.mnode, volname=cls.volname)
+ assert (ret == True), ("cleanup volume %s failed" % cls.volname)