# Copyright (C) 2016 Red Hat, Inc. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # """ Description: Module containing GlusterBaseClass which defines all the variables necessary for tests. """ import unittest import os import random import copy import datetime import socket from glusto.core import Glusto as g from glustolibs.gluster.exceptions import ConfigError from glustolibs.gluster.peer_ops import is_peer_connected, peer_status from glustolibs.gluster.volume_ops import set_volume_options from glustolibs.gluster.volume_libs import (setup_volume, cleanup_volume, log_volume_info_and_status) from glustolibs.gluster.volume_libs import ( wait_for_volume_process_to_be_online) from glustolibs.gluster.samba_libs import share_volume_over_smb from glustolibs.gluster.nfs_libs import export_volume_through_nfs from glustolibs.gluster.mount_ops import create_mount_objs from glustolibs.io.utils import log_mounts_info from glustolibs.gluster.lib_utils import inject_msg_in_logs class runs_on(g.CarteTestClass): """Decorator providing runs_on capability for standard unittest script""" def __init__(self, value): # the names of the class attributes set by the runs_on decorator self.axis_names = ['volume_type', 'mount_type'] # the options to replace 'ALL' in selections self.available_options = [['distributed', 'replicated', 'distributed-replicated', 'dispersed', 'distributed-dispersed'], ['glusterfs', 'nfs', 'cifs', 'smb']] # these are the volume and mount options to run and set in config # what do runs_on_volumes and runs_on_mounts need to be named???? run_on_volumes = self.available_options[0] run_on_mounts = self.available_options[1] if 'gluster' in g.config and g.config['gluster']: if ('running_on_volumes' in g.config['gluster'] and g.config['gluster']['running_on_volumes']): run_on_volumes = g.config['gluster']['running_on_volumes'] if ('running_on_mounts' in g.config['gluster'] and g.config['gluster']['running_on_mounts']): run_on_mounts = g.config['gluster']['running_on_mounts'] # selections is the above info from the run that is intersected with # the limits from the test script self.selections = [run_on_volumes, run_on_mounts] # value is the limits that are passed in by the decorator self.limits = value class GlusterBaseClass(unittest.TestCase): """GlusterBaseClass to be subclassed by Gluster Tests. This class reads the config for variable values that will be used in gluster tests. If variable values are not specified in the config file, the variable are defaulted to specific values. """ # these will be populated by either the runs_on decorator or # defaults in setUpClass() volume_type = None mount_type = None @classmethod def inject_msg_in_gluster_logs(cls, msg): """Inject all the gluster logs on servers, clients with msg Args: msg (str): Message string to be injected Returns: bool: True if injecting msg on the log files/dirs is successful. False Otherwise. """ _rc = True # Inject msg on server gluster logs ret = inject_msg_in_logs(cls.servers, log_msg=msg, list_of_dirs=cls.server_gluster_logs_dirs, list_of_files=cls.server_gluster_logs_files) if not ret: _rc = False if cls.mount_type is not None and "glusterfs" in cls.mount_type: ret = inject_msg_in_logs( cls.clients, log_msg=msg, list_of_dirs=cls.client_gluster_logs_dirs, list_of_files=cls.client_gluster_logs_files) if not ret: _rc = False return _rc @classmethod def get_ip_from_hostname(cls, nodes): """Returns list of IP's for the list of nodes in order. Args: nodes(list): List of nodes hostnames Returns: list: List of IP's corresponding to the hostnames of nodes. """ nodes_ips = [] if isinstance(nodes, str): nodes = [nodes] for node in nodes: try: ip = socket.gethostbyname(node) except socket.gaierror as e: g.log.error("Failed to get the IP of Host: %s : %s", node, e.strerror) ip = None nodes_ips.append(ip) return nodes_ips @classmethod def validate_peers_are_connected(cls): """Validate whether each server in the cluster is connected to all other servers in cluster. Returns (bool): True if all peers are in connected with other peers. False otherwise. """ # Validate if peer is connected from all the servers g.log.info("Validating if servers %s are connected from other servers " "in the cluster", cls.servers) for server in cls.servers: g.log.info("Validate servers %s are in connected from node %s", cls.servers, server) ret = is_peer_connected(server, cls.servers) if not ret: g.log.error("Some or all servers %s are not in connected " "state from node %s", cls.servers, server) return False g.log.info("Successfully validated servers %s are all in " "connected state from node %s", cls.servers, server) g.log.info("Successfully validated all servers %s are in connected " "state from other servers in the cluster", cls.servers) # Peer Status from mnode peer_status(cls.mnode) return True @classmethod def setup_volume(cls, volume_create_force=False): """Setup the volume: - Create the volume, Start volume, Set volume options, enable snapshot/quota/tier if specified in the config file. - Wait for volume processes to be online - Export volume as NFS/SMB share if mount_type is NFS or SMB - Log volume info and status Args: volume_create_force(bool): True if create_volume should be executed with 'force' option. Returns (bool): True if all the steps mentioned in the descriptions passes. False otherwise. """ force_volume_create = False if volume_create_force or cls.volume_create_force: force_volume_create = True # Validate peers before setting up volume g.log.info("Validate peers before setting up volume ") ret = cls.validate_peers_are_connected() if not ret: g.log.error("Failed to validate peers are in connected state " "before setting up volume") return False g.log.info("Successfully validated peers are in connected state " "before setting up volume") # Setup Volume g.log.info("Setting up volume %s", cls.volname) ret = setup_volume(mnode=cls.mnode, all_servers_info=cls.all_servers_info, volume_config=cls.volume, force=force_volume_create) if not ret: g.log.error("Failed to Setup volume %s", cls.volname) return False g.log.info("Successful in setting up volume %s", cls.volname) # Wait for volume processes to be online g.log.info("Wait for volume %s processes to be online", cls.volname) ret = wait_for_volume_process_to_be_online(cls.mnode, cls.volname) if not ret: g.log.error("Failed to wait for volume %s processes to " "be online", cls.volname) return False g.log.info("Successful in waiting for volume %s processes to be " "online", cls.volname) # Export/Share the volume based on mount_type if cls.mount_type != "glusterfs": g.log.info("Export/Sharing the volume %s", cls.volname) if "nfs" in cls.mount_type: ret = export_volume_through_nfs( mnode=cls.mnode, volname=cls.volname, enable_ganesha=cls.enable_nfs_ganesha) if not ret: g.log.error("Failed to export volume %s " "as NFS export", cls.volname) return False g.log.info("Successful in exporting the volume %s " "as NFS export", cls.volname) # Set NFS-Ganesha specific volume options if cls.enable_nfs_ganesha and cls.nfs_ganesha_export_options: g.log.info("Setting NFS-Ganesha export specific " "volume options on volume %s", cls.volname) ret = set_volume_options( mnode=cls.mnode, volname=cls.volname, options=cls.nfs_ganesha_export_options) if not ret: g.log.error("Failed to set NFS-Ganesha " "export specific options on " "volume %s", cls.volname) return False g.log.info("Successful in setting NFS-Ganesha export " "specific volume options on volume %s", cls.volname) if "smb" in cls.mount_type or "cifs" in cls.mount_type: ret = share_volume_over_smb(mnode=cls.mnode, volname=cls.volname, smb_users_info=cls.smb_users_info) if not ret: g.log.error("Failed to export volume %s " "as SMB Share", cls.volname) return False g.log.info("Successful in exporting volume %s as SMB Share", cls.volname) # Set SMB share specific volume options if cls.smb_share_options: g.log.info("Setting SMB share specific volume options " "on volume %s", cls.volname) ret = set_volume_options(mnode=cls.mnode, volname=cls.volname, options=cls.smb_share_options) if not ret: g.log.error("Failed to set SMB share " "specific options " "on volume %s", cls.volname) return False g.log.info("Successful in setting SMB share specific " "volume options on volume %s", cls.volname) # Log Volume Info and Status g.log.info("Log Volume %s Info and Status", cls.volname) ret = log_volume_info_and_status(cls.mnode, cls.volname) if not ret: g.log.error("Logging volume %s info and status failed", cls.volname) return False g.log.info("Successful in logging volume %s info and status", cls.volname) return True @classmethod def mount_volume(cls, mounts): """Mount volume Args: mounts(list): List of mount_objs Returns (bool): True if mounting the volume for a mount obj is successful. False otherwise """ g.log.info("Starting to mount volume %s", cls.volname) for mount_obj in mounts: g.log.info("Mounting volume '%s:%s' on '%s:%s'", mount_obj.server_system, mount_obj.volname, mount_obj.client_system, mount_obj.mountpoint) ret = mount_obj.mount() if not ret: g.log.error("Failed to mount volume '%s:%s' on '%s:%s'", mount_obj.server_system, mount_obj.volname, mount_obj.client_system, mount_obj.mountpoint) return False else: g.log.info("Successful in mounting volume '%s:%s' on " "'%s:%s'", mount_obj.server_system, mount_obj.volname, mount_obj.client_system, mount_obj.mountpoint) g.log.info("Successful in mounting all mount objs for the volume %s", cls.volname) # Get mounts info g.log.info("Get mounts Info:") log_mounts_info(mounts) return True @classmethod def setup_volume_and_mount_volume(cls, mounts, volume_create_force=False): """Setup the volume and mount the volume Args: mounts(list): List of mount_objs volume_create_force(bool): True if create_volume should be executed with 'force' option. Returns (bool): True if setting up volume and mounting the volume for a mount obj is successful. False otherwise """ # Setup Volume _rc = cls.setup_volume(volume_create_force) if not _rc: return _rc # Mount Volume _rc = cls.mount_volume(mounts) if not _rc: return _rc return True @classmethod def unmount_volume(cls, mounts): """Unmount all mounts for the volume Args: mounts(list): List of mount_objs Returns (bool): True if unmounting the volume for a mount obj is successful. False otherwise """ # Unmount volume g.log.info("Starting to UnMount Volume %s", cls.volname) for mount_obj in mounts: g.log.info("UnMounting volume '%s:%s' on '%s:%s'", mount_obj.server_system, mount_obj.volname, mount_obj.client_system, mount_obj.mountpoint) ret = mount_obj.unmount() if not ret: g.log.error("Failed to unmount volume '%s:%s' on '%s:%s'", mount_obj.server_system, mount_obj.volname, mount_obj.client_system, mount_obj.mountpoint) # Get mounts info g.log.info("Get mounts Info:") log_mounts_info(cls.mounts) return False else: g.log.info("Successful in unmounting volume '%s:%s' on " "'%s:%s'", mount_obj.server_system, mount_obj.volname, mount_obj.client_system, mount_obj.mountpoint) g.log.info("Successful in unmounting all mount objs for the volume %s", cls.volname) # Get mounts info g.log.info("Get mounts Info:") log_mounts_info(mounts) return True @classmethod def cleanup_volume(cls): """Cleanup the volume Returns (bool): True if cleanup volume is successful. False otherwise. """ g.log.info("Cleanup Volume %s", cls.volname) ret = cleanup_volume(mnode=cls.mnode, volname=cls.volname) if not ret: g.log.error("cleanup of volume %s failed", cls.volname) else: g.log.info("Successfully cleaned-up volume %s", cls.volname) # Log Volume Info and Status g.log.info("Log Volume %s Info and Status", cls.volname) log_volume_info_and_status(cls.mnode, cls.volname) return ret @classmethod def unmount_volume_and_cleanup_volume(cls, mounts): """Unmount the volume and cleanup volume Args: mounts(list): List of mount_objs Returns (bool): True if unmounting the volume for the mounts and cleaning up volume is successful. False otherwise """ # UnMount Volume _rc = cls.unmount_volume(mounts) if not _rc: return _rc # Setup Volume _rc = cls.cleanup_volume() if not _rc: return _rc return True @classmethod def setUpClass(cls): """Initialize all the variables necessary for testing Gluster """ # Get all servers cls.all_servers = None if 'servers' in g.config and g.config['servers']: cls.all_servers = g.config['servers'] cls.servers = cls.all_servers else: raise ConfigError("'servers' not defined in the global config") # Get all clients cls.all_clients = None if 'clients' in g.config and g.config['clients']: cls.all_clients = g.config['clients'] cls.clients = cls.all_clients else: raise ConfigError("'clients' not defined in the global config") # Get all servers info cls.all_servers_info = None if 'servers_info' in g.config and g.config['servers_info']: cls.all_servers_info = g.config['servers_info'] else: raise ConfigError("'servers_info' not defined in the global " "config") # All clients_info cls.all_clients_info = None if 'clients_info' in g.config and g.config['clients_info']: cls.all_clients_info = g.config['clients_info'] else: raise ConfigError("'clients_info' not defined in the global " "config") # Set mnode : Node on which gluster commands are executed cls.mnode = cls.all_servers[0] # Server IP's cls.servers_ips = [] cls.servers_ips = cls.get_ip_from_hostname(cls.servers) # SMB Cluster info try: cls.smb_users_info = ( g.config['gluster']['cluster_config']['smb']['users_info']) except KeyError: cls.smb_users_info = {} cls.smb_users_info['root'] = {} cls.smb_users_info['root']['password'] = 'foobar' cls.smb_users_info['root']['acl'] = 'rwx' # NFS-Ganesha Cluster info try: cls.enable_nfs_ganesha = bool(g.config['gluster']['cluster_config'] ['nfs_ganesha']['enable']) cls.num_of_nfs_ganesha_nodes = (g.config['gluster'] ['cluster_config']['nfs_ganesha'] ['num_of_nfs_ganesha_nodes']) cls.vips = (g.config['gluster']['cluster_config']['nfs_ganesha'] ['vips']) except KeyError: cls.enable_nfs_ganesha = False cls.num_of_nfs_ganesha_nodes = None cls.vips = [] # Defining default volume_types configuration. cls.default_volume_type_config = { 'replicated': { 'type': 'replicated', 'replica_count': 3, 'transport': 'tcp' }, 'dispersed': { 'type': 'dispersed', 'disperse_count': 6, 'redundancy_count': 2, 'transport': 'tcp' }, 'distributed': { 'type': 'distributed', 'dist_count': 4, 'transport': 'tcp' }, 'distributed-replicated': { 'type': 'distributed-replicated', 'dist_count': 2, 'replica_count': 3, 'transport': 'tcp' }, 'distributed-dispersed': { 'type': 'distributed-dispersed', 'dist_count': 2, 'disperse_count': 6, 'redundancy_count': 2, 'transport': 'tcp' } } # Check if default volume_type configuration is provided in # config yml if (g.config.get('gluster') and g.config['gluster'].get('volume_types')): default_volume_type_from_config = ( g.config['gluster']['volume_types']) for volume_type in default_volume_type_from_config.keys(): if default_volume_type_from_config[volume_type]: if volume_type in cls.default_volume_type_config: cls.default_volume_type_config[volume_type] = ( default_volume_type_from_config[volume_type]) # Create Volume with force option cls.volume_create_force = False if (g.config.get('gluster') and g.config['gluster'].get('volume_create_force')): cls.volume_create_force = ( g.config['gluster']['volume_create_force']) # Default volume options which is applicable for all the volumes cls.volume_options = {} if (g.config.get('gluster') and g.config['gluster'].get('volume_options')): cls.volume_options = g.config['gluster']['volume_options'] # If the volume is exported as SMB Share, then set the following # volume options on the share. cls.smb_share_options = {} if (g.config.get('gluster') and g.config['gluster'].get('smb_share_options')): cls.smb_share_options = ( g.config['gluster']['smb_share_options']) # If the volume is exported as NFS-Ganesha export, # then set the following volume options on the export. cls.nfs_ganesha_export_options = {} if (g.config.get('gluster') and g.config['gluster'].get('nfs_ganesha_export_options')): cls.nfs_ganesha_export_options = ( g.config['gluster']['nfs_ganesha_export_options']) # Get the volume configuration. cls.volume = {} if cls.volume_type: found_volume = False if 'gluster' in g.config: if 'volumes' in g.config['gluster']: for volume in g.config['gluster']['volumes']: if volume['voltype']['type'] == cls.volume_type: cls.volume = copy.deepcopy(volume) found_volume = True break if found_volume: if 'name' not in cls.volume: cls.volume['name'] = 'testvol_%s' % cls.volume_type if 'servers' not in cls.volume: cls.volume['servers'] = cls.all_servers if not found_volume: try: if g.config['gluster']['volume_types'][cls.volume_type]: cls.volume['voltype'] = (g.config['gluster'] ['volume_types'] [cls.volume_type]) except KeyError: try: cls.volume['voltype'] = (cls.default_volume_type_config [cls.volume_type]) except KeyError: raise ConfigError("Unable to get configs of volume " "type: %s", cls.volume_type) cls.volume['name'] = 'testvol_%s' % cls.volume_type cls.volume['servers'] = cls.all_servers # Set volume options if 'options' not in cls.volume: cls.volume['options'] = cls.volume_options # Define Volume Useful Variables. cls.volname = cls.volume['name'] cls.voltype = cls.volume['voltype']['type'] cls.servers = cls.volume['servers'] cls.mnode = cls.servers[0] cls.vol_options = cls.volume['options'] # Get the mount configuration. cls.mounts = [] if cls.mount_type: cls.mounts_dict_list = [] found_mount = False if 'gluster' in g.config: if 'mounts' in g.config['gluster']: for mount in g.config['gluster']['mounts']: if mount['protocol'] == cls.mount_type: temp_mount = {} temp_mount['protocol'] = cls.mount_type if 'volname' in mount and mount['volname']: if mount['volname'] == cls.volname: temp_mount = copy.deepcopy(mount) else: continue else: temp_mount['volname'] = cls.volname if ('server' not in mount or (not mount['server'])): temp_mount['server'] = cls.mnode else: temp_mount['server'] = mount['server'] if ('mountpoint' not in mount or (not mount['mountpoint'])): temp_mount['mountpoint'] = (os.path.join( "/mnt", '_'.join([cls.volname, cls.mount_type]))) else: temp_mount['mountpoint'] = mount['mountpoint'] if ('client' not in mount or (not mount['client'])): temp_mount['client'] = ( cls.all_clients_info[ random.choice( cls.all_clients_info.keys())] ) else: temp_mount['client'] = mount['client'] if 'options' in mount and mount['options']: temp_mount['options'] = mount['options'] else: temp_mount['options'] = '' cls.mounts_dict_list.append(temp_mount) found_mount = True if not found_mount: for client in cls.all_clients_info.keys(): mount = { 'protocol': cls.mount_type, 'server': cls.mnode, 'volname': cls.volname, 'client': cls.all_clients_info[client], 'mountpoint': (os.path.join( "/mnt", '_'.join([cls.volname, cls.mount_type]))), 'options': '' } cls.mounts_dict_list.append(mount) if cls.mount_type == 'cifs' or cls.mount_type == 'smb': for mount in cls.mounts_dict_list: if 'smbuser' not in mount: mount['smbuser'] = random.choice( cls.smb_users_info.keys()) mount['smbpasswd'] = ( cls.smb_users_info[mount['smbuser']]['password']) cls.mounts = create_mount_objs(cls.mounts_dict_list) # Defining clients from mounts. cls.clients = [] for mount in cls.mounts_dict_list: cls.clients.append(mount['client']['host']) cls.clients = list(set(cls.clients)) # Gluster Logs info cls.server_gluster_logs_dirs = ["/var/log/glusterfs", "/var/log/samba"] cls.server_gluster_logs_files = ["/var/log/ganesha.log", "/var/log/ganesha-gfapi.log"] if ('gluster' in g.config and 'server_gluster_logs_info' in g.config['gluster']): server_gluster_logs_info = ( g.config['gluster']['server_gluster_logs_info']) if ('dirs' in server_gluster_logs_info and server_gluster_logs_info['dirs']): cls.server_gluster_logs_dirs = ( server_gluster_logs_info['dirs']) if ('files' in server_gluster_logs_info and server_gluster_logs_info['files']): cls.server_gluster_logs_files = ( server_gluster_logs_info['files']) cls.client_gluster_logs_dirs = ["/var/log/glusterfs"] cls.client_gluster_logs_files = [] if ('gluster' in g.config and 'client_gluster_logs_info' in g.config['gluster']): client_gluster_logs_info = ( g.config['gluster']['client_gluster_logs_info']) if ('dirs' in client_gluster_logs_info and client_gluster_logs_info['dirs']): cls.client_gluster_logs_dirs = ( client_gluster_logs_info['dirs']) if ('files' in client_gluster_logs_info and client_gluster_logs_info['files']): cls.client_gluster_logs_files = ( client_gluster_logs_info['files']) # Have a unique string to recognize the test run for logging in # gluster logs if 'glustotest_run_id' not in g.config: g.config['glustotest_run_id'] = ( datetime.datetime.now().strftime('%H_%M_%d_%m_%Y')) cls.glustotest_run_id = g.config['glustotest_run_id'] msg = "Setupclass: %s : %s" % (cls.__name__, cls.glustotest_run_id) g.log.info(msg) cls.inject_msg_in_gluster_logs(msg) # Log the baseclass variables for debugging purposes g.log.debug("GlusterBaseClass Variables:\n %s", cls.__dict__) def setUp(self): msg = "Starting Test : %s : %s" % (self.id(), self.glustotest_run_id) g.log.info(msg) self.inject_msg_in_gluster_logs(msg) def tearDown(self): msg = "Ending Test: %s : %s" % (self.id(), self.glustotest_run_id) g.log.info(msg) self.inject_msg_in_gluster_logs(msg) @classmethod def tearDownClass(cls): msg = "Teardownclass: %s : %s" % (cls.__name__, cls.glustotest_run_id) g.log.info(msg) cls.inject_msg_in_gluster_logs(msg)