summaryrefslogtreecommitdiffstats
path: root/tests
diff options
context:
space:
mode:
authorShwetha Panduranga <spandura@redhat.com>2016-05-12 19:31:41 +0530
committerM S Vishwanath Bhat <vbhat@redhat.com>2016-05-19 05:59:24 -0700
commit8816d97faca31026c01c8c60241ebddcb855880e (patch)
tree71d61da48613299773863b7019bebca32e9b0348 /tests
parent5bd9d42c997695d0c673ae50c1dbb9f4ab23695d (diff)
Adding proper doc_strings to the functions and
modifying the existing functions in peer_ops Change-Id: I819293c3191040232b6c9d30877f11b2857f3931 BUG: 1335494 Signed-off-by: Shwetha Panduranga <spandura@redhat.com> Reviewed-on: http://review.gluster.org/14308 Smoke: Gluster Build System <jenkins@build.gluster.com> Reviewed-by: M S Vishwanath Bhat <vbhat@redhat.com> Tested-by: M S Vishwanath Bhat <vbhat@redhat.com> NetBSD-regression: NetBSD Build System <jenkins@build.gluster.org> CentOS-regression: Gluster Build System <jenkins@build.gluster.com>
Diffstat (limited to 'tests')
-rw-r--r--tests/distaf/distaf_libs/distaflibs-gluster/distaflibs/gluster/peer_ops.py517
1 files changed, 381 insertions, 136 deletions
diff --git a/tests/distaf/distaf_libs/distaflibs-gluster/distaflibs/gluster/peer_ops.py b/tests/distaf/distaf_libs/distaflibs-gluster/distaflibs/gluster/peer_ops.py
index f2c0d0c0009..18938f41224 100644
--- a/tests/distaf/distaf_libs/distaflibs-gluster/distaflibs/gluster/peer_ops.py
+++ b/tests/distaf/distaf_libs/distaflibs-gluster/distaflibs/gluster/peer_ops.py
@@ -21,184 +21,429 @@
"""
+from distaf.util import tc
import re
import time
-from distaf.util import tc
+import socket
+try:
+ import xml.etree.cElementTree as etree
+except ImportError:
+ import xml.etree.ElementTree as etree
+
+
+def peer_probe(server, mnode=None):
+ """Probe the specified server.
+
+ Args:
+ server (str): Server to be peer probed.
+
+ Kwargs:
+ mnode (str): Node on which command has to be executed.
+ If None, defaults to nodes[0].
+ Returns:
+ tuple: Tuple containing three elements (ret, out, err).
+ The first element 'ret' is of type 'int' and is the return value
+ of command execution.
-def peer_probe(pnode='', servers='', timeout=10):
+ The second element 'out' is of type 'str' and is the stdout value
+ of the command execution.
+
+ The third element 'err' is of type 'str' and is the stderr value
+ of the command execution.
"""
- Does peer probe and validates the same
- Returns True on success and False on failure
- Note: Input for parameter 'servers' should be in list format
+ if mnode is None:
+ mnode = tc.nodes[0]
+
+ cmd = "gluster peer probe %s" % server
+ return tc.run(mnode, cmd)
+
+def peer_detach(server, force=False, mnode=None):
+ """Detach the specified server.
+
+ Args:
+ server (str): Server to be peer detached.
+
+ Kwargs:
+ force (bool): option to detach peer.
+ Defaults to False.
+ mnode (str): Node on which command has to be executed.
+ If None, defaults to nodes[0].
+
+ Returns:
+ tuple: Tuple containing three elements (ret, out, err).
+ The first element 'ret' is of type 'int' and is the return value
+ of command execution.
+
+ The second element 'out' is of type 'str' and is the stdout value
+ of the command execution.
+
+ The third element 'err' is of type 'str' and is the stderr value
+ of the command execution.
"""
- if pnode == '':
- pnode = tc.nodes[0]
- if servers == '':
- servers = tc.nodes[1:]
+ if mnode is None:
+ mnode = tc.nodes[0]
+
+ if force:
+ cmd = "gluster peer detach %s force" % server
+ else:
+ cmd = "gluster peer detach %s" % server
+ return tc.run(mnode, cmd)
+
- nodes_pool_list = nodes_from_pool_list(pnode)
- if not nodes_pool_list:
+def peer_status(mnode=None):
+ """Runs 'gluster peer status' on specified node.
+
+ Kwargs:
+ mnode (str): Node on which command has to be executed.
+ If None, defaults to nodes[0].
+
+ Returns:
+ tuple: Tuple containing three elements (ret, out, err).
+ The first element 'ret' is of type 'int' and is the return value
+ of command execution.
+
+ The second element 'out' is of type 'str' and is the stdout value
+ of the command execution.
+
+ The third element 'err' is of type 'str' and is the stderr value
+ of the command execution.
+ """
+ tc.logger.info("Inside peer status")
+ if mnode is None:
+ mnode = tc.nodes[0]
+
+ cmd = "gluster peer status"
+ return tc.run(mnode, cmd)
+
+
+def pool_list(mnode=None):
+ """Runs 'gluster pool list' command on the specified node.
+
+ Kwargs:
+ mnode (str): Node on which command has to be executed.
+ If None, defaults to nodes[0].
+
+ Returns:
+ tuple: Tuple containing three elements (ret, out, err).
+ The first element 'ret' is of type 'int' and is the return value
+ of command execution.
+
+ The second element 'out' is of type 'str' and is the stdout value
+ of the command execution.
+
+ The third element 'err' is of type 'str' and is the stderr value
+ of the command execution.
+ """
+ if mnode is None:
+ mnode = tc.nodes[0]
+
+ cmd = "gluster pool list"
+ return tc.run(mnode, cmd)
+
+
+def peer_probe_servers(servers=None, validate=True, time_delay=10, mnode=None):
+ """Probe specified servers and validate whether probed servers
+ are in cluster and connected state if validate is set to True.
+
+ Kwargs:
+ servers (list): List of servers to be peer probed.
+ If None, defaults to nodes.
+ validate (bool): True to validate if probed peer is in cluster and
+ connected state. False otherwise. Defaults to True.
+ time_delay (int): time delay before validating peer status.
+ Defaults to 10 seconds.
+ mnode (str): Node on which command has to be executed.
+ If None, defaults to nodes[0].
+
+ Returns:
+ bool: True on success and False on failure.
+ """
+ if servers is None:
+ servers = tc.nodes[:]
+ if not isinstance(servers, list):
+ servers = [servers]
+ if mnode is None:
+ mnode = tc.nodes[0]
+ if mnode in servers:
+ servers.remove(mnode)
+
+ # Get list of nodes from 'gluster pool list'
+ nodes_in_pool_list = nodes_from_pool_list(mnode)
+ if nodes_in_pool_list is None:
+ tc.logger.error("Unable to get nodes from gluster pool list. "
+ "Failing peer probe.")
return False
+
for server in servers:
- if server not in nodes_pool_list:
- ret = tc.run(pnode, "gluster peer probe %s" % server)
- if ret[0] != 0 or \
- re.search(r'^peer\sprobe\:\ssuccess(.*)', ret[1]) is None:
- tc.logger.error("Failed to do peer probe for node %s" % server)
+ if server not in nodes_in_pool_list:
+ ret, out, _ = peer_probe(server, mnode)
+ if (ret != 0 or
+ re.search(r'^peer\sprobe\:\ssuccess(.*)', out) is None):
+ tc.logger.error("Failed to peer probe the node '%s'.", server)
return False
- time.sleep(timeout)
+ else:
+ tc.logger.info("Successfully peer probed the node '%s'.",
+ server)
- #Validating whether peer probe is successful
- if not validate_peer_status(pnode, servers):
- tc.logger.error("peer probe validation failed")
- return False
+ # Validating whether peer is in connected state after peer probe
+ if validate:
+ time.sleep(time_delay)
+ if not is_peer_connected(servers, mnode):
+ tc.logger.error("Validation after peer probe failed.")
+ return False
+ else:
+ tc.logger.info("Validation after peer probe is successful.")
return True
+def peer_detach_servers(servers=None, force=False, validate=True,
+ time_delay=10, mnode=None):
+ """Detach peers and validate status of peer if validate is set to True.
-def peer_detach(pnode='', servers='', force=False, timeout=10):
- """
- Does peer detach and validates the same
- Returns True on success and False on failure
- Note: Input for parameter 'servers' should be in list format
+ Kwargs:
+ servers (list): List of servers to be peer detached.
+ If None, defaults to nodes.
+ force (bool): option to detach peer.
+ Defaults to False.
+ validate (bool): True if status of the peer needs to be validated,
+ False otherwise. Defaults to True.
+ time_delay (int): time delay before executing validating peer.
+ status. Defaults to 10 seconds.
+ mnode (str): Node on which command has to be executed.
+ If None, defaults to nodes[0].
+
+ Returns:
+ bool: True on success and False on failure.
"""
- if pnode == '':
- pnode = tc.nodes[0]
- if servers == '':
- servers = tc.nodes[1:]
+ if servers is None:
+ servers = tc.nodes[:]
+ if not isinstance(servers, list):
+ servers = [servers]
+ if mnode is None:
+ mnode = tc.nodes[0]
+ if mnode in servers:
+ servers.remove(mnode)
for server in servers:
- if force:
- cmd = "gluster peer detach %s force" % server
- else:
- cmd = "gluster peer detach %s" % server
- ret = tc.run(pnode, cmd)
- if ret[0] != 0 or re.search(r'^peer\sdetach\:\ssuccess(.*)', ret[1]) \
- is None:
- tc.logger.error("Failed to do peer detach for node %s" % server)
+ ret, out, _ = peer_detach(server, force, mnode)
+ if (ret != 0 or
+ re.search(r'^peer\sdetach\:\ssuccess(.*)', out) is None):
+ tc.logger.error("Failed to peer detach the node '%s'.", server)
return False
- time.sleep(timeout)
- #Validating whether peer detach is successful
- if validate_peer_status(pnode, servers):
- tc.logger.error("peer detach validatiom failed")
- return False
+ # Validating whether peer detach is successful
+ if validate:
+ time.sleep(time_delay)
+ if is_peer_connected(servers, mnode):
+ tc.logger.error("Validation after peer detach failed.")
+ return False
+ else:
+ tc.logger.info("Validation after peer detach is successful")
return True
+def nodes_from_pool_list(mnode=None):
+ """Return list of nodes from the 'gluster pool list'.
-def peer_status(pnode=''):
- """
- Does peer status on the given node
- Returns: On success, peer status information in list of dicts
- If there is only one peer, then an empty list is sent
- On failure, None
+ Kwargs:
+ mnode (str): Node on which command has to be executed.
+ If None, defaults to nodes[0].
+
+ Returns:
+ NoneType: None if command execution fails.
+ list: List of nodes in pool on Success, Empty list on failure.
"""
- if pnode == '':
- pnode = tc.nodes[0]
- ret = tc.run(pnode, "gluster peer status")
- if ret[0] != 0:
- tc.logger.error("Failed to execute peer status command in node %s" \
- % pnode)
+ if mnode is None:
+ mnode = tc.nodes[0]
+
+ pool_list_data = get_pool_list(mnode)
+ if pool_list_data is None:
+ tc.logger.error("Unable to get Nodes from the pool list command.")
return None
- status_list = ret[1].strip().split('\n')
- if "Number of Peers: 0" == status_list[0] and len(status_list) == 1:
- tc.logger.debug("Only one server in the cluster")
- return status_list[1:]
+ nodes = []
+ for item in pool_list_data:
+ nodes.append(item['hostname'])
+ return nodes
- status_list = status_list[1:]
- peer_list = []
- for status in status_list:
- stat = [stat for stat in status.split('\n') if stat != '']
- temp_dict = {}
- for element in stat:
- elmt = element.split(':')
- temp_dict[elmt[0].strip()] = elmt[1].strip()
- peer_list.append(temp_dict)
- return peer_list
+def get_peer_status(mnode=None):
+ """Parse the output of command 'gluster peer status'.
-def validate_peer_status(pnode='', servers=''):
- """
- Checks whether peer probe succeeds using peer status command
- Returns True on success and False on failure
- Note: Input for parameter 'servers' should be in list format
+ Kwargs:
+ mnode (str): Node on which command has to be executed.
+ if None, defaults to nodes[0].
+
+ Returns:
+ NoneType: None if command execution fails or parse errors.
+ list: list of dicts on success.
+
+ Examples:
+ >>> get_peer_status(mnode = 'abc.lab.eng.xyz.com')
+ [{'uuid': '77dc299a-32f7-43d8-9977-7345a344c398',
+ 'hostname': 'ijk.lab.eng.xyz.com',
+ 'state': '3',
+ 'hostnames' : ['ijk.lab.eng.xyz.com'],
+ 'connected': '1',
+ 'stateStr': 'Peer in Cluster'},
+
+ {'uuid': 'b15b8337-9f8e-4ec3-8bdb-200d6a67ae12',
+ 'hostname': 'def.lab.eng.xyz.com',
+ 'state': '3',
+ 'hostnames': ['def.lab.eng.xyz.com'],
+ 'connected': '1',
+ 'stateStr': 'Peer in Cluster'}
+ ]
"""
- if pnode == '':
- pnode = tc.nodes[0]
- if servers == '':
- servers = tc.nodes[1:]
-
- check_flag = False
- status = peer_status(pnode)
- if status == []:
- return True
- if status == None:
- return False
+ if mnode is None:
+ mnode = tc.nodes[0]
- for stat in status:
- if stat['Hostname'] in servers:
- if re.match(r'([0-9a-f]{8})(?:-[0-9a-f]{4}){3}-[0-9a-f]{12}', \
- stat['Uuid'], re.I) is None or \
- stat['State'] != "Peer in Cluster (Connected)":
- tc.logger.error("peer probe unsuccessful for node %s" % \
- stat['Hostname'])
- check_flag = True
-
- if check_flag or not len(set(servers).intersection([stat_key['Hostname'] \
- for stat_key in status])):
- return False
+ ret, out, _ = tc.run(mnode, "gluster peer status --xml", verbose=False)
+ if ret != 0:
+ tc.logger.error("Failed to execute peer status command on node '%s'. "
+ "Hence failed to parse the peer status.", mnode)
+ return None
- return True
+ try:
+ root = etree.XML(out)
+ except etree.ParseError:
+ tc.logger.error("Failed to parse the gluster peer status xml output.")
+ return None
+ peer_status_list = []
+ for peer in root.findall("peerStatus/peer"):
+ peer_dict = {}
+ for element in peer.getchildren():
+ if element.tag == "hostnames":
+ hostnames_list = []
+ for hostname in element.getchildren():
+ hostnames_list.append(hostname.text)
+ element.text = hostnames_list
+ peer_dict[element.tag] = element.text
+ peer_status_list.append(peer_dict)
+ return peer_status_list
-def pool_list(pnode=''):
- """
- Does pool list on the given node
- Returns: On success, pool list information in list of dictionary format
- On Failure, None
+
+def get_pool_list(mnode=None):
+ """Parse the output of 'gluster pool list' command.
+
+ Kwargs:
+ mnode (str): Node on which command has to be executed.
+ If None, defaults to nodes[0].
+
+ Returns:
+ NoneType: None if command execution fails, parse errors.
+ list: list of dicts on success.
+
+ Examples:
+ >>> get_pool_list(mnode = 'abc.lab.eng.xyz.com')
+ [{'uuid': 'a2b88b10-eba2-4f97-add2-8dc37df08b27',
+ 'hostname': 'abc.lab.eng.xyz.com',
+ 'state': '3',
+ 'connected': '1',
+ 'stateStr': 'Peer in Cluster'},
+
+ {'uuid': 'b15b8337-9f8e-4ec3-8bdb-200d6a67ae12',
+ 'hostname': 'def.lab.eng.xyz.com',
+ 'state': '3',
+ 'hostnames': ['def.lab.eng.xyz.com'],
+ 'connected': '1',
+ 'stateStr': 'Peer in Cluster'}
+ ]
"""
- if pnode == '':
- pnode = tc.nodes[0]
+ if mnode is None:
+ mnode = tc.nodes[0]
- ret = tc.run(pnode, "gluster pool list")
- if ret[0] != 0:
- tc.logger.error("Failed to execute pool list in node %s" % pnode)
+ ret, out, _ = tc.run(mnode, "gluster pool list --xml", verbose=False)
+ if ret != 0:
+ tc.logger.error("Failed to execute 'pool list' on node %s. "
+ "Hence failed to parse the pool list.", mnode)
return None
- pool_info = []
- for index, item in enumerate(ret[1].split('\n')[:-1]):
- match = re.search(r'(\S+)\s*\t*\s*(\S+)\s*\t*\s*(\S+)\s*', item)
- if match is not None:
- if index == 0:
- keys = match.groups()
- else:
- temp_dict = {}
- for num, element in enumerate(match.groups()):
- temp_dict[keys[num]] = element
- pool_info.append(temp_dict)
+ try:
+ root = etree.XML(out)
+ except etree.ParseError:
+ tc.logger.error("Failed to parse the gluster pool list xml output.")
+ return None
- return pool_info
+ pool_list_list = []
+ for peer in root.findall("peerStatus/peer"):
+ peer_dict = {}
+ for element in peer.getchildren():
+ if element.tag == "hostname" and element.text == 'localhost':
+ element.text = mnode
+ if element.tag == "hostnames":
+ hostnames_list = []
+ for hostname in element.getchildren():
+ hostnames_list.append(hostname.text)
+ element.text = hostnames_list
+ peer_dict[element.tag] = element.text
+ pool_list_list.append(peer_dict)
+ return pool_list_list
-def nodes_from_pool_list(pnode=''):
- """
- Runs the "gluster pool list" on the specified server
- and returns a list of servers from the pool list
- Returns False on failure and list of servers on success
+def is_peer_connected(servers=None, mnode=None):
+ """Checks whether specified peers are in cluster and 'Connected' state.
+
+ Kwargs:
+ servers (list): List of servers to be validated.
+ If None, defaults to nodes.
+ mnode (str): Node from which peer probe has to be executed.
+ If None, defaults to nodes[0].
+
+ Returns
+ bool : True on success (peer in cluster and connected), False on
+ failure.
"""
- if pnode == '':
- pnode = tc.nodes[0]
- servers = []
- nodes_from_pool = pool_list(pnode)
- if not nodes_from_pool:
+ if servers is None:
+ servers = tc.nodes[:]
+ if not isinstance(servers, list):
+ servers = [servers]
+ if mnode is None:
+ mnode = tc.nodes[0]
+ if mnode in servers:
+ servers.remove(mnode)
+
+ peer_connected = True
+ peer_status_list = get_peer_status(mnode)
+ if peer_status_list is None:
+ tc.logger.error("Failed to parse the peer status. Hence failed to "
+ "validate the peer connected state.")
return False
- for server in nodes_from_pool:
- if server.get('Hostname') == "localhost":
- servers.insert(0, pnode)
- else:
- servers.append(server.get("Hostname"))
- return servers
+ if peer_status_list == []:
+ tc.logger.error("No peers present in the pool. Servers are not yet "
+ "connected.")
+ return False
+
+ server_ips = []
+ for server in servers:
+ server_ips.append(socket.gethostbyname(server))
+
+ is_connected = True
+ for peer_stat in peer_status_list:
+ if socket.gethostbyname(peer_stat['hostname']) in servers:
+ if (re.match(r'([0-9a-f]{8})(?:-[0-9a-f]{4}){3}-[0-9a-f]{12}',
+ peer_stat['uuid'], re.I) is None):
+ tc.logger.error("Invalid UUID for the node '%s'",
+ peer_stat['hostname'])
+ is_connected = False
+ if (peer_stat['stateStr'] != "Peer in Cluster" or
+ peer_stat['connected'] != '1'):
+ tc.logger.error("Peer '%s' not in connected state",
+ peer_stat['Hostname'])
+ is_connected = False
+
+ if not is_connected:
+ return False
+
+ peer_nodes = [stat_key['hostname'] for stat_key in peer_status_list]
+ if not (set(servers).issubset(peer_nodes)):
+ tc.logger.error("Servers: '%s' not yet added to the pool.",
+ (list(set(servers).difference(peer_nodes))))
+ return False
+
+ tc.logger.info("Servers: '%s' are all 'Peer in Cluster' and 'Connected' "
+ "state.", servers)
+ return True