1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
|
import ddt
import pytest
from openshiftstoragelibs.baseclass import BaseClass
from openshiftstoragelibs import exceptions
from openshiftstoragelibs import heketi_ops
from openshiftstoragelibs import heketi_version
from openshiftstoragelibs import node_ops
from openshiftstoragelibs import openshift_ops
@ddt.ddt
class TestHeketiServerStateExamineGluster(BaseClass):
def setUp(self):
super(TestHeketiServerStateExamineGluster, self).setUp()
self.node = self.ocp_master_node[0]
version = heketi_version.get_heketi_version(self.heketi_client_node)
if version < '8.0.0-7':
self.skipTest("heketi-client package %s does not support server "
"state examine gluster" % version.v_str)
@pytest.mark.tier1
def test_volume_inconsistencies(self):
# Examine Gluster cluster and Heketi that there is no inconsistencies
out = heketi_ops.heketi_examine_gluster(
self.heketi_client_node, self.heketi_server_url)
if ("heketi volume list matches with volume list of all nodes"
not in out['report']):
self.skipTest(
"heketi and Gluster are inconsistent to each other")
# create volume
vol = heketi_ops.heketi_volume_create(
self.heketi_client_node, self.heketi_server_url, 1, json=True)
self.addCleanup(
heketi_ops.heketi_volume_delete, self.heketi_client_node,
self.heketi_server_url, vol['id'])
# delete volume from gluster cluster directly
openshift_ops.cmd_run_on_gluster_pod_or_node(
self.node,
"gluster vol stop %s force --mode=script" % vol['name'])
openshift_ops.cmd_run_on_gluster_pod_or_node(
self.node,
"gluster vol delete %s --mode=script" % vol['name'])
# verify that heketi is reporting inconsistencies
out = heketi_ops.heketi_examine_gluster(
self.heketi_client_node, self.heketi_server_url)
self.assertNotIn(
"heketi volume list matches with volume list of all nodes",
out['report'])
@pytest.mark.tier0
@ddt.data('', 'block')
def test_compare_real_vol_count_with_db_check_info(self, vol_type):
"""Validate file/block volumes using heketi db check."""
# Create File/Block volume
block_vol = getattr(heketi_ops, 'heketi_%svolume_create' % vol_type)(
self.heketi_client_node, self.heketi_server_url, 1, json=True)
self.addCleanup(
getattr(heketi_ops, 'heketi_%svolume_delete' % vol_type),
self.heketi_client_node, self.heketi_server_url, block_vol["id"])
# Check Heketi DB using Heketi CLI
db_result = heketi_ops.heketi_db_check(
self.heketi_client_node, self.heketi_server_url)
vol_count = db_result["%svolumes" % vol_type]["total"]
vol_list = getattr(heketi_ops, 'heketi_%svolume_list' % vol_type)(
self.heketi_client_node, self.heketi_server_url, json=True)
count = len(vol_list["%svolumes" % vol_type])
self.assertEqual(
count, vol_count,
"%svolume count doesn't match expected "
"result %s, actual result is %s" % (vol_type, count, vol_count))
@pytest.mark.tier0
@ddt.data('device_count', 'node_count', 'bricks_count')
def test_verify_db_check(self, count_type):
"""Validate the nodes, devices and bricks count in heketi db"""
# Get the total number of nodes, devices and bricks from db check
db_info = heketi_ops.heketi_db_check(
self.heketi_client_node, self.heketi_server_url)
db_devices_count = db_info["devices"]["total"]
db_nodes_count = db_info["nodes"]["total"]
db_bricks_count = db_info["bricks"]["total"]
# Get the total number of nodes, devices and bricks from topology info
topology_info = heketi_ops.heketi_topology_info(
self.heketi_client_node, self.heketi_server_url, json=True)
topology_devices_count, topology_nodes_count = 0, 0
topology_bricks_count = 0
for cluster in topology_info['clusters']:
topology_nodes_count += len(cluster['nodes'])
if count_type == 'bricks_count' or 'device_count':
for node in cluster['nodes']:
topology_devices_count += len(node['devices'])
if count_type == 'bricks_count':
for device in node['devices']:
topology_bricks_count += len(device['bricks'])
# Compare the device count
if count_type == 'device_count':
msg = ("Devices count in db check {} and in topology info {} is "
"not same".format(db_devices_count, topology_devices_count))
self.assertEqual(topology_devices_count, db_devices_count, msg)
# Compare the node count
elif count_type == 'node_count':
msg = (
"Nodes count in db check {} and nodes count in topology info "
"{} is not same".format(db_nodes_count, topology_nodes_count))
self.assertEqual(topology_nodes_count, db_nodes_count, msg)
# Compare the bricks count
elif count_type == 'bricks_count':
msg = ("Bricks count in db check {} and bricks count in topology "
"info {} is not same".format(
db_bricks_count, topology_bricks_count))
self.assertEqual(topology_bricks_count, db_bricks_count, msg)
@pytest.mark.tier1
@ddt.data('', 'block')
def test_compare_heketi_volumes(self, vol_type):
"""Validate file/block volume count using heketi gluster examine"""
# Create some file/block volumes
vol_size = 1
h_node, h_url = self.heketi_client_node, self.heketi_server_url
for i in range(5):
volume = eval(
"heketi_ops.heketi_{}volume_create".format(vol_type))(
h_node, h_url, vol_size, json=True)['id']
self.addCleanup(
eval("heketi_ops.heketi_{}volume_delete".format(vol_type)),
h_node, h_url, volume)
# Get the list of file/block volumes from heketi gluster examine
out = heketi_ops.heketi_examine_gluster(
self.heketi_client_node, self.heketi_server_url)
examine_volumes, clusters = [], out['heketidb']['clusterentries']
for cluster in clusters.values():
examine_volumes += cluster['Info']['{}volumes'.format(vol_type)]
# Get list of file/block volume from heketi blockvolume list
heketi_volumes = eval(
"heketi_ops.heketi_{}volume_list".format(vol_type))(
h_node, h_url, json=True)['{}volumes'.format(vol_type)]
# Compare file/block volume list
self.assertEqual(
heketi_volumes,
examine_volumes,
"Heketi {}volume list {} and list of blockvolumes in heketi "
"gluster examine {} are not same".format(
vol_type, heketi_volumes, examine_volumes))
@pytest.mark.tier2
def test_validate_report_after_node_poweroff(self):
"""Validate node report in heketi gluster examine after poweroff"""
# Skip test if not able to connect to Cloud Provider
try:
node_ops.find_vm_name_by_ip_or_hostname(self.node)
except (NotImplementedError, exceptions.ConfigError) as err:
self.skipTest(err)
# Power off one of the gluster node
g_node = list(self.gluster_servers_info.values())[0]['manage']
vm_name = node_ops.find_vm_name_by_ip_or_hostname(g_node)
self.power_off_gluster_node_vm(vm_name, g_node)
# Check the information of offline node in gluster examine output
msg = "could not fetch data from node {}".format(g_node)
examine_msg = heketi_ops.heketi_examine_gluster(
self.heketi_client_node, self.heketi_server_url)['report'][1]
self.assertEqual(
examine_msg, msg, "Failed to generate error report for node {} in"
" gluster examine output".format(g_node))
|