summaryrefslogtreecommitdiffstats
path: root/tests/functional
diff options
context:
space:
mode:
authorNigel Babu <nigelb@redhat.com>2018-07-27 13:02:05 +0530
committerNigel Babu <nigelb@redhat.com>2018-08-07 19:48:57 +0530
commit82c94a926c1f1961932798257d18a2fc306f9393 (patch)
treea5d5aa43613182aa4d545bc4c0d10c123cf16fa2 /tests/functional
parente0bb79503c37843d02082b93425230f2afbbbde2 (diff)
Fix spelling mistake across the codebase
Change-Id: I46fc2feffe6443af6913785d67bf310838532421
Diffstat (limited to 'tests/functional')
-rw-r--r--tests/functional/afr/heal/test_self_heal_daemon_process.py18
-rw-r--r--tests/functional/afr/test_client_side_quorum.py48
-rw-r--r--tests/functional/afr/test_conservative_merge_of_files_heal_command.py2
-rw-r--r--tests/functional/afr/test_gfid_assignment_on_dist_rep_vol.py2
-rw-r--r--tests/functional/afr/test_gfid_heal.py2
-rw-r--r--tests/functional/afr/test_gfid_split_brain_resolution.py4
-rw-r--r--tests/functional/afr/test_quota_limit_entry_heal.py2
-rw-r--r--tests/functional/afr/test_self_heal_when_dir_quota_exceeded.py4
-rw-r--r--tests/functional/afr/test_self_heal_with_quota_object_limit.py4
-rwxr-xr-xtests/functional/afr/test_write_io_mount_point_resumed_quorum_restored.py2
-rwxr-xr-xtests/functional/arbiter/brick_cases/test_brickcases.py2
-rw-r--r--tests/functional/arbiter/brick_cases/test_replica3_to_arbiter.py4
-rwxr-xr-xtests/functional/arbiter/test_mount_point_while_deleting_files.py2
-rw-r--r--tests/functional/authentication/test_fusereject.py2
-rw-r--r--tests/functional/authentication/test_vol_auth.py2
-rw-r--r--tests/functional/bvt/test_cvt.py2
-rw-r--r--tests/functional/dht/test_dht_hash_value.py4
-rw-r--r--tests/functional/dht/test_stop_glusterd_while_rebalance_in_progress.py4
-rw-r--r--tests/functional/disperse/test_ec_validate_volume_creation.py14
-rw-r--r--tests/functional/fuse_subdir/test_fusesubdir_with_quota_objects.py6
-rw-r--r--tests/functional/fuse_subdir/test_fusesubdir_with_removebrick.py2
-rw-r--r--tests/functional/fuse_subdir/test_leveldownsubdir_with_multiclient.py4
-rw-r--r--tests/functional/fuse_subdir/test_multisubdir_with_multiclient.py4
-rw-r--r--tests/functional/fuse_subdir/test_quota_limits_fusesubdir_.py4
-rw-r--r--tests/functional/glusterd/test_brick_port_after_stop_glusterd_modify_volume.py2
-rw-r--r--tests/functional/glusterd/test_brick_status_when_quorum_not_met.py2
-rw-r--r--tests/functional/glusterd/test_concurrent_set.py4
-rw-r--r--tests/functional/glusterd/test_create_vol_with_used_bricks.py2
-rw-r--r--tests/functional/glusterd/test_nfs_quorum.py2
-rw-r--r--tests/functional/glusterd/test_peer_detach.py2
-rw-r--r--tests/functional/glusterd/test_peer_probe_while_snapd_running.py2
-rw-r--r--tests/functional/glusterd/test_probe_glusterd.py2
-rw-r--r--tests/functional/glusterd/test_probe_hostname.py12
-rw-r--r--tests/functional/glusterd/test_quorum_remove_brick.py2
-rw-r--r--tests/functional/glusterd/test_rebalance_hang.py4
-rw-r--r--tests/functional/glusterd/test_rebalance_spurious.py4
-rw-r--r--tests/functional/glusterd/test_remove_brick_after_restart_glusterd.py4
-rw-r--r--tests/functional/glusterd/test_volume_get.py4
-rw-r--r--tests/functional/glusterd/test_volume_network_ping_timeout.py2
-rw-r--r--tests/functional/glusterd/test_volume_reset.py6
-rw-r--r--tests/functional/glusterd/test_volume_status.py8
-rw-r--r--tests/functional/glusterd/test_volume_status_fd.py2
-rw-r--r--tests/functional/nfs_ganesha/test_nfs_ganesha_sanity.py2
-rw-r--r--tests/functional/nfs_ganesha/test_nfs_ganesha_volume_exports.py4
-rw-r--r--tests/functional/quota/list_path_values.py2
-rw-r--r--tests/functional/quota/test_limit_usage_deep_dir.py2
-rw-r--r--tests/functional/quota/test_multi_value_limit.py6
-rw-r--r--tests/functional/quota/test_quota_add_brick.py2
-rwxr-xr-xtests/functional/quota/test_quota_limit_dir_breadth.py2
-rw-r--r--tests/functional/quota/test_quota_unique_soft_limit.py2
-rw-r--r--tests/functional/quota/test_quota_volume_subdir_limits.py2
-rw-r--r--tests/functional/snapshot/test_256_snapshots.py6
-rw-r--r--tests/functional/snapshot/test_clone_snap.py4
-rw-r--r--tests/functional/snapshot/test_mount_snap.py2
-rw-r--r--tests/functional/snapshot/test_snap_delete_existing_scheduler.py2
-rw-r--r--tests/functional/snapshot/test_snap_list_after_restart.py2
-rw-r--r--tests/functional/snapshot/test_snap_uss.py6
-rw-r--r--tests/functional/snapshot/test_snapshot_create.py4
-rw-r--r--tests/functional/snapshot/test_snapshot_restore.py4
-rw-r--r--tests/functional/snapshot/test_uss_brick_down.py4
-rw-r--r--tests/functional/snapshot/test_uss_snap_active_deactive.py2
-rw-r--r--tests/functional/snapshot/test_validate_snap_scheduler.py4
-rw-r--r--tests/functional/snapshot/test_validate_snapshot_max_limit.py10
-rw-r--r--tests/functional/snapshot/test_validate_snapshot_rebalance.py4
64 files changed, 144 insertions, 144 deletions
diff --git a/tests/functional/afr/heal/test_self_heal_daemon_process.py b/tests/functional/afr/heal/test_self_heal_daemon_process.py
index ed71e4f..15cd439 100644
--- a/tests/functional/afr/heal/test_self_heal_daemon_process.py
+++ b/tests/functional/afr/heal/test_self_heal_daemon_process.py
@@ -176,7 +176,7 @@ class SelfHealDaemonProcessTests(GlusterBaseClass):
self.all_servers_info)
self.assertTrue(ret, ("Failed to add bricks to "
"volume %s " % self.volname))
- g.log.info("Add brick successfull")
+ g.log.info("Add brick successful")
# Log Volume Info and Status after expanding the volume
g.log.info("Logging volume info and Status after expanding volume")
@@ -231,11 +231,11 @@ class SelfHealDaemonProcessTests(GlusterBaseClass):
ret, pids = get_self_heal_daemon_pid(nodes)
self.assertTrue(ret, ("Either No self heal daemon process found or "
"more than One self heal daemon process found"))
- g.log.info("Successfull in getting self-heal daemon process "
+ g.log.info("Successful in getting self-heal daemon process "
"on nodes %s", nodes)
glustershd_pids_after_expanding = pids
- g.log.info("Self Heal Daemon Process ID's afetr expanding "
+ g.log.info("Self Heal Daemon Process ID's after expanding "
"volume: %s", glustershd_pids_after_expanding)
self.assertNotEqual(glustershd_pids,
@@ -389,7 +389,7 @@ class SelfHealDaemonProcessTests(GlusterBaseClass):
ret, pids = get_self_heal_daemon_pid(nodes)
self.assertTrue(ret, ("Either No self heal daemon process found or "
"more than One self heal daemon process found"))
- g.log.info("Succesfull in getting self heal daemon pids")
+ g.log.info("Successful in getting self heal daemon pids")
glustershd_pids = pids
# get the bricks for the volume
@@ -461,7 +461,7 @@ class SelfHealDaemonProcessTests(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", bricks_to_bring_offline)
+ "for %s successfully", bricks_to_bring_offline)
# restart glusterd after brought down the brick
g.log.info("Restart glusterd on all servers %s", nodes)
@@ -922,7 +922,7 @@ class SelfHealDaemonProcessTestsWithHealing(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", bricks_to_bring_offline)
+ "for %s successfully", bricks_to_bring_offline)
# get the bricks which are running
g.log.info("getting the brick list which are online")
@@ -996,9 +996,9 @@ class SelfHealDaemonProcessTestsWithHealing(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", bricks_to_bring_offline)
+ "for %s successfully", bricks_to_bring_offline)
- # wait for 60 sec and brought up the brick agian
+ # wait for 60 sec and brought up the brick again
time.sleep(60)
g.log.info("Bring bricks: %s online", bricks_to_bring_offline)
ret = bring_bricks_online(self.mnode, self.volname,
@@ -1182,7 +1182,7 @@ class SelfHealDaemonProcessTestsWithMultipleVolumes(GlusterBaseClass):
self.all_servers_info)
self.assertTrue(ret, ("Failed to add bricks to "
"volume %s " % volume))
- g.log.info("Add brick successfull")
+ g.log.info("Add brick successful")
# Log Volume Info and Status after expanding the volume
g.log.info("Logging volume info and Status after "
diff --git a/tests/functional/afr/test_client_side_quorum.py b/tests/functional/afr/test_client_side_quorum.py
index 0432a13..1dcb96e 100644
--- a/tests/functional/afr/test_client_side_quorum.py
+++ b/tests/functional/afr/test_client_side_quorum.py
@@ -114,7 +114,7 @@ class ClientSideQuorumTests(GlusterBaseClass):
ret = set_volume_options(self.mnode, self.volname, options)
self.assertTrue(ret, ("Unable to set volume option %s for"
"volume %s" % (options, self.volname)))
- g.log.info("Sucessfully set %s for volume %s", options, self.volname)
+ g.log.info("Successfully set %s for volume %s", options, self.volname)
# write files on all mounts
g.log.info("Starting IO on all mounts...")
@@ -155,7 +155,7 @@ class ClientSideQuorumTests(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", bricks_to_bring_offline)
+ "for %s successfully", bricks_to_bring_offline)
# create 2 files named newfile0.txt and newfile1.txt
g.log.info("Start creating 2 files on all mounts...")
@@ -172,7 +172,7 @@ class ClientSideQuorumTests(GlusterBaseClass):
g.log.info("Validating whether IO failed with read-only filesystem")
ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
self.mounts)
- self.assertTrue(ret, ("Unexpected error and IO successfull"
+ self.assertTrue(ret, ("Unexpected error and IO successful"
" on read-only filesystem"))
g.log.info("EXPECTED: Read-only file system in IO while creating file")
@@ -190,7 +190,7 @@ class ClientSideQuorumTests(GlusterBaseClass):
g.log.info("Validating whether IO failed with read-only filesystem")
ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
self.mounts)
- self.assertTrue(ret, ("Unexpected error and IO successfull"
+ self.assertTrue(ret, ("Unexpected error and IO successful"
" on read-only filesystem"))
g.log.info("EXPECTED: Read-only file system in IO while"
" creating directory")
@@ -282,7 +282,7 @@ class ClientSideQuorumTests(GlusterBaseClass):
ret, _, err = g.run(mount_obj.client_system, cmd)
self.assertFalse(ret, ("Unexpected error and stat on file fails"
" on read-only filesystem"))
- g.log.info("stat on file is successfull on read-only filesystem")
+ g.log.info("stat on file is successful on read-only filesystem")
# stat on dir
g.log.info("stat on directory on all mounts")
@@ -292,7 +292,7 @@ class ClientSideQuorumTests(GlusterBaseClass):
ret, _, err = g.run(mount_obj.client_system, cmd)
self.assertFalse(ret, ("Unexpected error and stat on directory"
" fails on read-only filesystem"))
- g.log.info("stat on dir is successfull on read-only filesystem")
+ g.log.info("stat on dir is successful on read-only filesystem")
# ls on mount point
g.log.info("ls on mount point on all mounts")
@@ -302,7 +302,7 @@ class ClientSideQuorumTests(GlusterBaseClass):
ret, _, err = g.run(mount_obj.client_system, cmd)
self.assertFalse(ret, ("Unexpected error and listing file fails"
" on read-only filesystem"))
- g.log.info("listing files is successfull on read-only filesystem")
+ g.log.info("listing files is successful on read-only filesystem")
def test_client_side_quorum_with_fixed_validate_max_bricks(self):
"""
@@ -402,7 +402,7 @@ class ClientSideQuorumTests(GlusterBaseClass):
self.assertEqual(option_dict['cluster.quorum-count'], '(null)',
("Default value for %s is not null"
" for volume %s" % (option, self.volname)))
- g.log.info("Succesfull in getting %s for the volume %s",
+ g.log.info("Successful in getting %s for the volume %s",
option, self.volname)
# set cluster.quorum-type to fixed and cluster.quorum-count to 1
@@ -452,7 +452,7 @@ class ClientSideQuorumTests(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", bricks_to_bring_offline)
+ "for %s successfully", bricks_to_bring_offline)
# create files
g.log.info("Starting IO on all mounts...")
@@ -609,7 +609,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass):
ret = set_volume_options(self.mnode, self.volname, options)
self.assertTrue(ret, ("Unable to set volume option %s for"
"volume %s" % (options, self.volname)))
- g.log.info("Sucessfully set %s for volume %s", options, self.volname)
+ g.log.info("Successfully set %s for volume %s", options, self.volname)
# Start IO on mounts
g.log.info("Starting IO on all mounts...")
@@ -991,7 +991,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass):
g.log.info("Validating whether IO failed with Read Only File System")
ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
self.mounts)
- self.assertTrue(ret, ("Unexpected Error and IO successfull"
+ self.assertTrue(ret, ("Unexpected Error and IO successful"
" on Read-Only File System"))
g.log.info("EXPECTED Read-only file system in IO while creating file")
@@ -1062,7 +1062,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", subvolumes_second_brick_list)
+ "for %s successfully", subvolumes_second_brick_list)
# start I/0 ( write and read ) - read must pass, write will fail
g.log.info("Start creating files on all mounts...")
@@ -1079,7 +1079,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass):
g.log.info("Validating whether IO failed with Read Only File System")
ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
self.mounts)
- self.assertTrue(ret, ("Unexpected Error and IO successfull"
+ self.assertTrue(ret, ("Unexpected Error and IO successful"
" on Read-Only File System"))
g.log.info("EXPECTED Read-only file system in IO while creating file")
@@ -1203,7 +1203,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", subvolumes_first_brick_list)
+ "for %s successfully", subvolumes_first_brick_list)
# start I/0 ( write and read ) - read must pass, write will fail
g.log.info("Start creating files on all mounts...")
@@ -1220,7 +1220,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass):
g.log.info("Validating whether IO failed with Read Only File System")
ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
self.mounts)
- self.assertTrue(ret, ("Unexpected Error and IO successfull"
+ self.assertTrue(ret, ("Unexpected Error and IO successful"
" on Read-Only File System"))
g.log.info("EXPECTED Read-only file system in IO while creating file")
@@ -1508,7 +1508,7 @@ class ClientSideQuorumTestsMultipleVols(GlusterBaseClass):
ret = set_volume_options(self.mnode, vol_name, options)
self.assertTrue(ret, ("Unable to set volume option %s for "
"volume %s" % (options, vol_name)))
- g.log.info("Sucessfully set %s for volume %s", options, vol_name)
+ g.log.info("Successfully set %s for volume %s", options, vol_name)
# check is options are set correctly
volume_list = get_volume_list(self.mnode)
@@ -1612,7 +1612,7 @@ class ClientSideQuorumTestsMultipleVols(GlusterBaseClass):
g.log.info("Validating if IO failed with read-only filesystem")
ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
self.mounts)
- self.assertTrue(ret, ("Unexpected error and IO successfull"
+ self.assertTrue(ret, ("Unexpected error and IO successful"
" on read-only filesystem"))
g.log.info("EXPECTED: "
"Read-only file system in IO while creating file")
@@ -1803,7 +1803,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", brick_to_bring_offline1)
+ "for %s successfully", brick_to_bring_offline1)
offline_brick1_from_replicasets.append(brick_to_bring_offline1)
# start I/0 ( write and read ) - must succeed
@@ -1853,7 +1853,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", brick_to_bring_offline2)
+ "for %s successfully", brick_to_bring_offline2)
offline_brick2_from_replicasets.append(brick_to_bring_offline2)
# start I/0 ( write and read ) - must succeed
@@ -1957,7 +1957,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass):
g.log.info("Validating whether IO failed with Read Only File System")
ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
self.mounts)
- self.assertTrue(ret, ("Unexpected Error and IO successfull"
+ self.assertTrue(ret, ("Unexpected Error and IO successful"
" on Read-Only File System"))
g.log.info("EXPECTED Read-only file system in IO while creating file")
@@ -2120,7 +2120,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", bricks_to_bring_offline)
+ "for %s successfully", bricks_to_bring_offline)
# start I/0 ( write and read ) - read must pass, write will fail
g.log.info("Start creating files on all mounts...")
@@ -2137,7 +2137,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass):
g.log.info("Validating whether IO failed with Read-only file system")
ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
self.mounts)
- self.assertTrue(ret, ("Unexpected error and IO successfull"
+ self.assertTrue(ret, ("Unexpected error and IO successful"
" on Read-only file system"))
g.log.info("EXPECTED: Read-only file system in IO while creating file")
@@ -2181,7 +2181,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass):
g.log.info("Validating whether IO failed with Read-only file system")
ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
self.mounts)
- self.assertTrue(ret, ("Unexpected error and IO successfull"
+ self.assertTrue(ret, ("Unexpected error and IO successful"
" on Read-only file system"))
g.log.info("EXPECTED: Read-only file system in IO while creating file")
@@ -2225,7 +2225,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass):
g.log.info("Validating whether IO failed with Read-only file system")
ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs,
self.mounts)
- self.assertTrue(ret, ("Unexpected error and IO successfull"
+ self.assertTrue(ret, ("Unexpected error and IO successful"
" on Read-only file system"))
g.log.info("EXPECTED: Read-only file system in IO while creating file")
diff --git a/tests/functional/afr/test_conservative_merge_of_files_heal_command.py b/tests/functional/afr/test_conservative_merge_of_files_heal_command.py
index 3a6c620..06514b9 100644
--- a/tests/functional/afr/test_conservative_merge_of_files_heal_command.py
+++ b/tests/functional/afr/test_conservative_merge_of_files_heal_command.py
@@ -139,7 +139,7 @@ class VerifySelfHealTriggersHealCommand(GlusterBaseClass):
ret = set_volume_options(self.mnode, self.volname, options)
self.assertTrue(ret, ("Unable to set volume option %s for"
"volume %s" % (options, self.volname)))
- g.log.info("Sucessfully set %s for volume %s", options, self.volname)
+ g.log.info("Successfully set %s for volume %s", options, self.volname)
# Bring brick 0 offline
g.log.info('Bringing bricks %s offline', bricks_list[0])
diff --git a/tests/functional/afr/test_gfid_assignment_on_dist_rep_vol.py b/tests/functional/afr/test_gfid_assignment_on_dist_rep_vol.py
index e815fa0..0015f07 100644
--- a/tests/functional/afr/test_gfid_assignment_on_dist_rep_vol.py
+++ b/tests/functional/afr/test_gfid_assignment_on_dist_rep_vol.py
@@ -112,7 +112,7 @@ class AssignGfidsOnAllSubvols(GlusterBaseClass):
# Verify gfids are same on all the bricks
self.verify_gfid("dir1")
- # Creat a new directory on all the bricks directly
+ # Create a new directory on all the bricks directly
bricks_list = get_all_bricks(self.mnode, self.volname)
for brick in bricks_list:
brick_node, brick_path = brick.split(":")
diff --git a/tests/functional/afr/test_gfid_heal.py b/tests/functional/afr/test_gfid_heal.py
index 76d423e..589a420 100644
--- a/tests/functional/afr/test_gfid_heal.py
+++ b/tests/functional/afr/test_gfid_heal.py
@@ -129,7 +129,7 @@ class HealGfidTest(GlusterBaseClass):
- Create a 1x3 volume and fuse mount it.
- Create 1 directory with 1 file inside it directly on each brick.
- Access the directories from the mount.
- - Launch heals ans verify that the heals are over.
+ - Launch heals and verify that the heals are over.
- Verify that the files and directories have gfid assigned.
"""
# pylint: disable=too-many-statements
diff --git a/tests/functional/afr/test_gfid_split_brain_resolution.py b/tests/functional/afr/test_gfid_split_brain_resolution.py
index 0d6b0e2..a73ee40 100644
--- a/tests/functional/afr/test_gfid_split_brain_resolution.py
+++ b/tests/functional/afr/test_gfid_split_brain_resolution.py
@@ -104,7 +104,7 @@ class TestSelfHeal(GlusterBaseClass):
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
g.log.info("Brought down the brick process "
- "for %s succesfully", brick_list)
+ "for %s successfully", brick_list)
ret = are_bricks_offline(self.mnode, self.volname, brick_list)
self.assertTrue(ret, 'Bricks %s are not offline' % brick_list)
@@ -157,7 +157,7 @@ class TestSelfHeal(GlusterBaseClass):
ret = set_volume_options(self.mnode, self.volname, options)
self.assertTrue(ret, ("Unable to set volume option %s for "
"volume %s" % (options, self.volname)))
- g.log.info("Sucessfully set %s for volume %s", options, self.volname)
+ g.log.info("Successfully set %s for volume %s", options, self.volname)
# Create dir inside which I/O will be performed.
ret = mkdir(self.mounts[0].client_system, "%s/test_gfid_split_brain"
diff --git a/tests/functional/afr/test_quota_limit_entry_heal.py b/tests/functional/afr/test_quota_limit_entry_heal.py
index 56388c1..033d326 100644
--- a/tests/functional/afr/test_quota_limit_entry_heal.py
+++ b/tests/functional/afr/test_quota_limit_entry_heal.py
@@ -130,7 +130,7 @@ class QuotaEntrySelfHealTest(GlusterBaseClass):
ret = set_volume_options(self.mnode, self.volname, options)
self.assertTrue(ret, ("Unable to set volume option %s for "
"volume %s" % (options, self.volname)))
- g.log.info("Sucessfully set %s for volume %s", options, self.volname)
+ g.log.info("Successfully set %s for volume %s", options, self.volname)
# Create directory on mount
ret = mkdir(self.mounts[0].client_system, "%s/dir"
diff --git a/tests/functional/afr/test_self_heal_when_dir_quota_exceeded.py b/tests/functional/afr/test_self_heal_when_dir_quota_exceeded.py
index 92568ba..4648c0f 100644
--- a/tests/functional/afr/test_self_heal_when_dir_quota_exceeded.py
+++ b/tests/functional/afr/test_self_heal_when_dir_quota_exceeded.py
@@ -101,7 +101,7 @@ class HealFilesWhenDirQuotaExceeded(GlusterBaseClass):
path))
ret = g.run(mount_object.client_system, cmd)
self.assertTrue(ret, "Failed to create directory on mountpoint")
- g.log.info("Directory created succesfully on mountpoint")
+ g.log.info("Directory created successfully on mountpoint")
# Enable Quota
g.log.info("Enabling quota on the volume %s", self.volname)
@@ -142,7 +142,7 @@ class HealFilesWhenDirQuotaExceeded(GlusterBaseClass):
"count=20; done" % (mount_object.mountpoint, path))
ret, _, _ = g.run(mount_object.client_system, cmd)
self.assertEqual(ret, 0, ("Failed to create files on %s", path))
- g.log.info("Files created succesfully on mountpoint")
+ g.log.info("Files created successfully on mountpoint")
bricks_list = get_all_bricks(self.mnode, self.volname)
diff --git a/tests/functional/afr/test_self_heal_with_quota_object_limit.py b/tests/functional/afr/test_self_heal_with_quota_object_limit.py
index 6fe45f7..ff308c3 100644
--- a/tests/functional/afr/test_self_heal_with_quota_object_limit.py
+++ b/tests/functional/afr/test_self_heal_with_quota_object_limit.py
@@ -101,7 +101,7 @@ class HealFilesWhenQuotaObjectLimitExceeded(GlusterBaseClass):
path))
ret = g.run(mount_object.client_system, cmd)
self.assertTrue(ret, "Failed to create directory on mountpoint")
- g.log.info("Directory created succesfully on mountpoint")
+ g.log.info("Directory created successfully on mountpoint")
# Enable Quota
g.log.info("Enabling quota on the volume %s", self.volname)
@@ -141,7 +141,7 @@ class HealFilesWhenQuotaObjectLimitExceeded(GlusterBaseClass):
% (self.script_upload_path, mount_object.mountpoint, path))
ret, _, _ = g.run(mount_object.client_system, cmd)
self.assertEqual(ret, 0, ("Failed to create files on %s", path))
- g.log.info("Files created succesfully on mountpoint")
+ g.log.info("Files created successfully on mountpoint")
bricks_list = get_all_bricks(self.mnode, self.volname)
diff --git a/tests/functional/afr/test_write_io_mount_point_resumed_quorum_restored.py b/tests/functional/afr/test_write_io_mount_point_resumed_quorum_restored.py
index 96d5d78..1812593 100755
--- a/tests/functional/afr/test_write_io_mount_point_resumed_quorum_restored.py
+++ b/tests/functional/afr/test_write_io_mount_point_resumed_quorum_restored.py
@@ -142,7 +142,7 @@ class ClientSideQuorumRestored(GlusterBaseClass):
ret = set_volume_options(self.mnode, self.volname, options)
self.assertTrue(ret, ("Unable to set volume option %s for"
"volume %s" % (options, self.volname)))
- g.log.info("Sucessfully set %s for volume %s",
+ g.log.info("Successfully set %s for volume %s",
options, self.volname)
# Creating files on client side
diff --git a/tests/functional/arbiter/brick_cases/test_brickcases.py b/tests/functional/arbiter/brick_cases/test_brickcases.py
index 9cce9af..a5e33f0 100755
--- a/tests/functional/arbiter/brick_cases/test_brickcases.py
+++ b/tests/functional/arbiter/brick_cases/test_brickcases.py
@@ -177,7 +177,7 @@ class GlusterArbiterVolumeTypeChangeClass(GlusterBaseClass):
self.all_servers_info, replica_count=1,
arbiter_count=1)
self.assertTrue(ret, ("Failed to expand the volume %s", self.volname))
- g.log.info("Changing volume to arbiter volume is successfull %s",
+ g.log.info("Changing volume to arbiter volume is successful %s",
self.volname)
# Log Volume Info and Status after changing the volume type from
diff --git a/tests/functional/arbiter/brick_cases/test_replica3_to_arbiter.py b/tests/functional/arbiter/brick_cases/test_replica3_to_arbiter.py
index 4fe2a8b..8d4130d 100644
--- a/tests/functional/arbiter/brick_cases/test_replica3_to_arbiter.py
+++ b/tests/functional/arbiter/brick_cases/test_replica3_to_arbiter.py
@@ -124,14 +124,14 @@ class GlusterArbiterVolumeTypeClass(GlusterBaseClass):
self.volname))
g.log.info("Volume %s : All process are online", self.volname)
- # Adding the bricks to amke arbiter brick
+ # Adding the bricks to make arbiter brick
g.log.info("Adding bricks to convert to Arbiter Volume")
replica_arbiter = {'replica_count': 3, 'arbiter_count': 1}
ret = expand_volume(self.mnode, self.volname, self.servers[2:],
self.all_servers_info, add_to_hot_tier=False,
**replica_arbiter)
self.assertTrue(ret, ("Failed to expand the volume %s", self.volname))
- g.log.info("Changing volume to arbiter volume is successfull %s",
+ g.log.info("Changing volume to arbiter volume is successful %s",
self.volname)
# Log Volume Info and Status after expanding the volume
diff --git a/tests/functional/arbiter/test_mount_point_while_deleting_files.py b/tests/functional/arbiter/test_mount_point_while_deleting_files.py
index c8e4804..1bbdf27 100755
--- a/tests/functional/arbiter/test_mount_point_while_deleting_files.py
+++ b/tests/functional/arbiter/test_mount_point_while_deleting_files.py
@@ -229,7 +229,7 @@ class VolumeSetDataSelfHealTests(GlusterBaseClass):
ret = bring_bricks_offline(volname, bricks_to_bring_offline)
self.assertTrue(ret, ("Failed to bring down the bricks. Please "
"check the log file for more details."))
- g.log.info("Brought down the brick process for %s succesfully",
+ g.log.info("Brought down the brick process for %s successfully",
bricks_to_bring_offline)
# delete files on all mounts
diff --git a/tests/functional/authentication/test_fusereject.py b/tests/functional/authentication/test_fusereject.py
index 44410a5..5b403f4 100644
--- a/tests/functional/authentication/test_fusereject.py
+++ b/tests/functional/authentication/test_fusereject.py
@@ -129,7 +129,7 @@ class AuthRejectVol(GlusterBaseClass):
if (ret == 0) & (not out):
g.log.error("Mount executed successfully due to bug 1586036")
elif (ret == 1) & (not out):
- g.log.info("Expected:Mounting has failed sucessfully")
+ g.log.info("Expected:Mounting has failed successfully")
else:
raise ExecutionError("Unexpected Mounting of Volume %s successful"
% self.volname)
diff --git a/tests/functional/authentication/test_vol_auth.py b/tests/functional/authentication/test_vol_auth.py
index e8180d1..bb3da6f 100644
--- a/tests/functional/authentication/test_vol_auth.py
+++ b/tests/functional/authentication/test_vol_auth.py
@@ -118,7 +118,7 @@ class AuthRejectVol(GlusterBaseClass):
if (ret == 0) & (not out):
g.log.error("Mount executed successfully due to bug 1586036")
elif (ret == 1) & (not out):
- g.log.info("Expected:Mounting has failed sucessfully")
+ g.log.info("Expected:Mounting has failed successfully")
else:
raise ExecutionError("Unexpected Mounting of Volume %s"
"successful" % self.volname)
diff --git a/tests/functional/bvt/test_cvt.py b/tests/functional/bvt/test_cvt.py
index f1250b7..05578b4 100644
--- a/tests/functional/bvt/test_cvt.py
+++ b/tests/functional/bvt/test_cvt.py
@@ -663,7 +663,7 @@ class TestGlusterReplaceBrickSanity(GlusterBasicFeaturesSanityBaseClass):
# This test is disabled on nfs because of bug 1473668. A patch to apply the
-# workaround mentiond on the bug could not make this test green either.
+# workaround mentioned on the bug could not make this test green either.
@runs_on([['replicated', 'distributed-replicated', 'dispersed',
'distributed-dispersed'],
['glusterfs', 'cifs']])
diff --git a/tests/functional/dht/test_dht_hash_value.py b/tests/functional/dht/test_dht_hash_value.py
index fd83576..c5f08dc 100644
--- a/tests/functional/dht/test_dht_hash_value.py
+++ b/tests/functional/dht/test_dht_hash_value.py
@@ -187,12 +187,12 @@ class TestDHTHashValue(GlusterBaseClass):
mountpoint)
attributes = get_fattr_list(client_host, mountpoint)
self.assertFalse('trusted.gfid' in attributes,
- "Expected: Mount point should't display xattr:"
+ "Expected: Mount point shouldn't display xattr:"
"{xattr}. Actual: xattrs {xattr} is "
"presented on mount point".
format(xattr='trusted.gfid'))
self.assertFalse('trusted.glusterfs.dht' in attributes,
- "Expected: Mount point should't display xattr:"
+ "Expected: Mount point shouldn't display xattr:"
"{xattr}. Actual: xattrs {xattr} is "
"presented on mount point".
format(xattr='trusted.glusterfs.dht'))
diff --git a/tests/functional/dht/test_stop_glusterd_while_rebalance_in_progress.py b/tests/functional/dht/test_stop_glusterd_while_rebalance_in_progress.py
index bbe4eb5..c9b6010 100644
--- a/tests/functional/dht/test_stop_glusterd_while_rebalance_in_progress.py
+++ b/tests/functional/dht/test_stop_glusterd_while_rebalance_in_progress.py
@@ -134,7 +134,7 @@ class RebalanceValidation(GlusterBaseClass):
self.volname))
g.log.info("Volume %s: Started rebalance", self.volname)
- # Wait for atleast one file to be lookedup/scanned on the nodes
+ # Wait for at least one file to be lookedup/scanned on the nodes
status_info = get_rebalance_status(self.mnode, self.volname)
count = 0
while count < 100:
@@ -146,7 +146,7 @@ class RebalanceValidation(GlusterBaseClass):
lookups_start_count += 1
sleep(5)
if lookups_start_count == len(self.servers):
- g.log.info("Volume %s: Atleast one file is lookedup/scanned "
+ g.log.info("Volume %s: At least one file is lookedup/scanned "
"on all nodes", self.volname)
break
count += 1
diff --git a/tests/functional/disperse/test_ec_validate_volume_creation.py b/tests/functional/disperse/test_ec_validate_volume_creation.py
index 99637f2..d9e490c 100644
--- a/tests/functional/disperse/test_ec_validate_volume_creation.py
+++ b/tests/functional/disperse/test_ec_validate_volume_creation.py
@@ -96,7 +96,7 @@ class EcValidateVolumeCreate(GlusterBaseClass):
# Setup Volume and Mount Volume
g.log.info("Starting to Setup Volume and Mount Volume")
ret = self.setup_volume_and_mount_volume(mounts=self.mounts)
- self.assertFalse(ret, ("Volume Setup and Mount succeded"
+ self.assertFalse(ret, ("Volume Setup and Mount succeeded"
" for volume %s", self.volname))
g.log.info("Successfully verified invalid input parameters")
@@ -108,8 +108,8 @@ class EcValidateVolumeCreate(GlusterBaseClass):
# Setup Volume and Mount Volume
g.log.info("Starting to Setup Volume and Mount Volume")
ret = self.setup_volume_and_mount_volume(mounts=self.mounts)
- self.assertFalse(ret, ("Volume Setup and Mount succeded for volume %s",
- self.volname))
+ self.assertFalse(ret, ("Volume Setup and Mount succeeded for volume "
+ "%s", self.volname))
g.log.info("Successfully verified invalid input parameters")
def test_invalid_usecase_three(self):
@@ -120,8 +120,8 @@ class EcValidateVolumeCreate(GlusterBaseClass):
# Setup Volume and Mount Volume
g.log.info("Starting to Setup Volume and Mount Volume")
ret = self.setup_volume_and_mount_volume(mounts=self.mounts)
- self.assertFalse(ret, ("Volume Setup and Mount succeded for volume %s",
- self.volname))
+ self.assertFalse(ret, ("Volume Setup and Mount succeeded for volume "
+ "%s", self.volname))
g.log.info("Successfully verified invalid input parameters")
def test_invalid_usecase_four(self):
@@ -132,6 +132,6 @@ class EcValidateVolumeCreate(GlusterBaseClass):
# Setup Volume and Mount Volume
g.log.info("Starting to Setup Volume and Mount Volume")
ret = self.setup_volume_and_mount_volume(mounts=self.mounts)
- self.assertFalse(ret, ("Volume Setup and Mount succeded for volume %s",
- self.volname))
+ self.assertFalse(ret, ("Volume Setup and Mount succeeded for volume "
+ "%s", self.volname))
g.log.info("Successfully verified invalid input parameters")
diff --git a/tests/functional/fuse_subdir/test_fusesubdir_with_quota_objects.py b/tests/functional/fuse_subdir/test_fusesubdir_with_quota_objects.py
index 10dc0dd..01ca474 100644
--- a/tests/functional/fuse_subdir/test_fusesubdir_with_quota_objects.py
+++ b/tests/functional/fuse_subdir/test_fusesubdir_with_quota_objects.py
@@ -171,7 +171,7 @@ class SubdirWithQuotaObject(GlusterBaseClass):
% (mount_object.mountpoint))
ret, _, _ = g.run(mount_object.client_system, cmd)
self.assertEqual(ret, 0, "Failed to create files on mountpoint")
- g.log.info("Files created succesfully on mountpoint")
+ g.log.info("Files created successfully on mountpoint")
# Fetch Quota List object on the volume
g.log.info("Get Quota list on the volume %s",
@@ -215,7 +215,7 @@ class SubdirWithQuotaObject(GlusterBaseClass):
% (mount_object.mountpoint, i), parents=True)
self.assertTrue(ret, "Failed to create directories"
"on mountpoint")
- g.log.info("Directories created succesfully on mountpoint")
+ g.log.info("Directories created successfully on mountpoint")
# Get Quota List on the volume
g.log.info("Get Quota list on the volume %s",
@@ -240,7 +240,7 @@ class SubdirWithQuotaObject(GlusterBaseClass):
else:
self.assertTrue(ret, "Directory creation got failed"
"on volume")
- g.log.info("Direction creation successfull on volume")
+ g.log.info("Direction creation successful on volume")
def tearDown(self):
"""
diff --git a/tests/functional/fuse_subdir/test_fusesubdir_with_removebrick.py b/tests/functional/fuse_subdir/test_fusesubdir_with_removebrick.py
index 8e06027..30ac9ee 100644
--- a/tests/functional/fuse_subdir/test_fusesubdir_with_removebrick.py
+++ b/tests/functional/fuse_subdir/test_fusesubdir_with_removebrick.py
@@ -220,7 +220,7 @@ class SubdirWithRemoveBrick(GlusterBaseClass):
ret, _, _ = umount_volume(client, self.mpoint,
self.mount_type)
if ret != 0:
- raise ExecutionError("Unmounting the moint point %s failed"
+ raise ExecutionError("Unmounting the mount point %s failed"
% self.mpoint)
g.log.info("Unmount Volume Successful")
cmd = ("rm -rf %s") % self.mpoint
diff --git a/tests/functional/fuse_subdir/test_leveldownsubdir_with_multiclient.py b/tests/functional/fuse_subdir/test_leveldownsubdir_with_multiclient.py
index f7115ce..eef0589 100644
--- a/tests/functional/fuse_subdir/test_leveldownsubdir_with_multiclient.py
+++ b/tests/functional/fuse_subdir/test_leveldownsubdir_with_multiclient.py
@@ -60,7 +60,7 @@ class SubdirLevelDownDirMapping(GlusterBaseClass):
ret, _, _ = g.run(self.mounts[0].client_system, cmd)
self.assertEqual(ret, 0, "Failed to create Nested directories"
"on mountpoint")
- g.log.info("Nested Directories created succesfully on mountpoint")
+ g.log.info("Nested Directories created successfully on mountpoint")
# unmount volume
ret = self.unmount_volume(self.mounts)
@@ -134,7 +134,7 @@ class SubdirLevelDownDirMapping(GlusterBaseClass):
ret, _, _ = umount_volume(client, self.mpoint,
self.mount_type)
if ret == 1:
- raise ExecutionError("Unmounting the moint point %s failed"
+ raise ExecutionError("Unmounting the mount point %s failed"
% self.mpoint)
g.log.info("Unmount Volume Successful")
cmd = ("rm -rf %s") % self.mpoint
diff --git a/tests/functional/fuse_subdir/test_multisubdir_with_multiclient.py b/tests/functional/fuse_subdir/test_multisubdir_with_multiclient.py
index deb5952..ccd513e 100644
--- a/tests/functional/fuse_subdir/test_multisubdir_with_multiclient.py
+++ b/tests/functional/fuse_subdir/test_multisubdir_with_multiclient.py
@@ -59,7 +59,7 @@ class MultipleDirMappingClient(GlusterBaseClass):
self.mounts[0].mountpoint))
ret, _, _ = g.run(self.mounts[0].client_system, cmd)
self.assertEqual(ret, 0, "Failed to create directories on mountpoint")
- g.log.info("Directories created succesfully on mountpoint")
+ g.log.info("Directories created successfully on mountpoint")
# unmount volume
ret = self.unmount_volume(self.mounts)
@@ -127,7 +127,7 @@ class MultipleDirMappingClient(GlusterBaseClass):
ret, _, _ = umount_volume(client, self.mpoint,
self.mount_type)
if ret == 1:
- raise ExecutionError("Unmounting the moint point %s failed"
+ raise ExecutionError("Unmounting the mount point %s failed"
% self.mpoint)
g.log.info("Unmount Volume Successful")
cmd = ("rm -rf %s") % self.mpoint
diff --git a/tests/functional/fuse_subdir/test_quota_limits_fusesubdir_.py b/tests/functional/fuse_subdir/test_quota_limits_fusesubdir_.py
index 4c6e0cf..b8b9206 100644
--- a/tests/functional/fuse_subdir/test_quota_limits_fusesubdir_.py
+++ b/tests/functional/fuse_subdir/test_quota_limits_fusesubdir_.py
@@ -200,7 +200,7 @@ class FuseSubdirQuotaTest(GlusterBaseClass):
"count=1;done" % (mount_object.mountpoint))
ret, _, _ = g.run(mount_object.client_system, cmd)
self.assertEqual(ret, 0, "Failed to create files on mountpoint")
- g.log.info("Files created succesfully on mountpoint")
+ g.log.info("Files created successfully on mountpoint")
# Again Get Quota List on the volume
@@ -254,7 +254,7 @@ class FuseSubdirQuotaTest(GlusterBaseClass):
ret, _, _ = g.run(self.clients[1], cmd)
self.assertEqual(ret, 0, ("Failed to create files on %s"
% self.clients[1]))
- g.log.info("Files created succesfully on %s:%s",
+ g.log.info("Files created successfully on %s:%s",
self.clients[1], self.subdir_mounts[1].mountpoint)
def tearDown(self):
diff --git a/tests/functional/glusterd/test_brick_port_after_stop_glusterd_modify_volume.py b/tests/functional/glusterd/test_brick_port_after_stop_glusterd_modify_volume.py
index da80f67..e44514a 100644
--- a/tests/functional/glusterd/test_brick_port_after_stop_glusterd_modify_volume.py
+++ b/tests/functional/glusterd/test_brick_port_after_stop_glusterd_modify_volume.py
@@ -104,7 +104,7 @@ class TestBrickPortAfterModifyVolume(GlusterBaseClass):
bricks_list, force=False)
self.assertEqual(ret[0], 0, ("Unable"
"to create volume %s" % self.volname))
- g.log.info("Volume created successfuly %s", self.volname)
+ g.log.info("Volume created successfully %s", self.volname)
ret, _, _ = volume_start(self.mnode, self.volname)
self.assertEqual(ret, 0, ("Failed to start the "
diff --git a/tests/functional/glusterd/test_brick_status_when_quorum_not_met.py b/tests/functional/glusterd/test_brick_status_when_quorum_not_met.py
index 6cb3ee0..cc77b3e 100644
--- a/tests/functional/glusterd/test_brick_status_when_quorum_not_met.py
+++ b/tests/functional/glusterd/test_brick_status_when_quorum_not_met.py
@@ -132,7 +132,7 @@ class TestBrickStatusWhenQuorumNotMet(GlusterBaseClass):
# Verfiying node count in volume status after glusterd
# started on servers, Its not possible to check the brick status
- # immediately after glusterd start, thats why verifying that all
+ # immediately after glusterd start, that's why verifying that all
# glusterd started nodes available in gluster volume status or not
count = 0
while count < 50:
diff --git a/tests/functional/glusterd/test_concurrent_set.py b/tests/functional/glusterd/test_concurrent_set.py
index 7c753ea..4b432b7 100644
--- a/tests/functional/glusterd/test_concurrent_set.py
+++ b/tests/functional/glusterd/test_concurrent_set.py
@@ -60,7 +60,7 @@ class TestConcurrentSet(GlusterBaseClass):
self.brick_list, force=False)
self.assertEqual(ret[0], 0, ("Unable"
"to create volume %s" % self.volname))
- g.log.info("Volume created successfuly %s", self.volname)
+ g.log.info("Volume created successfully %s", self.volname)
# Create a volume
self.volname = "second-vol"
@@ -72,7 +72,7 @@ class TestConcurrentSet(GlusterBaseClass):
self.brick_list, force=False)
self.assertEqual(ret[0], 0, ("Unable"
"to create volume %s" % self.volname))
- g.log.info("Volume created successfuly %s", self.volname)
+ g.log.info("Volume created successfully %s", self.volname)
cmd1 = ("for i in `seq 1 100`; do gluster volume set first-vol "
"read-ahead on; done")
diff --git a/tests/functional/glusterd/test_create_vol_with_used_bricks.py b/tests/functional/glusterd/test_create_vol_with_used_bricks.py
index 940d7a4..1be34f7 100644
--- a/tests/functional/glusterd/test_create_vol_with_used_bricks.py
+++ b/tests/functional/glusterd/test_create_vol_with_used_bricks.py
@@ -111,7 +111,7 @@ class TestCreateVolWithUsedBricks(GlusterBaseClass):
# Mounting volume
ret = self.mount_volume(self.mounts)
self.assertTrue(ret, "Volume mount failed for %s" % self.volname)
- g.log.info("Volume mounted sucessfully : %s", self.volname)
+ g.log.info("Volume mounted successfully : %s", self.volname)
# run IOs
g.log.info("Starting IO on all mounts...")
diff --git a/tests/functional/glusterd/test_nfs_quorum.py b/tests/functional/glusterd/test_nfs_quorum.py
index ced5b71..62d2ce2 100644
--- a/tests/functional/glusterd/test_nfs_quorum.py
+++ b/tests/functional/glusterd/test_nfs_quorum.py
@@ -82,7 +82,7 @@ class TestNfsMountAndServerQuorumSettings(GlusterBaseClass):
# Mounting a NFS volume
ret = self.mount_volume(self.mounts)
self.assertTrue(ret, "NFS volume mount failed for %s" % self.volname)
- g.log.info("Volume mounted sucessfully : %s", self.volname)
+ g.log.info("Volume mounted successfully : %s", self.volname)
# unmounting NFS Volume
ret = self.unmount_volume(self.mounts)
diff --git a/tests/functional/glusterd/test_peer_detach.py b/tests/functional/glusterd/test_peer_detach.py
index 2bae76d..6330369 100644
--- a/tests/functional/glusterd/test_peer_detach.py
+++ b/tests/functional/glusterd/test_peer_detach.py
@@ -68,7 +68,7 @@ class PeerDetachVerification(GlusterBaseClass):
# Assigning non existing host to variable
self.non_exist_host = '256.256.256.256'
- # Assigning invalid ip to vaiable
+ # Assigning invalid ip to variable
self.invalid_ip = '10.11.a'
# Peer detach to specified server
diff --git a/tests/functional/glusterd/test_peer_probe_while_snapd_running.py b/tests/functional/glusterd/test_peer_probe_while_snapd_running.py
index c35b3ea..aff0156 100644
--- a/tests/functional/glusterd/test_peer_probe_while_snapd_running.py
+++ b/tests/functional/glusterd/test_peer_probe_while_snapd_running.py
@@ -97,7 +97,7 @@ class TestPeerProbeWhileSnapdRunning(GlusterBaseClass):
# Checking snapd running or not
ret = is_snapd_running(self.mnode, self.volname)
- self.assertTrue(ret, "Snapd not runnig for volume %s" % self.volname)
+ self.assertTrue(ret, "Snapd not running for volume %s" % self.volname)
g.log.info("snapd running for volume %s", self.volname)
# Probing new node
diff --git a/tests/functional/glusterd/test_probe_glusterd.py b/tests/functional/glusterd/test_probe_glusterd.py
index d14991d..54b99ee 100644
--- a/tests/functional/glusterd/test_probe_glusterd.py
+++ b/tests/functional/glusterd/test_probe_glusterd.py
@@ -62,7 +62,7 @@ class PeerProbeInvalidIpNonExistingHost(GlusterBaseClass):
# Assigning non existing ip to variable
self.non_exist_ip = '256.256.256.256'
- # Assigning invalid ip to vaiable
+ # Assigning invalid ip to variable
self.invalid_ip = '10.11.a'
# Assigning non existing host to variable
diff --git a/tests/functional/glusterd/test_probe_hostname.py b/tests/functional/glusterd/test_probe_hostname.py
index 55476ed..6e7d87b 100644
--- a/tests/functional/glusterd/test_probe_hostname.py
+++ b/tests/functional/glusterd/test_probe_hostname.py
@@ -103,14 +103,14 @@ class TestPeerProbe(GlusterBaseClass):
self.brick_list, force=False)
self.assertEqual(ret, 0, "Unable"
"to create volume % s" % self.volname)
- g.log.info("Volume created successfuly % s", self.volname)
+ g.log.info("Volume created successfully % s", self.volname)
# Start a volume
g.log.info("Start a volume")
ret, _, _ = volume_start(self.mnode, self.volname)
self.assertEqual(ret, 0, "Unable"
"to start volume % s" % self.volname)
- g.log.info("Volume started successfuly % s", self.volname)
+ g.log.info("Volume started successfully % s", self.volname)
# Get volume info
g.log.info("get volume info")
@@ -127,7 +127,7 @@ class TestPeerProbe(GlusterBaseClass):
ret, _, _ = volume_stop(self.mnode, self.volname)
self.assertEqual(ret, 0, "Unable"
"to stop volume % s" % self.volname)
- g.log.info("Volume stopped successfuly % s", self.volname)
+ g.log.info("Volume stopped successfully % s", self.volname)
# Create a volume
self.volname = "test-vol-fqdn"
@@ -154,14 +154,14 @@ class TestPeerProbe(GlusterBaseClass):
my_brick_list, force=False)
self.assertEqual(ret, 0, "Unable"
"to create volume % s" % self.volname)
- g.log.info("Volume created successfuly % s", self.volname)
+ g.log.info("Volume created successfully % s", self.volname)
# Start a volume
g.log.info("Start a volume")
ret, _, _ = volume_start(self.mnode, self.volname)
self.assertEqual(ret, 0, "Unable"
"to start volume % s" % self.volname)
- g.log.info("Volume started successfuly % s", self.volname)
+ g.log.info("Volume started successfully % s", self.volname)
# Get volume info
g.log.info("get volume info")
@@ -178,4 +178,4 @@ class TestPeerProbe(GlusterBaseClass):
ret, _, _ = volume_stop(self.mnode, self.volname)
self.assertEqual(ret, 0, "Unable"
"to stop volume % s" % self.volname)
- g.log.info("Volume stopped successfuly % s", self.volname)
+ g.log.info("Volume stopped successfully % s", self.volname)
diff --git a/tests/functional/glusterd/test_quorum_remove_brick.py b/tests/functional/glusterd/test_quorum_remove_brick.py
index 6d5e45b..4429d82 100644
--- a/tests/functional/glusterd/test_quorum_remove_brick.py
+++ b/tests/functional/glusterd/test_quorum_remove_brick.py
@@ -120,7 +120,7 @@ class TestServerQuorumNotMet(GlusterBaseClass):
% self.random_server)
g.log.info("Glusterd stopped successfully on %s", self.random_server)
- # Forming brick list for perfroming remove brick operation
+ # Forming brick list for performing remove brick operation
remove_brick_list = form_bricks_list_to_remove_brick(self.mnode,
self.volname)
self.assertIsNotNone(remove_brick_list, "Failed to get brick list for "
diff --git a/tests/functional/glusterd/test_rebalance_hang.py b/tests/functional/glusterd/test_rebalance_hang.py
index d96a404..a826703 100644
--- a/tests/functional/glusterd/test_rebalance_hang.py
+++ b/tests/functional/glusterd/test_rebalance_hang.py
@@ -114,7 +114,7 @@ class TestRebalanceHang(GlusterBaseClass):
bricks_list, force=False)
self.assertEqual(ret, 0, ("Unable"
"to create volume %s" % self.volname))
- g.log.info("Volume created successfuly %s", self.volname)
+ g.log.info("Volume created successfully %s", self.volname)
ret, _, _ = volume_start(self.mnode, self.volname, False)
self.assertEqual(ret, 0, ("Failed to start the "
@@ -130,7 +130,7 @@ class TestRebalanceHang(GlusterBaseClass):
mserver=self.mnode,
mclient=self.mounts[0].client_system)
self.assertEqual(ret, 0, ("Volume %s is not mounted") % self.volname)
- g.log.info("Volume mounted sucessfully : %s", self.volname)
+ g.log.info("Volume mounted successfully : %s", self.volname)
self.all_mounts_procs = []
# Creating files
diff --git a/tests/functional/glusterd/test_rebalance_spurious.py b/tests/functional/glusterd/test_rebalance_spurious.py
index 9b73188..ad16a00 100644
--- a/tests/functional/glusterd/test_rebalance_spurious.py
+++ b/tests/functional/glusterd/test_rebalance_spurious.py
@@ -114,7 +114,7 @@ class TestSpuriousRebalance(GlusterBaseClass):
bricks_list, force=False)
self.assertEqual(ret, 0, ("Unable"
"to create volume %s" % self.volname))
- g.log.info("Volume created successfuly %s", self.volname)
+ g.log.info("Volume created successfully %s", self.volname)
ret, _, _ = volume_start(self.mnode, self.volname, False)
self.assertEqual(ret, 0, ("Failed to start the "
@@ -130,7 +130,7 @@ class TestSpuriousRebalance(GlusterBaseClass):
mserver=self.mnode,
mclient=self.mounts[0].client_system)
self.assertEqual(ret, 0, ("Volume %s is not mounted") % self.volname)
- g.log.info("Volume mounted sucessfully : %s", self.volname)
+ g.log.info("Volume mounted successfully : %s", self.volname)
remove_brick_list = []
remove_brick_list.append(bricks_list[2])
ret, _, _ = remove_brick(self.mnode, self.volname, remove_brick_list,
diff --git a/tests/functional/glusterd/test_remove_brick_after_restart_glusterd.py b/tests/functional/glusterd/test_remove_brick_after_restart_glusterd.py
index 217eae5..f84c7db 100644
--- a/tests/functional/glusterd/test_remove_brick_after_restart_glusterd.py
+++ b/tests/functional/glusterd/test_remove_brick_after_restart_glusterd.py
@@ -112,7 +112,7 @@ class TestRemoveBrickAfterRestartGlusterd(GlusterBaseClass):
bricks_list, force=False, **kwargs)
self.assertEqual(ret[0], 0, ("Unable"
"to create volume %s" % self.volname))
- g.log.info("Volume created successfuly %s", self.volname)
+ g.log.info("Volume created successfully %s", self.volname)
ret, _, _ = volume_start(self.mnode, self.volname, False)
self.assertEqual(ret, 0, ("Failed to start the "
@@ -128,7 +128,7 @@ class TestRemoveBrickAfterRestartGlusterd(GlusterBaseClass):
mserver=self.mnode,
mclient=self.mounts[0].client_system)
self.assertEqual(ret, 0, ("Volume %s is not mounted") % self.volname)
- g.log.info("Volume mounted sucessfully : %s", self.volname)
+ g.log.info("Volume mounted successfully : %s", self.volname)
self.all_mounts_procs = []
# Creating files
diff --git a/tests/functional/glusterd/test_volume_get.py b/tests/functional/glusterd/test_volume_get.py
index 228b152..5746d5d 100644
--- a/tests/functional/glusterd/test_volume_get.py
+++ b/tests/functional/glusterd/test_volume_get.py
@@ -161,7 +161,7 @@ class TestVolumeGet(GlusterBaseClass):
"existing volume %s with non existing option",
self.volname)
- # perfroming gluster volume get volname all
+ # performing gluster volume get volname all
ret = get_volume_options(self.mnode, self.volname, "all")
self.assertIsNotNone(ret, "gluster volume get %s all command "
@@ -194,7 +194,7 @@ class TestVolumeGet(GlusterBaseClass):
"performance.low-prio-threads value is not correct")
g.log.info("performance.low-prio-threads value is correct")
- # perfroming gluster volume get volname all
+ # performing gluster volume get volname all
ret = get_volume_options(self.mnode, self.volname, "all")
self.assertIsNotNone(ret, "gluster volume get %s all command "
"failed" % self.volname)
diff --git a/tests/functional/glusterd/test_volume_network_ping_timeout.py b/tests/functional/glusterd/test_volume_network_ping_timeout.py
index b897c4a..7d72d8a 100644
--- a/tests/functional/glusterd/test_volume_network_ping_timeout.py
+++ b/tests/functional/glusterd/test_volume_network_ping_timeout.py
@@ -96,7 +96,7 @@ class CheckVolumeChecksumAfterChangingNetworkPingTimeOut(GlusterBaseClass):
# Mounting volume as glusterfs
ret = self.mount_volume(self.mounts)
self.assertTrue(ret, "volume mount failed for %s" % self.volname)
- g.log.info("Volume mounted sucessfully : %s", self.volname)
+ g.log.info("Volume mounted successfully : %s", self.volname)
# Checking volume mounted or not
ret = is_mounted(self.volname, self.mounts[0].mountpoint, self.mnode,
diff --git a/tests/functional/glusterd/test_volume_reset.py b/tests/functional/glusterd/test_volume_reset.py
index f61fdaa..9720c3e 100644
--- a/tests/functional/glusterd/test_volume_reset.py
+++ b/tests/functional/glusterd/test_volume_reset.py
@@ -124,10 +124,10 @@ class GlusterdVolumeReset(GlusterBaseClass):
cmd = "gluster volume reset " + self.volname
ret, _, _ = g.run(self.mnode, cmd)
self.assertEqual(ret, 0, "volume reset failed for : %s" % self.volname)
- g.log.info("volume resetted succefully :%s", self.volname)
+ g.log.info("Volume reset successfully :%s", self.volname)
# After volume reset snap daemon will not be running,
- # bitd and scrub deamons will be in running state.
+ # bitd and scrub daemons will be in running state.
g.log.info("checking snapshot, scrub and bitrot daemons\
running or not after volume reset")
for mnode in self.servers:
@@ -157,7 +157,7 @@ class GlusterdVolumeReset(GlusterBaseClass):
ret, _, _ = g.run(self.mnode, cmd)
self.assertEqual(ret, 0, "volume reset fail\
for : %s" % self.volname)
- g.log.info("Volume reset sucessfully with force option :%s",
+ g.log.info("Volume reset successfully with force option :%s",
self.volname)
# After volume reset bitd, snapd, scrub daemons will not be running,
diff --git a/tests/functional/glusterd/test_volume_status.py b/tests/functional/glusterd/test_volume_status.py
index ff5d575..109586f 100644
--- a/tests/functional/glusterd/test_volume_status.py
+++ b/tests/functional/glusterd/test_volume_status.py
@@ -110,17 +110,17 @@ class VolumeStatusWhenIOInProgress(GlusterBaseClass):
command on all clusters randomly.
"gluster volume status volname inode" command should not get
hang while IOs in progress.
- Then check that IOs completed successfullly or not on mount point.
+ Then check that IOs completed successfully or not on mount point.
Check that files in mount point listing properly or not.
'''
# Mounting a volume
ret = self.mount_volume(self.mounts)
self.assertTrue(ret, "Volume mount failed for %s" % self.volname)
- g.log.info("Volume mounted sucessfully : %s", self.volname)
+ g.log.info("Volume mounted successfully : %s", self.volname)
- # After Mounting immediately writting IO's are failing some times,
- # thats why keeping sleep for 10 secs
+ # After Mounting immediately writing IO's are failing some times,
+ # that's why keeping sleep for 10 secs
sleep(10)
# run IOs
diff --git a/tests/functional/glusterd/test_volume_status_fd.py b/tests/functional/glusterd/test_volume_status_fd.py
index 2765325..415c96d 100644
--- a/tests/functional/glusterd/test_volume_status_fd.py
+++ b/tests/functional/glusterd/test_volume_status_fd.py
@@ -94,7 +94,7 @@ class VolumeStatusFdWhenIOInProgress(GlusterBaseClass):
-> Mount the volume on 2 clients
-> Run I/O's on mountpoint
-> While I/O's are in progress
- -> Perfrom gluster volume status fd repeatedly
+ -> Perform gluster volume status fd repeatedly
-> List all files and dirs listed
'''
diff --git a/tests/functional/nfs_ganesha/test_nfs_ganesha_sanity.py b/tests/functional/nfs_ganesha/test_nfs_ganesha_sanity.py
index a00c22a..05caf4a 100644
--- a/tests/functional/nfs_ganesha/test_nfs_ganesha_sanity.py
+++ b/tests/functional/nfs_ganesha/test_nfs_ganesha_sanity.py
@@ -75,7 +75,7 @@ class TestNfsGaneshaSanity(NfsGaneshaVolumeBaseClass):
"tar xvf linux-2.6.1.tar.gz" % (mount_obj.mountpoint))
ret, out, err = g.run(mount_obj.client_system, cmd)
if ret == 0:
- g.log.info("Succesfully untared the tarball!")
+ g.log.info("Successfully untared the tarball!")
else:
g.log.error("ERROR ! Kernel untar errored out!")
self.assertEqual(ret, 0, "Kernel untar failed!")
diff --git a/tests/functional/nfs_ganesha/test_nfs_ganesha_volume_exports.py b/tests/functional/nfs_ganesha/test_nfs_ganesha_volume_exports.py
index 8a8b28c..ead84db 100644
--- a/tests/functional/nfs_ganesha/test_nfs_ganesha_volume_exports.py
+++ b/tests/functional/nfs_ganesha/test_nfs_ganesha_volume_exports.py
@@ -69,7 +69,7 @@ class TestNfsGaneshaVolumeExports(NfsGaneshaVolumeBaseClass):
g.log.info("Testing nfs ganesha export after volume stop/start."
"Count : %s", str(i))
- # Stoping volume
+ # Stopping volume
ret = volume_stop(self.mnode, self.volname)
self.assertTrue(ret, ("Failed to stop volume %s" % self.volname))
@@ -485,7 +485,7 @@ class TestNfsGaneshaSubDirExportsWithIO(NfsGaneshaIOBaseClass):
export_file)))
# Stop and start volume to take the modified export file to effect.
- # Stoping volume
+ # Stopping volume
ret = volume_stop(self.mnode, self.volname)
self.assertTrue(ret, ("Failed to stop volume %s" % self.volname))
diff --git a/tests/functional/quota/list_path_values.py b/tests/functional/quota/list_path_values.py
index 5c9bf81..2923704 100644
--- a/tests/functional/quota/list_path_values.py
+++ b/tests/functional/quota/list_path_values.py
@@ -112,7 +112,7 @@ class QuotaListPathValues(GlusterBaseClass):
"count=1;done" % (mount_object.mountpoint))
ret, _, _ = g.run(mount_object.client_system, cmd)
self.assertEqual(ret, 0, "Failed to create files on mountpoint")
- g.log.info("Files created succesfully on mountpoint")
+ g.log.info("Files created successfully on mountpoint")
# Get Quota list without specifying the path
g.log.info("Get Quota list for the volume %s", self.volname)
diff --git a/tests/functional/quota/test_limit_usage_deep_dir.py b/tests/functional/quota/test_limit_usage_deep_dir.py
index f066441..974249c 100644
--- a/tests/functional/quota/test_limit_usage_deep_dir.py
+++ b/tests/functional/quota/test_limit_usage_deep_dir.py
@@ -200,7 +200,7 @@ class LimitUsageDeepDir(GlusterBaseClass):
dir_name)
g.log.info("Quota list validate and file created successful "
"for %s", dir_name)
- g.log.info("Files created and quota validated succesfully")
+ g.log.info("Files created and quota validated successfully")
# Deleting data and validating quota
self.all_mounts_procs = []
diff --git a/tests/functional/quota/test_multi_value_limit.py b/tests/functional/quota/test_multi_value_limit.py
index 08bf6d2..3991c61 100644
--- a/tests/functional/quota/test_multi_value_limit.py
+++ b/tests/functional/quota/test_multi_value_limit.py
@@ -154,7 +154,7 @@ class QuotaMultiValueLimits(GlusterBaseClass):
"count=1 ; "
"done" % mount_dir)
ret, _, _ = g.run(client, cmd)
- self.assertEqual(ret, 1, "Failed: Files created successfully inspite "
+ self.assertEqual(ret, 1, "Failed: Files created successfully in spite "
"of crossing hard-limit")
g.log.info("Files creation stopped on mountpoint once exceeded "
"hard limit")
@@ -202,7 +202,7 @@ class QuotaMultiValueLimits(GlusterBaseClass):
"count=1 ; "
"done" % mount_dir)
ret, _, _ = g.run(client, cmd)
- self.assertEqual(ret, 1, "Failed: Files created successfully inspite "
+ self.assertEqual(ret, 1, "Failed: Files created successfully in spite "
"of crossing hard-limit")
g.log.info("Files creation stopped on mountpoint once exceeded "
"hard limit")
@@ -250,7 +250,7 @@ class QuotaMultiValueLimits(GlusterBaseClass):
"count=1 ; "
"done" % mount_dir)
ret, _, _ = g.run(client, cmd)
- self.assertEqual(ret, 1, "Failed: Files created successfully inspite "
+ self.assertEqual(ret, 1, "Failed: Files created successfully in spite "
"of crossing hard-limit")
g.log.info("Files creation stopped on mountpoint once exceeded "
"hard limit")
diff --git a/tests/functional/quota/test_quota_add_brick.py b/tests/functional/quota/test_quota_add_brick.py
index 58e9610..1078cef 100644
--- a/tests/functional/quota/test_quota_add_brick.py
+++ b/tests/functional/quota/test_quota_add_brick.py
@@ -120,7 +120,7 @@ class QuotaAddBrick(GlusterBaseClass):
% (mount_dir))
ret, _, _ = g.run(client, cmd)
self.assertEqual(ret, 0, ("Failed to create files"))
- g.log.info("Files created succesfully")
+ g.log.info("Files created successfully")
# Quota validate
ret = quota_validate(self.mnode, self.volname,
diff --git a/tests/functional/quota/test_quota_limit_dir_breadth.py b/tests/functional/quota/test_quota_limit_dir_breadth.py
index b70c118..952e341 100755
--- a/tests/functional/quota/test_quota_limit_dir_breadth.py
+++ b/tests/functional/quota/test_quota_limit_dir_breadth.py
@@ -165,7 +165,7 @@ class QuotaLimitDirBreadth(GlusterBaseClass):
ret, _, _ = g.run(mount_object.client_system, cmd)
self.assertFalse(ret, "Failed to create files in %s"
% dir_name)
- g.log.info("Files created succesfully in %s", dir_name)
+ g.log.info("Files created successfully in %s", dir_name)
# Get Quota list on Volume
g.log.info("Get Quota list for every directory on the volume %s",
diff --git a/tests/functional/quota/test_quota_unique_soft_limit.py b/tests/functional/quota/test_quota_unique_soft_limit.py
index f7f726d..b175f12 100644
--- a/tests/functional/quota/test_quota_unique_soft_limit.py
+++ b/tests/functional/quota/test_quota_unique_soft_limit.py
@@ -156,7 +156,7 @@ class QuotaUniqueSoftLimit(GlusterBaseClass):
ret, _, _ = g.run(client, cmd)
self.assertEqual(ret, 0, ("Failed to create files on %s",
dir_name))
- g.log.info("Files created succesfully on %s", dir_name)
+ g.log.info("Files created successfully on %s", dir_name)
time.sleep(1)
diff --git a/tests/functional/quota/test_quota_volume_subdir_limits.py b/tests/functional/quota/test_quota_volume_subdir_limits.py
index 1291796..7e4245c 100644
--- a/tests/functional/quota/test_quota_volume_subdir_limits.py
+++ b/tests/functional/quota/test_quota_volume_subdir_limits.py
@@ -153,7 +153,7 @@ class QuotaVolumeAndSubdirLimits(GlusterBaseClass):
ret, _, _ = g.run(client, cmd)
self.assertEqual(ret, 0, ("Failed to create files in /foo%s",
var1))
- g.log.info("Files created succesfully in /foo%s", var1)
+ g.log.info("Files created successfully in /foo%s", var1)
g.log.info("Files creation is successful on all directories of the "
"volume %s", self.volname)
diff --git a/tests/functional/snapshot/test_256_snapshots.py b/tests/functional/snapshot/test_256_snapshots.py
index f3e6e4b..9c5ceaf 100644
--- a/tests/functional/snapshot/test_256_snapshots.py
+++ b/tests/functional/snapshot/test_256_snapshots.py
@@ -38,7 +38,7 @@ class SanpCreate256(GlusterBaseClass):
4. Varify IO
5. modify max snap limit to default to 256.
6. Create 256 snapshots
- 7. Varify 256 created sucessfully
+ 7. Varify 256 created successfully
8. Create 257th snapshot - check for failure
-- it should fail.
9. Cleanup
@@ -92,7 +92,7 @@ class SanpCreate256(GlusterBaseClass):
ret = cls.unmount_volume_and_cleanup_volume(cls.mounts)
if not ret:
raise ExecutionError("Failed to cleanup volume and mount")
- g.log.info("Cleanup successfull for the volume and mount")
+ g.log.info("Cleanup successful for the volume and mount")
GlusterBaseClass.tearDownClass.im_func(cls)
@@ -155,7 +155,7 @@ class SanpCreate256(GlusterBaseClass):
for i in range(1, 257, 1):
self.assertTrue(("snapy%s" % i in snap_list), "%s snap not "
"found " % ("snapy%s" % i))
- g.log.info("Sucessfully validated names of snap")
+ g.log.info("Successfully validated names of snap")
# Try to create 257th snapshot
cmd_str = "gluster snapshot create %s %s %s" % ("snap", self.volname,
diff --git a/tests/functional/snapshot/test_clone_snap.py b/tests/functional/snapshot/test_clone_snap.py
index a7ebd96..6311a28 100644
--- a/tests/functional/snapshot/test_clone_snap.py
+++ b/tests/functional/snapshot/test_clone_snap.py
@@ -198,7 +198,7 @@ class CloneSnapshot(GlusterBaseClass):
"do touch file$i; done; cd;") % mount_obj
proc = g.run(self.clients[0], cmd)
all_mounts_procs.append(proc)
- g.log.info("I/O on mountpoint is Successfull")
+ g.log.info("I/O on mountpoint is successful")
# create snapshot
g.log.info("Starting to Create snapshot of clone volume")
@@ -226,7 +226,7 @@ class CloneSnapshot(GlusterBaseClass):
ret, _, _ = umount_volume(self.clients[0], self.mpoint,
self.mount_type)
if ret == 1:
- raise ExecutionError("Unmounting the moint point %s failed"
+ raise ExecutionError("Unmounting the mount point %s failed"
% self.mpoint)
g.log.info("Mount point %s deleted successfully", self.mpoint)
g.log.info("Unmount Volume Successful")
diff --git a/tests/functional/snapshot/test_mount_snap.py b/tests/functional/snapshot/test_mount_snap.py
index e01be9d..67b8a43 100644
--- a/tests/functional/snapshot/test_mount_snap.py
+++ b/tests/functional/snapshot/test_mount_snap.py
@@ -177,7 +177,7 @@ class TestSnapMountSnapshot(GlusterBaseClass):
# validate io should fail
self.assertFalse(
validate_io_procs(all_mounts_procs, self.mounts),
- "Unexpected: IO Successfull on all clients"
+ "Unexpected: IO Successful on all clients"
)
g.log.info("Expected: IO failed on clients")
diff --git a/tests/functional/snapshot/test_snap_delete_existing_scheduler.py b/tests/functional/snapshot/test_snap_delete_existing_scheduler.py
index 436e7f2..a321dcd 100644
--- a/tests/functional/snapshot/test_snap_delete_existing_scheduler.py
+++ b/tests/functional/snapshot/test_snap_delete_existing_scheduler.py
@@ -186,7 +186,7 @@ class SnapshotDeleteExistingScheduler(GlusterBaseClass):
count += 2
self.assertEqual(status.strip().split(":")[2], ' Enabled',
"Failed to check status of scheduler")
- g.log.info("Successfuly checked scheduler status")
+ g.log.info("Successfully checked scheduler status")
# write files on all mounts
g.log.info("Starting IO on all mounts...")
diff --git a/tests/functional/snapshot/test_snap_list_after_restart.py b/tests/functional/snapshot/test_snap_list_after_restart.py
index 1eb8375..52f34c1 100644
--- a/tests/functional/snapshot/test_snap_list_after_restart.py
+++ b/tests/functional/snapshot/test_snap_list_after_restart.py
@@ -122,7 +122,7 @@ class SnapshotGlusterddown(GlusterBaseClass):
g.log.info("Starting to list all snapshots")
for server in self.servers[0:]:
out = get_snap_list(server)
- self.assertIsNotNone(out, "Falied to list snap in node"
+ self.assertIsNotNone(out, "Failed to list snap in node"
"%s" % server)
self.assertEqual(len(out), 2, "Failed to validate snap list"
"on node %s" % server)
diff --git a/tests/functional/snapshot/test_snap_uss.py b/tests/functional/snapshot/test_snap_uss.py
index aad6d89..60e06f3 100644
--- a/tests/functional/snapshot/test_snap_uss.py
+++ b/tests/functional/snapshot/test_snap_uss.py
@@ -197,7 +197,7 @@ class SnapshotUssSnap(GlusterBaseClass):
for mount_obj in self.mounts:
ret, _, _ = uss_list_snaps(mount_obj.client_system,
mount_obj.mountpoint)
- self.assertEqual(ret, 0, "Falied to list snapshot Informations")
+ self.assertEqual(ret, 0, "Failed to list snapshot information")
g.log.info("Successfully Listed snapshots Created")
# Disable USS running
@@ -211,8 +211,8 @@ class SnapshotUssSnap(GlusterBaseClass):
for mount_obj in self.mounts:
ret, out, _ = uss_list_snaps(mount_obj.client_system,
mount_obj.mountpoint)
- self.assertEqual(ret, 0, "Failed to list snapshot Informations")
- g.log.info("Successfully Listed snapshots Created")
+ self.assertEqual(ret, 0, "Failed to list snapshot information")
+ g.log.info("Successfully listed snapshots Created")
# Validate after disabling USS, all files should be same
g.log.info("Validate files after disabling uss")
diff --git a/tests/functional/snapshot/test_snapshot_create.py b/tests/functional/snapshot/test_snapshot_create.py
index 5ffb916..e1bc9c2 100644
--- a/tests/functional/snapshot/test_snapshot_create.py
+++ b/tests/functional/snapshot/test_snapshot_create.py
@@ -101,7 +101,7 @@ class SnapCreate(GlusterBaseClass):
ret = cls.unmount_volume_and_cleanup_volume(cls.mounts)
if not ret:
raise ExecutionError("Failed to cleanup volume and mount")
- g.log.info("Cleanup successfull for the volume and mount")
+ g.log.info("Cleanup successful for the volume and mount")
GlusterBaseClass.tearDownClass.im_func(cls)
@@ -202,4 +202,4 @@ class SnapCreate(GlusterBaseClass):
for i in range(0, 5):
self.assertIn("snapy%s" % i, snap_list, "%s snap not "
"found " % ("snapy%s" % i))
- g.log.info("Sucessfully validated names of snap")
+ g.log.info("Successfully validated names of snap")
diff --git a/tests/functional/snapshot/test_snapshot_restore.py b/tests/functional/snapshot/test_snapshot_restore.py
index cbec67e..966551b 100644
--- a/tests/functional/snapshot/test_snapshot_restore.py
+++ b/tests/functional/snapshot/test_snapshot_restore.py
@@ -106,7 +106,7 @@ class SnapRestore(GlusterBaseClass):
ret = cls.unmount_volume_and_cleanup_volume(cls.mounts)
if not ret:
raise ExecutionError("Failed to cleanup volume and mount")
- g.log.info("Cleanup successfull for the volume and mount")
+ g.log.info("Cleanup successful for the volume and mount")
GlusterBaseClass.tearDownClass.im_func(cls)
@@ -157,7 +157,7 @@ class SnapRestore(GlusterBaseClass):
% self.volname))
g.log.info("Volume options for%s is set successfully", self.volname)
- # Get brick list befor taking snap_restore
+ # Get brick list before taking snap_restore
bricks_before_snap_restore = get_all_bricks(self.mnode, self.volname)
g.log.info("Brick List before snap restore "
"volume: %s", bricks_before_snap_restore)
diff --git a/tests/functional/snapshot/test_uss_brick_down.py b/tests/functional/snapshot/test_uss_brick_down.py
index 94b48c0..172b0f2 100644
--- a/tests/functional/snapshot/test_uss_brick_down.py
+++ b/tests/functional/snapshot/test_uss_brick_down.py
@@ -98,7 +98,7 @@ class SnapUssBrickDown(GlusterBaseClass):
* Mount volume
* Perform I/O on mounts
* Bring down one brick
- * Enbale USS
+ * Enable USS
* Validate USS is enabled
* Bring the brick online using gluster v start force
* Create 2 snapshots snapy1 & snapy2
@@ -164,7 +164,7 @@ class SnapUssBrickDown(GlusterBaseClass):
g.log.info("Bring the brick online using gluster v start force")
ret, _, _ = volume_start(self.mnode, self.volname, force=True)
self.assertEqual(ret, 0, "Volume start with force failed")
- g.log.info("Volume start with force successfull")
+ g.log.info("Volume start with force successful")
# Create 2 snapshot
g.log.info("Creating 2 snapshots for volume %s", self.volname)
diff --git a/tests/functional/snapshot/test_uss_snap_active_deactive.py b/tests/functional/snapshot/test_uss_snap_active_deactive.py
index c28cc1d..93c0231 100644
--- a/tests/functional/snapshot/test_uss_snap_active_deactive.py
+++ b/tests/functional/snapshot/test_uss_snap_active_deactive.py
@@ -100,7 +100,7 @@ class SnapUssActiveD(GlusterBaseClass):
* Perform I/O on mounts
* Create 2 snapshots snapy1 & snapy2
* Validate snap created
- * Enbale USS
+ * Enable USS
* Validate USS is enabled
* Validate snapd is running
* Activate snapy1 & snapy2
diff --git a/tests/functional/snapshot/test_validate_snap_scheduler.py b/tests/functional/snapshot/test_validate_snap_scheduler.py
index ea54f26..c8513ca 100644
--- a/tests/functional/snapshot/test_validate_snap_scheduler.py
+++ b/tests/functional/snapshot/test_validate_snap_scheduler.py
@@ -127,7 +127,7 @@ class SnapshotSchedulerBehaviour(GlusterBaseClass):
count += 2
self.assertEqual(status.strip().split(":")[2], ' Enabled',
"Failed to check status of scheduler")
- g.log.info("Successfuly checked scheduler status")
+ g.log.info("Successfully checked scheduler status")
# Disable snap scheduler
g.log.info("Starting to disable snapshot scheduler on all nodes")
@@ -148,4 +148,4 @@ class SnapshotSchedulerBehaviour(GlusterBaseClass):
count += 2
self.assertEqual(status.strip().split(":")[2], ' Disabled',
"Failed to check status of scheduler")
- g.log.info("Successfuly checked scheduler status")
+ g.log.info("Successfully checked scheduler status")
diff --git a/tests/functional/snapshot/test_validate_snapshot_max_limit.py b/tests/functional/snapshot/test_validate_snapshot_max_limit.py
index 75582e4..3df2a98 100644
--- a/tests/functional/snapshot/test_validate_snapshot_max_limit.py
+++ b/tests/functional/snapshot/test_validate_snapshot_max_limit.py
@@ -26,7 +26,7 @@ Steps :
5. modify max snap limit to default to 10.
6. modify soft-limit to 50%
6. Create 5 snapshots
- 7. Varify 5 created sucessfully
+ 7. Varify 5 created successfully
8. Create 6th snapshot - check for warning
-- it should not fail.
9. modify soft-limit to 100%
@@ -65,7 +65,7 @@ class SnapCreateMax(GlusterBaseClass):
5. modify max snap limit to default to 10.
6. modify soft-limit to 50%
6. Create 5 snapshots
- 7. Varify 5 created sucessfully
+ 7. Varify 5 created successfully
8. Create 6th snapshot - check for warning
-- it should not fail.
9. modify soft-limit to 100%
@@ -125,7 +125,7 @@ class SnapCreateMax(GlusterBaseClass):
ret = self.unmount_volume_and_cleanup_volume(self.mounts)
if not ret:
raise ExecutionError("Failed to cleanup volume and mount")
- g.log.info("Cleanup successfull for the volume and mount")
+ g.log.info("Cleanup successful for the volume and mount")
def test_validate_snaps_max_limit(self):
# pylint: disable=too-many-statements
@@ -194,7 +194,7 @@ class SnapCreateMax(GlusterBaseClass):
for i in range(1, 6):
self.assertTrue(("snapy%s" % i in snap_list), "%s snap not "
"found " % ("snapy%s" % i))
- g.log.info("Sucessfully validated names of snapshots")
+ g.log.info("Successfully validated names of snapshots")
# create 6th snapshot
cmd_str = "gluster snapshot create %s %s %s" % ("snapy6", self.volname,
@@ -246,7 +246,7 @@ class SnapCreateMax(GlusterBaseClass):
cmd_str = "gluster snapshot create %s %s %s" % ("snap", self.volname,
"no-timestamp")
ret, _, _ = g.run(self.mnode, cmd_str)
- self.assertNotEqual(ret, 0, ("Unexpected: sucessfully created 'snap' "
+ self.assertNotEqual(ret, 0, ("Unexpected: successfully created 'snap' "
"for %s" % self.volname))
g.log.info("Expected: Snapshot 'snap' not created as it is 11th snap")
diff --git a/tests/functional/snapshot/test_validate_snapshot_rebalance.py b/tests/functional/snapshot/test_validate_snapshot_rebalance.py
index 7eb6744..f9e018d 100644
--- a/tests/functional/snapshot/test_validate_snapshot_rebalance.py
+++ b/tests/functional/snapshot/test_validate_snapshot_rebalance.py
@@ -55,7 +55,7 @@ class SnapCreateRebal(GlusterBaseClass):
during rebalance
9. After rebalance is completed, create snapshots with the same name as
in Step 7
- -- this operation should be sucessfull
+ -- this operation should be successful
10. Cleanup
"""
@@ -103,7 +103,7 @@ class SnapCreateRebal(GlusterBaseClass):
ret = self.unmount_volume_and_cleanup_volume(self.mounts)
if not ret:
raise ExecutionError("Failed to cleanup volume and mount")
- g.log.info("Cleanup successfull for the volume and mount")
+ g.log.info("Cleanup successful for the volume and mount")
def test_snapshot_while_rebalance(self):
# pylint: disable=too-many-statements, missing-docstring