From 82c94a926c1f1961932798257d18a2fc306f9393 Mon Sep 17 00:00:00 2001 From: Nigel Babu Date: Fri, 27 Jul 2018 13:02:05 +0530 Subject: Fix spelling mistake across the codebase Change-Id: I46fc2feffe6443af6913785d67bf310838532421 --- .../afr/heal/test_self_heal_daemon_process.py | 18 ++++---- tests/functional/afr/test_client_side_quorum.py | 48 +++++++++++----------- ...est_conservative_merge_of_files_heal_command.py | 2 +- .../afr/test_gfid_assignment_on_dist_rep_vol.py | 2 +- tests/functional/afr/test_gfid_heal.py | 2 +- .../afr/test_gfid_split_brain_resolution.py | 4 +- .../functional/afr/test_quota_limit_entry_heal.py | 2 +- .../afr/test_self_heal_when_dir_quota_exceeded.py | 4 +- .../afr/test_self_heal_with_quota_object_limit.py | 4 +- ...write_io_mount_point_resumed_quorum_restored.py | 2 +- .../arbiter/brick_cases/test_brickcases.py | 2 +- .../brick_cases/test_replica3_to_arbiter.py | 4 +- .../test_mount_point_while_deleting_files.py | 2 +- tests/functional/authentication/test_fusereject.py | 2 +- tests/functional/authentication/test_vol_auth.py | 2 +- tests/functional/bvt/test_cvt.py | 2 +- tests/functional/dht/test_dht_hash_value.py | 4 +- ...st_stop_glusterd_while_rebalance_in_progress.py | 4 +- .../disperse/test_ec_validate_volume_creation.py | 14 +++---- .../test_fusesubdir_with_quota_objects.py | 6 +-- .../test_fusesubdir_with_removebrick.py | 2 +- .../test_leveldownsubdir_with_multiclient.py | 4 +- .../test_multisubdir_with_multiclient.py | 4 +- .../fuse_subdir/test_quota_limits_fusesubdir_.py | 4 +- ...brick_port_after_stop_glusterd_modify_volume.py | 2 +- .../test_brick_status_when_quorum_not_met.py | 2 +- tests/functional/glusterd/test_concurrent_set.py | 4 +- .../glusterd/test_create_vol_with_used_bricks.py | 2 +- tests/functional/glusterd/test_nfs_quorum.py | 2 +- tests/functional/glusterd/test_peer_detach.py | 2 +- .../test_peer_probe_while_snapd_running.py | 2 +- tests/functional/glusterd/test_probe_glusterd.py | 2 +- tests/functional/glusterd/test_probe_hostname.py | 12 +++--- .../glusterd/test_quorum_remove_brick.py | 2 +- tests/functional/glusterd/test_rebalance_hang.py | 4 +- .../functional/glusterd/test_rebalance_spurious.py | 4 +- .../test_remove_brick_after_restart_glusterd.py | 4 +- tests/functional/glusterd/test_volume_get.py | 4 +- .../glusterd/test_volume_network_ping_timeout.py | 2 +- tests/functional/glusterd/test_volume_reset.py | 6 +-- tests/functional/glusterd/test_volume_status.py | 8 ++-- tests/functional/glusterd/test_volume_status_fd.py | 2 +- .../nfs_ganesha/test_nfs_ganesha_sanity.py | 2 +- .../nfs_ganesha/test_nfs_ganesha_volume_exports.py | 4 +- tests/functional/quota/list_path_values.py | 2 +- .../functional/quota/test_limit_usage_deep_dir.py | 2 +- tests/functional/quota/test_multi_value_limit.py | 6 +-- tests/functional/quota/test_quota_add_brick.py | 2 +- .../quota/test_quota_limit_dir_breadth.py | 2 +- .../quota/test_quota_unique_soft_limit.py | 2 +- .../quota/test_quota_volume_subdir_limits.py | 2 +- tests/functional/snapshot/test_256_snapshots.py | 6 +-- tests/functional/snapshot/test_clone_snap.py | 4 +- tests/functional/snapshot/test_mount_snap.py | 2 +- .../test_snap_delete_existing_scheduler.py | 2 +- .../snapshot/test_snap_list_after_restart.py | 2 +- tests/functional/snapshot/test_snap_uss.py | 6 +-- tests/functional/snapshot/test_snapshot_create.py | 4 +- tests/functional/snapshot/test_snapshot_restore.py | 4 +- tests/functional/snapshot/test_uss_brick_down.py | 4 +- .../snapshot/test_uss_snap_active_deactive.py | 2 +- .../snapshot/test_validate_snap_scheduler.py | 4 +- .../snapshot/test_validate_snapshot_max_limit.py | 10 ++--- .../snapshot/test_validate_snapshot_rebalance.py | 4 +- 64 files changed, 144 insertions(+), 144 deletions(-) (limited to 'tests/functional') diff --git a/tests/functional/afr/heal/test_self_heal_daemon_process.py b/tests/functional/afr/heal/test_self_heal_daemon_process.py index ed71e4f2b..15cd43951 100644 --- a/tests/functional/afr/heal/test_self_heal_daemon_process.py +++ b/tests/functional/afr/heal/test_self_heal_daemon_process.py @@ -176,7 +176,7 @@ class SelfHealDaemonProcessTests(GlusterBaseClass): self.all_servers_info) self.assertTrue(ret, ("Failed to add bricks to " "volume %s " % self.volname)) - g.log.info("Add brick successfull") + g.log.info("Add brick successful") # Log Volume Info and Status after expanding the volume g.log.info("Logging volume info and Status after expanding volume") @@ -231,11 +231,11 @@ class SelfHealDaemonProcessTests(GlusterBaseClass): ret, pids = get_self_heal_daemon_pid(nodes) self.assertTrue(ret, ("Either No self heal daemon process found or " "more than One self heal daemon process found")) - g.log.info("Successfull in getting self-heal daemon process " + g.log.info("Successful in getting self-heal daemon process " "on nodes %s", nodes) glustershd_pids_after_expanding = pids - g.log.info("Self Heal Daemon Process ID's afetr expanding " + g.log.info("Self Heal Daemon Process ID's after expanding " "volume: %s", glustershd_pids_after_expanding) self.assertNotEqual(glustershd_pids, @@ -389,7 +389,7 @@ class SelfHealDaemonProcessTests(GlusterBaseClass): ret, pids = get_self_heal_daemon_pid(nodes) self.assertTrue(ret, ("Either No self heal daemon process found or " "more than One self heal daemon process found")) - g.log.info("Succesfull in getting self heal daemon pids") + g.log.info("Successful in getting self heal daemon pids") glustershd_pids = pids # get the bricks for the volume @@ -461,7 +461,7 @@ class SelfHealDaemonProcessTests(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", bricks_to_bring_offline) + "for %s successfully", bricks_to_bring_offline) # restart glusterd after brought down the brick g.log.info("Restart glusterd on all servers %s", nodes) @@ -922,7 +922,7 @@ class SelfHealDaemonProcessTestsWithHealing(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", bricks_to_bring_offline) + "for %s successfully", bricks_to_bring_offline) # get the bricks which are running g.log.info("getting the brick list which are online") @@ -996,9 +996,9 @@ class SelfHealDaemonProcessTestsWithHealing(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", bricks_to_bring_offline) + "for %s successfully", bricks_to_bring_offline) - # wait for 60 sec and brought up the brick agian + # wait for 60 sec and brought up the brick again time.sleep(60) g.log.info("Bring bricks: %s online", bricks_to_bring_offline) ret = bring_bricks_online(self.mnode, self.volname, @@ -1182,7 +1182,7 @@ class SelfHealDaemonProcessTestsWithMultipleVolumes(GlusterBaseClass): self.all_servers_info) self.assertTrue(ret, ("Failed to add bricks to " "volume %s " % volume)) - g.log.info("Add brick successfull") + g.log.info("Add brick successful") # Log Volume Info and Status after expanding the volume g.log.info("Logging volume info and Status after " diff --git a/tests/functional/afr/test_client_side_quorum.py b/tests/functional/afr/test_client_side_quorum.py index 0432a13ab..1dcb96ef5 100644 --- a/tests/functional/afr/test_client_side_quorum.py +++ b/tests/functional/afr/test_client_side_quorum.py @@ -114,7 +114,7 @@ class ClientSideQuorumTests(GlusterBaseClass): ret = set_volume_options(self.mnode, self.volname, options) self.assertTrue(ret, ("Unable to set volume option %s for" "volume %s" % (options, self.volname))) - g.log.info("Sucessfully set %s for volume %s", options, self.volname) + g.log.info("Successfully set %s for volume %s", options, self.volname) # write files on all mounts g.log.info("Starting IO on all mounts...") @@ -155,7 +155,7 @@ class ClientSideQuorumTests(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", bricks_to_bring_offline) + "for %s successfully", bricks_to_bring_offline) # create 2 files named newfile0.txt and newfile1.txt g.log.info("Start creating 2 files on all mounts...") @@ -172,7 +172,7 @@ class ClientSideQuorumTests(GlusterBaseClass): g.log.info("Validating whether IO failed with read-only filesystem") ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs, self.mounts) - self.assertTrue(ret, ("Unexpected error and IO successfull" + self.assertTrue(ret, ("Unexpected error and IO successful" " on read-only filesystem")) g.log.info("EXPECTED: Read-only file system in IO while creating file") @@ -190,7 +190,7 @@ class ClientSideQuorumTests(GlusterBaseClass): g.log.info("Validating whether IO failed with read-only filesystem") ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs, self.mounts) - self.assertTrue(ret, ("Unexpected error and IO successfull" + self.assertTrue(ret, ("Unexpected error and IO successful" " on read-only filesystem")) g.log.info("EXPECTED: Read-only file system in IO while" " creating directory") @@ -282,7 +282,7 @@ class ClientSideQuorumTests(GlusterBaseClass): ret, _, err = g.run(mount_obj.client_system, cmd) self.assertFalse(ret, ("Unexpected error and stat on file fails" " on read-only filesystem")) - g.log.info("stat on file is successfull on read-only filesystem") + g.log.info("stat on file is successful on read-only filesystem") # stat on dir g.log.info("stat on directory on all mounts") @@ -292,7 +292,7 @@ class ClientSideQuorumTests(GlusterBaseClass): ret, _, err = g.run(mount_obj.client_system, cmd) self.assertFalse(ret, ("Unexpected error and stat on directory" " fails on read-only filesystem")) - g.log.info("stat on dir is successfull on read-only filesystem") + g.log.info("stat on dir is successful on read-only filesystem") # ls on mount point g.log.info("ls on mount point on all mounts") @@ -302,7 +302,7 @@ class ClientSideQuorumTests(GlusterBaseClass): ret, _, err = g.run(mount_obj.client_system, cmd) self.assertFalse(ret, ("Unexpected error and listing file fails" " on read-only filesystem")) - g.log.info("listing files is successfull on read-only filesystem") + g.log.info("listing files is successful on read-only filesystem") def test_client_side_quorum_with_fixed_validate_max_bricks(self): """ @@ -402,7 +402,7 @@ class ClientSideQuorumTests(GlusterBaseClass): self.assertEqual(option_dict['cluster.quorum-count'], '(null)', ("Default value for %s is not null" " for volume %s" % (option, self.volname))) - g.log.info("Succesfull in getting %s for the volume %s", + g.log.info("Successful in getting %s for the volume %s", option, self.volname) # set cluster.quorum-type to fixed and cluster.quorum-count to 1 @@ -452,7 +452,7 @@ class ClientSideQuorumTests(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", bricks_to_bring_offline) + "for %s successfully", bricks_to_bring_offline) # create files g.log.info("Starting IO on all mounts...") @@ -609,7 +609,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass): ret = set_volume_options(self.mnode, self.volname, options) self.assertTrue(ret, ("Unable to set volume option %s for" "volume %s" % (options, self.volname))) - g.log.info("Sucessfully set %s for volume %s", options, self.volname) + g.log.info("Successfully set %s for volume %s", options, self.volname) # Start IO on mounts g.log.info("Starting IO on all mounts...") @@ -991,7 +991,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass): g.log.info("Validating whether IO failed with Read Only File System") ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs, self.mounts) - self.assertTrue(ret, ("Unexpected Error and IO successfull" + self.assertTrue(ret, ("Unexpected Error and IO successful" " on Read-Only File System")) g.log.info("EXPECTED Read-only file system in IO while creating file") @@ -1062,7 +1062,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", subvolumes_second_brick_list) + "for %s successfully", subvolumes_second_brick_list) # start I/0 ( write and read ) - read must pass, write will fail g.log.info("Start creating files on all mounts...") @@ -1079,7 +1079,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass): g.log.info("Validating whether IO failed with Read Only File System") ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs, self.mounts) - self.assertTrue(ret, ("Unexpected Error and IO successfull" + self.assertTrue(ret, ("Unexpected Error and IO successful" " on Read-Only File System")) g.log.info("EXPECTED Read-only file system in IO while creating file") @@ -1203,7 +1203,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", subvolumes_first_brick_list) + "for %s successfully", subvolumes_first_brick_list) # start I/0 ( write and read ) - read must pass, write will fail g.log.info("Start creating files on all mounts...") @@ -1220,7 +1220,7 @@ class ClientSideQuorumCross2Tests(GlusterBaseClass): g.log.info("Validating whether IO failed with Read Only File System") ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs, self.mounts) - self.assertTrue(ret, ("Unexpected Error and IO successfull" + self.assertTrue(ret, ("Unexpected Error and IO successful" " on Read-Only File System")) g.log.info("EXPECTED Read-only file system in IO while creating file") @@ -1508,7 +1508,7 @@ class ClientSideQuorumTestsMultipleVols(GlusterBaseClass): ret = set_volume_options(self.mnode, vol_name, options) self.assertTrue(ret, ("Unable to set volume option %s for " "volume %s" % (options, vol_name))) - g.log.info("Sucessfully set %s for volume %s", options, vol_name) + g.log.info("Successfully set %s for volume %s", options, vol_name) # check is options are set correctly volume_list = get_volume_list(self.mnode) @@ -1612,7 +1612,7 @@ class ClientSideQuorumTestsMultipleVols(GlusterBaseClass): g.log.info("Validating if IO failed with read-only filesystem") ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs, self.mounts) - self.assertTrue(ret, ("Unexpected error and IO successfull" + self.assertTrue(ret, ("Unexpected error and IO successful" " on read-only filesystem")) g.log.info("EXPECTED: " "Read-only file system in IO while creating file") @@ -1803,7 +1803,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", brick_to_bring_offline1) + "for %s successfully", brick_to_bring_offline1) offline_brick1_from_replicasets.append(brick_to_bring_offline1) # start I/0 ( write and read ) - must succeed @@ -1853,7 +1853,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", brick_to_bring_offline2) + "for %s successfully", brick_to_bring_offline2) offline_brick2_from_replicasets.append(brick_to_bring_offline2) # start I/0 ( write and read ) - must succeed @@ -1957,7 +1957,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass): g.log.info("Validating whether IO failed with Read Only File System") ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs, self.mounts) - self.assertTrue(ret, ("Unexpected Error and IO successfull" + self.assertTrue(ret, ("Unexpected Error and IO successful" " on Read-Only File System")) g.log.info("EXPECTED Read-only file system in IO while creating file") @@ -2120,7 +2120,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", bricks_to_bring_offline) + "for %s successfully", bricks_to_bring_offline) # start I/0 ( write and read ) - read must pass, write will fail g.log.info("Start creating files on all mounts...") @@ -2137,7 +2137,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass): g.log.info("Validating whether IO failed with Read-only file system") ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs, self.mounts) - self.assertTrue(ret, ("Unexpected error and IO successfull" + self.assertTrue(ret, ("Unexpected error and IO successful" " on Read-only file system")) g.log.info("EXPECTED: Read-only file system in IO while creating file") @@ -2181,7 +2181,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass): g.log.info("Validating whether IO failed with Read-only file system") ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs, self.mounts) - self.assertTrue(ret, ("Unexpected error and IO successfull" + self.assertTrue(ret, ("Unexpected error and IO successful" " on Read-only file system")) g.log.info("EXPECTED: Read-only file system in IO while creating file") @@ -2225,7 +2225,7 @@ class ClientSideQuorumTestsWithSingleVolumeCross3(GlusterBaseClass): g.log.info("Validating whether IO failed with Read-only file system") ret, _ = is_io_procs_fail_with_rofs(self, all_mounts_procs, self.mounts) - self.assertTrue(ret, ("Unexpected error and IO successfull" + self.assertTrue(ret, ("Unexpected error and IO successful" " on Read-only file system")) g.log.info("EXPECTED: Read-only file system in IO while creating file") diff --git a/tests/functional/afr/test_conservative_merge_of_files_heal_command.py b/tests/functional/afr/test_conservative_merge_of_files_heal_command.py index 3a6c62069..06514b972 100644 --- a/tests/functional/afr/test_conservative_merge_of_files_heal_command.py +++ b/tests/functional/afr/test_conservative_merge_of_files_heal_command.py @@ -139,7 +139,7 @@ class VerifySelfHealTriggersHealCommand(GlusterBaseClass): ret = set_volume_options(self.mnode, self.volname, options) self.assertTrue(ret, ("Unable to set volume option %s for" "volume %s" % (options, self.volname))) - g.log.info("Sucessfully set %s for volume %s", options, self.volname) + g.log.info("Successfully set %s for volume %s", options, self.volname) # Bring brick 0 offline g.log.info('Bringing bricks %s offline', bricks_list[0]) diff --git a/tests/functional/afr/test_gfid_assignment_on_dist_rep_vol.py b/tests/functional/afr/test_gfid_assignment_on_dist_rep_vol.py index e815fa0b6..0015f079e 100644 --- a/tests/functional/afr/test_gfid_assignment_on_dist_rep_vol.py +++ b/tests/functional/afr/test_gfid_assignment_on_dist_rep_vol.py @@ -112,7 +112,7 @@ class AssignGfidsOnAllSubvols(GlusterBaseClass): # Verify gfids are same on all the bricks self.verify_gfid("dir1") - # Creat a new directory on all the bricks directly + # Create a new directory on all the bricks directly bricks_list = get_all_bricks(self.mnode, self.volname) for brick in bricks_list: brick_node, brick_path = brick.split(":") diff --git a/tests/functional/afr/test_gfid_heal.py b/tests/functional/afr/test_gfid_heal.py index 76d423e2d..589a420a0 100644 --- a/tests/functional/afr/test_gfid_heal.py +++ b/tests/functional/afr/test_gfid_heal.py @@ -129,7 +129,7 @@ class HealGfidTest(GlusterBaseClass): - Create a 1x3 volume and fuse mount it. - Create 1 directory with 1 file inside it directly on each brick. - Access the directories from the mount. - - Launch heals ans verify that the heals are over. + - Launch heals and verify that the heals are over. - Verify that the files and directories have gfid assigned. """ # pylint: disable=too-many-statements diff --git a/tests/functional/afr/test_gfid_split_brain_resolution.py b/tests/functional/afr/test_gfid_split_brain_resolution.py index 0d6b0e220..a73ee407d 100644 --- a/tests/functional/afr/test_gfid_split_brain_resolution.py +++ b/tests/functional/afr/test_gfid_split_brain_resolution.py @@ -104,7 +104,7 @@ class TestSelfHeal(GlusterBaseClass): self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) g.log.info("Brought down the brick process " - "for %s succesfully", brick_list) + "for %s successfully", brick_list) ret = are_bricks_offline(self.mnode, self.volname, brick_list) self.assertTrue(ret, 'Bricks %s are not offline' % brick_list) @@ -157,7 +157,7 @@ class TestSelfHeal(GlusterBaseClass): ret = set_volume_options(self.mnode, self.volname, options) self.assertTrue(ret, ("Unable to set volume option %s for " "volume %s" % (options, self.volname))) - g.log.info("Sucessfully set %s for volume %s", options, self.volname) + g.log.info("Successfully set %s for volume %s", options, self.volname) # Create dir inside which I/O will be performed. ret = mkdir(self.mounts[0].client_system, "%s/test_gfid_split_brain" diff --git a/tests/functional/afr/test_quota_limit_entry_heal.py b/tests/functional/afr/test_quota_limit_entry_heal.py index 56388c157..033d326f4 100644 --- a/tests/functional/afr/test_quota_limit_entry_heal.py +++ b/tests/functional/afr/test_quota_limit_entry_heal.py @@ -130,7 +130,7 @@ class QuotaEntrySelfHealTest(GlusterBaseClass): ret = set_volume_options(self.mnode, self.volname, options) self.assertTrue(ret, ("Unable to set volume option %s for " "volume %s" % (options, self.volname))) - g.log.info("Sucessfully set %s for volume %s", options, self.volname) + g.log.info("Successfully set %s for volume %s", options, self.volname) # Create directory on mount ret = mkdir(self.mounts[0].client_system, "%s/dir" diff --git a/tests/functional/afr/test_self_heal_when_dir_quota_exceeded.py b/tests/functional/afr/test_self_heal_when_dir_quota_exceeded.py index 92568baa9..4648c0f68 100644 --- a/tests/functional/afr/test_self_heal_when_dir_quota_exceeded.py +++ b/tests/functional/afr/test_self_heal_when_dir_quota_exceeded.py @@ -101,7 +101,7 @@ class HealFilesWhenDirQuotaExceeded(GlusterBaseClass): path)) ret = g.run(mount_object.client_system, cmd) self.assertTrue(ret, "Failed to create directory on mountpoint") - g.log.info("Directory created succesfully on mountpoint") + g.log.info("Directory created successfully on mountpoint") # Enable Quota g.log.info("Enabling quota on the volume %s", self.volname) @@ -142,7 +142,7 @@ class HealFilesWhenDirQuotaExceeded(GlusterBaseClass): "count=20; done" % (mount_object.mountpoint, path)) ret, _, _ = g.run(mount_object.client_system, cmd) self.assertEqual(ret, 0, ("Failed to create files on %s", path)) - g.log.info("Files created succesfully on mountpoint") + g.log.info("Files created successfully on mountpoint") bricks_list = get_all_bricks(self.mnode, self.volname) diff --git a/tests/functional/afr/test_self_heal_with_quota_object_limit.py b/tests/functional/afr/test_self_heal_with_quota_object_limit.py index 6fe45f7b5..ff308c3f6 100644 --- a/tests/functional/afr/test_self_heal_with_quota_object_limit.py +++ b/tests/functional/afr/test_self_heal_with_quota_object_limit.py @@ -101,7 +101,7 @@ class HealFilesWhenQuotaObjectLimitExceeded(GlusterBaseClass): path)) ret = g.run(mount_object.client_system, cmd) self.assertTrue(ret, "Failed to create directory on mountpoint") - g.log.info("Directory created succesfully on mountpoint") + g.log.info("Directory created successfully on mountpoint") # Enable Quota g.log.info("Enabling quota on the volume %s", self.volname) @@ -141,7 +141,7 @@ class HealFilesWhenQuotaObjectLimitExceeded(GlusterBaseClass): % (self.script_upload_path, mount_object.mountpoint, path)) ret, _, _ = g.run(mount_object.client_system, cmd) self.assertEqual(ret, 0, ("Failed to create files on %s", path)) - g.log.info("Files created succesfully on mountpoint") + g.log.info("Files created successfully on mountpoint") bricks_list = get_all_bricks(self.mnode, self.volname) diff --git a/tests/functional/afr/test_write_io_mount_point_resumed_quorum_restored.py b/tests/functional/afr/test_write_io_mount_point_resumed_quorum_restored.py index 96d5d7864..18125933a 100755 --- a/tests/functional/afr/test_write_io_mount_point_resumed_quorum_restored.py +++ b/tests/functional/afr/test_write_io_mount_point_resumed_quorum_restored.py @@ -142,7 +142,7 @@ class ClientSideQuorumRestored(GlusterBaseClass): ret = set_volume_options(self.mnode, self.volname, options) self.assertTrue(ret, ("Unable to set volume option %s for" "volume %s" % (options, self.volname))) - g.log.info("Sucessfully set %s for volume %s", + g.log.info("Successfully set %s for volume %s", options, self.volname) # Creating files on client side diff --git a/tests/functional/arbiter/brick_cases/test_brickcases.py b/tests/functional/arbiter/brick_cases/test_brickcases.py index 9cce9af03..a5e33f018 100755 --- a/tests/functional/arbiter/brick_cases/test_brickcases.py +++ b/tests/functional/arbiter/brick_cases/test_brickcases.py @@ -177,7 +177,7 @@ class GlusterArbiterVolumeTypeChangeClass(GlusterBaseClass): self.all_servers_info, replica_count=1, arbiter_count=1) self.assertTrue(ret, ("Failed to expand the volume %s", self.volname)) - g.log.info("Changing volume to arbiter volume is successfull %s", + g.log.info("Changing volume to arbiter volume is successful %s", self.volname) # Log Volume Info and Status after changing the volume type from diff --git a/tests/functional/arbiter/brick_cases/test_replica3_to_arbiter.py b/tests/functional/arbiter/brick_cases/test_replica3_to_arbiter.py index 4fe2a8ba8..8d4130d1f 100644 --- a/tests/functional/arbiter/brick_cases/test_replica3_to_arbiter.py +++ b/tests/functional/arbiter/brick_cases/test_replica3_to_arbiter.py @@ -124,14 +124,14 @@ class GlusterArbiterVolumeTypeClass(GlusterBaseClass): self.volname)) g.log.info("Volume %s : All process are online", self.volname) - # Adding the bricks to amke arbiter brick + # Adding the bricks to make arbiter brick g.log.info("Adding bricks to convert to Arbiter Volume") replica_arbiter = {'replica_count': 3, 'arbiter_count': 1} ret = expand_volume(self.mnode, self.volname, self.servers[2:], self.all_servers_info, add_to_hot_tier=False, **replica_arbiter) self.assertTrue(ret, ("Failed to expand the volume %s", self.volname)) - g.log.info("Changing volume to arbiter volume is successfull %s", + g.log.info("Changing volume to arbiter volume is successful %s", self.volname) # Log Volume Info and Status after expanding the volume diff --git a/tests/functional/arbiter/test_mount_point_while_deleting_files.py b/tests/functional/arbiter/test_mount_point_while_deleting_files.py index c8e4804ee..1bbdf279f 100755 --- a/tests/functional/arbiter/test_mount_point_while_deleting_files.py +++ b/tests/functional/arbiter/test_mount_point_while_deleting_files.py @@ -229,7 +229,7 @@ class VolumeSetDataSelfHealTests(GlusterBaseClass): ret = bring_bricks_offline(volname, bricks_to_bring_offline) self.assertTrue(ret, ("Failed to bring down the bricks. Please " "check the log file for more details.")) - g.log.info("Brought down the brick process for %s succesfully", + g.log.info("Brought down the brick process for %s successfully", bricks_to_bring_offline) # delete files on all mounts diff --git a/tests/functional/authentication/test_fusereject.py b/tests/functional/authentication/test_fusereject.py index 44410a5d9..5b403f4ed 100644 --- a/tests/functional/authentication/test_fusereject.py +++ b/tests/functional/authentication/test_fusereject.py @@ -129,7 +129,7 @@ class AuthRejectVol(GlusterBaseClass): if (ret == 0) & (not out): g.log.error("Mount executed successfully due to bug 1586036") elif (ret == 1) & (not out): - g.log.info("Expected:Mounting has failed sucessfully") + g.log.info("Expected:Mounting has failed successfully") else: raise ExecutionError("Unexpected Mounting of Volume %s successful" % self.volname) diff --git a/tests/functional/authentication/test_vol_auth.py b/tests/functional/authentication/test_vol_auth.py index e8180d1e4..bb3da6f96 100644 --- a/tests/functional/authentication/test_vol_auth.py +++ b/tests/functional/authentication/test_vol_auth.py @@ -118,7 +118,7 @@ class AuthRejectVol(GlusterBaseClass): if (ret == 0) & (not out): g.log.error("Mount executed successfully due to bug 1586036") elif (ret == 1) & (not out): - g.log.info("Expected:Mounting has failed sucessfully") + g.log.info("Expected:Mounting has failed successfully") else: raise ExecutionError("Unexpected Mounting of Volume %s" "successful" % self.volname) diff --git a/tests/functional/bvt/test_cvt.py b/tests/functional/bvt/test_cvt.py index f1250b764..05578b400 100644 --- a/tests/functional/bvt/test_cvt.py +++ b/tests/functional/bvt/test_cvt.py @@ -663,7 +663,7 @@ class TestGlusterReplaceBrickSanity(GlusterBasicFeaturesSanityBaseClass): # This test is disabled on nfs because of bug 1473668. A patch to apply the -# workaround mentiond on the bug could not make this test green either. +# workaround mentioned on the bug could not make this test green either. @runs_on([['replicated', 'distributed-replicated', 'dispersed', 'distributed-dispersed'], ['glusterfs', 'cifs']]) diff --git a/tests/functional/dht/test_dht_hash_value.py b/tests/functional/dht/test_dht_hash_value.py index fd83576cf..c5f08dcfa 100644 --- a/tests/functional/dht/test_dht_hash_value.py +++ b/tests/functional/dht/test_dht_hash_value.py @@ -187,12 +187,12 @@ class TestDHTHashValue(GlusterBaseClass): mountpoint) attributes = get_fattr_list(client_host, mountpoint) self.assertFalse('trusted.gfid' in attributes, - "Expected: Mount point should't display xattr:" + "Expected: Mount point shouldn't display xattr:" "{xattr}. Actual: xattrs {xattr} is " "presented on mount point". format(xattr='trusted.gfid')) self.assertFalse('trusted.glusterfs.dht' in attributes, - "Expected: Mount point should't display xattr:" + "Expected: Mount point shouldn't display xattr:" "{xattr}. Actual: xattrs {xattr} is " "presented on mount point". format(xattr='trusted.glusterfs.dht')) diff --git a/tests/functional/dht/test_stop_glusterd_while_rebalance_in_progress.py b/tests/functional/dht/test_stop_glusterd_while_rebalance_in_progress.py index bbe4eb52e..c9b60103a 100644 --- a/tests/functional/dht/test_stop_glusterd_while_rebalance_in_progress.py +++ b/tests/functional/dht/test_stop_glusterd_while_rebalance_in_progress.py @@ -134,7 +134,7 @@ class RebalanceValidation(GlusterBaseClass): self.volname)) g.log.info("Volume %s: Started rebalance", self.volname) - # Wait for atleast one file to be lookedup/scanned on the nodes + # Wait for at least one file to be lookedup/scanned on the nodes status_info = get_rebalance_status(self.mnode, self.volname) count = 0 while count < 100: @@ -146,7 +146,7 @@ class RebalanceValidation(GlusterBaseClass): lookups_start_count += 1 sleep(5) if lookups_start_count == len(self.servers): - g.log.info("Volume %s: Atleast one file is lookedup/scanned " + g.log.info("Volume %s: At least one file is lookedup/scanned " "on all nodes", self.volname) break count += 1 diff --git a/tests/functional/disperse/test_ec_validate_volume_creation.py b/tests/functional/disperse/test_ec_validate_volume_creation.py index 99637f20f..d9e490caf 100644 --- a/tests/functional/disperse/test_ec_validate_volume_creation.py +++ b/tests/functional/disperse/test_ec_validate_volume_creation.py @@ -96,7 +96,7 @@ class EcValidateVolumeCreate(GlusterBaseClass): # Setup Volume and Mount Volume g.log.info("Starting to Setup Volume and Mount Volume") ret = self.setup_volume_and_mount_volume(mounts=self.mounts) - self.assertFalse(ret, ("Volume Setup and Mount succeded" + self.assertFalse(ret, ("Volume Setup and Mount succeeded" " for volume %s", self.volname)) g.log.info("Successfully verified invalid input parameters") @@ -108,8 +108,8 @@ class EcValidateVolumeCreate(GlusterBaseClass): # Setup Volume and Mount Volume g.log.info("Starting to Setup Volume and Mount Volume") ret = self.setup_volume_and_mount_volume(mounts=self.mounts) - self.assertFalse(ret, ("Volume Setup and Mount succeded for volume %s", - self.volname)) + self.assertFalse(ret, ("Volume Setup and Mount succeeded for volume " + "%s", self.volname)) g.log.info("Successfully verified invalid input parameters") def test_invalid_usecase_three(self): @@ -120,8 +120,8 @@ class EcValidateVolumeCreate(GlusterBaseClass): # Setup Volume and Mount Volume g.log.info("Starting to Setup Volume and Mount Volume") ret = self.setup_volume_and_mount_volume(mounts=self.mounts) - self.assertFalse(ret, ("Volume Setup and Mount succeded for volume %s", - self.volname)) + self.assertFalse(ret, ("Volume Setup and Mount succeeded for volume " + "%s", self.volname)) g.log.info("Successfully verified invalid input parameters") def test_invalid_usecase_four(self): @@ -132,6 +132,6 @@ class EcValidateVolumeCreate(GlusterBaseClass): # Setup Volume and Mount Volume g.log.info("Starting to Setup Volume and Mount Volume") ret = self.setup_volume_and_mount_volume(mounts=self.mounts) - self.assertFalse(ret, ("Volume Setup and Mount succeded for volume %s", - self.volname)) + self.assertFalse(ret, ("Volume Setup and Mount succeeded for volume " + "%s", self.volname)) g.log.info("Successfully verified invalid input parameters") diff --git a/tests/functional/fuse_subdir/test_fusesubdir_with_quota_objects.py b/tests/functional/fuse_subdir/test_fusesubdir_with_quota_objects.py index 10dc0ddb3..01ca474f2 100644 --- a/tests/functional/fuse_subdir/test_fusesubdir_with_quota_objects.py +++ b/tests/functional/fuse_subdir/test_fusesubdir_with_quota_objects.py @@ -171,7 +171,7 @@ class SubdirWithQuotaObject(GlusterBaseClass): % (mount_object.mountpoint)) ret, _, _ = g.run(mount_object.client_system, cmd) self.assertEqual(ret, 0, "Failed to create files on mountpoint") - g.log.info("Files created succesfully on mountpoint") + g.log.info("Files created successfully on mountpoint") # Fetch Quota List object on the volume g.log.info("Get Quota list on the volume %s", @@ -215,7 +215,7 @@ class SubdirWithQuotaObject(GlusterBaseClass): % (mount_object.mountpoint, i), parents=True) self.assertTrue(ret, "Failed to create directories" "on mountpoint") - g.log.info("Directories created succesfully on mountpoint") + g.log.info("Directories created successfully on mountpoint") # Get Quota List on the volume g.log.info("Get Quota list on the volume %s", @@ -240,7 +240,7 @@ class SubdirWithQuotaObject(GlusterBaseClass): else: self.assertTrue(ret, "Directory creation got failed" "on volume") - g.log.info("Direction creation successfull on volume") + g.log.info("Direction creation successful on volume") def tearDown(self): """ diff --git a/tests/functional/fuse_subdir/test_fusesubdir_with_removebrick.py b/tests/functional/fuse_subdir/test_fusesubdir_with_removebrick.py index 8e060271d..30ac9eefb 100644 --- a/tests/functional/fuse_subdir/test_fusesubdir_with_removebrick.py +++ b/tests/functional/fuse_subdir/test_fusesubdir_with_removebrick.py @@ -220,7 +220,7 @@ class SubdirWithRemoveBrick(GlusterBaseClass): ret, _, _ = umount_volume(client, self.mpoint, self.mount_type) if ret != 0: - raise ExecutionError("Unmounting the moint point %s failed" + raise ExecutionError("Unmounting the mount point %s failed" % self.mpoint) g.log.info("Unmount Volume Successful") cmd = ("rm -rf %s") % self.mpoint diff --git a/tests/functional/fuse_subdir/test_leveldownsubdir_with_multiclient.py b/tests/functional/fuse_subdir/test_leveldownsubdir_with_multiclient.py index f7115cedd..eef05896c 100644 --- a/tests/functional/fuse_subdir/test_leveldownsubdir_with_multiclient.py +++ b/tests/functional/fuse_subdir/test_leveldownsubdir_with_multiclient.py @@ -60,7 +60,7 @@ class SubdirLevelDownDirMapping(GlusterBaseClass): ret, _, _ = g.run(self.mounts[0].client_system, cmd) self.assertEqual(ret, 0, "Failed to create Nested directories" "on mountpoint") - g.log.info("Nested Directories created succesfully on mountpoint") + g.log.info("Nested Directories created successfully on mountpoint") # unmount volume ret = self.unmount_volume(self.mounts) @@ -134,7 +134,7 @@ class SubdirLevelDownDirMapping(GlusterBaseClass): ret, _, _ = umount_volume(client, self.mpoint, self.mount_type) if ret == 1: - raise ExecutionError("Unmounting the moint point %s failed" + raise ExecutionError("Unmounting the mount point %s failed" % self.mpoint) g.log.info("Unmount Volume Successful") cmd = ("rm -rf %s") % self.mpoint diff --git a/tests/functional/fuse_subdir/test_multisubdir_with_multiclient.py b/tests/functional/fuse_subdir/test_multisubdir_with_multiclient.py index deb59520d..ccd513e15 100644 --- a/tests/functional/fuse_subdir/test_multisubdir_with_multiclient.py +++ b/tests/functional/fuse_subdir/test_multisubdir_with_multiclient.py @@ -59,7 +59,7 @@ class MultipleDirMappingClient(GlusterBaseClass): self.mounts[0].mountpoint)) ret, _, _ = g.run(self.mounts[0].client_system, cmd) self.assertEqual(ret, 0, "Failed to create directories on mountpoint") - g.log.info("Directories created succesfully on mountpoint") + g.log.info("Directories created successfully on mountpoint") # unmount volume ret = self.unmount_volume(self.mounts) @@ -127,7 +127,7 @@ class MultipleDirMappingClient(GlusterBaseClass): ret, _, _ = umount_volume(client, self.mpoint, self.mount_type) if ret == 1: - raise ExecutionError("Unmounting the moint point %s failed" + raise ExecutionError("Unmounting the mount point %s failed" % self.mpoint) g.log.info("Unmount Volume Successful") cmd = ("rm -rf %s") % self.mpoint diff --git a/tests/functional/fuse_subdir/test_quota_limits_fusesubdir_.py b/tests/functional/fuse_subdir/test_quota_limits_fusesubdir_.py index 4c6e0cf4a..b8b9206e4 100644 --- a/tests/functional/fuse_subdir/test_quota_limits_fusesubdir_.py +++ b/tests/functional/fuse_subdir/test_quota_limits_fusesubdir_.py @@ -200,7 +200,7 @@ class FuseSubdirQuotaTest(GlusterBaseClass): "count=1;done" % (mount_object.mountpoint)) ret, _, _ = g.run(mount_object.client_system, cmd) self.assertEqual(ret, 0, "Failed to create files on mountpoint") - g.log.info("Files created succesfully on mountpoint") + g.log.info("Files created successfully on mountpoint") # Again Get Quota List on the volume @@ -254,7 +254,7 @@ class FuseSubdirQuotaTest(GlusterBaseClass): ret, _, _ = g.run(self.clients[1], cmd) self.assertEqual(ret, 0, ("Failed to create files on %s" % self.clients[1])) - g.log.info("Files created succesfully on %s:%s", + g.log.info("Files created successfully on %s:%s", self.clients[1], self.subdir_mounts[1].mountpoint) def tearDown(self): diff --git a/tests/functional/glusterd/test_brick_port_after_stop_glusterd_modify_volume.py b/tests/functional/glusterd/test_brick_port_after_stop_glusterd_modify_volume.py index da80f67f4..e44514aff 100644 --- a/tests/functional/glusterd/test_brick_port_after_stop_glusterd_modify_volume.py +++ b/tests/functional/glusterd/test_brick_port_after_stop_glusterd_modify_volume.py @@ -104,7 +104,7 @@ class TestBrickPortAfterModifyVolume(GlusterBaseClass): bricks_list, force=False) self.assertEqual(ret[0], 0, ("Unable" "to create volume %s" % self.volname)) - g.log.info("Volume created successfuly %s", self.volname) + g.log.info("Volume created successfully %s", self.volname) ret, _, _ = volume_start(self.mnode, self.volname) self.assertEqual(ret, 0, ("Failed to start the " diff --git a/tests/functional/glusterd/test_brick_status_when_quorum_not_met.py b/tests/functional/glusterd/test_brick_status_when_quorum_not_met.py index 6cb3ee075..cc77b3ea5 100644 --- a/tests/functional/glusterd/test_brick_status_when_quorum_not_met.py +++ b/tests/functional/glusterd/test_brick_status_when_quorum_not_met.py @@ -132,7 +132,7 @@ class TestBrickStatusWhenQuorumNotMet(GlusterBaseClass): # Verfiying node count in volume status after glusterd # started on servers, Its not possible to check the brick status - # immediately after glusterd start, thats why verifying that all + # immediately after glusterd start, that's why verifying that all # glusterd started nodes available in gluster volume status or not count = 0 while count < 50: diff --git a/tests/functional/glusterd/test_concurrent_set.py b/tests/functional/glusterd/test_concurrent_set.py index 7c753ea78..4b432b784 100644 --- a/tests/functional/glusterd/test_concurrent_set.py +++ b/tests/functional/glusterd/test_concurrent_set.py @@ -60,7 +60,7 @@ class TestConcurrentSet(GlusterBaseClass): self.brick_list, force=False) self.assertEqual(ret[0], 0, ("Unable" "to create volume %s" % self.volname)) - g.log.info("Volume created successfuly %s", self.volname) + g.log.info("Volume created successfully %s", self.volname) # Create a volume self.volname = "second-vol" @@ -72,7 +72,7 @@ class TestConcurrentSet(GlusterBaseClass): self.brick_list, force=False) self.assertEqual(ret[0], 0, ("Unable" "to create volume %s" % self.volname)) - g.log.info("Volume created successfuly %s", self.volname) + g.log.info("Volume created successfully %s", self.volname) cmd1 = ("for i in `seq 1 100`; do gluster volume set first-vol " "read-ahead on; done") diff --git a/tests/functional/glusterd/test_create_vol_with_used_bricks.py b/tests/functional/glusterd/test_create_vol_with_used_bricks.py index 940d7a451..1be34f734 100644 --- a/tests/functional/glusterd/test_create_vol_with_used_bricks.py +++ b/tests/functional/glusterd/test_create_vol_with_used_bricks.py @@ -111,7 +111,7 @@ class TestCreateVolWithUsedBricks(GlusterBaseClass): # Mounting volume ret = self.mount_volume(self.mounts) self.assertTrue(ret, "Volume mount failed for %s" % self.volname) - g.log.info("Volume mounted sucessfully : %s", self.volname) + g.log.info("Volume mounted successfully : %s", self.volname) # run IOs g.log.info("Starting IO on all mounts...") diff --git a/tests/functional/glusterd/test_nfs_quorum.py b/tests/functional/glusterd/test_nfs_quorum.py index ced5b719f..62d2ce24a 100644 --- a/tests/functional/glusterd/test_nfs_quorum.py +++ b/tests/functional/glusterd/test_nfs_quorum.py @@ -82,7 +82,7 @@ class TestNfsMountAndServerQuorumSettings(GlusterBaseClass): # Mounting a NFS volume ret = self.mount_volume(self.mounts) self.assertTrue(ret, "NFS volume mount failed for %s" % self.volname) - g.log.info("Volume mounted sucessfully : %s", self.volname) + g.log.info("Volume mounted successfully : %s", self.volname) # unmounting NFS Volume ret = self.unmount_volume(self.mounts) diff --git a/tests/functional/glusterd/test_peer_detach.py b/tests/functional/glusterd/test_peer_detach.py index 2bae76d2a..633036927 100644 --- a/tests/functional/glusterd/test_peer_detach.py +++ b/tests/functional/glusterd/test_peer_detach.py @@ -68,7 +68,7 @@ class PeerDetachVerification(GlusterBaseClass): # Assigning non existing host to variable self.non_exist_host = '256.256.256.256' - # Assigning invalid ip to vaiable + # Assigning invalid ip to variable self.invalid_ip = '10.11.a' # Peer detach to specified server diff --git a/tests/functional/glusterd/test_peer_probe_while_snapd_running.py b/tests/functional/glusterd/test_peer_probe_while_snapd_running.py index c35b3eaeb..aff015638 100644 --- a/tests/functional/glusterd/test_peer_probe_while_snapd_running.py +++ b/tests/functional/glusterd/test_peer_probe_while_snapd_running.py @@ -97,7 +97,7 @@ class TestPeerProbeWhileSnapdRunning(GlusterBaseClass): # Checking snapd running or not ret = is_snapd_running(self.mnode, self.volname) - self.assertTrue(ret, "Snapd not runnig for volume %s" % self.volname) + self.assertTrue(ret, "Snapd not running for volume %s" % self.volname) g.log.info("snapd running for volume %s", self.volname) # Probing new node diff --git a/tests/functional/glusterd/test_probe_glusterd.py b/tests/functional/glusterd/test_probe_glusterd.py index d14991dbd..54b99eec2 100644 --- a/tests/functional/glusterd/test_probe_glusterd.py +++ b/tests/functional/glusterd/test_probe_glusterd.py @@ -62,7 +62,7 @@ class PeerProbeInvalidIpNonExistingHost(GlusterBaseClass): # Assigning non existing ip to variable self.non_exist_ip = '256.256.256.256' - # Assigning invalid ip to vaiable + # Assigning invalid ip to variable self.invalid_ip = '10.11.a' # Assigning non existing host to variable diff --git a/tests/functional/glusterd/test_probe_hostname.py b/tests/functional/glusterd/test_probe_hostname.py index 55476edc0..6e7d87b53 100644 --- a/tests/functional/glusterd/test_probe_hostname.py +++ b/tests/functional/glusterd/test_probe_hostname.py @@ -103,14 +103,14 @@ class TestPeerProbe(GlusterBaseClass): self.brick_list, force=False) self.assertEqual(ret, 0, "Unable" "to create volume % s" % self.volname) - g.log.info("Volume created successfuly % s", self.volname) + g.log.info("Volume created successfully % s", self.volname) # Start a volume g.log.info("Start a volume") ret, _, _ = volume_start(self.mnode, self.volname) self.assertEqual(ret, 0, "Unable" "to start volume % s" % self.volname) - g.log.info("Volume started successfuly % s", self.volname) + g.log.info("Volume started successfully % s", self.volname) # Get volume info g.log.info("get volume info") @@ -127,7 +127,7 @@ class TestPeerProbe(GlusterBaseClass): ret, _, _ = volume_stop(self.mnode, self.volname) self.assertEqual(ret, 0, "Unable" "to stop volume % s" % self.volname) - g.log.info("Volume stopped successfuly % s", self.volname) + g.log.info("Volume stopped successfully % s", self.volname) # Create a volume self.volname = "test-vol-fqdn" @@ -154,14 +154,14 @@ class TestPeerProbe(GlusterBaseClass): my_brick_list, force=False) self.assertEqual(ret, 0, "Unable" "to create volume % s" % self.volname) - g.log.info("Volume created successfuly % s", self.volname) + g.log.info("Volume created successfully % s", self.volname) # Start a volume g.log.info("Start a volume") ret, _, _ = volume_start(self.mnode, self.volname) self.assertEqual(ret, 0, "Unable" "to start volume % s" % self.volname) - g.log.info("Volume started successfuly % s", self.volname) + g.log.info("Volume started successfully % s", self.volname) # Get volume info g.log.info("get volume info") @@ -178,4 +178,4 @@ class TestPeerProbe(GlusterBaseClass): ret, _, _ = volume_stop(self.mnode, self.volname) self.assertEqual(ret, 0, "Unable" "to stop volume % s" % self.volname) - g.log.info("Volume stopped successfuly % s", self.volname) + g.log.info("Volume stopped successfully % s", self.volname) diff --git a/tests/functional/glusterd/test_quorum_remove_brick.py b/tests/functional/glusterd/test_quorum_remove_brick.py index 6d5e45b11..4429d8231 100644 --- a/tests/functional/glusterd/test_quorum_remove_brick.py +++ b/tests/functional/glusterd/test_quorum_remove_brick.py @@ -120,7 +120,7 @@ class TestServerQuorumNotMet(GlusterBaseClass): % self.random_server) g.log.info("Glusterd stopped successfully on %s", self.random_server) - # Forming brick list for perfroming remove brick operation + # Forming brick list for performing remove brick operation remove_brick_list = form_bricks_list_to_remove_brick(self.mnode, self.volname) self.assertIsNotNone(remove_brick_list, "Failed to get brick list for " diff --git a/tests/functional/glusterd/test_rebalance_hang.py b/tests/functional/glusterd/test_rebalance_hang.py index d96a4043a..a826703c1 100644 --- a/tests/functional/glusterd/test_rebalance_hang.py +++ b/tests/functional/glusterd/test_rebalance_hang.py @@ -114,7 +114,7 @@ class TestRebalanceHang(GlusterBaseClass): bricks_list, force=False) self.assertEqual(ret, 0, ("Unable" "to create volume %s" % self.volname)) - g.log.info("Volume created successfuly %s", self.volname) + g.log.info("Volume created successfully %s", self.volname) ret, _, _ = volume_start(self.mnode, self.volname, False) self.assertEqual(ret, 0, ("Failed to start the " @@ -130,7 +130,7 @@ class TestRebalanceHang(GlusterBaseClass): mserver=self.mnode, mclient=self.mounts[0].client_system) self.assertEqual(ret, 0, ("Volume %s is not mounted") % self.volname) - g.log.info("Volume mounted sucessfully : %s", self.volname) + g.log.info("Volume mounted successfully : %s", self.volname) self.all_mounts_procs = [] # Creating files diff --git a/tests/functional/glusterd/test_rebalance_spurious.py b/tests/functional/glusterd/test_rebalance_spurious.py index 9b7318812..ad16a0039 100644 --- a/tests/functional/glusterd/test_rebalance_spurious.py +++ b/tests/functional/glusterd/test_rebalance_spurious.py @@ -114,7 +114,7 @@ class TestSpuriousRebalance(GlusterBaseClass): bricks_list, force=False) self.assertEqual(ret, 0, ("Unable" "to create volume %s" % self.volname)) - g.log.info("Volume created successfuly %s", self.volname) + g.log.info("Volume created successfully %s", self.volname) ret, _, _ = volume_start(self.mnode, self.volname, False) self.assertEqual(ret, 0, ("Failed to start the " @@ -130,7 +130,7 @@ class TestSpuriousRebalance(GlusterBaseClass): mserver=self.mnode, mclient=self.mounts[0].client_system) self.assertEqual(ret, 0, ("Volume %s is not mounted") % self.volname) - g.log.info("Volume mounted sucessfully : %s", self.volname) + g.log.info("Volume mounted successfully : %s", self.volname) remove_brick_list = [] remove_brick_list.append(bricks_list[2]) ret, _, _ = remove_brick(self.mnode, self.volname, remove_brick_list, diff --git a/tests/functional/glusterd/test_remove_brick_after_restart_glusterd.py b/tests/functional/glusterd/test_remove_brick_after_restart_glusterd.py index 217eae5dc..f84c7dba8 100644 --- a/tests/functional/glusterd/test_remove_brick_after_restart_glusterd.py +++ b/tests/functional/glusterd/test_remove_brick_after_restart_glusterd.py @@ -112,7 +112,7 @@ class TestRemoveBrickAfterRestartGlusterd(GlusterBaseClass): bricks_list, force=False, **kwargs) self.assertEqual(ret[0], 0, ("Unable" "to create volume %s" % self.volname)) - g.log.info("Volume created successfuly %s", self.volname) + g.log.info("Volume created successfully %s", self.volname) ret, _, _ = volume_start(self.mnode, self.volname, False) self.assertEqual(ret, 0, ("Failed to start the " @@ -128,7 +128,7 @@ class TestRemoveBrickAfterRestartGlusterd(GlusterBaseClass): mserver=self.mnode, mclient=self.mounts[0].client_system) self.assertEqual(ret, 0, ("Volume %s is not mounted") % self.volname) - g.log.info("Volume mounted sucessfully : %s", self.volname) + g.log.info("Volume mounted successfully : %s", self.volname) self.all_mounts_procs = [] # Creating files diff --git a/tests/functional/glusterd/test_volume_get.py b/tests/functional/glusterd/test_volume_get.py index 228b15209..5746d5d59 100644 --- a/tests/functional/glusterd/test_volume_get.py +++ b/tests/functional/glusterd/test_volume_get.py @@ -161,7 +161,7 @@ class TestVolumeGet(GlusterBaseClass): "existing volume %s with non existing option", self.volname) - # perfroming gluster volume get volname all + # performing gluster volume get volname all ret = get_volume_options(self.mnode, self.volname, "all") self.assertIsNotNone(ret, "gluster volume get %s all command " @@ -194,7 +194,7 @@ class TestVolumeGet(GlusterBaseClass): "performance.low-prio-threads value is not correct") g.log.info("performance.low-prio-threads value is correct") - # perfroming gluster volume get volname all + # performing gluster volume get volname all ret = get_volume_options(self.mnode, self.volname, "all") self.assertIsNotNone(ret, "gluster volume get %s all command " "failed" % self.volname) diff --git a/tests/functional/glusterd/test_volume_network_ping_timeout.py b/tests/functional/glusterd/test_volume_network_ping_timeout.py index b897c4a7a..7d72d8ab2 100644 --- a/tests/functional/glusterd/test_volume_network_ping_timeout.py +++ b/tests/functional/glusterd/test_volume_network_ping_timeout.py @@ -96,7 +96,7 @@ class CheckVolumeChecksumAfterChangingNetworkPingTimeOut(GlusterBaseClass): # Mounting volume as glusterfs ret = self.mount_volume(self.mounts) self.assertTrue(ret, "volume mount failed for %s" % self.volname) - g.log.info("Volume mounted sucessfully : %s", self.volname) + g.log.info("Volume mounted successfully : %s", self.volname) # Checking volume mounted or not ret = is_mounted(self.volname, self.mounts[0].mountpoint, self.mnode, diff --git a/tests/functional/glusterd/test_volume_reset.py b/tests/functional/glusterd/test_volume_reset.py index f61fdaaba..9720c3e92 100644 --- a/tests/functional/glusterd/test_volume_reset.py +++ b/tests/functional/glusterd/test_volume_reset.py @@ -124,10 +124,10 @@ class GlusterdVolumeReset(GlusterBaseClass): cmd = "gluster volume reset " + self.volname ret, _, _ = g.run(self.mnode, cmd) self.assertEqual(ret, 0, "volume reset failed for : %s" % self.volname) - g.log.info("volume resetted succefully :%s", self.volname) + g.log.info("Volume reset successfully :%s", self.volname) # After volume reset snap daemon will not be running, - # bitd and scrub deamons will be in running state. + # bitd and scrub daemons will be in running state. g.log.info("checking snapshot, scrub and bitrot daemons\ running or not after volume reset") for mnode in self.servers: @@ -157,7 +157,7 @@ class GlusterdVolumeReset(GlusterBaseClass): ret, _, _ = g.run(self.mnode, cmd) self.assertEqual(ret, 0, "volume reset fail\ for : %s" % self.volname) - g.log.info("Volume reset sucessfully with force option :%s", + g.log.info("Volume reset successfully with force option :%s", self.volname) # After volume reset bitd, snapd, scrub daemons will not be running, diff --git a/tests/functional/glusterd/test_volume_status.py b/tests/functional/glusterd/test_volume_status.py index ff5d5752f..109586f35 100644 --- a/tests/functional/glusterd/test_volume_status.py +++ b/tests/functional/glusterd/test_volume_status.py @@ -110,17 +110,17 @@ class VolumeStatusWhenIOInProgress(GlusterBaseClass): command on all clusters randomly. "gluster volume status volname inode" command should not get hang while IOs in progress. - Then check that IOs completed successfullly or not on mount point. + Then check that IOs completed successfully or not on mount point. Check that files in mount point listing properly or not. ''' # Mounting a volume ret = self.mount_volume(self.mounts) self.assertTrue(ret, "Volume mount failed for %s" % self.volname) - g.log.info("Volume mounted sucessfully : %s", self.volname) + g.log.info("Volume mounted successfully : %s", self.volname) - # After Mounting immediately writting IO's are failing some times, - # thats why keeping sleep for 10 secs + # After Mounting immediately writing IO's are failing some times, + # that's why keeping sleep for 10 secs sleep(10) # run IOs diff --git a/tests/functional/glusterd/test_volume_status_fd.py b/tests/functional/glusterd/test_volume_status_fd.py index 2765325c5..415c96de7 100644 --- a/tests/functional/glusterd/test_volume_status_fd.py +++ b/tests/functional/glusterd/test_volume_status_fd.py @@ -94,7 +94,7 @@ class VolumeStatusFdWhenIOInProgress(GlusterBaseClass): -> Mount the volume on 2 clients -> Run I/O's on mountpoint -> While I/O's are in progress - -> Perfrom gluster volume status fd repeatedly + -> Perform gluster volume status fd repeatedly -> List all files and dirs listed ''' diff --git a/tests/functional/nfs_ganesha/test_nfs_ganesha_sanity.py b/tests/functional/nfs_ganesha/test_nfs_ganesha_sanity.py index a00c22a5a..05caf4a43 100644 --- a/tests/functional/nfs_ganesha/test_nfs_ganesha_sanity.py +++ b/tests/functional/nfs_ganesha/test_nfs_ganesha_sanity.py @@ -75,7 +75,7 @@ class TestNfsGaneshaSanity(NfsGaneshaVolumeBaseClass): "tar xvf linux-2.6.1.tar.gz" % (mount_obj.mountpoint)) ret, out, err = g.run(mount_obj.client_system, cmd) if ret == 0: - g.log.info("Succesfully untared the tarball!") + g.log.info("Successfully untared the tarball!") else: g.log.error("ERROR ! Kernel untar errored out!") self.assertEqual(ret, 0, "Kernel untar failed!") diff --git a/tests/functional/nfs_ganesha/test_nfs_ganesha_volume_exports.py b/tests/functional/nfs_ganesha/test_nfs_ganesha_volume_exports.py index 8a8b28cc3..ead84dbc5 100644 --- a/tests/functional/nfs_ganesha/test_nfs_ganesha_volume_exports.py +++ b/tests/functional/nfs_ganesha/test_nfs_ganesha_volume_exports.py @@ -69,7 +69,7 @@ class TestNfsGaneshaVolumeExports(NfsGaneshaVolumeBaseClass): g.log.info("Testing nfs ganesha export after volume stop/start." "Count : %s", str(i)) - # Stoping volume + # Stopping volume ret = volume_stop(self.mnode, self.volname) self.assertTrue(ret, ("Failed to stop volume %s" % self.volname)) @@ -485,7 +485,7 @@ class TestNfsGaneshaSubDirExportsWithIO(NfsGaneshaIOBaseClass): export_file))) # Stop and start volume to take the modified export file to effect. - # Stoping volume + # Stopping volume ret = volume_stop(self.mnode, self.volname) self.assertTrue(ret, ("Failed to stop volume %s" % self.volname)) diff --git a/tests/functional/quota/list_path_values.py b/tests/functional/quota/list_path_values.py index 5c9bf81d4..292370473 100644 --- a/tests/functional/quota/list_path_values.py +++ b/tests/functional/quota/list_path_values.py @@ -112,7 +112,7 @@ class QuotaListPathValues(GlusterBaseClass): "count=1;done" % (mount_object.mountpoint)) ret, _, _ = g.run(mount_object.client_system, cmd) self.assertEqual(ret, 0, "Failed to create files on mountpoint") - g.log.info("Files created succesfully on mountpoint") + g.log.info("Files created successfully on mountpoint") # Get Quota list without specifying the path g.log.info("Get Quota list for the volume %s", self.volname) diff --git a/tests/functional/quota/test_limit_usage_deep_dir.py b/tests/functional/quota/test_limit_usage_deep_dir.py index f066441e2..974249c79 100644 --- a/tests/functional/quota/test_limit_usage_deep_dir.py +++ b/tests/functional/quota/test_limit_usage_deep_dir.py @@ -200,7 +200,7 @@ class LimitUsageDeepDir(GlusterBaseClass): dir_name) g.log.info("Quota list validate and file created successful " "for %s", dir_name) - g.log.info("Files created and quota validated succesfully") + g.log.info("Files created and quota validated successfully") # Deleting data and validating quota self.all_mounts_procs = [] diff --git a/tests/functional/quota/test_multi_value_limit.py b/tests/functional/quota/test_multi_value_limit.py index 08bf6d259..3991c614b 100644 --- a/tests/functional/quota/test_multi_value_limit.py +++ b/tests/functional/quota/test_multi_value_limit.py @@ -154,7 +154,7 @@ class QuotaMultiValueLimits(GlusterBaseClass): "count=1 ; " "done" % mount_dir) ret, _, _ = g.run(client, cmd) - self.assertEqual(ret, 1, "Failed: Files created successfully inspite " + self.assertEqual(ret, 1, "Failed: Files created successfully in spite " "of crossing hard-limit") g.log.info("Files creation stopped on mountpoint once exceeded " "hard limit") @@ -202,7 +202,7 @@ class QuotaMultiValueLimits(GlusterBaseClass): "count=1 ; " "done" % mount_dir) ret, _, _ = g.run(client, cmd) - self.assertEqual(ret, 1, "Failed: Files created successfully inspite " + self.assertEqual(ret, 1, "Failed: Files created successfully in spite " "of crossing hard-limit") g.log.info("Files creation stopped on mountpoint once exceeded " "hard limit") @@ -250,7 +250,7 @@ class QuotaMultiValueLimits(GlusterBaseClass): "count=1 ; " "done" % mount_dir) ret, _, _ = g.run(client, cmd) - self.assertEqual(ret, 1, "Failed: Files created successfully inspite " + self.assertEqual(ret, 1, "Failed: Files created successfully in spite " "of crossing hard-limit") g.log.info("Files creation stopped on mountpoint once exceeded " "hard limit") diff --git a/tests/functional/quota/test_quota_add_brick.py b/tests/functional/quota/test_quota_add_brick.py index 58e9610ed..1078ceff1 100644 --- a/tests/functional/quota/test_quota_add_brick.py +++ b/tests/functional/quota/test_quota_add_brick.py @@ -120,7 +120,7 @@ class QuotaAddBrick(GlusterBaseClass): % (mount_dir)) ret, _, _ = g.run(client, cmd) self.assertEqual(ret, 0, ("Failed to create files")) - g.log.info("Files created succesfully") + g.log.info("Files created successfully") # Quota validate ret = quota_validate(self.mnode, self.volname, diff --git a/tests/functional/quota/test_quota_limit_dir_breadth.py b/tests/functional/quota/test_quota_limit_dir_breadth.py index b70c1187f..952e34134 100755 --- a/tests/functional/quota/test_quota_limit_dir_breadth.py +++ b/tests/functional/quota/test_quota_limit_dir_breadth.py @@ -165,7 +165,7 @@ class QuotaLimitDirBreadth(GlusterBaseClass): ret, _, _ = g.run(mount_object.client_system, cmd) self.assertFalse(ret, "Failed to create files in %s" % dir_name) - g.log.info("Files created succesfully in %s", dir_name) + g.log.info("Files created successfully in %s", dir_name) # Get Quota list on Volume g.log.info("Get Quota list for every directory on the volume %s", diff --git a/tests/functional/quota/test_quota_unique_soft_limit.py b/tests/functional/quota/test_quota_unique_soft_limit.py index f7f726df8..b175f12a2 100644 --- a/tests/functional/quota/test_quota_unique_soft_limit.py +++ b/tests/functional/quota/test_quota_unique_soft_limit.py @@ -156,7 +156,7 @@ class QuotaUniqueSoftLimit(GlusterBaseClass): ret, _, _ = g.run(client, cmd) self.assertEqual(ret, 0, ("Failed to create files on %s", dir_name)) - g.log.info("Files created succesfully on %s", dir_name) + g.log.info("Files created successfully on %s", dir_name) time.sleep(1) diff --git a/tests/functional/quota/test_quota_volume_subdir_limits.py b/tests/functional/quota/test_quota_volume_subdir_limits.py index 1291796a1..7e4245c16 100644 --- a/tests/functional/quota/test_quota_volume_subdir_limits.py +++ b/tests/functional/quota/test_quota_volume_subdir_limits.py @@ -153,7 +153,7 @@ class QuotaVolumeAndSubdirLimits(GlusterBaseClass): ret, _, _ = g.run(client, cmd) self.assertEqual(ret, 0, ("Failed to create files in /foo%s", var1)) - g.log.info("Files created succesfully in /foo%s", var1) + g.log.info("Files created successfully in /foo%s", var1) g.log.info("Files creation is successful on all directories of the " "volume %s", self.volname) diff --git a/tests/functional/snapshot/test_256_snapshots.py b/tests/functional/snapshot/test_256_snapshots.py index f3e6e4b0e..9c5ceaf99 100644 --- a/tests/functional/snapshot/test_256_snapshots.py +++ b/tests/functional/snapshot/test_256_snapshots.py @@ -38,7 +38,7 @@ class SanpCreate256(GlusterBaseClass): 4. Varify IO 5. modify max snap limit to default to 256. 6. Create 256 snapshots - 7. Varify 256 created sucessfully + 7. Varify 256 created successfully 8. Create 257th snapshot - check for failure -- it should fail. 9. Cleanup @@ -92,7 +92,7 @@ class SanpCreate256(GlusterBaseClass): ret = cls.unmount_volume_and_cleanup_volume(cls.mounts) if not ret: raise ExecutionError("Failed to cleanup volume and mount") - g.log.info("Cleanup successfull for the volume and mount") + g.log.info("Cleanup successful for the volume and mount") GlusterBaseClass.tearDownClass.im_func(cls) @@ -155,7 +155,7 @@ class SanpCreate256(GlusterBaseClass): for i in range(1, 257, 1): self.assertTrue(("snapy%s" % i in snap_list), "%s snap not " "found " % ("snapy%s" % i)) - g.log.info("Sucessfully validated names of snap") + g.log.info("Successfully validated names of snap") # Try to create 257th snapshot cmd_str = "gluster snapshot create %s %s %s" % ("snap", self.volname, diff --git a/tests/functional/snapshot/test_clone_snap.py b/tests/functional/snapshot/test_clone_snap.py index a7ebd965c..6311a282a 100644 --- a/tests/functional/snapshot/test_clone_snap.py +++ b/tests/functional/snapshot/test_clone_snap.py @@ -198,7 +198,7 @@ class CloneSnapshot(GlusterBaseClass): "do touch file$i; done; cd;") % mount_obj proc = g.run(self.clients[0], cmd) all_mounts_procs.append(proc) - g.log.info("I/O on mountpoint is Successfull") + g.log.info("I/O on mountpoint is successful") # create snapshot g.log.info("Starting to Create snapshot of clone volume") @@ -226,7 +226,7 @@ class CloneSnapshot(GlusterBaseClass): ret, _, _ = umount_volume(self.clients[0], self.mpoint, self.mount_type) if ret == 1: - raise ExecutionError("Unmounting the moint point %s failed" + raise ExecutionError("Unmounting the mount point %s failed" % self.mpoint) g.log.info("Mount point %s deleted successfully", self.mpoint) g.log.info("Unmount Volume Successful") diff --git a/tests/functional/snapshot/test_mount_snap.py b/tests/functional/snapshot/test_mount_snap.py index e01be9dd8..67b8a43cd 100644 --- a/tests/functional/snapshot/test_mount_snap.py +++ b/tests/functional/snapshot/test_mount_snap.py @@ -177,7 +177,7 @@ class TestSnapMountSnapshot(GlusterBaseClass): # validate io should fail self.assertFalse( validate_io_procs(all_mounts_procs, self.mounts), - "Unexpected: IO Successfull on all clients" + "Unexpected: IO Successful on all clients" ) g.log.info("Expected: IO failed on clients") diff --git a/tests/functional/snapshot/test_snap_delete_existing_scheduler.py b/tests/functional/snapshot/test_snap_delete_existing_scheduler.py index 436e7f214..a321dcd39 100644 --- a/tests/functional/snapshot/test_snap_delete_existing_scheduler.py +++ b/tests/functional/snapshot/test_snap_delete_existing_scheduler.py @@ -186,7 +186,7 @@ class SnapshotDeleteExistingScheduler(GlusterBaseClass): count += 2 self.assertEqual(status.strip().split(":")[2], ' Enabled', "Failed to check status of scheduler") - g.log.info("Successfuly checked scheduler status") + g.log.info("Successfully checked scheduler status") # write files on all mounts g.log.info("Starting IO on all mounts...") diff --git a/tests/functional/snapshot/test_snap_list_after_restart.py b/tests/functional/snapshot/test_snap_list_after_restart.py index 1eb83751c..52f34c14a 100644 --- a/tests/functional/snapshot/test_snap_list_after_restart.py +++ b/tests/functional/snapshot/test_snap_list_after_restart.py @@ -122,7 +122,7 @@ class SnapshotGlusterddown(GlusterBaseClass): g.log.info("Starting to list all snapshots") for server in self.servers[0:]: out = get_snap_list(server) - self.assertIsNotNone(out, "Falied to list snap in node" + self.assertIsNotNone(out, "Failed to list snap in node" "%s" % server) self.assertEqual(len(out), 2, "Failed to validate snap list" "on node %s" % server) diff --git a/tests/functional/snapshot/test_snap_uss.py b/tests/functional/snapshot/test_snap_uss.py index aad6d897c..60e06f359 100644 --- a/tests/functional/snapshot/test_snap_uss.py +++ b/tests/functional/snapshot/test_snap_uss.py @@ -197,7 +197,7 @@ class SnapshotUssSnap(GlusterBaseClass): for mount_obj in self.mounts: ret, _, _ = uss_list_snaps(mount_obj.client_system, mount_obj.mountpoint) - self.assertEqual(ret, 0, "Falied to list snapshot Informations") + self.assertEqual(ret, 0, "Failed to list snapshot information") g.log.info("Successfully Listed snapshots Created") # Disable USS running @@ -211,8 +211,8 @@ class SnapshotUssSnap(GlusterBaseClass): for mount_obj in self.mounts: ret, out, _ = uss_list_snaps(mount_obj.client_system, mount_obj.mountpoint) - self.assertEqual(ret, 0, "Failed to list snapshot Informations") - g.log.info("Successfully Listed snapshots Created") + self.assertEqual(ret, 0, "Failed to list snapshot information") + g.log.info("Successfully listed snapshots Created") # Validate after disabling USS, all files should be same g.log.info("Validate files after disabling uss") diff --git a/tests/functional/snapshot/test_snapshot_create.py b/tests/functional/snapshot/test_snapshot_create.py index 5ffb91681..e1bc9c272 100644 --- a/tests/functional/snapshot/test_snapshot_create.py +++ b/tests/functional/snapshot/test_snapshot_create.py @@ -101,7 +101,7 @@ class SnapCreate(GlusterBaseClass): ret = cls.unmount_volume_and_cleanup_volume(cls.mounts) if not ret: raise ExecutionError("Failed to cleanup volume and mount") - g.log.info("Cleanup successfull for the volume and mount") + g.log.info("Cleanup successful for the volume and mount") GlusterBaseClass.tearDownClass.im_func(cls) @@ -202,4 +202,4 @@ class SnapCreate(GlusterBaseClass): for i in range(0, 5): self.assertIn("snapy%s" % i, snap_list, "%s snap not " "found " % ("snapy%s" % i)) - g.log.info("Sucessfully validated names of snap") + g.log.info("Successfully validated names of snap") diff --git a/tests/functional/snapshot/test_snapshot_restore.py b/tests/functional/snapshot/test_snapshot_restore.py index cbec67e04..966551b58 100644 --- a/tests/functional/snapshot/test_snapshot_restore.py +++ b/tests/functional/snapshot/test_snapshot_restore.py @@ -106,7 +106,7 @@ class SnapRestore(GlusterBaseClass): ret = cls.unmount_volume_and_cleanup_volume(cls.mounts) if not ret: raise ExecutionError("Failed to cleanup volume and mount") - g.log.info("Cleanup successfull for the volume and mount") + g.log.info("Cleanup successful for the volume and mount") GlusterBaseClass.tearDownClass.im_func(cls) @@ -157,7 +157,7 @@ class SnapRestore(GlusterBaseClass): % self.volname)) g.log.info("Volume options for%s is set successfully", self.volname) - # Get brick list befor taking snap_restore + # Get brick list before taking snap_restore bricks_before_snap_restore = get_all_bricks(self.mnode, self.volname) g.log.info("Brick List before snap restore " "volume: %s", bricks_before_snap_restore) diff --git a/tests/functional/snapshot/test_uss_brick_down.py b/tests/functional/snapshot/test_uss_brick_down.py index 94b48c043..172b0f291 100644 --- a/tests/functional/snapshot/test_uss_brick_down.py +++ b/tests/functional/snapshot/test_uss_brick_down.py @@ -98,7 +98,7 @@ class SnapUssBrickDown(GlusterBaseClass): * Mount volume * Perform I/O on mounts * Bring down one brick - * Enbale USS + * Enable USS * Validate USS is enabled * Bring the brick online using gluster v start force * Create 2 snapshots snapy1 & snapy2 @@ -164,7 +164,7 @@ class SnapUssBrickDown(GlusterBaseClass): g.log.info("Bring the brick online using gluster v start force") ret, _, _ = volume_start(self.mnode, self.volname, force=True) self.assertEqual(ret, 0, "Volume start with force failed") - g.log.info("Volume start with force successfull") + g.log.info("Volume start with force successful") # Create 2 snapshot g.log.info("Creating 2 snapshots for volume %s", self.volname) diff --git a/tests/functional/snapshot/test_uss_snap_active_deactive.py b/tests/functional/snapshot/test_uss_snap_active_deactive.py index c28cc1d60..93c0231ee 100644 --- a/tests/functional/snapshot/test_uss_snap_active_deactive.py +++ b/tests/functional/snapshot/test_uss_snap_active_deactive.py @@ -100,7 +100,7 @@ class SnapUssActiveD(GlusterBaseClass): * Perform I/O on mounts * Create 2 snapshots snapy1 & snapy2 * Validate snap created - * Enbale USS + * Enable USS * Validate USS is enabled * Validate snapd is running * Activate snapy1 & snapy2 diff --git a/tests/functional/snapshot/test_validate_snap_scheduler.py b/tests/functional/snapshot/test_validate_snap_scheduler.py index ea54f26d2..c8513ca22 100644 --- a/tests/functional/snapshot/test_validate_snap_scheduler.py +++ b/tests/functional/snapshot/test_validate_snap_scheduler.py @@ -127,7 +127,7 @@ class SnapshotSchedulerBehaviour(GlusterBaseClass): count += 2 self.assertEqual(status.strip().split(":")[2], ' Enabled', "Failed to check status of scheduler") - g.log.info("Successfuly checked scheduler status") + g.log.info("Successfully checked scheduler status") # Disable snap scheduler g.log.info("Starting to disable snapshot scheduler on all nodes") @@ -148,4 +148,4 @@ class SnapshotSchedulerBehaviour(GlusterBaseClass): count += 2 self.assertEqual(status.strip().split(":")[2], ' Disabled', "Failed to check status of scheduler") - g.log.info("Successfuly checked scheduler status") + g.log.info("Successfully checked scheduler status") diff --git a/tests/functional/snapshot/test_validate_snapshot_max_limit.py b/tests/functional/snapshot/test_validate_snapshot_max_limit.py index 75582e4ed..3df2a98dc 100644 --- a/tests/functional/snapshot/test_validate_snapshot_max_limit.py +++ b/tests/functional/snapshot/test_validate_snapshot_max_limit.py @@ -26,7 +26,7 @@ Steps : 5. modify max snap limit to default to 10. 6. modify soft-limit to 50% 6. Create 5 snapshots - 7. Varify 5 created sucessfully + 7. Varify 5 created successfully 8. Create 6th snapshot - check for warning -- it should not fail. 9. modify soft-limit to 100% @@ -65,7 +65,7 @@ class SnapCreateMax(GlusterBaseClass): 5. modify max snap limit to default to 10. 6. modify soft-limit to 50% 6. Create 5 snapshots - 7. Varify 5 created sucessfully + 7. Varify 5 created successfully 8. Create 6th snapshot - check for warning -- it should not fail. 9. modify soft-limit to 100% @@ -125,7 +125,7 @@ class SnapCreateMax(GlusterBaseClass): ret = self.unmount_volume_and_cleanup_volume(self.mounts) if not ret: raise ExecutionError("Failed to cleanup volume and mount") - g.log.info("Cleanup successfull for the volume and mount") + g.log.info("Cleanup successful for the volume and mount") def test_validate_snaps_max_limit(self): # pylint: disable=too-many-statements @@ -194,7 +194,7 @@ class SnapCreateMax(GlusterBaseClass): for i in range(1, 6): self.assertTrue(("snapy%s" % i in snap_list), "%s snap not " "found " % ("snapy%s" % i)) - g.log.info("Sucessfully validated names of snapshots") + g.log.info("Successfully validated names of snapshots") # create 6th snapshot cmd_str = "gluster snapshot create %s %s %s" % ("snapy6", self.volname, @@ -246,7 +246,7 @@ class SnapCreateMax(GlusterBaseClass): cmd_str = "gluster snapshot create %s %s %s" % ("snap", self.volname, "no-timestamp") ret, _, _ = g.run(self.mnode, cmd_str) - self.assertNotEqual(ret, 0, ("Unexpected: sucessfully created 'snap' " + self.assertNotEqual(ret, 0, ("Unexpected: successfully created 'snap' " "for %s" % self.volname)) g.log.info("Expected: Snapshot 'snap' not created as it is 11th snap") diff --git a/tests/functional/snapshot/test_validate_snapshot_rebalance.py b/tests/functional/snapshot/test_validate_snapshot_rebalance.py index 7eb674475..f9e018d64 100644 --- a/tests/functional/snapshot/test_validate_snapshot_rebalance.py +++ b/tests/functional/snapshot/test_validate_snapshot_rebalance.py @@ -55,7 +55,7 @@ class SnapCreateRebal(GlusterBaseClass): during rebalance 9. After rebalance is completed, create snapshots with the same name as in Step 7 - -- this operation should be sucessfull + -- this operation should be successful 10. Cleanup """ @@ -103,7 +103,7 @@ class SnapCreateRebal(GlusterBaseClass): ret = self.unmount_volume_and_cleanup_volume(self.mounts) if not ret: raise ExecutionError("Failed to cleanup volume and mount") - g.log.info("Cleanup successfull for the volume and mount") + g.log.info("Cleanup successful for the volume and mount") def test_snapshot_while_rebalance(self): # pylint: disable=too-many-statements, missing-docstring -- cgit