summaryrefslogtreecommitdiffstats
path: root/tests/functional
Commit message (Collapse)AuthorAgeFilesLines
...
* [Snapshot] Change shared volume mount validation functionVinayak Papnoi2019-08-081-65/+51
| | | | | | | | | | | The test case makes use of 'is_shared_volume_unmounted' function which is redundant as the purpose can be solved via a very similar function 'is_shared_volume_mounted'. This change has been implemented. There are various cosmetic changes that have been implemented as well. Change-Id: I560b464b4bcc436658db49c0a5ed8c7aadacfb6a Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* Fix AFR test case tearDownVinayak Papnoi2019-08-081-6/+18
| | | | | | | | | | | | The test case 'test_client_side_quorum_with_fixed_for_cross3' does not include the tearDown part where the volume options which have been set inside the test case have to be reset to default. This fix includes the necessary tearDown steps along with a few cosmetic changes. Change-Id: I86187cef4523492ec97707ff93d0eca365293008 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* Adding test case to mount volume, remove /var/log/glusterfs/ and remount the ↵kshithijiyer2019-08-061-0/+60
| | | | | | | | | | | | | | | volume. Test case: 1. Create all types of volumes and start them. 2. Mount all volumes on clients. 3. Delete /var/log/glusterfs folder on client. 4. Run IO on all the mount points. 5. Unmount and remount all volumes. 6. Check if logs are regenerated or not. Change-Id: I4f90d709c4da6e1c73cf95f4075c50aa44cdd811 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Enable client-side heal as client-side healing is diabled in RHGS 3.5Anees Patel2019-08-051-0/+10
| | | | | Change-Id: I0992b1b9af4e12f4e20d7a5dc184048de104d89d Signed-off-by: Anees Patel <anepatel@redhat.com>
* Optimized test case: tests/functional/glusterd/test_remove_brick_scenarios.pyhadarsharon2019-08-051-16/+28
| | | | | | | | | | Worked on the following: Improved I/O performance of test (writing 100k files to a mounted volume) by applying the following changes: 1. Modified the touch command to write as many files as possible per process, thus requiring less processes to write the 100k files 2. Using Threads to parallelize the touch processes from within the test, for better efficiency Change-Id: Id969f387f4b7b8e88daf688f7bada950cff2c412 Signed-off-by: hadarsharon <hsharon@redhat.com>
* Adding test case to enable brickmux, create start and stop 3 volumes.kshithijiyer2019-07-291-0/+115
| | | | | | | | | | | Test Case: 1.Set cluster.brick-multiplex to enabled. 2.Create three 1x3 replica volumes. 3.Start all the three volumes. 4.Stop three volumes one by one. Change-Id: Ibf3e81e7424d6a429da0aa12efeae7fffd3338f2 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* [EC]Fix a script issue to change permission of the directoryubansal2019-07-291-2/+7
| | | | | | | | Added a line to change permission of the directory so that client side healing happens for the directory also Change-Id: If4a24f2dbd6c9c85d4cb2944d1ad4795dbc39adb Signed-off-by: ubansal <ubansal@redhat.com>
* Enable client-side heal as client-side healing is disabled by default in ↵Anees Patel2019-07-291-0/+10
| | | | | | | RHGS 3.5 Change-Id: I500912b5217b675f9fdff4fe1cb518b465de245c Signed-off-by: Anees Patel <anepatel@redhat.com>
* [glusterfind] Addition of 'Test' in the class name of glusterfind create testVinayak Papnoi2019-07-291-32/+24
| | | | | | | | | | | | | | The class name was missing 'Test' at the beginning. Changing the name of the class to 'TestGlusterFindCreateCLI' from 'GlusterFindCreateCLI'. Removing the setUpClass and tearDownClass and replacing them with setUp and tearDown. Changing the variable names to be intelligible. Change-Id: Ibb5d9c6ef75ef11960aad35d65c343fa08fc9de1 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* Adding testcase to remove /var/log/glusterfs and mounting volume.kshithijiyer2019-07-231-0/+125
| | | | | | | | | | | | | Test Case: 1. Create all types of volumes. 2. Start all volumes. 3. Delete /var/log/glusterfs folder on the client. 4. Mount all the volumes one by one. 5. Run IO on all the mount points. 6. Check if logs are generated in /var/log/glusterfs/. Change-Id: I7a3275aad940116c3506b22b13a670e455d9ef00 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Adding code to not stop glusterd on mnode.kshithijiyer2019-07-081-6/+28
| | | | | | | | | | | | | This test case was failing in a test run as the mnode was not removed from the list self.servers because of which there were runs where glusterd was stoppped instead and command was executed on mnod.As well as adding code to check and start glusterd on the node in instances where the test case fails. Change-Id: Id203102d3f0ec82af0ac215f0ecaf7ae22b630f5 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Modifying test_enabling_gluster_debug_mode to do more operations.kshithijiyer2019-07-021-3/+7
| | | | | | | | | | | | While running test_enabling_gluster_debug_mode through jenkins it was observed that running volume operation once wasn't generating enough of logs by the time the logs were checked which lead to failure of the test case in the jenkins run. So modifying the logic which generates logs to run operation in a loop to generated a good amount of logs. Change-Id: Id7a12c86a04dc86d4856dbe30d945e70e64ea4f7 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Converting text files to markdown files.kshithijiyer2019-06-302-21/+20
| | | | | | | | | It's better to use markdown files instead of text files for readme files. Hence converting readme files to readme.md files. Change-Id: I41c1b2f065895d885f4b1fabdc9b9e4051810e80 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* [glusterfind] Test case to verify the 'glusterfind pre' functionalityVinayak Papnoi2019-06-301-0/+204
| | | | | | | | | | | | | | | | | | | | | | | Verifying the 'glusterfind pre' command functionality with valid and invalid values for the required and optional parameters. * Create a session on the volume * Perform some I/O from the mount point * Perform glusterfind pre with the following combinations: - Valid values for required parameters - Invalid values for required parameters - Valid values for optional parameters - Invalid values for optional parameters * Perform glusterfind post Where Required parameters: volname, sessname and outfile Optional parameters: full, debug, gftype, tagforfullfind, namespace, noencode, disablepartial, regenoutfile, outprefix, fieldsep Change-Id: Id795d632ebb8f43f49a9d15d83aced87670007b6 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [glusterfind] Test case to verify the 'glusterfind post' functionalityVinayak Papnoi2019-06-301-0/+199
| | | | | | | | | | | | | | | | | | | | | Verifying the 'glusterfind post' command functionality with valid and invalid values for the required and optional parameters. * Create a session on the volume * Perform some I/O from the mount point * Perform glusterfind pre - Perform glusterfind post with the following combinations: - Valid values for required parameters - Invalid values for required parameters - Valid values for optional parameters - Invalid values for optional parameters Where Required parameters: volname and sessname Optional parameters: debug Change-Id: Ia4c59b5ea3626d83bd6cedfda44e04ddbfa1567d Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [glusterfind] Test case to verify the 'glusterfind query' functionalityVinayak Papnoi2019-06-301-0/+164
| | | | | | | | | | | | | | | | | | | | | | | Verifying the 'glusterfind query' command functionality with valid and invalid values for the required and optional parameters. *** glusterfind query command does not require a session to be created *** * Perform some I/O from the mount point * Perform glusterfind query with the following combinations: - Valid values for required parameters - Invalid values for required parameters - Valid values for optional parameters - Invalid values for optional parameters Where Required parameters: volname and outfile Optional parameters: full, debug, gftype, tagforfullfind, since, namespace, noencode, disablepartial, end, regenoutfile, outprefix, fieldsep Change-Id: I5b9dbc987cec0f4ff9e460f875da72de7c9fe1c5 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [glusterfind] Test case to verify the glusterfind delete functionalityVinayak Papnoi2019-06-301-0/+145
| | | | | | | | | | | | | | | | | | | | | | Verifying the glusterfind delete command functionality with valid and invalid values for the required and optional parameters. * Create a volume * Create a session on the volume * Perform glusterfind list to check if session is created * Delete the glusterfind session with the following combinations: - Valid values for required parameters - Invalid values for required parameters - Valid values for optional parameters - Invalid values for optional parameters * Perform glusterfind list to check if session is deleted Where: Required parameters: volname and sessname Optional parameters: debug Change-Id: I8240ccfd728cd7aa138103ffe1c24a705aa1b51c Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [glusterfind] Test case to verify the glusterfind create functionalityVinayak Papnoi2019-06-172-0/+138
| | | | | | | | | | | | | | | | | | Verifying the glusterfind create command functionality with valid and invalid values for the required and optional parameters. * Create a volume * Create a session on the volume with the following combinations: - Valid values for required parameters - Invalid values for required parameters - Valid values for optional parameters - Invalid values for optional parameters Where: Required parameters: volname and sessname Optional parameters: debug, force, reset-session-time Change-Id: Ic46863f63679e3c26a95b1fa96138cb7d0c27a3d Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* Adding code to cleanup bricks in test_add_brick.pykshithijiyer2019-06-131-9/+18
| | | | | | | | | | | While executing the test suit it was observerd that the test case test_add_remove_brick was failing due to remains from the test case test_add_brick_functionality. Hence adding the code to clean all the briks post test in test_add_brick.py. Change-Id: Iace9e51582ab4fa1f0f184283e6205aa6140b4a2 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Optimized test case: tests/functional/glusterd/test_add_brickhadarsharon2019-06-091-65/+54
| | | | | | | | | | | Worked on the following: 1. Removed redundant throwaway variables (ret, _, _) 2. More consistent exceptions 3. Added comments within code 4. Clarified Error messages in case of Assertion Errors Change-Id: I8ca0acce848bd9a8a5d217b5a4e247590177154d Signed-off-by: hadarsharon <hsharon@redhat.com>
* [Snapshot] Fixing Jira ticket RHGSQE-233Vinayak Papnoi2019-06-061-56/+66
| | | | | | | | | Removing 'ExecutionError' raised inside the test case function. Adding the description for the test case function. Fixing the typo errors and various cosmetic changes. Change-Id: Ica505417f482c7a622be5da70c1053913cf3052c Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [Snapshot] Fixing Jira ticket RHGSQE-234Vinayak Papnoi2019-06-051-60/+71
| | | | | | | | | Removing 'ExecutionError' raised inside the test case function. Adding the description for the test case function. Fixing the typo errors and various cosmetic changes. Change-Id: I496dc1df067870d43e5c3c97fc27da4294862eba Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* glusto-tests/glusterd: enable glusterd in debug modekshithijiyer2019-05-291-0/+169
| | | | | | | | | | | | | | | | | In this test case we will enable glusterd in debug mode and check the glusterd log for the debug messages. Steps followed: 1. Stop glusterd. 2. Change log level to DEBUG in /usr/local/lib/systemd/system/glusterd.service. 3. Remove glusterd log. 4. Start glusterd. 5. Issue some gluster commands. 6. Check for debug messages in glusterd log. Change-Id: Id1173be6da2ef1c2233459fb23f4b27308c923f2 Signed-off-by: Sanju Rakonde <srakonde@redhat.com> Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* This TC is to ammend the previous TC to prevent it from failing - picking up ↵sayalee2019-05-281-7/+29
| | | | | | | random servers which are not part of volume Change-Id: I8dcb8fcd8fd0bfaefcaf24a3f2aa222b44ff7f24 Signed-off-by: sayalee <root@localhost.localdomain>
* Adding the assert statement removed during changes.kshithijiyer2019-05-231-0/+2
| | | | | | | | | It seems like an assert statement got missed duing the recent changes. Adding back the assert and submitting a patch Signed-off-by: kshithijiyer <kshithij.ki@gmail.com> Change-Id: Ic8ab0d6e54da510faf479cd09cf122ccf8cedfbb
* nfs-ganesha: Test to verify mounting volume on another client while IO is ↵Jilju Joy2019-05-231-0/+151
| | | | | | | progressing on one client Change-Id: I42dceccaa0f257e25a4e5f2bc8d21f84e94216e0 Signed-off-by: Jilju Joy <jijoy@redhat.com>
* nfs-ganesha: Changes in test cases based on latest libraryJilju Joy2019-05-231-211/+428
| | | | | Change-Id: I78f987383725ef43caa62ad084dd64ec6be2363b Signed-off-by: Jilju Joy <jijoy@redhat.com>
* Changing systemctl to service to fix jira issue RHGSQE-197kshithijiyer2019-05-231-2/+4
| | | | | | | | Bug https://bugzilla.redhat.com/show_bug.cgi?id=1690254 has to be fixed before merging this patch. Change-Id: I90e669269fafa9d0a064a64883c3e4b88080d25f Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Changing error messages to be checked as per new messages.kshithijiyer2019-05-081-2/+18
| | | | | | | | | | | | | | | | | | | | Changing error message displayed when peer detach is issued with bricks are present on the node which is being detached. Adding a logic to handle both the new as well as the old error message. Old msg: peer detach: failed: Brick(s) with the peer <my_server> exist in cluster New msg: peer detach: failed: Peer <my_server> hosts one or more bricks. If the peer is in not recoverable state then use either replace-brick or remove-brick command with force to remove all bricks from the peer and attempt the peer detach again. Change-Id: I3d8fdac2c33638ecc2a8b5782c68caebbf17cf41 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Changing error message to be check in test_volume_getkshithijiyer2019-05-061-0/+2
| | | | | | | | | | | | | | The error message which is displayed when we do a gluster v get for options which don't exist has been changed. Adding a if based logic which can check for the old as well as the new error message. Old msg: volume get option: failed: Did you mean auth.allow or ...reject? New msg: volume get option: failed: Did you mean ctime.noatime? Signed-off-by: kshithijiyer <kshithij.ki@gmail.com> Change-Id: I9496d391a7da9dba64d3426a024c2b1b68455f20
* Adding test to validate output of profile infokshithijiyer2019-05-031-0/+223
| | | | | | | | | | | | | | | Test Case: 1) Create a volume and start it. 2) Mount volume on client and start IO. 3) Start profile info on the volume. 4) Run profile info with different parameters and see if all bricks are present or not. 5) Stop profile on the volume. 6) Create another volume. 7) Start profile without starting the volume. Change-Id: I6e8ec9285d48c1c828cd1d20bff6ea8f3de064f7 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Adding test for profile operations with one node downkshithijiyer2019-04-291-0/+220
| | | | | | | | | | | | | | Test Case: 1) Create a volume and start it. 2) Mount volume on client and start IO. 3) Start profile info on the volume. 4) Stop glusterd on one node. 5) Run profile info with different parameters and see if all bricks are present or not. 6) Stop profile on the volume. Change-Id: Ie573414816362ebbe30d2c419fd0e348522ceaec Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test case to detach node used for mountingkshithijiyer2019-04-171-0/+219
| | | | | | | | | | | | | | | | | | | Test case: 1.Create a 1X3 volume with only 3 nodes from the cluster. 2.Mount volume on client node using the ip of the fourth node. 3.Write IOs to the volume. 4.Detach node N4 from cluster. 5.Create a new directory on the mount point. 6.Create a few files using the same command used in step 3. 7.Add three more bricks to make the volume 2x3 using add-brick command. 8.Do a gluster volume rebalance on the volume. 9.Create more files from the client on the mount point. 10.Check for files on bricks from both replica sets. 11.Create a new directory from the client on the mount point. 12.Check for directory in both replica sets. Change-Id: I228b79955dca565a40994919b2903e59cad7d8f5 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test to profile start when quorum not metBala Konda Reddy M2019-04-161-0/+147
| | | | | | | | | | | | | 1. Create a volume 2. Set the quorum type to server and ratio to 90 3. Stop glusterd randomly on one of the node 4. Start profile on the volume 5. Start glusterd on the node where it is stopped 6. Start profile on the volume 7. Stop profile on the volume where it is started Change-Id: Ifeb9fddf6f1a14c9df73ed2f0453636d2853e944 Signed-off-by: Bala Konda Reddy M <bmekala@redhat.com>
* Adding test to run gluster commands when glusterd is down on one nodekshithijiyer2019-04-121-0/+102
| | | | | Change-Id: Ibf41c11a4e98baeaad658ee10ba8a807318504be Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Validating whether peers are connected or not before volume creationBala Konda Reddy M2019-04-121-1/+17
| | | | | | | | | | In jenkins this case is failing with peers are not connected while volume creation. Now having a check before creating the volume to make sure that peers are in cluster and in connected state after peer probe. Change-Id: I8aa9d2c4d1669475dd8867d42752a31604ff572f Signed-off-by: Bala Konda Reddy M <bmekala@redhat.com>
* Adding code for cleanup of all bricks on each serverkshithijiyer2019-04-091-2/+18
| | | | | Change-Id: I405843e0093ddb7138ee0a8afbfd4cd2f91e6284 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Checking if peers are connected after peer probekshithijiyer2019-04-091-0/+24
| | | | | Change-Id: I252ab0c0f6248b9a5c1d7977146c15876e144b38 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* This TC is to check for File operations in an EC volumeubansal2019-04-081-0/+337
| | | | | Change-Id: Ib1aff1c1bf843dddac5862e55a049d7b47603049 Signed-off-by: ubansal <ubansal@redhat.com>
* Removing test_volume_sanity TCAkarsha2019-04-011-106/+0
| | | | | Change-Id: I82c76f26f4bf359ce153ff35850ce5c349bc759b Signed-off-by: Akarsha <akrai@redhat.com>
* Adding code to check if peers are connected in test_spurious_rebalancekshithijiyer2019-03-291-1/+13
| | | | | Change-Id: I4a1097fbdebd49555fffcfa5fe609f4070e39182 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Validating peer is in connected or notBala Konda Reddy M2019-03-281-0/+15
| | | | | | | | | In jenkins right after peer probe, add brick function is failing with peer not in cluster. So having check for the peer to be connected or not then proceed to next step. Change-Id: I73bf92819ad44f7a6a14795ab07c45d260cd04eb Signed-off-by: Bala Konda Reddy M <bmekala@redhat.com>
* Test case to change reserve limit with remove brick in progresskshithijiyer2019-03-131-0/+270
| | | | | Change-Id: I53fb7f4cceae395698568129669dc5f3a9a5e4bb Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Fixing busy umount errorsAnees Patel2019-02-251-0/+239
| | | | | Change-Id: I7f7b5cfdee09067d8d96bfcf56ce8a3372ca9368 Signed-off-by: Anees Patel <anepatel@redhat.com>
* Test to check remove brick scenarios wrt to glusterdkshithijiyer2019-02-201-0/+188
| | | | | Change-Id: I1bfa2fb3ae4ff1fc247b40c73f4fade9a3afeede Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test to reduce volume from 2x3 to 2x2 and to distkshithijiyer2019-02-121-0/+115
| | | | | Change-Id: I64309c3b46dc9087eeb3181acba63b981b2ecc6f Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test to check volume create with glusterd restart on one nodekshithijiyer2019-02-121-0/+116
| | | | | Change-Id: Ica0771bdee1e96e9d6bb5157fb6c2125a4b419f1 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test case to check peer detach warningkshithijiyer2019-02-111-0/+83
| | | | | Change-Id: Idc379ad7f31274cc63f384d7223bf769bb89ace3 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Fixes for checking uid, gid and permissionVitalii Koriakov2019-02-081-60/+79
| | | | | Change-Id: I683e6ff47120b7db8ee6ae02ed83eba19e6ac4c9 Signed-off-by: Vitalii Koriakov <vkoriako@redhat.com>
* This Tc is used to check remove brick on a dispersed volumeubansal2019-02-071-0/+166
| | | | | Change-Id: Id8cfc0dd31cf4f6f381ec7bb07d4aba06d52b43e Signed-off-by: ubansal <ubansal@redhat.com>