summaryrefslogtreecommitdiffstats
path: root/tests
Commit message (Collapse)AuthorAgeFilesLines
* [Fix] Fixing string formatting errors and client heal errorskshithijiyer2019-09-191-37/+51
| | | | | | Change-Id: Ifef2ffe022accf59edcbc949c505f47931b19fe4 Signed-off-by: Anees Patel <anepatel@redhat.com> Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* nfs-ganesha: Modifying test case based on latest libraryJilju Joy2019-09-191-45/+94
| | | | | Change-Id: I5372caf852b41e127c46f6afa697435dcde9dbf6 Signed-off-by: Jilju Joy <jijoy@redhat.com>
* nfs-ganesha: Test case to verify replace brick operation while IO is in progressJilju Joy2019-09-191-0/+176
| | | | | Change-Id: I40e36e1d4d404fcb9709fa9e50d33cad77172350 Signed-off-by: Jilju Joy <jijoy@redhat.com>
* nfs_ganesha: Test case to validate add brick operation while IO is in progressJilju Joy2019-09-191-0/+185
| | | | | | | | | | | Steps 1. Start IO on mount points 2. Add bricks to expand the volume 3. Start rebalance and wait for its completion 4. Validate IOs Change-Id: I97c09c9fc226afeff1446d225959730715f89aef Signed-off-by: Jilju Joy <jijoy@redhat.com>
* nfs-ganesha:Testcase to create and export new volume while IOs are in progressJilju Joy2019-09-191-0/+250
| | | | | | | | | | | Steps 1. Start IO on mount points 2. Create another volume 'volume_new' 3. Export volume_new through nfs-ganesha 4. Mount the volume on clients Change-Id: I2c4fe59e9a85e6668672d31a0a6c27d11c7f03f8 Signed-off-by: Jilju Joy <jijoy@redhat.com>
* Cthon test case for NFS GaneshaArjun Sharma2019-09-121-0/+121
| | | | | Change-Id: I3fb826bd0ecbe46bee4b9f8594b23f16921adbec Signed-off-by: Arjun Sharma <arjsharm@redhat.com>
* Add 7 negative scenarios for EC volume createubansal2019-09-111-1/+73
| | | | | | | | | With redundancy count as negative or disperse count with negative(different permutations) and disperse data count equal to disperse count Change-Id: I761851c64833256532464f56a9a78e20ceb8a4e1 Signed-off-by: ubansal <ubansal@redhat.com>
* Fix AFR test case tearDown and library importVinayak Papnoi2019-09-111-3/+16
| | | | | | | | | | | | | The test case 'test_client_side_quorum_with_fixed_validate_max_bricks' does not have a tearDown part where the volume options which have been set inside the test case have not been reset to default. The library function 'set_volume_options' was being imported from a wrong library. This fix includes this change along with the tearDown steps. Change-Id: Ic57494e7a7e8a25303b7979f98cc2dfbc9a7d7b6 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* Checking quota daemon and self heal daemon process after rebootkshithijiyer2019-09-111-0/+167
| | | | | | | | | | | | | For each node should have one self heal daemon and one quota daemon should be up running that means total number of self heal daemon and quota daemon to be up and ruuning is (number of nodes *2), in code i am checking that count should be equalent to (number of nodes * 2) Change-Id: I79d40467edc255a479a369f19a6fd1fec9111f53 Signed-off-by: Rajesh Madaka <rmadaka@redhat.com> Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* [DHT] : Modifying the messages in g.log.info and ExecutionErrorsayaleeraut2019-09-111-4/+4
| | | | | | | | | | | | The ExecutionError msg at line 66 should be "failed to clean-up volume" because it is under tearDownClass. The g.log.info msg at line 123 should be "Checking if gfid xattr of directories is displayed and is same on all the bricks on the server node" as the code below it checks for the gfid xattr on the bricks on server node and not if it is displayed on mount point. Change-Id: If4e20e3487a44c1cc7047504d19cc9859424ccd4 Signed-off-by: sayaleeraut <saraut@redhat.com>
* do getfattr in brick_path rather than m_pointyinkui2019-09-111-2/+3
| | | | | Change-Id: I240ecb0b4a9c99134b7a5cd237a59c2857d0fb7b Signed-off-by: yinkui <13965432176@163.com>
* access self.mnod so we must get layout in mount_point/dirpath, and change ↵yinkui2019-09-041-2/+2
| | | | | | | the log. Change-Id: I361b2c59108b19480906f6dfd49b023ed1eb05cd Signed-off-by: yinkui <13965432176@163.com>
* Added a library for daemon reload and fixing testcaseBala Konda Reddy M2019-09-041-8/+26
| | | | | | | | | | | | | | After changing the type of unit file from INFO to DEBUG. Performing daemon reload. Earlier using running commands continuosly to generated debug messages instead of running continuosly, restarted glusterd in one of the nodes so that while handshake the logs will be in Debug mode. After validating reverting back the unit file to INFO and daemon reload Change-Id: I8c99407eff2ea98a836f37fc2d89bb99f7eeccb7 Signed-off-by: Bala Konda Reddy M <bmekala@redhat.com>
* Fixes: changed triggering heal by getting md5sumkshithijiyer2019-09-041-8/+7
| | | | | | Change-Id: I4d056b94b4ea59beee7eb24e7e5d5f65d7256b4a Signed-off-by: Vitalii Koriakov <vkoriako@redhat.com> Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* [glusterfind] Test case to verify file renamesVinayak Papnoi2019-08-201-0/+231
| | | | | | | | | | | | | | | | | | Verifying the glusterfind functionality with renames of files. * Create a session on the volume * Create various files from mount point * Perform glusterfind pre * Perform glusterfind post * Check the contents of outfile * Rename the files created from mount point * Perform glusterfind pre * Perform glusterfind post * Check the contents of outfile Files renamed must be listed Change-Id: Ib7682e86d59f0519b267ec01cda999920a30de86 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [glusterfind] Test case to verify file deletesVinayak Papnoi2019-08-201-0/+239
| | | | | | | | | | | | | | | | | | Verifying the glusterfind functionality with deletion of files. * Create a session on the volume * Create various files from mount point * Perform glusterfind pre * Perform glusterfind post * Check the contents of outfile * Delete the files created from mount point * Perform glusterfind pre * Perform glusterfind post * Check the contents of outfile Files deleted must be listed Change-Id: I2ee05a2c97983fb521648e372db21c8361a2c835 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* glusterd test cases: Enabling and disabling shared storagekshithijiyer2019-08-191-0/+192
| | | | | | | | | | | | | | | | | | | | Steps: -> Enable a shared storage -> Disable a shared storage -> Create volume of any type with name gluster_shared_storage -> Disable the shared storage -> Check, volume created in step-3 is not deleted -> Delete the volume -> Enable the shared storage -> Check volume with name gluster_shared_storage is created -> Disable the shared storage Change-Id: I1fd29d51e32cadd7978771f4a37ac87176d90372 Signed-off-by: Gaurav Yadav <gyadav@redhat.com> Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* [Snapshot] Change shared volume mount validation functionVinayak Papnoi2019-08-081-65/+51
| | | | | | | | | | | The test case makes use of 'is_shared_volume_unmounted' function which is redundant as the purpose can be solved via a very similar function 'is_shared_volume_mounted'. This change has been implemented. There are various cosmetic changes that have been implemented as well. Change-Id: I560b464b4bcc436658db49c0a5ed8c7aadacfb6a Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* Fix AFR test case tearDownVinayak Papnoi2019-08-081-6/+18
| | | | | | | | | | | | The test case 'test_client_side_quorum_with_fixed_for_cross3' does not include the tearDown part where the volume options which have been set inside the test case have to be reset to default. This fix includes the necessary tearDown steps along with a few cosmetic changes. Change-Id: I86187cef4523492ec97707ff93d0eca365293008 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* Adding test case to mount volume, remove /var/log/glusterfs/ and remount the ↵kshithijiyer2019-08-061-0/+60
| | | | | | | | | | | | | | | volume. Test case: 1. Create all types of volumes and start them. 2. Mount all volumes on clients. 3. Delete /var/log/glusterfs folder on client. 4. Run IO on all the mount points. 5. Unmount and remount all volumes. 6. Check if logs are regenerated or not. Change-Id: I4f90d709c4da6e1c73cf95f4075c50aa44cdd811 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Enable client-side heal as client-side healing is diabled in RHGS 3.5Anees Patel2019-08-051-0/+10
| | | | | Change-Id: I0992b1b9af4e12f4e20d7a5dc184048de104d89d Signed-off-by: Anees Patel <anepatel@redhat.com>
* Optimized test case: tests/functional/glusterd/test_remove_brick_scenarios.pyhadarsharon2019-08-051-16/+28
| | | | | | | | | | Worked on the following: Improved I/O performance of test (writing 100k files to a mounted volume) by applying the following changes: 1. Modified the touch command to write as many files as possible per process, thus requiring less processes to write the 100k files 2. Using Threads to parallelize the touch processes from within the test, for better efficiency Change-Id: Id969f387f4b7b8e88daf688f7bada950cff2c412 Signed-off-by: hadarsharon <hsharon@redhat.com>
* Adding test case to enable brickmux, create start and stop 3 volumes.kshithijiyer2019-07-291-0/+115
| | | | | | | | | | | Test Case: 1.Set cluster.brick-multiplex to enabled. 2.Create three 1x3 replica volumes. 3.Start all the three volumes. 4.Stop three volumes one by one. Change-Id: Ibf3e81e7424d6a429da0aa12efeae7fffd3338f2 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* [EC]Fix a script issue to change permission of the directoryubansal2019-07-291-2/+7
| | | | | | | | Added a line to change permission of the directory so that client side healing happens for the directory also Change-Id: If4a24f2dbd6c9c85d4cb2944d1ad4795dbc39adb Signed-off-by: ubansal <ubansal@redhat.com>
* Enable client-side heal as client-side healing is disabled by default in ↵Anees Patel2019-07-291-0/+10
| | | | | | | RHGS 3.5 Change-Id: I500912b5217b675f9fdff4fe1cb518b465de245c Signed-off-by: Anees Patel <anepatel@redhat.com>
* [glusterfind] Addition of 'Test' in the class name of glusterfind create testVinayak Papnoi2019-07-291-32/+24
| | | | | | | | | | | | | | The class name was missing 'Test' at the beginning. Changing the name of the class to 'TestGlusterFindCreateCLI' from 'GlusterFindCreateCLI'. Removing the setUpClass and tearDownClass and replacing them with setUp and tearDown. Changing the variable names to be intelligible. Change-Id: Ibb5d9c6ef75ef11960aad35d65c343fa08fc9de1 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* Adding testcase to remove /var/log/glusterfs and mounting volume.kshithijiyer2019-07-231-0/+125
| | | | | | | | | | | | | Test Case: 1. Create all types of volumes. 2. Start all volumes. 3. Delete /var/log/glusterfs folder on the client. 4. Mount all the volumes one by one. 5. Run IO on all the mount points. 6. Check if logs are generated in /var/log/glusterfs/. Change-Id: I7a3275aad940116c3506b22b13a670e455d9ef00 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Adding code to not stop glusterd on mnode.kshithijiyer2019-07-081-6/+28
| | | | | | | | | | | | | This test case was failing in a test run as the mnode was not removed from the list self.servers because of which there were runs where glusterd was stoppped instead and command was executed on mnod.As well as adding code to check and start glusterd on the node in instances where the test case fails. Change-Id: Id203102d3f0ec82af0ac215f0ecaf7ae22b630f5 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Modifying test_enabling_gluster_debug_mode to do more operations.kshithijiyer2019-07-021-3/+7
| | | | | | | | | | | | While running test_enabling_gluster_debug_mode through jenkins it was observed that running volume operation once wasn't generating enough of logs by the time the logs were checked which lead to failure of the test case in the jenkins run. So modifying the logic which generates logs to run operation in a loop to generated a good amount of logs. Change-Id: Id7a12c86a04dc86d4856dbe30d945e70e64ea4f7 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Converting text files to markdown files.kshithijiyer2019-06-302-21/+20
| | | | | | | | | It's better to use markdown files instead of text files for readme files. Hence converting readme files to readme.md files. Change-Id: I41c1b2f065895d885f4b1fabdc9b9e4051810e80 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* [glusterfind] Test case to verify the 'glusterfind pre' functionalityVinayak Papnoi2019-06-301-0/+204
| | | | | | | | | | | | | | | | | | | | | | | Verifying the 'glusterfind pre' command functionality with valid and invalid values for the required and optional parameters. * Create a session on the volume * Perform some I/O from the mount point * Perform glusterfind pre with the following combinations: - Valid values for required parameters - Invalid values for required parameters - Valid values for optional parameters - Invalid values for optional parameters * Perform glusterfind post Where Required parameters: volname, sessname and outfile Optional parameters: full, debug, gftype, tagforfullfind, namespace, noencode, disablepartial, regenoutfile, outprefix, fieldsep Change-Id: Id795d632ebb8f43f49a9d15d83aced87670007b6 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [glusterfind] Test case to verify the 'glusterfind post' functionalityVinayak Papnoi2019-06-301-0/+199
| | | | | | | | | | | | | | | | | | | | | Verifying the 'glusterfind post' command functionality with valid and invalid values for the required and optional parameters. * Create a session on the volume * Perform some I/O from the mount point * Perform glusterfind pre - Perform glusterfind post with the following combinations: - Valid values for required parameters - Invalid values for required parameters - Valid values for optional parameters - Invalid values for optional parameters Where Required parameters: volname and sessname Optional parameters: debug Change-Id: Ia4c59b5ea3626d83bd6cedfda44e04ddbfa1567d Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [glusterfind] Test case to verify the 'glusterfind query' functionalityVinayak Papnoi2019-06-301-0/+164
| | | | | | | | | | | | | | | | | | | | | | | Verifying the 'glusterfind query' command functionality with valid and invalid values for the required and optional parameters. *** glusterfind query command does not require a session to be created *** * Perform some I/O from the mount point * Perform glusterfind query with the following combinations: - Valid values for required parameters - Invalid values for required parameters - Valid values for optional parameters - Invalid values for optional parameters Where Required parameters: volname and outfile Optional parameters: full, debug, gftype, tagforfullfind, since, namespace, noencode, disablepartial, end, regenoutfile, outprefix, fieldsep Change-Id: I5b9dbc987cec0f4ff9e460f875da72de7c9fe1c5 Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [glusterfind] Test case to verify the glusterfind delete functionalityVinayak Papnoi2019-06-301-0/+145
| | | | | | | | | | | | | | | | | | | | | | Verifying the glusterfind delete command functionality with valid and invalid values for the required and optional parameters. * Create a volume * Create a session on the volume * Perform glusterfind list to check if session is created * Delete the glusterfind session with the following combinations: - Valid values for required parameters - Invalid values for required parameters - Valid values for optional parameters - Invalid values for optional parameters * Perform glusterfind list to check if session is deleted Where: Required parameters: volname and sessname Optional parameters: debug Change-Id: I8240ccfd728cd7aa138103ffe1c24a705aa1b51c Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [glusterfind] Test case to verify the glusterfind create functionalityVinayak Papnoi2019-06-172-0/+138
| | | | | | | | | | | | | | | | | | Verifying the glusterfind create command functionality with valid and invalid values for the required and optional parameters. * Create a volume * Create a session on the volume with the following combinations: - Valid values for required parameters - Invalid values for required parameters - Valid values for optional parameters - Invalid values for optional parameters Where: Required parameters: volname and sessname Optional parameters: debug, force, reset-session-time Change-Id: Ic46863f63679e3c26a95b1fa96138cb7d0c27a3d Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* Adding code to cleanup bricks in test_add_brick.pykshithijiyer2019-06-131-9/+18
| | | | | | | | | | | While executing the test suit it was observerd that the test case test_add_remove_brick was failing due to remains from the test case test_add_brick_functionality. Hence adding the code to clean all the briks post test in test_add_brick.py. Change-Id: Iace9e51582ab4fa1f0f184283e6205aa6140b4a2 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Optimized test case: tests/functional/glusterd/test_add_brickhadarsharon2019-06-091-65/+54
| | | | | | | | | | | Worked on the following: 1. Removed redundant throwaway variables (ret, _, _) 2. More consistent exceptions 3. Added comments within code 4. Clarified Error messages in case of Assertion Errors Change-Id: I8ca0acce848bd9a8a5d217b5a4e247590177154d Signed-off-by: hadarsharon <hsharon@redhat.com>
* [Snapshot] Fixing Jira ticket RHGSQE-233Vinayak Papnoi2019-06-061-56/+66
| | | | | | | | | Removing 'ExecutionError' raised inside the test case function. Adding the description for the test case function. Fixing the typo errors and various cosmetic changes. Change-Id: Ica505417f482c7a622be5da70c1053913cf3052c Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [Snapshot] Fixing Jira ticket RHGSQE-234Vinayak Papnoi2019-06-051-60/+71
| | | | | | | | | Removing 'ExecutionError' raised inside the test case function. Adding the description for the test case function. Fixing the typo errors and various cosmetic changes. Change-Id: I496dc1df067870d43e5c3c97fc27da4294862eba Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* glusto-tests/glusterd: enable glusterd in debug modekshithijiyer2019-05-291-0/+169
| | | | | | | | | | | | | | | | | In this test case we will enable glusterd in debug mode and check the glusterd log for the debug messages. Steps followed: 1. Stop glusterd. 2. Change log level to DEBUG in /usr/local/lib/systemd/system/glusterd.service. 3. Remove glusterd log. 4. Start glusterd. 5. Issue some gluster commands. 6. Check for debug messages in glusterd log. Change-Id: Id1173be6da2ef1c2233459fb23f4b27308c923f2 Signed-off-by: Sanju Rakonde <srakonde@redhat.com> Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* This TC is to ammend the previous TC to prevent it from failing - picking up ↵sayalee2019-05-281-7/+29
| | | | | | | random servers which are not part of volume Change-Id: I8dcb8fcd8fd0bfaefcaf24a3f2aa222b44ff7f24 Signed-off-by: sayalee <root@localhost.localdomain>
* Adding the assert statement removed during changes.kshithijiyer2019-05-231-0/+2
| | | | | | | | | It seems like an assert statement got missed duing the recent changes. Adding back the assert and submitting a patch Signed-off-by: kshithijiyer <kshithij.ki@gmail.com> Change-Id: Ic8ab0d6e54da510faf479cd09cf122ccf8cedfbb
* nfs-ganesha: Test to verify mounting volume on another client while IO is ↵Jilju Joy2019-05-231-0/+151
| | | | | | | progressing on one client Change-Id: I42dceccaa0f257e25a4e5f2bc8d21f84e94216e0 Signed-off-by: Jilju Joy <jijoy@redhat.com>
* nfs-ganesha: Changes in test cases based on latest libraryJilju Joy2019-05-231-211/+428
| | | | | Change-Id: I78f987383725ef43caa62ad084dd64ec6be2363b Signed-off-by: Jilju Joy <jijoy@redhat.com>
* Changing systemctl to service to fix jira issue RHGSQE-197kshithijiyer2019-05-231-2/+4
| | | | | | | | Bug https://bugzilla.redhat.com/show_bug.cgi?id=1690254 has to be fixed before merging this patch. Change-Id: I90e669269fafa9d0a064a64883c3e4b88080d25f Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Changing error messages to be checked as per new messages.kshithijiyer2019-05-081-2/+18
| | | | | | | | | | | | | | | | | | | | Changing error message displayed when peer detach is issued with bricks are present on the node which is being detached. Adding a logic to handle both the new as well as the old error message. Old msg: peer detach: failed: Brick(s) with the peer <my_server> exist in cluster New msg: peer detach: failed: Peer <my_server> hosts one or more bricks. If the peer is in not recoverable state then use either replace-brick or remove-brick command with force to remove all bricks from the peer and attempt the peer detach again. Change-Id: I3d8fdac2c33638ecc2a8b5782c68caebbf17cf41 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Changing error message to be check in test_volume_getkshithijiyer2019-05-061-0/+2
| | | | | | | | | | | | | | The error message which is displayed when we do a gluster v get for options which don't exist has been changed. Adding a if based logic which can check for the old as well as the new error message. Old msg: volume get option: failed: Did you mean auth.allow or ...reject? New msg: volume get option: failed: Did you mean ctime.noatime? Signed-off-by: kshithijiyer <kshithij.ki@gmail.com> Change-Id: I9496d391a7da9dba64d3426a024c2b1b68455f20
* Adding test to validate output of profile infokshithijiyer2019-05-031-0/+223
| | | | | | | | | | | | | | | Test Case: 1) Create a volume and start it. 2) Mount volume on client and start IO. 3) Start profile info on the volume. 4) Run profile info with different parameters and see if all bricks are present or not. 5) Stop profile on the volume. 6) Create another volume. 7) Start profile without starting the volume. Change-Id: I6e8ec9285d48c1c828cd1d20bff6ea8f3de064f7 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Adding test for profile operations with one node downkshithijiyer2019-04-291-0/+220
| | | | | | | | | | | | | | Test Case: 1) Create a volume and start it. 2) Mount volume on client and start IO. 3) Start profile info on the volume. 4) Stop glusterd on one node. 5) Run profile info with different parameters and see if all bricks are present or not. 6) Stop profile on the volume. Change-Id: Ie573414816362ebbe30d2c419fd0e348522ceaec Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test case to detach node used for mountingkshithijiyer2019-04-171-0/+219
| | | | | | | | | | | | | | | | | | | Test case: 1.Create a 1X3 volume with only 3 nodes from the cluster. 2.Mount volume on client node using the ip of the fourth node. 3.Write IOs to the volume. 4.Detach node N4 from cluster. 5.Create a new directory on the mount point. 6.Create a few files using the same command used in step 3. 7.Add three more bricks to make the volume 2x3 using add-brick command. 8.Do a gluster volume rebalance on the volume. 9.Create more files from the client on the mount point. 10.Check for files on bricks from both replica sets. 11.Create a new directory from the client on the mount point. 12.Check for directory in both replica sets. Change-Id: I228b79955dca565a40994919b2903e59cad7d8f5 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>