summaryrefslogtreecommitdiffstats
path: root/tests
Commit message (Collapse)AuthorAgeFilesLines
* [Snapshot] Fixing Jira ticket RHGSQE-233Vinayak Papnoi2019-06-061-56/+66
| | | | | | | | | Removing 'ExecutionError' raised inside the test case function. Adding the description for the test case function. Fixing the typo errors and various cosmetic changes. Change-Id: Ica505417f482c7a622be5da70c1053913cf3052c Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* [Snapshot] Fixing Jira ticket RHGSQE-234Vinayak Papnoi2019-06-051-60/+71
| | | | | | | | | Removing 'ExecutionError' raised inside the test case function. Adding the description for the test case function. Fixing the typo errors and various cosmetic changes. Change-Id: I496dc1df067870d43e5c3c97fc27da4294862eba Signed-off-by: Vinayak Papnoi <vpapnoi@redhat.com>
* glusto-tests/glusterd: enable glusterd in debug modekshithijiyer2019-05-291-0/+169
| | | | | | | | | | | | | | | | | In this test case we will enable glusterd in debug mode and check the glusterd log for the debug messages. Steps followed: 1. Stop glusterd. 2. Change log level to DEBUG in /usr/local/lib/systemd/system/glusterd.service. 3. Remove glusterd log. 4. Start glusterd. 5. Issue some gluster commands. 6. Check for debug messages in glusterd log. Change-Id: Id1173be6da2ef1c2233459fb23f4b27308c923f2 Signed-off-by: Sanju Rakonde <srakonde@redhat.com> Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* This TC is to ammend the previous TC to prevent it from failing - picking up ↵sayalee2019-05-281-7/+29
| | | | | | | random servers which are not part of volume Change-Id: I8dcb8fcd8fd0bfaefcaf24a3f2aa222b44ff7f24 Signed-off-by: sayalee <root@localhost.localdomain>
* Adding the assert statement removed during changes.kshithijiyer2019-05-231-0/+2
| | | | | | | | | It seems like an assert statement got missed duing the recent changes. Adding back the assert and submitting a patch Signed-off-by: kshithijiyer <kshithij.ki@gmail.com> Change-Id: Ic8ab0d6e54da510faf479cd09cf122ccf8cedfbb
* nfs-ganesha: Test to verify mounting volume on another client while IO is ↵Jilju Joy2019-05-231-0/+151
| | | | | | | progressing on one client Change-Id: I42dceccaa0f257e25a4e5f2bc8d21f84e94216e0 Signed-off-by: Jilju Joy <jijoy@redhat.com>
* nfs-ganesha: Changes in test cases based on latest libraryJilju Joy2019-05-231-211/+428
| | | | | Change-Id: I78f987383725ef43caa62ad084dd64ec6be2363b Signed-off-by: Jilju Joy <jijoy@redhat.com>
* Changing systemctl to service to fix jira issue RHGSQE-197kshithijiyer2019-05-231-2/+4
| | | | | | | | Bug https://bugzilla.redhat.com/show_bug.cgi?id=1690254 has to be fixed before merging this patch. Change-Id: I90e669269fafa9d0a064a64883c3e4b88080d25f Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Changing error messages to be checked as per new messages.kshithijiyer2019-05-081-2/+18
| | | | | | | | | | | | | | | | | | | | Changing error message displayed when peer detach is issued with bricks are present on the node which is being detached. Adding a logic to handle both the new as well as the old error message. Old msg: peer detach: failed: Brick(s) with the peer <my_server> exist in cluster New msg: peer detach: failed: Peer <my_server> hosts one or more bricks. If the peer is in not recoverable state then use either replace-brick or remove-brick command with force to remove all bricks from the peer and attempt the peer detach again. Change-Id: I3d8fdac2c33638ecc2a8b5782c68caebbf17cf41 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Changing error message to be check in test_volume_getkshithijiyer2019-05-061-0/+2
| | | | | | | | | | | | | | The error message which is displayed when we do a gluster v get for options which don't exist has been changed. Adding a if based logic which can check for the old as well as the new error message. Old msg: volume get option: failed: Did you mean auth.allow or ...reject? New msg: volume get option: failed: Did you mean ctime.noatime? Signed-off-by: kshithijiyer <kshithij.ki@gmail.com> Change-Id: I9496d391a7da9dba64d3426a024c2b1b68455f20
* Adding test to validate output of profile infokshithijiyer2019-05-031-0/+223
| | | | | | | | | | | | | | | Test Case: 1) Create a volume and start it. 2) Mount volume on client and start IO. 3) Start profile info on the volume. 4) Run profile info with different parameters and see if all bricks are present or not. 5) Stop profile on the volume. 6) Create another volume. 7) Start profile without starting the volume. Change-Id: I6e8ec9285d48c1c828cd1d20bff6ea8f3de064f7 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Adding test for profile operations with one node downkshithijiyer2019-04-291-0/+220
| | | | | | | | | | | | | | Test Case: 1) Create a volume and start it. 2) Mount volume on client and start IO. 3) Start profile info on the volume. 4) Stop glusterd on one node. 5) Run profile info with different parameters and see if all bricks are present or not. 6) Stop profile on the volume. Change-Id: Ie573414816362ebbe30d2c419fd0e348522ceaec Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test case to detach node used for mountingkshithijiyer2019-04-171-0/+219
| | | | | | | | | | | | | | | | | | | Test case: 1.Create a 1X3 volume with only 3 nodes from the cluster. 2.Mount volume on client node using the ip of the fourth node. 3.Write IOs to the volume. 4.Detach node N4 from cluster. 5.Create a new directory on the mount point. 6.Create a few files using the same command used in step 3. 7.Add three more bricks to make the volume 2x3 using add-brick command. 8.Do a gluster volume rebalance on the volume. 9.Create more files from the client on the mount point. 10.Check for files on bricks from both replica sets. 11.Create a new directory from the client on the mount point. 12.Check for directory in both replica sets. Change-Id: I228b79955dca565a40994919b2903e59cad7d8f5 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test to profile start when quorum not metBala Konda Reddy M2019-04-161-0/+147
| | | | | | | | | | | | | 1. Create a volume 2. Set the quorum type to server and ratio to 90 3. Stop glusterd randomly on one of the node 4. Start profile on the volume 5. Start glusterd on the node where it is stopped 6. Start profile on the volume 7. Stop profile on the volume where it is started Change-Id: Ifeb9fddf6f1a14c9df73ed2f0453636d2853e944 Signed-off-by: Bala Konda Reddy M <bmekala@redhat.com>
* Modificatons to base class to include slave infoRoch-elle2019-04-121-0/+110
| | | | | | | Added new config file for georep Change-Id: I544ee35f8cd6a2f2ed744f72caa28fa4fd63e5b3 Signed-off-by: Roch-elle <rallan@redhat.com>
* Adding test to run gluster commands when glusterd is down on one nodekshithijiyer2019-04-121-0/+102
| | | | | Change-Id: Ibf41c11a4e98baeaad658ee10ba8a807318504be Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Validating whether peers are connected or not before volume creationBala Konda Reddy M2019-04-121-1/+17
| | | | | | | | | | In jenkins this case is failing with peers are not connected while volume creation. Now having a check before creating the volume to make sure that peers are in cluster and in connected state after peer probe. Change-Id: I8aa9d2c4d1669475dd8867d42752a31604ff572f Signed-off-by: Bala Konda Reddy M <bmekala@redhat.com>
* Adding code for cleanup of all bricks on each serverkshithijiyer2019-04-091-2/+18
| | | | | Change-Id: I405843e0093ddb7138ee0a8afbfd4cd2f91e6284 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Checking if peers are connected after peer probekshithijiyer2019-04-091-0/+24
| | | | | Change-Id: I252ab0c0f6248b9a5c1d7977146c15876e144b38 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* This TC is to check for File operations in an EC volumeubansal2019-04-081-0/+337
| | | | | Change-Id: Ib1aff1c1bf843dddac5862e55a049d7b47603049 Signed-off-by: ubansal <ubansal@redhat.com>
* Removing test_volume_sanity TCAkarsha2019-04-011-106/+0
| | | | | Change-Id: I82c76f26f4bf359ce153ff35850ce5c349bc759b Signed-off-by: Akarsha <akrai@redhat.com>
* Adding code to check if peers are connected in test_spurious_rebalancekshithijiyer2019-03-291-1/+13
| | | | | Change-Id: I4a1097fbdebd49555fffcfa5fe609f4070e39182 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Validating peer is in connected or notBala Konda Reddy M2019-03-281-0/+15
| | | | | | | | | In jenkins right after peer probe, add brick function is failing with peer not in cluster. So having check for the peer to be connected or not then proceed to next step. Change-Id: I73bf92819ad44f7a6a14795ab07c45d260cd04eb Signed-off-by: Bala Konda Reddy M <bmekala@redhat.com>
* Test case to change reserve limit with remove brick in progresskshithijiyer2019-03-131-0/+270
| | | | | Change-Id: I53fb7f4cceae395698568129669dc5f3a9a5e4bb Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Fixing busy umount errorsAnees Patel2019-02-251-0/+239
| | | | | Change-Id: I7f7b5cfdee09067d8d96bfcf56ce8a3372ca9368 Signed-off-by: Anees Patel <anepatel@redhat.com>
* Test to check remove brick scenarios wrt to glusterdkshithijiyer2019-02-201-0/+188
| | | | | Change-Id: I1bfa2fb3ae4ff1fc247b40c73f4fade9a3afeede Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test to reduce volume from 2x3 to 2x2 and to distkshithijiyer2019-02-121-0/+115
| | | | | Change-Id: I64309c3b46dc9087eeb3181acba63b981b2ecc6f Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test to check volume create with glusterd restart on one nodekshithijiyer2019-02-121-0/+116
| | | | | Change-Id: Ica0771bdee1e96e9d6bb5157fb6c2125a4b419f1 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Test case to check peer detach warningkshithijiyer2019-02-111-0/+83
| | | | | Change-Id: Idc379ad7f31274cc63f384d7223bf769bb89ace3 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Fixes for checking uid, gid and permissionVitalii Koriakov2019-02-081-60/+79
| | | | | Change-Id: I683e6ff47120b7db8ee6ae02ed83eba19e6ac4c9 Signed-off-by: Vitalii Koriakov <vkoriako@redhat.com>
* This Tc is used to check remove brick on a dispersed volumeubansal2019-02-071-0/+166
| | | | | Change-Id: Id8cfc0dd31cf4f6f381ec7bb07d4aba06d52b43e Signed-off-by: ubansal <ubansal@redhat.com>
* Test to check if setting volume level options at cluster level is crashing ↵kshithijiyer2019-02-071-0/+72
| | | | | | | glusterd Change-Id: I6ee034f019a4aa36a83e087f2d9fed007e4fd9d7 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* This testcase checks add-brick on a dispersed volumeubansal2019-02-071-0/+186
| | | | | Change-Id: I640f5c554fab791aa5f196415c5204f7cbca83a4 Signed-off-by: ubansal <ubansal@redhat.com>
* Increasing the wait time for rebalance after remove brick operationJilju Joy2019-02-051-1/+1
| | | | Change-Id: Idc864d0efb02d021d7e30908f1a9ef5553afee65
* Increase wait time for the completion of rebalance. Correction in assert ↵Jilju Joy2019-01-311-4/+4
| | | | | | statement Change-Id: I4333ab2eb63271c5c35756798ea515c7a4d4a3fb
* Delete quorum method from old file. Added fixes for flake8Vitalii Koriakov2019-01-303-801/+140
| | | | | Change-Id: I2acf835a4cf7301c64c4c8a9423f78672cdf9aa4 Signed-off-by: Vitalii Koriakov <vkoriako@redhat.com>
* New quorum behavior fixesVitalii Koriakov2019-01-281-0/+651
| | | | | Change-Id: I364054c35f623893700798bedef965fe05f6aabf Signed-off-by: Vitalii Koriakov <vkoriako@redhat.com>
* Fixed jira issue RHGSQE-33kshithijiyer2019-01-281-3/+10
| | | | | | | | Altered code to check for daemons only on servers where the bricks for a given volume are present. Change-Id: I79312f3b09fd5e1b0fdf6db40e29481662e56303 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Delete test_client_side_quorum_with_auto_option and ↵Vitalii Koriakov2019-01-251-360/+0
| | | | | | | test_client_side_quorum_with_auto_option_overwrite_fixed Change-Id: I5e22228eaf8574f2ccb1ae38cb98ec01e6493fdf Signed-off-by: Vitalii Koriakov <vkoriako@redhat.com>
* fixing the quorum changes according to new behaviourVijay Avuthu2019-01-251-3/+4
| | | | | | | If quorum is not met, reads/writes fails with transport end point not connected Change-Id: I219c99fc5b96147c059174daf0383454e1bd2831
* New quorum behavior fixesVitalii Koriakov2019-01-251-0/+333
| | | | | Change-Id: I0a2c0ba2e28fc23fe3ff2db57b4ba3c0f08993aa Signed-off-by: Vitalii Koriakov <vkoriako@redhat.com>
* New quorum behavior fixesVitalii Koriakov2019-01-251-0/+218
| | | | | Change-Id: I156e80e958d9e4c7aeec3a97bbcb16e8bfa36f30 Signed-off-by: Vitalii Koriakov <vkoriako@redhat.com>
* This testcase checks that EC version of bricks are not updated when non FOPs ↵ubansal2019-01-251-0/+115
| | | | | | operation are performed like chmod Change-Id: I797253cd4454359bd8f0596c322b2eb71a8a4751
* Correction in verification of volume/sub_directory unmountJilju Joy2019-01-243-6/+6
| | | | Change-Id: I2abad563b965d67520b8bea6da39b30422d7d028
* Fixed jira issue RHGSQE-29 and added quorum reset codekshithijiyer2019-01-242-30/+50
| | | | | Change-Id: Ibd50170d2c3172d7b98c2174630d31a066762f7c Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Fixed jira issue RHGSQE-31kshithijiyer2019-01-231-13/+16
| | | | | Change-Id: I627c78792c6c1ea12c4a023095a4a983f8cee9b0 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Reset quorum value to defaultkshithijiyer2019-01-231-0/+9
| | | | | Change-Id: I0486abff96ea3ea626ce4d18ac0c24f10ed6a846 Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Reverting quorum value back to 51%kshithijiyer2019-01-231-0/+7
| | | | | Change-Id: I9e61480ff93d0c16b66eeddacdaef715a0b47d1c Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Fix in teardown()kshithijiyer2019-01-221-10/+19
| | | | | Change-Id: Ie44703a9d114b9ecaa5bbce07a98c8a040393f2c Signed-off-by: kshithijiyer <kshithij.ki@gmail.com>
* Fix: Volume create should fail when a node is downBala Konda Reddy M2019-01-221-2/+6
| | | | | | | | | | | | Earlier while volume creation randomly selected one node and stopped gluster but the volume type used in the test is pure distribute and default glusto config takes 4 bricks. With the fix now randomly selecting only first 4 nodes for volume creation and it will fail. Change-Id: I3cf2fc8281c9747b190e1fe9ef471edbb2c4d2ca Signed-off-by: Bala Konda Reddy M <bmekala@redhat.com>