summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--cli/src/cli-rpc-ops.c23
-rw-r--r--tests/basic/tier/new-tier-cmds.t64
2 files changed, 63 insertions, 24 deletions
diff --git a/cli/src/cli-rpc-ops.c b/cli/src/cli-rpc-ops.c
index 9bef143574c..a2519e8c0e8 100644
--- a/cli/src/cli-rpc-ops.c
+++ b/cli/src/cli-rpc-ops.c
@@ -1557,10 +1557,18 @@ gf_cli_print_rebalance_status (dict_t *dict, enum gf_task_types task_type)
snprintf (key, 256, "status-%d", i);
ret = dict_get_int32 (dict, key, (int32_t *)&status_rcd);
- if (ret) {
+ if (ret == -ENOENT) {
gf_log ("cli", GF_LOG_TRACE, "count %d %d", count, i);
gf_log ("cli", GF_LOG_TRACE, "failed to get status");
- goto out;
+ gf_log ("cli", GF_LOG_ERROR, "node down and has failed"
+ " to set dict");
+ continue;
+ /* skip this node if value not available*/
+ } else if (ret) {
+ gf_log ("cli", GF_LOG_TRACE, "count %d %d", count, i);
+ gf_log ("cli", GF_LOG_TRACE, "failed to get status");
+ continue;
+ /* skip this node if value not available*/
}
if (GF_DEFRAG_STATUS_NOT_STARTED == status_rcd)
@@ -1682,10 +1690,17 @@ gf_cli_print_tier_status (dict_t *dict, enum gf_task_types task_type)
snprintf (key, 256, "status-%d", i);
ret = dict_get_int32 (dict, key, (int32_t *)&status_rcd);
- if (ret) {
+ if (ret == -ENOENT) {
gf_log ("cli", GF_LOG_TRACE, "count: %d, %d,"
"failed to get status", count, i);
- goto out;
+ gf_log ("cli", GF_LOG_ERROR, "node down and has failed"
+ " to set dict");
+ continue;
+ /*skipping this node as value unavailable*/
+ } else if (ret) {
+ gf_log ("cli", GF_LOG_TRACE, "count: %d, %d,"
+ "failed to get status", count, i);
+ continue;
}
if (GF_DEFRAG_STATUS_NOT_STARTED == status_rcd)
diff --git a/tests/basic/tier/new-tier-cmds.t b/tests/basic/tier/new-tier-cmds.t
index c38ecf9d7db..ce8dbae8660 100644
--- a/tests/basic/tier/new-tier-cmds.t
+++ b/tests/basic/tier/new-tier-cmds.t
@@ -3,33 +3,42 @@
. $(dirname $0)/../../include.rc
. $(dirname $0)/../../volume.rc
. $(dirname $0)/../../tier.rc
+. $(dirname $0)/../../cluster.rc
# Creates a tiered volume with pure distribute hot and cold tiers
# Both hot and cold tiers will have an equal number of bricks.
+function check_peers {
+ $CLI_1 peer status | grep 'Peer in Cluster (Connected)' | wc -l
+}
+
function create_dist_tier_vol () {
- mkdir $B0/cold
- mkdir $B0/hot
- TEST $CLI volume create $V0 disperse 6 disperse-data 4 $H0:$B0/cold/${V0}{1..12}
- TEST $CLI volume set $V0 performance.quick-read off
- TEST $CLI volume set $V0 performance.io-cache off
- TEST $CLI volume start $V0
- TEST $CLI volume attach-tier $V0 replica 2 $H0:$B0/hot/${V0}{0..5}
- TEST $CLI volume set $V0 cluster.tier-mode test
+ TEST $CLI_1 volume create $V0 $H1:$B1/${V0} $H2:$B2/${V0} $H3:$B3/${V0}
+ TEST $CLI_1 volume start $V0
+ TEST $CLI_1 volume attach-tier $V0 $H1:$B1/${V0}_h1 $H2:$B2/${V0}_h2 $H3:$B3/${V0}_h3
}
function tier_detach_commit () {
- $CLI volume tier $V0 detach commit | grep "success" | wc -l
+ $CLI_1 volume tier $V0 detach commit | grep "success" | wc -l
+}
+
+function tier_detach_status_node_down () {
+ $CLI_1 volume tier $V0 detach status | wc -l
+}
+
+function tier_status_node_down () {
+ $CLI_1 volume tier $V0 status | wc -l
}
cleanup;
-#Basic checks
-TEST glusterd
-TEST pidof glusterd
-TEST $CLI volume status
+#setup cluster and test volume
+TEST launch_cluster 3; # start 3-node virtual cluster
+TEST $CLI_1 peer probe $H2; # peer probe server 2 from server 1 cli
+TEST $CLI_1 peer probe $H3; # peer probe server 3 from server 1 cli
+EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers;
#Create and start a tiered volume
create_dist_tier_vol
@@ -37,21 +46,36 @@ create_dist_tier_vol
#Issue detach tier on the tiered volume
#Will throw error saying detach tier not started
-EXPECT "Tier command failed" $CLI volume tier $V0 detach status
+EXPECT "Tier command failed" $CLI_1 volume tier $V0 detach status
#after starting detach tier the detach tier status should display the status
-TEST $CLI volume tier $V0 detach start
+TEST $CLI_1 volume tier $V0 detach start
+
+TEST $CLI_1 volume tier $V0 detach status
+
+#kill a node
+TEST kill_node 2
+
+#check if we have the rest of the node available printed in the output of detach status
+EXPECT_WITHIN $PROCESS_UP_TIMEOUT "4" tier_detach_status_node_down
+
+#check if we have the rest of the node available printed in the output of tier status
+EXPECT_WITHIN $PROCESS_UP_TIMEOUT "5" tier_status_node_down
+
+TEST $glusterd_2;
+
+EXPECT_WITHIN $PROBE_TIMEOUT 2 check_peers;
-TEST $CLI volume tier $V0 detach status
+TEST $CLI_1 volume tier $V0 detach status
-TEST $CLI volume tier $V0 detach stop
+TEST $CLI_1 volume tier $V0 detach stop
#If detach tier is stopped the detach tier command will fail
-EXPECT "Tier command failed" $CLI volume tier $V0 detach status
+EXPECT "Tier command failed" $CLI_1 volume tier $V0 detach status
-TEST $CLI volume tier $V0 detach start
+TEST $CLI_1 volume tier $V0 detach start
#wait for the detach to complete
EXPECT_WITHIN $PROCESS_UP_TIMEOUT "1" tier_detach_commit
@@ -59,7 +83,7 @@ EXPECT_WITHIN $PROCESS_UP_TIMEOUT "1" tier_detach_commit
#If detach tier is committed then the detach status should fail throwing an error
#saying its not a tiered volume
-EXPECT "Tier command failed" $CLI volume tier $V0 detach status
+EXPECT "Tier command failed" $CLI_1 volume tier $V0 detach status
cleanup;