diff options
| -rw-r--r-- | tests/bugs/glusterd/bug-1238706-daemons-stop-on-peer-cleanup.t | 41 | ||||
| -rw-r--r-- | tests/volume.rc | 16 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-messages.h | 16 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-sm.c | 40 | 
4 files changed, 90 insertions, 23 deletions
diff --git a/tests/bugs/glusterd/bug-1238706-daemons-stop-on-peer-cleanup.t b/tests/bugs/glusterd/bug-1238706-daemons-stop-on-peer-cleanup.t new file mode 100644 index 00000000000..9ff1758f9c1 --- /dev/null +++ b/tests/bugs/glusterd/bug-1238706-daemons-stop-on-peer-cleanup.t @@ -0,0 +1,41 @@ +#!/bin/bash + +## Test case for stopping all running daemons service on peer detach. + +. $(dirname $0)/../../include.rc +. $(dirname $0)/../../volume.rc +. $(dirname $0)/../../cluster.rc + +cleanup; + + +## Start a 2 node virtual cluster +TEST launch_cluster 2; + +## Peer probe server 2 from server 1 cli +TEST $CLI_1 peer probe $H2; + +EXPECT_WITHIN $PROBE_TIMEOUT 1 peer_count + + +## Creating and starting volume +TEST $CLI_1 volume create $V0 $H1:$B1/${V0}0 $H1:$B1/${V0}1 +TEST $CLI_1 volume start $V0 + +## To Do: Add test case for quota and snapshot daemon. Currently quota +##        Daemon is not working in cluster framework. And sanpd daemon +##        Start only in one node in cluster framework. Add test case +##        once patch http://review.gluster.org/#/c/11666/ merged, + +## We are having 2 node "nfs" daemon should run on both node. +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "2" get_nfs_count + +## Detach 2nd node from the cluster. +TEST $CLI_1 peer detach $H2; + + +## After detaching 2nd node we will have only 1 nfs and quota daemon running. +EXPECT_WITHIN $PROCESS_UP_TIMEOUT "1" get_nfs_count + +cleanup; + diff --git a/tests/volume.rc b/tests/volume.rc index e397f093a13..a100bde55ae 100644 --- a/tests/volume.rc +++ b/tests/volume.rc @@ -547,6 +547,14 @@ function get_quotad_count {          ps auxww | grep glusterfs | grep quotad.pid | grep -v grep | wc -l  } +function get_nfs_count { +        ps auxww | grep glusterfs | grep nfs.pid | grep -v grep | wc -l +} + +function get_snapd_count { +        ps auxww | grep glusterfs | grep snapd.pid | grep -v grep | wc -l +} +  function drop_cache() {  	case $OSTYPE in  	Linux) @@ -601,12 +609,4 @@ function quota_hl_exceeded()  } -function get_nfs_count { -        ps auxww | grep glusterfs | grep nfs.pid | grep -v grep | wc -l -} - -function get_snapd_count { -        ps auxww | grep glusterfs | grep snapd.pid | grep -v grep | wc -l -} - diff --git a/xlators/mgmt/glusterd/src/glusterd-messages.h b/xlators/mgmt/glusterd/src/glusterd-messages.h index 1282787410c..c5603254aa9 100644 --- a/xlators/mgmt/glusterd/src/glusterd-messages.h +++ b/xlators/mgmt/glusterd/src/glusterd-messages.h @@ -4589,6 +4589,22 @@   */  #define GD_MSG_SVC_STOP_SUCCESS                     (GLUSTERD_COMP_BASE + 568) +/*! + * @messageid + * @diagnosis + * @recommendedaction + * + */ +#define GD_MSG_PARAM_NULL                           (GLUSTERD_COMP_BASE + 569) + +/*! + * @messageid + * @diagnosis + * @recommendedaction + * + */ +#define GD_MSG_SVC_STOP_FAIL                        (GLUSTERD_COMP_BASE + 570) +  /*------------*/  #define glfs_msg_end_x GLFS_MSGID_END, "Invalid: End of messages"  #endif /* !_GLUSTERD_MESSAGES_H_ */ diff --git a/xlators/mgmt/glusterd/src/glusterd-sm.c b/xlators/mgmt/glusterd/src/glusterd-sm.c index 0455c54964a..ce2110aef61 100644 --- a/xlators/mgmt/glusterd/src/glusterd-sm.c +++ b/xlators/mgmt/glusterd/src/glusterd-sm.c @@ -31,6 +31,7 @@  #include "glusterd-op-sm.h"  #include "glusterd-utils.h"  #include "glusterd-store.h" +#include "glusterd-svc-helper.h"  #include "glusterd-snapshot-utils.h"  #include "glusterd-server-quorum.h" @@ -596,17 +597,29 @@ out:  /* Clean up stale volumes on the peer being detached. The volumes which have   * bricks on other peers are stale with respect to the detached peer.   */ -static int +static void  glusterd_peer_detach_cleanup (glusterd_conf_t *priv)  { -        int                     ret = -1; -        glusterd_volinfo_t      *volinfo = NULL; +        int                     ret          = -1; +        glusterd_volinfo_t      *volinfo     = NULL;          glusterd_volinfo_t      *tmp_volinfo = NULL; +        glusterd_svc_t          *svc         = NULL;          GF_ASSERT (priv);          cds_list_for_each_entry_safe (volinfo, tmp_volinfo, &priv->volumes,                                        vol_list) { +                /* Stop snapd daemon service if snapd daemon is running*/ +                if (!volinfo->is_snap_volume) { +                        svc = &(volinfo->snapd.svc); +                        ret = svc->stop (svc, SIGTERM); +                        if (ret) { +                                gf_msg (THIS->name, GF_LOG_ERROR, 0, +                                        GD_MSG_SVC_STOP_FAIL, "Failed to " +                                        "stop snapd daemon service."); +                        } +                } +                  /* The peer detach checks make sure that, at this point in the                   * detach process, there are only volumes contained completely                   * within or completely outside the detached peer. @@ -623,14 +636,17 @@ glusterd_peer_detach_cleanup (glusterd_conf_t *priv)                                  gf_msg (THIS->name, GF_LOG_ERROR, 0,                                          GD_MSG_STALE_VOL_REMOVE_FAIL,                                          "Error deleting stale volume"); -                                goto out;                          }                  }          } -        ret = 0; -out: -        gf_msg_debug (THIS->name, 0, "Returning %d", ret); -        return ret; + +        /* Stop all daemon services of Detaching node once  peer detached */ +        ret = glusterd_svcs_stop (); +        if (ret) { +                gf_msg (THIS->name, GF_LOG_ERROR, 0, +                        GD_MSG_SVC_STOP_FAIL, +                        "Failed to stop all daemon services."); +        }  }  static int @@ -675,13 +691,7 @@ glusterd_ac_handle_friend_remove_req (glusterd_friend_sm_event_t *event,          }          rcu_read_unlock (); -        ret = glusterd_peer_detach_cleanup (priv); -        if (ret) { -                gf_msg (THIS->name, GF_LOG_WARNING, 0, -                        GD_MSG_PEER_DETACH_CLEANUP_FAIL, -                        "Peer detach cleanup was not successful"); -                ret = 0; -        } +        glusterd_peer_detach_cleanup (priv);  out:          if (new_event)                  GF_FREE (new_event->peername);  | 
