diff options
| author | Jeff Darcy <jdarcy@redhat.com> | 2016-12-08 16:24:15 -0500 | 
|---|---|---|
| committer | Vijay Bellur <vbellur@redhat.com> | 2017-01-30 19:13:58 -0500 | 
| commit | 1a95fc3036db51b82b6a80952f0908bc2019d24a (patch) | |
| tree | b983ac196a8165d5cb5e860a5ef97d3e9a41b5c9 /xlators/mgmt/glusterd/src/glusterd-replace-brick.c | |
| parent | 7f7d7a939e46b330a084d974451eee4757ba61b4 (diff) | |
core: run many bricks within one glusterfsd process
This patch adds support for multiple brick translator stacks running
in a single brick server process.  This reduces our per-brick memory usage by
approximately 3x, and our appetite for TCP ports even more.  It also creates
potential to avoid process/thread thrashing, and to improve QoS by scheduling
more carefully across the bricks, but realizing that potential will require
further work.
Multiplexing is controlled by the "cluster.brick-multiplex" global option.  By
default it's off, and bricks are started in separate processes as before.  If
multiplexing is enabled, then *compatible* bricks (mostly those with the same
transport options) will be started in the same process.
Change-Id: I45059454e51d6f4cbb29a4953359c09a408695cb
BUG: 1385758
Signed-off-by: Jeff Darcy <jdarcy@redhat.com>
Reviewed-on: https://review.gluster.org/14763
Smoke: Gluster Build System <jenkins@build.gluster.org>
NetBSD-regression: NetBSD Build System <jenkins@build.gluster.org>
CentOS-regression: Gluster Build System <jenkins@build.gluster.org>
Reviewed-by: Vijay Bellur <vbellur@redhat.com>
Diffstat (limited to 'xlators/mgmt/glusterd/src/glusterd-replace-brick.c')
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-replace-brick.c | 27 | 
1 files changed, 0 insertions, 27 deletions
diff --git a/xlators/mgmt/glusterd/src/glusterd-replace-brick.c b/xlators/mgmt/glusterd/src/glusterd-replace-brick.c index eb1a714bfd5..fb29c6efcfd 100644 --- a/xlators/mgmt/glusterd/src/glusterd-replace-brick.c +++ b/xlators/mgmt/glusterd/src/glusterd-replace-brick.c @@ -326,22 +326,6 @@ out:          return ret;  } -static int -rb_kill_destination_brick (glusterd_volinfo_t *volinfo, -                           glusterd_brickinfo_t *dst_brickinfo) -{ -        glusterd_conf_t  *priv               = NULL; -        char              pidfile[PATH_MAX]  = {0,}; - -        priv = THIS->private; - -        snprintf (pidfile, PATH_MAX, "%s/vols/%s/%s", -                  priv->workdir, volinfo->volname, -                  RB_DSTBRICK_PIDFILE); - -        return glusterd_service_stop ("brick", pidfile, SIGTERM, _gf_true); -} -  int  glusterd_op_perform_replace_brick (glusterd_volinfo_t  *volinfo, @@ -526,17 +510,6 @@ glusterd_op_replace_brick (dict_t *dict, dict_t *rsp_dict)                  goto out;          } -        if (gf_is_local_addr (dst_brickinfo->hostname)) { -                gf_msg_debug (this->name, 0, "I AM THE DESTINATION HOST"); -                ret = rb_kill_destination_brick (volinfo, dst_brickinfo); -                if (ret) { -                        gf_msg (this->name, GF_LOG_CRITICAL, 0, -                                GD_MSG_BRK_CLEANUP_FAIL, -                                "Unable to cleanup dst brick"); -                        goto out; -                } -        } -          ret = glusterd_svcs_stop (volinfo);          if (ret) {                  gf_msg (this->name, GF_LOG_ERROR, 0,  | 
