summaryrefslogtreecommitdiffstats
path: root/doc/admin-guide/en-US/markdown/admin_geo-replication.md
blob: 849957244a9da0b7843b9d122b65eb51e490bb07 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
Managing Geo-replication
========================

Geo-replication provides a continuous, asynchronous, and incremental
replication service from one site to another over Local Area Networks
(LANs), Wide Area Network (WANs), and across the Internet.

Geo-replication uses a master–slave model, whereby replication and
mirroring occurs between the following partners:

-   Master – a GlusterFS volume

-   Slave – a slave which can be of the following types:

    -   A local directory which can be represented as file URL like
        `file:///path/to/dir`. You can use shortened form, for example,
        ` /path/to/dir`.

    -   A GlusterFS Volume - Slave volume can be either a local volume
        like `gluster://localhost:volname` (shortened form - `:volname`)
        or a volume served by different host like
        `gluster://host:volname` (shortened form - `host:volname`).

    > **Note**
    >
    > Both of the above types can be accessed remotely using SSH tunnel.
    > To use SSH, add an SSH prefix to either a file URL or gluster type
    > URL. For example, ` ssh://root@remote-host:/path/to/dir`
    > (shortened form - `root@remote-host:/path/to/dir`) or
    > `ssh://root@remote-host:gluster://localhost:volname` (shortened
    > from - `root@remote-host::volname`).

This section introduces Geo-replication, illustrates the various
deployment scenarios, and explains how to configure the system to
provide replication and mirroring in your environment.

Replicated Volumes vs Geo-replication
=====================================

The following table lists the difference between replicated volumes and
geo-replication:

  Replicated Volumes                                                                      Geo-replication
  --------------------------------------------------------------------------------------- -----------------------------------------------------------------------------------------------------------------
  Mirrors data across clusters                                                            Mirrors data across geographically distributed clusters
  Provides high-availability                                                              Ensures backing up of data for disaster recovery
  Synchronous replication (each and every file operation is sent across all the bricks)   Asynchronous replication (checks for the changes in files periodically and syncs them on detecting differences)

Preparing to Deploy Geo-replication
===================================

This section provides an overview of the Geo-replication deployment
scenarios, describes how you can check the minimum system requirements,
and explores common deployment scenarios.

-   ?

-   ?

-   ?

-   ?

-   ?

Exploring Geo-replication Deployment Scenarios
----------------------------------------------

Geo-replication provides an incremental replication service over Local
Area Networks (LANs), Wide Area Network (WANs), and across the Internet.
This section illustrates the most common deployment scenarios for
Geo-replication, including the following:

-   Geo-replication over LAN

-   Geo-replication over WAN

-   Geo-replication over the Internet

-   Multi-site cascading Geo-replication

**Geo-replication over LAN**

You can configure Geo-replication to mirror data over a Local Area
Network.

![ Geo-replication over LAN ][]

**Geo-replication over WAN**

You can configure Geo-replication to replicate data over a Wide Area
Network.

![ Geo-replication over WAN ][]

**Geo-replication over Internet**

You can configure Geo-replication to mirror data over the Internet.

![ Geo-replication over Internet ][]

**Multi-site cascading Geo-replication**

You can configure Geo-replication to mirror data in a cascading fashion
across multiple sites.

![ Multi-site cascading Geo-replication ][]

Geo-replication Deployment Overview
-----------------------------------

Deploying Geo-replication involves the following steps:

1.  Verify that your environment matches the minimum system requirement.
    For more information, see ?.

2.  Determine the appropriate deployment scenario. For more information,
    see ?.

3.  Start Geo-replication on master and slave systems, as required. For
    more information, see ?.

Checking Geo-replication Minimum Requirements
---------------------------------------------

Before deploying GlusterFS Geo-replication, verify that your systems
match the minimum requirements.

The following table outlines the minimum requirements for both master
and slave nodes within your environment:

  Component                Master                                                                Slave
  ------------------------ --------------------------------------------------------------------- --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
  Operating System         GNU/Linux                                                             GNU/Linux
  Filesystem               GlusterFS 3.2 or higher                                               GlusterFS 3.2 or higher (GlusterFS needs to be installed, but does not need to be running), ext3, ext4, or XFS (any other POSIX compliant file system would work, but has not been tested extensively)
  Python                   Python 2.4 (with ctypes external module), or Python 2.5 (or higher)   Python 2.4 (with ctypes external module), or Python 2.5 (or higher)
  Secure shell             OpenSSH version 4.0 (or higher)                                       SSH2-compliant daemon
  Remote synchronization   rsync 3.0.7 or higher                                                 rsync 3.0.7 or higher
  FUSE                     GlusterFS supported versions                                          GlusterFS supported versions

Setting Up the Environment for Geo-replication
----------------------------------------------

**Time Synchronization**

-   On bricks of a geo-replication master volume, all the servers' time
    must be uniform. You are recommended to set up NTP (Network Time
    Protocol) service to keep the bricks sync in time and avoid
    out-of-time sync effect.

    For example: In a Replicated volume where brick1 of the master is at
    12.20 hrs and brick 2 of the master is at 12.10 hrs with 10 minutes
    time lag, all the changes in brick2 between this period may go
    unnoticed during synchronization of files with Slave.

    For more information on setting up NTP, see [][].

**To setup Geo-replication for SSH**

Password-less login has to be set up between the host machine (where
geo-replication Start command will be issued) and the remote machine
(where slave process should be launched through SSH).

1.  On the node where geo-replication sessions are to be set up, run the
    following command:

    `# ssh-keygen -f /var/lib/glusterd/geo-replication/secret.pem`

    Press Enter twice to avoid passphrase.

2.  Run the following command on master for all the slave hosts:

    `# ssh-copy-id -i /var/lib/glusterd/geo-replication/secret.pem.pub @`

Setting Up the Environment for a Secure Geo-replication Slave
-------------------------------------------------------------

You can configure a secure slave using SSH so that master is granted a
restricted access. With GlusterFS, you need not specify configuration
parameters regarding the slave on the master-side configuration. For
example, the master does not require the location of the rsync program
on slave but the slave must ensure that rsync is in the PATH of the user
which the master connects using SSH. The only information that master
and slave have to negotiate are the slave-side user account, slave's
resources that master uses as slave resources, and the master's public
key. Secure access to the slave can be established using the following
options:

-   Restricting Remote Command Execution

-   Using `Mountbroker` for Slaves

-   Using IP based Access Control

**Backward Compatibility**

Your existing Ge-replication environment will work with GlusterFS,
except for the following:

-   The process of secure reconfiguration affects only the glusterfs
    instance on slave. The changes are transparent to master with the
    exception that you may have to change the SSH target to an
    unprivileged account on slave.

-   The following are the some exceptions where this might not work:

    -   Geo-replication URLs which specify the slave resource when
        configuring master will include the following special
        characters: space, \*, ?, [;

    -   Slave must have a running instance of glusterd, even if there is
        no gluster volume among the mounted slave resources (that is,
        file tree slaves are used exclusively) .

### Restricting Remote Command Execution

If you restrict remote command execution, then the Slave audits commands
coming from the master and the commands related to the given
geo-replication session is allowed. The Slave also provides access only
to the files within the slave resource which can be read or manipulated
by the Master.

To restrict remote command execution:

1.  Identify the location of the gsyncd helper utility on Slave. This
    utility is installed in `PREFIX/libexec/glusterfs/gsyncd`, where
    PREFIX is a compile-time parameter of glusterfs. For example,
    `--prefix=PREFIX` to the configure script with the following common
    values` /usr, /usr/local, and /opt/glusterfs/glusterfs_version`.

2.  Ensure that command invoked from master to slave passed through the
    slave's gsyncd utility.

    You can use either of the following two options:

    -   Set gsyncd with an absolute path as the shell for the account
        which the master connects through SSH. If you need to use a
        privileged account, then set it up by creating a new user with
        UID 0.

    -   Setup key authentication with command enforcement to gsyncd. You
        must prefix the copy of master's public key in the Slave
        account's `authorized_keys` file with the following command:

        `command=<path to gsyncd>`.

        For example,
        `command="PREFIX/glusterfs/gsyncd" ssh-rsa AAAAB3Nza....`

### Using Mountbroker for Slaves

`mountbroker` is a new service of glusterd. This service allows an
unprivileged process to own a GlusterFS mount by registering a label
(and DSL (Domain-specific language) options ) with glusterd through a
glusterd volfile. Using CLI, you can send a mount request to glusterd to
receive an alias (symlink) of the mounted volume.

A request from the agent , the unprivileged slave agents use the
mountbroker service of glusterd to set up an auxiliary gluster mount for
the agent in a special environment which ensures that the agent is only
allowed to access with special parameters that provide administrative
level access to the particular volume.

**To setup an auxiliary gluster mount for the agent**:

1.  Create a new group. For example, `geogroup`.

2.  Create a unprivileged account. For example, ` geoaccount`. Make it a
    member of ` geogroup`.

3.  Create a new directory owned by root and with permissions *0711.*
    For example, create a create mountbroker-root directory
    `/var/mountbroker-root`.

4.  Add the following options to the glusterd volfile, assuming the name
    of the slave gluster volume as `slavevol`:

    `option mountbroker-root /var/mountbroker-root `

    `option mountbroker-geo-replication.geoaccount slavevol`

    `option geo-replication-log-group geogroup`

    If you are unable to locate the glusterd volfile at
    `/etc/glusterfs/glusterd.vol`, you can create a volfile containing
    both the default configuration and the above options and place it at
    `/etc/glusterfs/`.

    A sample glusterd volfile along with default options:

        volume management
            type mgmt/glusterd
            option working-directory /var/lib/glusterd
            option transport-type socket,rdma
            option transport.socket.keepalive-time 10
            option transport.socket.keepalive-interval 2
            option transport.socket.read-fail-log off

            option mountbroker-root /var/mountbroker-root 
            option mountbroker-geo-replication.geoaccount slavevol
            option geo-replication-log-group geogroup
        end-volume

    If you host multiple slave volumes on Slave, you can repeat step 2.
    for each of them and add the following options to the `volfile`:

        option mountbroker-geo-replication.geoaccount2 slavevol2
        option mountbroker-geo-replication.geoaccount3 slavevol3

5.  Setup Master to access Slave as `geoaccount@Slave`.

    You can add multiple slave volumes within the same account
    (geoaccount) by providing comma-separated list (without spaces) as
    the argument of `mountbroker-geo-replication.geogroup`. You can also
    have multiple options of the form `mountbroker-geo-replication.*`.
    It is recommended to use one service account per Master machine. For
    example, if there are multiple slave volumes on Slave for the master
    machines Master1, Master2, and Master3, then create a dedicated
    service user on Slave for them by repeating Step 2. for each (like
    geogroup1, geogroup2, and geogroup3), and then add the following
    corresponding options to the volfile:

    `option mountbroker-geo-replication.geoaccount1 slavevol11,slavevol12,slavevol13`

    `option mountbroker-geo-replication.geoaccount2 slavevol21,slavevol22`

    `option mountbroker-geo-replication.geoaccount3 slavevol31`

    Now set up Master1 to ssh to geoaccount1@Slave, etc.

    You must restart glusterd after making changes in the configuration
    to effect the updates.

### Using IP based Access Control

You can use IP based access control method to provide access control for
the slave resources using IP address. You can use method for both Slave
and file tree slaves, but in the section, we are focusing on file tree
slaves using this method.

To set access control based on IP address for file tree slaves:

1.  Set a general restriction for accessibility of file tree resources:

    `# gluster volume geo-replication '/*' config allow-network ::1,127.0.0.1 `

    This will refuse all requests for spawning slave agents except for
    requests initiated locally.

2.  If you want the to lease file tree at `/data/slave-tree` to Master,
    enter the following command:

    `# gluster volume geo-replicationconfig allow-network `

    `MasterIP` is the IP address of Master. The slave agent spawn
    request from master will be accepted if it is executed at
    `/data/slave-tree`.

If the Master side network configuration does not enable the Slave to
recognize the exact IP address of Master, you can use CIDR notation to
specify a subnet instead of a single IP address as MasterIP or even
comma-separated lists of CIDR subnets.

If you want to extend IP based access control to gluster slaves, use the
following command:

`# gluster volume geo-replication '*' config allow-network ::1,127.0.0.1`

Starting Geo-replication
========================

This section describes how to configure and start Gluster
Geo-replication in your storage environment, and verify that it is
functioning correctly.

-   ?

-   ?

-   ?

-   ?

-   ?

Starting Geo-replication
------------------------

To start Gluster Geo-replication

-   Start geo-replication between the hosts using the following command:

    `# gluster volume geo-replication  start`

    For example:

        # gluster volume geo-replication Volume1 example.com:/data/remote_dir start
        Starting geo-replication session between Volume1
        example.com:/data/remote_dir has been successful

    > **Note**
    >
    > You may need to configure the service before starting Gluster
    > Geo-replication. For more information, see ?.

Verifying Successful Deployment
-------------------------------

You can use the gluster command to verify the status of Gluster
Geo-replication in your environment.

**To verify the status Gluster Geo-replication**

-   Verify the status by issuing the following command on host:

    `# gluster volume geo-replication  status`

    For example:

    `# gluster volume geo-replication Volume1 example.com:/data/remote_dir status`

        # gluster volume geo-replication Volume1 example.com:/data/remote_dir status

        MASTER    SLAVE                            STATUS
        ______    ______________________________   ____________
        Volume1 root@example.com:/data/remote_dir  Starting....

Displaying Geo-replication Status Information
---------------------------------------------

You can display status information about a specific geo-replication
master session, or a particular master-slave session, or all
geo-replication sessions, as needed.

**To display geo-replication status information**

-   Display information of all geo-replication sessions using the
    following command:

        # gluster volume geo-replication Volume1 example.com:/data/remote_dir status

        MASTER    SLAVE                            STATUS
        ______    ______________________________   ____________
        Volume1 root@example.com:/data/remote_dir  Starting....

-   Display information of a particular master slave session using the
    following command:

    `# gluster volume geo-replication  status`

    For example, to display information of Volume1 and
    example.com:/data/remote\_dir

    `# gluster volume geo-replication Volume1 example.com:/data/remote_dir status`

    The status of the geo-replication between Volume1 and
    example.com:/data/remote\_dir is displayed.

-   Display information of all geo-replication sessions belonging to a
    master

    `# gluster volume geo-replication MASTER status`

    For example, to display information of Volume1

        # gluster volume geo-replication Volume1 example.com:/data/remote_dir status

        MASTER    SLAVE                            STATUS
        ______    ______________________________   ____________
        Volume1 ssh://example.com:gluster://127.0.0.1:remove_volume  OK

        Volume1 ssh://example.com:file:///data/remote_dir  OK

    The status of a session could be one of the following four:

-   **Starting**: This is the initial phase of the Geo-replication
    session; it remains in this state for a minute, to make sure no
    abnormalities are present.

-   **OK**: The geo-replication session is in a stable state.

-   **Faulty**: The geo-replication session has witnessed some
    abnormality and the situation has to be investigated further. For
    further information, see ? section.

-   **Corrupt**: The monitor thread which is monitoring the
    geo-replication session has died. This situation should not occur
    normally, if it persists contact Red Hat Support[][1].

Configuring Geo-replication
---------------------------

To configure Gluster Geo-replication

-   Use the following command at the Gluster command line:

    `# gluster volume geo-replication  config [options]`

    For more information about the options, see ?.

    For example:

    To view list of all option/value pair, use the following command:

    `# gluster volume geo-replication Volume1 example.com:/data/remote_dir config`

Stopping Geo-replication
------------------------

You can use the gluster command to stop Gluster Geo-replication (syncing
of data from Master to Slave) in your environment.

**To stop Gluster Geo-replication**

-   Stop geo-replication between the hosts using the following command:

    `# gluster volume geo-replication  stop `

    For example:

        # gluster volume geo-replication Volume1 example.com:/data/remote_dir stop
        Stopping geo-replication session between Volume1 and
        example.com:/data/remote_dir has been successful

    See ? for more information about the gluster command.

Restoring Data from the Slave
=============================

You can restore data from the slave to the master volume, whenever the
master volume becomes faulty for reasons like hardware failure.

The example in this section assumes that you are using the Master Volume
(Volume1) with the following configuration:

    machine1# gluster volume info
    Type: Distribute
    Status: Started
    Number of Bricks: 2
    Transport-type: tcp
    Bricks:
    Brick1: machine1:/export/dir16
    Brick2: machine2:/export/dir16
    Options Reconfigured:
    geo-replication.indexing: on

The data is syncing from master volume (Volume1) to slave directory
(example.com:/data/remote\_dir). To view the status of this
geo-replication session run the following command on Master:

    # gluster volume geo-replication Volume1 root@example.com:/data/remote_dir status

    MASTER    SLAVE                             STATUS
    ______    ______________________________    ____________
    Volume1  root@example.com:/data/remote_dir   OK

**Before Failure**

Assume that the Master volume had 100 files and was mounted at
/mnt/gluster on one of the client machines (client). Run the following
command on Client machine to view the list of files:

    client# ls /mnt/gluster | wc –l
    100

The slave directory (example.com) will have same data as in the master
volume and same can be viewed by running the following command on slave:

    example.com# ls /data/remote_dir/ | wc –l
    100

**After Failure**

If one of the bricks (machine2) fails, then the status of
Geo-replication session is changed from "OK" to "Faulty". To view the
status of this geo-replication session run the following command on
Master:

    # gluster volume geo-replication Volume1 root@example.com:/data/remote_dir status

    MASTER    SLAVE                              STATUS
    ______    ______________________________     ____________
    Volume1   root@example.com:/data/remote_dir  Faulty

Machine2 is failed and now you can see discrepancy in number of files
between master and slave. Few files will be missing from the master
volume but they will be available only on slave as shown below.

Run the following command on Client:

    client # ls /mnt/gluster | wc –l
    52

Run the following command on slave (example.com):

    Example.com# # ls /data/remote_dir/ | wc –l
    100

**To restore data from the slave machine**

1.  Stop all Master's geo-replication sessions using the following
    command:

    `# gluster volume geo-replication  stop`

    For example:

        machine1# gluster volume geo-replication Volume1
        example.com:/data/remote_dir stop

        Stopping geo-replication session between Volume1 &
        example.com:/data/remote_dir has been successful

    > **Note**
    >
    > Repeat `# gluster volume geo-replication  stop `command on all
    > active geo-replication sessions of master volume.

2.  Replace the faulty brick in the master by using the following
    command:

    `# gluster volume replace-brick  start`

    For example:

        machine1# gluster volume replace-brick Volume1 machine2:/export/dir16 machine3:/export/dir16 start
        Replace-brick started successfully

3.  Commit the migration of data using the following command:

    `# gluster volume replace-brick  commit force `

    For example:

        machine1# gluster volume replace-brick Volume1 machine2:/export/dir16 machine3:/export/dir16 commit force
        Replace-brick commit successful

4.  Verify the migration of brick by viewing the volume info using the
    following command:

    `# gluster volume info `

    For example:

        machine1# gluster volume info
        Volume Name: Volume1
        Type: Distribute
        Status: Started
        Number of Bricks: 2
        Transport-type: tcp
        Bricks:
        Brick1: machine1:/export/dir16
        Brick2: machine3:/export/dir16
        Options Reconfigured:
        geo-replication.indexing: on

5.  Run rsync command manually to sync data from slave to master
    volume's client (mount point).

    For example:

    `example.com# rsync -PavhS --xattrs --ignore-existing /data/remote_dir/ client:/mnt/gluster`

    Verify that the data is synced by using the following command:

    On master volume, run the following command:

        Client # ls | wc –l
        100

    On the Slave run the following command:

        example.com# ls /data/remote_dir/ | wc –l
        100

    Now Master volume and Slave directory is synced.

6.  Restart geo-replication session from master to slave using the
    following command:

    `# gluster volume geo-replication  start `

    For example:

        machine1# gluster volume geo-replication Volume1
        example.com:/data/remote_dir start
        Starting geo-replication session between Volume1 &
        example.com:/data/remote_dir has been successful

Best Practices
==============

**Manually Setting Time**

If you have to change the time on your bricks manually, then you must
set uniform time on all bricks. This avoids the out-of-time sync issue
described in ?. Setting time backward corrupts the geo-replication
index, so the recommended way to set the time manually is:

1.  Stop geo-replication between the master and slave using the
    following command:

    `# gluster volume geo-replication  sto`p

2.  Stop the geo-replication indexing using the following command:

    `# gluster volume set  geo-replication.indexing of`f

3.  Set uniform time on all bricks.s

4.  Restart your geo-replication sessions by using the following
    command:

    `# gluster volume geo-replication start `

**Running Geo-replication commands in one system**

It is advisable to run the geo-replication commands in one of the bricks
in the trusted storage pool. This is because, the log files for the
geo-replication session would be stored in the \*Server\* where the
Geo-replication start is initiated. Hence it would be easier to locate
the log-files when required.

**Isolation**

Geo-replication slave operation is not sandboxed as of now and is ran as
a privileged service. So for the security reason, it is advised to
create a sandbox environment (dedicated machine / dedicated virtual
machine / chroot/container type solution) by the administrator to run
the geo-replication slave in it. Enhancement in this regard will be
available in follow-up minor release.

  [ Geo-replication over LAN ]: images/Geo-Rep_LAN.png
  [ Geo-replication over WAN ]: images/Geo-Rep_WAN.png
  [ Geo-replication over Internet ]: images/Geo-Rep03_Internet.png
  [ Multi-site cascading Geo-replication ]: images/Geo-Rep04_Cascading.png
  []: http://docs.redhat.com/docs/en-US/Red_Hat_Enterprise_Linux/6/html/Migration_Planning_Guide/ch04s07.html
  [1]: www.redhat.com/support/