[Gluster-users] Single Point of failure in geo Replication

Aravinda Vishwanathapura Krishna Murthy avishwan at redhat.com
Thu Oct 17 03:02:03 UTC 2019


On Wed, Oct 16, 2019 at 11:08 PM deepu srinivasan <sdeepugd at gmail.com>
wrote:

> Hi Users
> Is there a single point of failure in GeoReplication for gluster?
> My Case:
> I Use 3 nodes in both master and slave volume.
> Master volume : Node1,Node2,Node3
> Slave Volume : Node4,Node5,Node6
> I tried to recreate the scenario to test a single point of failure.
>
> Geo-Replication Status:
>
> *Master Node         Slave Node         Status *
> Node1                   Node4                  Active
> Node2                   Node4                  Passive
> Node3                   Node4                  Passive
>
> Step 1: Stoped the glusterd daemon in Node4.
> Result: There were only two-node statuses like the one below.
>
> *Master Node         Slave Node         Status *
> Node2                   Node4                  Passive
> Node3                   Node4                  Passive
>
>
> Will the GeoReplication session goes down if the primary slave is down?
>


Hi Deepu,

Geo-replication depends on a primary slave node to get the information
about other nodes which are part of Slave Volume.

Once the workers are started, it is not dependent on the primary slave
node. Will not fail if a primary goes down. But if any other node goes down
then the worker will try to connect to some other node, for which it tries
to run Volume status command on the slave node using the following command.

```
ssh -i <georep-pem> <primary-node> gluster volume status <slavevol>
```

The above command will fail and Worker will not get the list of Slave nodes
to which it can connect to.

This is only a temporary failure until the primary node comes back online.
If the primary node is permanently down then run Geo-rep delete and Geo-rep
create command again with the new primary node. (Note: Geo-rep Delete and
Create will remember the last sync time and resume once it starts)

I will evaluate the possibility of caching a list of Slave nodes so that it
can be used as a backup primary node in case of failures. I will open
Github issue for the same.

Thanks for reporting the issue.

-- 
regards
Aravinda VK
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20191017/53f34e78/attachment.html>


More information about the Gluster-users mailing list