[Gluster-users] Geo-Replication Issue

Aravinda avishwan at redhat.com
Mon Dec 15 05:37:06 UTC 2014


Hi Dave,

Yes, it is normal that one among the replica pair will be active and 
participate in syncing, If active node goes down then other node will 
become active.

--
regards
Aravinda
http://aravindavk.in

On 12/14/2014 10:30 PM, David Gibbons wrote:
> Thank you for the advice. After re-compiling gluster with the xml 
> option, I was able to get geo-replication started!
>
> Is this output normal? This is a 2x2 distributed/replicated volume:
>
>         ]# gluster volume geo-rep shares gfs-a-bkp::bkpshares status
>
>
>         MASTER NODE                     MASTER VOL    MASTER BRICK    
>                 SLAVE                   STATUS     CHECKPOINT STATUS  
>          CRAWL STATUS
>
>         ----------------------------------------------------------------------------------------------------------------------------------------------------
>
>         gfs-a-2                         shares
>          /mnt/a-2-shares-brick-1/brick    gfs-a-bkp::bkpshares  Active
>             N/A                  Hybrid Crawl
>
>         gfs-a-2                         shares
>          /mnt/a-2-shares-brick-2/brick    gfs-a-bkp::bkpshares  Active
>             N/A                  Hybrid Crawl
>
>         gfs-a-2                         shares
>          /mnt/a-2-shares-brick-3/brick    gfs-a-bkp::bkpshares  Active
>             N/A                  Hybrid Crawl
>
>         gfs-a-2                         shares
>          /mnt/a-2-shares-brick-4/brick    gfs-a-bkp::bkpshares  Active
>             N/A                  Hybrid Crawl
>
>         gfs-a-3                         shares
>          /mnt/a-3-shares-brick-1/brick    gfs-a-bkp::bkpshares
>          Passive    N/A                  N/A
>
>         gfs-a-3                         shares
>          /mnt/a-3-shares-brick-2/brick    gfs-a-bkp::bkpshares
>          Passive    N/A                  N/A
>
>         gfs-a-3                         shares
>          /mnt/a-3-shares-brick-3/brick    gfs-a-bkp::bkpshares
>          Passive    N/A                  N/A
>
>         gfs-a-3                         shares
>          /mnt/a-3-shares-brick-4/brick    gfs-a-bkp::bkpshares
>          Passive    N/A                  N/A
>
>         gfs-a-4                         shares
>          /mnt/a-4-shares-brick-1/brick    gfs-a-bkp::bkpshares
>          Passive    N/A                  N/A
>
>         gfs-a-4                         shares
>          /mnt/a-4-shares-brick-2/brick    gfs-a-bkp::bkpshares
>          Passive    N/A                  N/A
>
>         gfs-a-4                         shares
>          /mnt/a-4-shares-brick-3/brick    gfs-a-bkp::bkpshares
>          Passive    N/A                  N/A
>
>         gfs-a-4                         shares
>          /mnt/a-4-shares-brick-4/brick    gfs-a-bkp::bkpshares
>          Passive    N/A                  N/A
>
>         gfs-a-1                         shares
>          /mnt/a-1-shares-brick-1/brick    gfs-a-bkp::bkpshares  Active
>             N/A                  Hybrid Crawl
>
>         gfs-a-1                         shares
>          /mnt/a-1-shares-brick-2/brick    gfs-a-bkp::bkpshares  Active
>             N/A                  Hybrid Crawl
>
>         gfs-a-1                         shares
>          /mnt/a-1-shares-brick-3/brick    gfs-a-bkp::bkpshares  Active
>             N/A                  Hybrid Crawl
>
>         gfs-a-1                         shares
>          /mnt/a-1-shares-brick-4/brick    gfs-a-bkp::bkpshares  Active
>             N/A                  Hybrid Crawl
>
>
> What I mean to say is, is it normal for two of the nodes to be in 
> active mode and two of the nodes to be in passive mode? I'm thinking 
> the answer is yes due to the distributed/replicated nature, but would 
> like some confirmation of that.
>
> Cheers,
> Dave
>
> On Thu, Dec 11, 2014 at 12:19 PM, Aravinda <avishwan at redhat.com 
> <mailto:avishwan at redhat.com>> wrote:
>
>     Geo-replication depends on the xml output of Gluster Cli commands.
>     For example, before connecting to slave nodes it gets the nodes
>     list from both master and slave using gluster volume info and
>     status commands with --xml.
>
>     The Python tracebacks you are seeing in logs are due to inability
>     to parse the output of gluster commands when xml is not supported.
>
>     --
>     regards
>     Aravinda
>     http://aravindavk.in
>
>
>
>     On 12/11/2014 07:56 PM, David Gibbons wrote:
>>     Thanks for the feedback, answers inline below:
>>
>>          Have you followed all the upgrade steps w.r.t geo-rep
>>            mentioned in the following link?
>>
>>
>>     I didn't upgrade geo-rep, I disconnected the old replicated
>>     server and started from scratch. So everything with regard to
>>     geo-rep is fresh/brand-new.
>>
>>         2. Does the output of command 'gluster vol info <vol-name>
>>         --xml' proper ?
>>            Please paste the output.
>>
>>
>>     I do not have gluster compiled with xml. Perhaps that is the
>>     problem. Here is the output of the command you referenced:
>>
>>         XML output not supported. Ignoring '--xml' option
>>
>>
>>     This is my config summary:
>>
>>         GlusterFS configure summary
>>         ===========================
>>         FUSE client          : yes
>>         Infiniband verbs     : no
>>         epoll IO multiplex   : yes
>>         argp-standalone      : no
>>         fusermount           : yes
>>         readline             : no
>>         georeplication       : yes
>>         Linux-AIO            : no
>>         Enable Debug         : no
>>         systemtap            : no
>>         Block Device xlator  : no
>>         glupy                : no
>>         Use syslog           : yes
>>         XML output           : no
>>         QEMU Block formats   : no
>>         Encryption xlator    : no
>>
>>
>>     Am I missing something that is require for geo-replication? I've
>>     found the documentation for those of us who are building the
>>     binaries to be a bit lacking with regard to dependencies within
>>     the project.
>>
>>     Cheers,
>>     Dave
>>
>>
>>
>>     _______________________________________________
>>     Gluster-users mailing list
>>     Gluster-users at gluster.org  <mailto:Gluster-users at gluster.org>
>>     http://supercolony.gluster.org/mailman/listinfo/gluster-users
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20141215/abb50e6f/attachment.html>


More information about the Gluster-users mailing list