[Gluster-users] Geo Replication stops replicating
deepu srinivasan
sdeepugd at gmail.com
Thu Jun 6 11:23:04 UTC 2019
Hi Sunny
I have changed the file in /usr/libexec/glusterfs/peer_mountbroker.py as
mentioned in the patch.
Now the "gluster-mountbroker status" command is working fine. But the
geo-replication seems to be in the faulty state still.
[image: Screen Shot 2019-06-06 at 4.50.30 PM.png]
[image: Screen Shot 2019-06-06 at 4.51.55 PM.png]
Thankyou
Deepak
On Thu, Jun 6, 2019 at 4:10 PM Sunny Kumar <sunkumar at redhat.com> wrote:
> Above error can be tracked here:
>
> https://bugzilla.redhat.com/show_bug.cgi?id=1709248
>
> and patch link:
> https://review.gluster.org/#/c/glusterfs/+/22716/
>
> You can apply patch and test it however its waiting on regression to
> pass and merge.
>
> -Sunny
>
>
> On Thu, Jun 6, 2019 at 4:00 PM deepu srinivasan <sdeepugd at gmail.com>
> wrote:
> >
> > Hi
> > I have followed the following steps to create the geo-replication but
> the status seems to be in a faulty state.
> >
> > Steps :
> >
> > Installed cluster version 5.6 in totally six nodes.
> >>
> >> glusterfs 5.6
> >>
> >> Repository revision: git://git.gluster.org/glusterfs.git
> >>
> >> Copyright (c) 2006-2016 Red Hat, Inc. <https://www.gluster.org/>
> >>
> >> GlusterFS comes with ABSOLUTELY NO WARRANTY.
> >>
> >> It is licensed to you under your choice of the GNU Lesser
> >>
> >> General Public License, version 3 or any later version (LGPLv3
> >>
> >> or later), or the GNU General Public License, version 2 (GPLv2),
> >>
> >> in all cases as published by the Free Software Foundation
> >
> >
> > peer_probed the first three nodes and second three nodes.
> >
> >
> >
> > Added new volume in both the clusters
> >
> >
> >
> > execute gluster-mountbroker commands and restarted glusterd.
> >>
> >> gluster-mountbroker setup /var/mountbroker-root sas
> >>
> >> gluster-mountbroker remove --volume code-misc --user sas
> >
> >
> > configured a passwordless sssh from master to slave
> >>
> >> ssh-keygen; ssh-copy-id sas at 192.168.185.107
> >
> > created a common pem pub file
> >>
> >> gluster system:: execute gsec_create
> >
> > created geo-replication session.
> >>
> >> gluster volume geo-replication code-misc sas at 192.168.185.107::code-misc
> create push-pem
> >
> > executed the following command in slave
> >>
> >> /usr/libexec/glusterfs/set_geo_rep_pem_keys.sh sas code-misc code-misc
> >
> > started the gluster geo-replication.
> >>
> >> gluster volume geo-replication code-misc sas at 192.168.185.107::code-misc
> start
> >
> >
> > Now the geo-replication works fine.
> > Tested with 2000 files All seems to sync finely.
> >
> > Now I updated all the node to version 6.2 by using rpms which were built
> by the source code in a docker container in my personal machine.
> >
> >
> >> gluster --version
> >>
> >> glusterfs 6.2
> >>
> >> Repository revision: git://git.gluster.org/glusterfs.git
> >>
> >> Copyright (c) 2006-2016 Red Hat, Inc. <https://www.gluster.org/>
> >>
> >> GlusterFS comes with ABSOLUTELY NO WARRANTY.
> >>
> >> It is licensed to you under your choice of the GNU Lesser
> >>
> >> General Public License, version 3 or any later version (LGPLv3
> >>
> >> or later), or the GNU General Public License, version 2 (GPLv2),
> >>
> >> in all cases as published by the Free Software Foundation.
> >
> >
> > I have stopped the glusterd daemons in all the node along with the
> volume and geo-replication.
> > Now I started the daemons, volume and geo-replication session the status
> seems to be faulty.
> > Also noted that the result of "gluster-mountbroker status" command
> always end in python exception like this
> >>
> >> Traceback (most recent call last):
> >>
> >> File "/usr/sbin/gluster-mountbroker", line 396, in <module>
> >>
> >> runcli()
> >>
> >> File "/usr/lib/python2.7/site-packages/gluster/cliutils/cliutils.py",
> line 225, in runcli
> >>
> >> cls.run(args)
> >>
> >> File "/usr/sbin/gluster-mountbroker", line 275, in run
> >>
> >> out = execute_in_peers("node-status")
> >>
> >> File "/usr/lib/python2.7/site-packages/gluster/cliutils/cliutils.py",
> line 127, in execute_in_peers
> >>
> >> raise GlusterCmdException((rc, out, err, " ".join(cmd)))
> >>
> >> gluster.cliutils.cliutils.GlusterCmdException: (1, '', 'Unable to end.
> Error : Success\n', 'gluster system:: execute mountbroker.py node-status')
> >
> >
> > Is it I or everyone gets an error for gluster-mountbroker command for
> gluster version greater than 6.0?. Please help.
> >
> > Thank you
> > Deepak
> >
> >
> > On Thu, Jun 6, 2019 at 10:35 AM Sunny Kumar <sunkumar at redhat.com> wrote:
> >>
> >> Hi,
> >>
> >> Updated link for documentation :
> >>
> >> --
> https://docs.gluster.org/en/latest/Administrator%20Guide/Geo%20Replication/
> >>
> >> You can use this tool as well:
> >> http://aravindavk.in/blog/gluster-georep-tools/
> >>
> >> -Sunny
> >>
> >> On Thu, Jun 6, 2019 at 10:29 AM Kotresh Hiremath Ravishankar
> >> <khiremat at redhat.com> wrote:
> >> >
> >> > Hi,
> >> >
> >> > I think the steps to setup non-root geo-rep is not followed properly.
> The following entry is missing in glusterd vol file which is required.
> >> >
> >> > The message "E [MSGID: 106061]
> [glusterd-mountbroker.c:555:glusterd_do_mount] 0-management: 'option
> mountbroker-root' missing in glusterd vol file" repeated 33 times between
> [2019-06-05 08:50:46.361384] and [2019-06-05 08:52:34.019757]
> >> >
> >> > Could you please the steps from below?
> >> >
> >> >
> https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.4/html-single/administration_guide/index#Setting_Up_the_Environment_for_a_Secure_Geo-replication_Slave
> >> >
> >> > And let us know if you still face the issue.
> >> >
> >> >
> >> >
> >> >
> >> > On Thu, Jun 6, 2019 at 10:24 AM deepu srinivasan <sdeepugd at gmail.com>
> wrote:
> >> >>
> >> >> Hi Kotresh, Sunny
> >> >> I Have mailed the logs I found in one of the slave machines. Is
> there anything to do with permission? Please help.
> >> >>
> >> >> On Wed, Jun 5, 2019 at 2:28 PM deepu srinivasan <sdeepugd at gmail.com>
> wrote:
> >> >>>
> >> >>> Hi Kotresh, Sunny
> >> >>> Found this log in the slave machine.
> >> >>>>
> >> >>>> [2019-06-05 08:49:10.632583] I [MSGID: 106488]
> [glusterd-handler.c:1559:__glusterd_handle_cli_get_volume] 0-management:
> Received get vol req
> >> >>>>
> >> >>>> The message "I [MSGID: 106488]
> [glusterd-handler.c:1559:__glusterd_handle_cli_get_volume] 0-management:
> Received get vol req" repeated 2 times between [2019-06-05 08:49:10.632583]
> and [2019-06-05 08:49:10.670863]
> >> >>>>
> >> >>>> The message "I [MSGID: 106496]
> [glusterd-handler.c:3187:__glusterd_handle_mount] 0-glusterd: Received
> mount req" repeated 34 times between [2019-06-05 08:48:41.005398] and
> [2019-06-05 08:50:37.254063]
> >> >>>>
> >> >>>> The message "E [MSGID: 106061]
> [glusterd-mountbroker.c:555:glusterd_do_mount] 0-management: 'option
> mountbroker-root' missing in glusterd vol file" repeated 34 times between
> [2019-06-05 08:48:41.005434] and [2019-06-05 08:50:37.254079]
> >> >>>>
> >> >>>> The message "W [MSGID: 106176]
> [glusterd-mountbroker.c:719:glusterd_do_mount] 0-management: unsuccessful
> mount request [No such file or directory]" repeated 34 times between
> [2019-06-05 08:48:41.005444] and [2019-06-05 08:50:37.254080]
> >> >>>>
> >> >>>> [2019-06-05 08:50:46.361347] I [MSGID: 106496]
> [glusterd-handler.c:3187:__glusterd_handle_mount] 0-glusterd: Received
> mount req
> >> >>>>
> >> >>>> [2019-06-05 08:50:46.361384] E [MSGID: 106061]
> [glusterd-mountbroker.c:555:glusterd_do_mount] 0-management: 'option
> mountbroker-root' missing in glusterd vol file
> >> >>>>
> >> >>>> [2019-06-05 08:50:46.361419] W [MSGID: 106176]
> [glusterd-mountbroker.c:719:glusterd_do_mount] 0-management: unsuccessful
> mount request [No such file or directory]
> >> >>>>
> >> >>>> The message "I [MSGID: 106496]
> [glusterd-handler.c:3187:__glusterd_handle_mount] 0-glusterd: Received
> mount req" repeated 33 times between [2019-06-05 08:50:46.361347] and
> [2019-06-05 08:52:34.019741]
> >> >>>>
> >> >>>> The message "E [MSGID: 106061]
> [glusterd-mountbroker.c:555:glusterd_do_mount] 0-management: 'option
> mountbroker-root' missing in glusterd vol file" repeated 33 times between
> [2019-06-05 08:50:46.361384] and [2019-06-05 08:52:34.019757]
> >> >>>>
> >> >>>> The message "W [MSGID: 106176]
> [glusterd-mountbroker.c:719:glusterd_do_mount] 0-management: unsuccessful
> mount request [No such file or directory]" repeated 33 times between
> [2019-06-05 08:50:46.361419] and [2019-06-05 08:52:34.019758]
> >> >>>>
> >> >>>> [2019-06-05 08:52:44.426839] I [MSGID: 106496]
> [glusterd-handler.c:3187:__glusterd_handle_mount] 0-glusterd: Received
> mount req
> >> >>>>
> >> >>>> [2019-06-05 08:52:44.426886] E [MSGID: 106061]
> [glusterd-mountbroker.c:555:glusterd_do_mount] 0-management: 'option
> mountbroker-root' missing in glusterd vol file
> >> >>>>
> >> >>>> [2019-06-05 08:52:44.426896] W [MSGID: 106176]
> [glusterd-mountbroker.c:719:glusterd_do_mount] 0-management: unsuccessful
> mount request [No such file or directory]
> >> >>>
> >> >>>
> >> >>> On Wed, Jun 5, 2019 at 1:06 AM deepu srinivasan <sdeepugd at gmail.com>
> wrote:
> >> >>>>
> >> >>>> Thankyou Kotresh
> >> >>>>
> >> >>>> On Tue, Jun 4, 2019, 11:20 PM Kotresh Hiremath Ravishankar <
> khiremat at redhat.com> wrote:
> >> >>>>>
> >> >>>>> Ccing Sunny, who was investing similar issue.
> >> >>>>>
> >> >>>>> On Tue, Jun 4, 2019 at 5:46 PM deepu srinivasan <
> sdeepugd at gmail.com> wrote:
> >> >>>>>>
> >> >>>>>> Have already added the path in bashrc . Still in faulty state
> >> >>>>>>
> >> >>>>>> On Tue, Jun 4, 2019, 5:27 PM Kotresh Hiremath Ravishankar <
> khiremat at redhat.com> wrote:
> >> >>>>>>>
> >> >>>>>>> could you please try adding /usr/sbin to $PATH for user 'sas'?
> If it's bash, add 'export PATH=/usr/sbin:$PATH' in
> >> >>>>>>> /home/sas/.bashrc
> >> >>>>>>>
> >> >>>>>>> On Tue, Jun 4, 2019 at 5:24 PM deepu srinivasan <
> sdeepugd at gmail.com> wrote:
> >> >>>>>>>>
> >> >>>>>>>> Hi Kortesh
> >> >>>>>>>> Please find the logs of the above error
> >> >>>>>>>> Master log snippet
> >> >>>>>>>>>
> >> >>>>>>>>> [2019-06-04 11:52:09.254731] I [resource(worker
> /home/sas/gluster/data/code-misc):1379:connect_remote] SSH: Initializing
> SSH connection between master and slave...
> >> >>>>>>>>> [2019-06-04 11:52:09.308923] D [repce(worker
> /home/sas/gluster/data/code-misc):196:push] RepceClient: call
> 89724:139652759443264:1559649129.31 __repce_version__() ...
> >> >>>>>>>>> [2019-06-04 11:52:09.602792] E [syncdutils(worker
> /home/sas/gluster/data/code-misc):311:log_raise_exception] <top>:
> connection to peer is broken
> >> >>>>>>>>> [2019-06-04 11:52:09.603312] E [syncdutils(worker
> /home/sas/gluster/data/code-misc):805:errlog] Popen: command returned
> error cmd=ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i
> /var/lib/ glusterd/geo-replication/secret.pem -p 22 -oControlMaster=auto -S
> /tmp/gsyncd-aux-ssh-4aL2tc/d893f66e0addc32f7d0080bb503f5185.sock
> sas at 192.168.185.107 /usr/libexec/glusterfs/gsyncd slave code-misc sas@
> 192.168.185.107::code-misc --master-node 192.168.185.106
> --master-node-id 851b64d0-d885-4ae9-9b38-ab5b15db0fec --master-brick
> /home/sas/gluster/data/code-misc --local-node 192.168.185.122
> --local-node- id bcaa7af6-c3a1-4411-8e99-4ebecb32eb6a --slave-timeout 120
> --slave-log-level DEBUG --slave-gluster-log-level INFO
> --slave-gluster-command-dir /usr/sbin error=1
> >> >>>>>>>>> [2019-06-04 11:52:09.614996] I [repce(agent
> /home/sas/gluster/data/code-misc):97:service_loop] RepceServer: terminating
> on reaching EOF.
> >> >>>>>>>>> [2019-06-04 11:52:09.615545] D
> [monitor(monitor):271:monitor] Monitor:
> worker(/home/sas/gluster/data/code-misc) connected
> >> >>>>>>>>> [2019-06-04 11:52:09.616528] I
> [monitor(monitor):278:monitor] Monitor: worker died in startup phase
> brick=/home/sas/gluster/data/code-misc
> >> >>>>>>>>> [2019-06-04 11:52:09.619391] I
> [gsyncdstatus(monitor):248:set_worker_status] GeorepStatus: Worker Status
> Change status=Faulty
> >> >>>>>>>>
> >> >>>>>>>>
> >> >>>>>>>> Slave log snippet
> >> >>>>>>>>>
> >> >>>>>>>>> [2019-06-04 11:50:09.782668] E [syncdutils(slave
> 192.168.185.106/home/sas/gluster/data/code-misc):809:logerr] Popen:
> /usr/sbin/gluster> 2 : failed with this errno (No such file or directory)
> >> >>>>>>>>> [2019-06-04 11:50:11.188167] W [gsyncd(slave
> 192.168.185.125/home/sas/gluster/data/code-misc):305:main] <top>: Session
> config file not exists, using the default config
> path=/var/lib/glusterd/geo-replication/code-misc_192.168.185.107_code-misc/gsyncd.conf
> >> >>>>>>>>> [2019-06-04 11:50:11.201070] I [resource(slave
> 192.168.185.125/home/sas/gluster/data/code-misc):1098:connect] GLUSTER:
> Mounting gluster volume locally...
> >> >>>>>>>>> [2019-06-04 11:50:11.271231] E [resource(slave
> 192.168.185.125/home/sas/gluster/data/code-misc):1006:handle_mounter]
> MountbrokerMounter: glusterd answered mnt=
> >> >>>>>>>>> [2019-06-04 11:50:11.271998] E [syncdutils(slave
> 192.168.185.125/home/sas/gluster/data/code-misc):805:errlog] Popen:
> command returned error cmd=/usr/sbin/gluster --remote-host=localhost
> system:: mount sas user-map-root=sas aux-gfid-mount acl log-level=INFO
> log-file=/var/log/glusterfs/geo-replication-slaves/code-misc_192.168.185.107_code-misc/mnt-192.168.185.125-home-sas-gluster-data-code-misc.log
> volfile-server=localhost volfile-id=code-misc client-pid=-1 error=1
> >> >>>>>>>>> [2019-06-04 11:50:11.272113] E [syncdutils(slave
> 192.168.185.125/home/sas/gluster/data/code-misc):809:logerr] Popen:
> /usr/sbin/gluster> 2 : failed with this errno (No such file or directory)
> >> >>>>>>>>
> >> >>>>>>>>
> >> >>>>>>>> On Tue, Jun 4, 2019 at 5:10 PM deepu srinivasan <
> sdeepugd at gmail.com> wrote:
> >> >>>>>>>>>
> >> >>>>>>>>> Hi
> >> >>>>>>>>> As discussed I have upgraded gluster from 4.1 to 6.2 version.
> But the Geo replication failed to start.
> >> >>>>>>>>> Stays in faulty state
> >> >>>>>>>>>
> >> >>>>>>>>> On Fri, May 31, 2019, 5:32 PM deepu srinivasan <
> sdeepugd at gmail.com> wrote:
> >> >>>>>>>>>>
> >> >>>>>>>>>> Checked the data. It remains in 2708. No progress.
> >> >>>>>>>>>>
> >> >>>>>>>>>> On Fri, May 31, 2019 at 4:36 PM Kotresh Hiremath Ravishankar
> <khiremat at redhat.com> wrote:
> >> >>>>>>>>>>>
> >> >>>>>>>>>>> That means it could be working and the defunct process
> might be some old zombie one. Could you check, that data progress ?
> >> >>>>>>>>>>>
> >> >>>>>>>>>>> On Fri, May 31, 2019 at 4:29 PM deepu srinivasan <
> sdeepugd at gmail.com> wrote:
> >> >>>>>>>>>>>>
> >> >>>>>>>>>>>> Hi
> >> >>>>>>>>>>>> When i change the rsync option the rsync process doesnt
> seem to start . Only a defunt process is listed in ps aux. Only when i set
> rsync option to " " and restart all the process the rsync process is listed
> in ps aux.
> >> >>>>>>>>>>>>
> >> >>>>>>>>>>>>
> >> >>>>>>>>>>>> On Fri, May 31, 2019 at 4:23 PM Kotresh Hiremath
> Ravishankar <khiremat at redhat.com> wrote:
> >> >>>>>>>>>>>>>
> >> >>>>>>>>>>>>> Yes, rsync config option should have fixed this issue.
> >> >>>>>>>>>>>>>
> >> >>>>>>>>>>>>> Could you share the output of the following?
> >> >>>>>>>>>>>>>
> >> >>>>>>>>>>>>> 1. gluster volume geo-replication <MASTERVOL>
> <SLAVEHOST>::<SLAVEVOL> config rsync-options
> >> >>>>>>>>>>>>> 2. ps -ef | grep rsync
> >> >>>>>>>>>>>>>
> >> >>>>>>>>>>>>> On Fri, May 31, 2019 at 4:11 PM deepu srinivasan <
> sdeepugd at gmail.com> wrote:
> >> >>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>> Done.
> >> >>>>>>>>>>>>>> We got the following result .
> >> >>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>> 1559298781.338234 write(2, "rsync: link_stat
> \"/tmp/gsyncd-aux-mount-EEJ_sY/.gfid/3fa6aed8-802e-4efe-9903-8bc171176d88\"
> failed: No such file or directory (2)", 128
> >> >>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>> seems like a file is missing ?
> >> >>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>> On Fri, May 31, 2019 at 3:25 PM Kotresh Hiremath
> Ravishankar <khiremat at redhat.com> wrote:
> >> >>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>> Hi,
> >> >>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>> Could you take the strace with with more string size?
> The argument strings are truncated.
> >> >>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>> strace -s 500 -ttt -T -p <rsync pid>
> >> >>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>> On Fri, May 31, 2019 at 3:17 PM deepu srinivasan <
> sdeepugd at gmail.com> wrote:
> >> >>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>> Hi Kotresh
> >> >>>>>>>>>>>>>>>> The above-mentioned work around did not work properly.
> >> >>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>> On Fri, May 31, 2019 at 3:16 PM deepu srinivasan <
> sdeepugd at gmail.com> wrote:
> >> >>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>> Hi Kotresh
> >> >>>>>>>>>>>>>>>>> We have tried the above-mentioned rsync option and we
> are planning to have the version upgrade to 6.0.
> >> >>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>> On Fri, May 31, 2019 at 11:04 AM Kotresh Hiremath
> Ravishankar <khiremat at redhat.com> wrote:
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>> Hi,
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>> This looks like the hang because stderr buffer
> filled up with errors messages and no one reading it.
> >> >>>>>>>>>>>>>>>>>> I think this issue is fixed in latest releases. As a
> workaround, you can do following and check if it works.
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>> Prerequisite:
> >> >>>>>>>>>>>>>>>>>> rsync version should be > 3.1.0
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>> Workaround:
> >> >>>>>>>>>>>>>>>>>> gluster volume geo-replication <MASTERVOL>
> <SLAVEHOST>::<SLAVEVOL> config rsync-options "--ignore-missing-args"
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>> Thanks,
> >> >>>>>>>>>>>>>>>>>> Kotresh HR
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>> On Thu, May 30, 2019 at 5:39 PM deepu srinivasan <
> sdeepugd at gmail.com> wrote:
> >> >>>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>> Hi
> >> >>>>>>>>>>>>>>>>>>> We were evaluating Gluster geo Replication between
> two DCs one is in US west and one is in US east. We took multiple trials
> for different file size.
> >> >>>>>>>>>>>>>>>>>>> The Geo Replication tends to stop replicating but
> while checking the status it appears to be in Active state. But the slave
> volume did not increase in size.
> >> >>>>>>>>>>>>>>>>>>> So we have restarted the geo-replication session
> and checked the status. The status was in an active state and it was in
> History Crawl for a long time. We have enabled the DEBUG mode in logging
> and checked for any error.
> >> >>>>>>>>>>>>>>>>>>> There was around 2000 file appeared for syncing
> candidate. The Rsync process starts but the rsync did not happen in the
> slave volume. Every time the rsync process appears in the "ps auxxx" list
> but the replication did not happen in the slave end. What would be the
> cause of this problem? Is there anyway to debug it?
> >> >>>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>> We have also checked the strace of the rync program.
> >> >>>>>>>>>>>>>>>>>>> it displays something like this
> >> >>>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>> "write(2, "rsync: link_stat \"/tmp/gsyncd-au"...,
> 128"
> >> >>>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>> We are using the below specs
> >> >>>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>> Gluster version - 4.1.7
> >> >>>>>>>>>>>>>>>>>>> Sync mode - rsync
> >> >>>>>>>>>>>>>>>>>>> Volume - 1x3 in each end (master and slave)
> >> >>>>>>>>>>>>>>>>>>> Intranet Bandwidth - 10 Gig
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>>>> --
> >> >>>>>>>>>>>>>>>>>> Thanks and Regards,
> >> >>>>>>>>>>>>>>>>>> Kotresh H R
> >> >>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>>
> >> >>>>>>>>>>>>>>> --
> >> >>>>>>>>>>>>>>> Thanks and Regards,
> >> >>>>>>>>>>>>>>> Kotresh H R
> >> >>>>>>>>>>>>>
> >> >>>>>>>>>>>>>
> >> >>>>>>>>>>>>>
> >> >>>>>>>>>>>>> --
> >> >>>>>>>>>>>>> Thanks and Regards,
> >> >>>>>>>>>>>>> Kotresh H R
> >> >>>>>>>>>>>
> >> >>>>>>>>>>>
> >> >>>>>>>>>>>
> >> >>>>>>>>>>> --
> >> >>>>>>>>>>> Thanks and Regards,
> >> >>>>>>>>>>> Kotresh H R
> >> >>>>>>>
> >> >>>>>>>
> >> >>>>>>>
> >> >>>>>>> --
> >> >>>>>>> Thanks and Regards,
> >> >>>>>>> Kotresh H R
> >> >>>>>
> >> >>>>>
> >> >>>>>
> >> >>>>> --
> >> >>>>> Thanks and Regards,
> >> >>>>> Kotresh H R
> >> >
> >> >
> >> >
> >> > --
> >> > Thanks and Regards,
> >> > Kotresh H R
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190606/436bb906/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: Screen Shot 2019-06-06 at 4.50.30 PM.png
Type: image/png
Size: 44504 bytes
Desc: not available
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190606/436bb906/attachment-0002.png>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: Screen Shot 2019-06-06 at 4.51.55 PM.png
Type: image/png
Size: 48203 bytes
Desc: not available
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190606/436bb906/attachment-0003.png>
More information about the Gluster-users
mailing list