[Gluster-users] Geo rep fail
anthony garnier
sokar6012 at hotmail.com
Thu Aug 2 08:55:18 UTC 2012
Hi Vijay,
Thx for your help, I tried with root and it worked !
Many thanks,
Anthony
> Date: Thu, 2 Aug 2012 02:28:27 -0400
> From: vkoppad at redhat.com
> To: sokar6012 at hotmail.com
> CC: gluster-users at gluster.org
> Subject: Re: [Gluster-users] Geo rep fail
>
> Hi anthony,
>
> What I understood from your invocation of geo-rep session is,
> you are trying to start geo-rep with slave as a normal-user.
> To successfully start geo-rep session , the slave need to be as a super user.
> Otherwise if you really want to have slave as a normal-user , you should
> set-up geo-rep through Mount-broker, the details of which you can get here,
>
> http://docs.redhat.com/docs/en-US/Red_Hat_Storage/2.0/html/Administration_Guide/chap-User_Guide-Geo_Rep-Preparation-Settingup_Slave.html
>
> Thanks,
> Vijaykumar
>
> ----- Original Message -----
> From: "anthony garnier" <sokar6012 at hotmail.com>
> To: vkoppad at redhat.com
> Cc: gluster-users at gluster.org
> Sent: Wednesday, August 1, 2012 6:28:37 PM
> Subject: Re: [Gluster-users] Geo rep fail
>
>
>
> Hi Vijay,
>
> Some complementary info :
>
> * SLES 11.2
> * 3.0.26-0.7-xen
> * glusterfs 3.3.0 built on Jul 16 2012 14:28:16
> * Python 2.6.8
> * rsync version 3.0.4
> * OpenSSH_4.3p2, OpenSSL 0.9.8a 11 Oct 2005
>
>
> *
> ssh command used : ssh -oPasswordAuthentication=no -oStrictHostKeyChecking=no -i /var/lib/glusterd/geo-replication/secret.pem <= key of user sshux
>
> I also changed 1 line in gconf.py because I was aving diffiulties with the control master option and -S option
>
> # cls.ssh_ctl_args = ["-oControlMaster=auto", "-S", os.path.join(ctld, "gsycnd-ssh-%r@%h:%p")]
> cls.ssh_ctl_args = ["-oControlMaster=no"]
>
> Gluster cmd :
>
> # gluster volume geo-replication test ssh://sshux@yval1020:/users/geo-rep start
>
>
> Thx for your help.
>
> Anthony
>
>
>
> > Date: Tue, 31 Jul 2012 08:18:52 -0400
> > From: vkoppad at redhat.com
> > To: sokar6012 at hotmail.com
> > CC: gluster-users at gluster.org
> > Subject: Re: [Gluster-users] Geo rep fail
> >
> > Thanks anthony, I'll try to reproduce that.
> >
> > -Vijaykumar
> >
> > ----- Original Message -----
> > From: "anthony garnier" <sokar6012 at hotmail.com>
> > To: vkoppad at redhat.com
> > Cc: gluster-users at gluster.org
> > Sent: Tuesday, July 31, 2012 5:13:13 PM
> > Subject: Re: [Gluster-users] Geo rep fail
> >
> >
> >
> > Hi Vijay,
> >
> > I used the tarball here : http://download.gluster.org/pub/gluster/glusterfs/LATEST/
> >
> >
> >
> >
> > > Date: Tue, 31 Jul 2012 07:39:51 -0400
> > > From: vkoppad at redhat.com
> > > To: sokar6012 at hotmail.com
> > > CC: gluster-users at gluster.org
> > > Subject: Re: [Gluster-users] Geo rep fail
> > >
> > > Hi anthony,
> > >
> > > By Glusterfs-3.3 version, you mean this rpm
> > > http://bits.gluster.com/pub/gluster/glusterfs/3.3.0/.
> > > or If you are working with git repo, can you give me branch and Head.
> > >
> > > -Vijaykumar
> > >
> > > ----- Original Message -----
> > > From: "anthony garnier" <sokar6012 at hotmail.com>
> > > To: gluster-users at gluster.org
> > > Sent: Tuesday, July 31, 2012 2:47:40 PM
> > > Subject: [Gluster-users] Geo rep fail
> > >
> > >
> > >
> > > Hello everyone,
> > >
> > > I'm using Glusterfs 3.3 and I have some difficulties to setup geo-replication over ssh.
> > >
> > > # gluster volume geo-replication test status
> > > MASTER SLAVE STATUS
> > > --------------------------------------------------------------------------------
> > > test ssh://sshux@yval1020:/users/geo-rep faulty
> > > test file:///users/geo-rep OK
> > >
> > > As you can see, the one in a local folder works fine.
> > >
> > > This is my config :
> > >
> > > Volume Name: test
> > > Type: Replicate
> > > Volume ID: 2f0b0eff-6166-4601-8667-6530561eea1c
> > > Status: Started
> > > Number of Bricks: 1 x 2 = 2
> > > Transport-type: tcp
> > > Bricks:
> > > Brick1: yval1010:/users/exp
> > > Brick2: yval1020:/users/exp
> > > Options Reconfigured:
> > > geo-replication.indexing: on
> > > cluster.eager-lock: on
> > > performance.cache-refresh-timeout: 60
> > > network.ping-timeout: 10
> > > performance.cache-size: 512MB
> > > performance.write-behind-window-size: 256MB
> > > features.quota-timeout: 30
> > > features.limit-usage: /:20GB,/kernel:5GB,/toto:2GB,/troll:1GB
> > > features.quota: on
> > > nfs.port: 2049
> > >
> > >
> > > This is the log :
> > >
> > > [2012-07-31 11:10:38.711314] I [monitor(monitor):81:monitor] Monitor: starting gsyncd worker
> > > [2012-07-31 11:10:38.844959] I [gsyncd:354:main_i] <top>: syncing: gluster://localhost:test -> ssh://sshux@yval1020:/users/geo-rep
> > > [2012-07-31 11:10:44.526469] I [master:284:crawl] GMaster: new master is 2f0b0eff-6166-4601-8667-6530561eea1c
> > > [2012-07-31 11:10:44.527038] I [master:288:crawl] GMaster: primary master with volume id 2f0b0eff-6166-4601-8667-6530561eea1c ...
> > > [2012-07-31 11:10:44.644319] E [repce:188:__call__] RepceClient: call 10810:140268954724096:1343725844.53 (xtime) failed on peer with OSError
> > > [2012-07-31 11:10:44.644629] E [syncdutils:184:log_raise_exception] <top>: FAIL:
> > > Traceback (most recent call last):
> > > File "/soft/GLUSTERFS//libexec/glusterfs/python/syncdaemon/gsyncd.py", line 115, in main
> > > main_i()
> > > File "/soft/GLUSTERFS//libexec/glusterfs/python/syncdaemon/gsyncd.py", line 365, in main_i
> > > local.service_loop(*[r for r in [remote] if r])
> > > File "/soft/GLUSTERFS/libexec/glusterfs/python/syncdaemon/resource.py", line 756, in service_loop
> > > GMaster(self, args[0]).crawl_loop()
> > > File "/soft/GLUSTERFS/libexec/glusterfs/python/syncdaemon/master.py", line 143, in crawl_loop
> > > self.crawl()
> > > File "/soft/GLUSTERFS/libexec/glusterfs/python/syncdaemon/master.py", line 308, in crawl
> > > xtr0 = self.xtime(path, self.slave)
> > > File "/soft/GLUSTERFS/libexec/glusterfs/python/syncdaemon/master.py", line 74, in xtime
> > > xt = rsc.server.xtime(path, self.uuid)
> > > File "/soft/GLUSTERFS/libexec/glusterfs/python/syncdaemon/repce.py", line 204, in __call__
> > > return self.ins(self.meth, *a)
> > > File "/soft/GLUSTERFS/libexec/glusterfs/python/syncdaemon/repce.py", line 189, in __call__
> > > raise res
> > > OSError: [Errno 95] Operation not supported
> > >
> > >
> > > Apparently there is some errors with xtime and yet I have extended attribute activated.
> > > Any help will be gladly appreciated.
> > >
> > > Anthony
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > _______________________________________________
> > > Gluster-users mailing list
> > > Gluster-users at gluster.org
> > > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
> >
> > _______________________________________________
> > Gluster-users mailing list
> > Gluster-users at gluster.org
> > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20120802/b1c2860c/attachment.html>
More information about the Gluster-users
mailing list