[Gluster-devel] AFR+locks bug?
Anand Avati
avati at zresearch.com
Tue Feb 12 16:42:20 UTC 2008
Szabolcs,
do you still face this issue?
avati
2008/1/19, Székelyi Szabolcs <cc at avaxio.hu>:
>
> Anand Avati wrote:
> > Szabolcs,
> > I suspect it might be an issue with 2.7.2-glfs8. We are seeing similar
> > issues with the 2.7.2 fuse. Please let us know if 2.7.0 works well for
> you.
>
> Well, with fuse-2.7.0-glfs7, the same happens.
>
> --
> Szabolcs
>
> > 2008/1/17, Székelyi Szabolcs <cc at avaxio.hu <mailto:cc at avaxio.hu>>:
> >
> > AFR with posix-locks behaves really strange nowadays... GlusterFS is
> a
> > fresh TLA checkout (patch-636), FUSE is brand the new 2.7.2-glfs8.
> >
> > I have 4 servers with a 4-way AFR on each and features/posix-locks
> > loaded just above storage/posix bricks. On each AFR, one replica is
> the
> > local storage, the remaining 3 are on the other 3 servers.
> >
> > The 4 AFR bricks are mounted on each server from 'localhost'.
> >
> > The machines are freshly booted. Basic FS functions (ls, copy, cat)
> work
> > fine.
> >
> > Now I run a distributed locking test using [1]. On the "master"
> locker I
> > get:
> >
> > > # /tmp/locktests -n 10 -c 3 -f /mnt/glusterfs/testfile
> > > Init
> > > process initalization
> > > ....................
> > > --------------------------------------
> > >
> > > TEST : TRY TO WRITE ON A READ LOCK:==========
> > > TEST : TRY TO WRITE ON A WRITE LOCK:==========
> > > TEST : TRY TO READ ON A READ LOCK:==========
> > > TEST : TRY TO READ ON A WRITE LOCK:==========
> > > TEST : TRY TO SET A READ LOCK ON A READ LOCK:
> >
> > After about 5 minutes, another
> >
> > > RDONLY: fcntl: Transport endpoint is not connected
> >
> > appears, and the locking processes exit on all slave servers, the
> master
> > blocks.
> >
> > The mount point locks up. Even an `ls` from a different terminal
> seems
> > to block forever.
> >
> > You can find my server config below. Client configs are simple, just
> a
> > protocol/client brick from localhost. I can provide server debug
> logs if
> > you need.
> >
> > Any idea?
> >
> > Thanks,
> > --
> > Szabolcs
> >
> >
> > [1]
> >
> http://nfsv4.bullopensource.org/tools/tests_tools/locktests-net.tar.gz
> >
> >
> > My server config (from a single node, lu1):
> >
> > volume data-posix
> > type storage/posix
> > option directory /srv/glusterfs
> > end-volume
> >
> > volume data1
> > type features/posix-locks
> > subvolumes data-posix
> > end-volume
> >
> > volume data2
> > type protocol/client
> > option transport-type tcp/client
> > option remote-host lu2
> > option remote-subvolume data2
> > end-volume
> >
> > volume data3
> > type protocol/client
> > option transport-type tcp/client
> > option remote-host lu3
> > option remote-subvolume data3
> > end-volume
> >
> > volume data4
> > type protocol/client
> > option transport-type tcp/client
> > option remote-host lu4
> > option remote-subvolume data4
> > end-volume
> >
> > volume data-afr
> > type cluster/afr
> > subvolumes data1 data2 data3 data4
> > end-volume
> >
> > volume server
> > type protocol/server
> > subvolumes data1 data-afr
> > option transport-type tcp/server
> > option auth.ip.data1.allow 10.0.0.*
> > option auth.ip.data-afr.allow 127.0.0.1 <http://127.0.0.1>,
> 10.0.0.*
> > end-volume
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at nongnu.org
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>
--
If I traveled to the end of the rainbow
As Dame Fortune did intend,
Murphy would be there to tell me
The pot's at the other end.
More information about the Gluster-devel
mailing list