[Gluster-devel] AFR+locks bug?

SZÉKELYI Szabolcs cc at avaxio.hu
Mon Mar 17 19:19:51 UTC 2008


Hi Anand,

Anand Avati wrote:
> Szabolcs,
>  do you still face this issue?

Sorry for the long long delay and the late reply.

I've seen that the posix-locks translator had received heavy development
and had gone under major redesign a pair of months ago. So I tried the
latest TLA patchset, but no luck.

However, the situation is somewhat different now. On the master locker
(server), I get:

> lusta1:~# locktests -n 10 -f /mnt/glusterfs/locktest-alone.vol -c 3
> Init
> process initalization
> ....................
> --------------------------------------
> 
> TEST : TRY TO WRITE ON A READ  LOCK:==========
> TEST : TRY TO WRITE ON A WRITE LOCK:==========
> TEST : TRY TO READ  ON A READ  LOCK:==========
> TEST : TRY TO READ  ON A WRITE LOCK:==========
> TEST : TRY TO SET A READ  LOCK ON A READ  LOCK:===xxxxx=x

And a mount point lockup on all machines.

FUSE was not upgraded this time, it's still 2.7.0-glfs7.

Thanks,
--
Szabolcs

> 2008/1/19, Székelyi Szabolcs <cc at avaxio.hu <mailto:cc at avaxio.hu>>:
> 
>     Anand Avati wrote:
>     > Szabolcs,
>     >  I suspect it might be an issue with 2.7.2-glfs8. We are seeing
>     similar
>     > issues with the 2.7.2 fuse. Please let us know if 2.7.0 works well
>     for you.
> 
>     Well, with fuse-2.7.0-glfs7, the same happens.
> 
>     --
>     Szabolcs
> 
>     > 2008/1/17, Székelyi Szabolcs <cc at avaxio.hu <mailto:cc at avaxio.hu>
>     <mailto:cc at avaxio.hu <mailto:cc at avaxio.hu>>>:
>     >
>     >     AFR with posix-locks behaves really strange nowadays...
>     GlusterFS is a
>     >     fresh TLA checkout (patch-636), FUSE is brand the new 2.7.2-glfs8.
>     >
>     >     I have 4 servers with a 4-way AFR on each and features/posix-locks
>     >     loaded just above storage/posix bricks. On each AFR, one
>     replica is the
>     >     local storage, the remaining 3 are on the other 3 servers.
>     >
>     >     The 4 AFR bricks are mounted on each server from 'localhost'.
>     >
>     >     The machines are freshly booted. Basic FS functions (ls, copy,
>     cat) work
>     >     fine.
>     >
>     >     Now I run a distributed locking test using [1]. On the
>     "master" locker I
>     >     get:
>     >
>     >     > # /tmp/locktests -n 10 -c 3  -f /mnt/glusterfs/testfile
>     >     > Init
>     >     > process initalization
>     >     > ....................
>     >     > --------------------------------------
>     >     >
>     >     > TEST : TRY TO WRITE ON A READ  LOCK:==========
>     >     > TEST : TRY TO WRITE ON A WRITE LOCK:==========
>     >     > TEST : TRY TO READ  ON A READ  LOCK:==========
>     >     > TEST : TRY TO READ  ON A WRITE LOCK:==========
>     >     > TEST : TRY TO SET A READ  LOCK ON A READ  LOCK:
>     >
>     >     After about 5 minutes, another
>     >
>     >     > RDONLY: fcntl: Transport endpoint is not connected
>     >
>     >     appears, and the locking processes exit on all slave servers,
>     the master
>     >     blocks.
>     >
>     >     The mount point locks up. Even an `ls` from a different
>     terminal seems
>     >     to block forever.
>     >
>     >     You can find my server config below. Client configs are
>     simple, just a
>     >     protocol/client brick from localhost. I can provide server
>     debug logs if
>     >     you need.
>     >
>     >     Any idea?
>     >
>     >     Thanks,
>     >     --
>     >     Szabolcs
>     >
>     >
>     >     [1]
>     >    
>     http://nfsv4.bullopensource.org/tools/tests_tools/locktests-net.tar.gz
>     >
>     >
>     >     My server config (from a single node, lu1):
>     >
>     >     volume data-posix
>     >       type storage/posix
>     >       option directory /srv/glusterfs
>     >     end-volume
>     >
>     >     volume data1
>     >       type features/posix-locks
>     >       subvolumes data-posix
>     >     end-volume
>     >
>     >     volume data2
>     >       type protocol/client
>     >       option transport-type tcp/client
>     >       option remote-host lu2
>     >       option remote-subvolume data2
>     >     end-volume
>     >
>     >     volume data3
>     >       type protocol/client
>     >       option transport-type tcp/client
>     >       option remote-host lu3
>     >       option remote-subvolume data3
>     >     end-volume
>     >
>     >     volume data4
>     >       type protocol/client
>     >       option transport-type tcp/client
>     >       option remote-host lu4
>     >       option remote-subvolume data4
>     >     end-volume
>     >
>     >     volume data-afr
>     >       type cluster/afr
>     >       subvolumes data1 data2 data3 data4
>     >     end-volume
>     >
>     >     volume server
>     >       type protocol/server
>     >       subvolumes data1 data-afr
>     >       option transport-type tcp/server
>     >       option auth.ip.data1.allow 10.0.0.*
>     >       option auth.ip.data-afr.allow 127.0.0.1 <http://127.0.0.1>
>     <http://127.0.0.1>,10.0.0.*
>     >     end-volume
> 
> 
>     _______________________________________________
>     Gluster-devel mailing list
>     Gluster-devel at nongnu.org <mailto:Gluster-devel at nongnu.org>
>     http://lists.nongnu.org/mailman/listinfo/gluster-devel
> 
> 
> 
> 
> -- 
> If I traveled to the end of the rainbow
> As Dame Fortune did intend,
> Murphy would be there to tell me
> The pot's at the other end.






More information about the Gluster-devel mailing list