[Gluster-devel] svn co / cp fail

Pooya Woodcock pooya at packetcloud.net
Thu Apr 26 18:24:01 UTC 2007


Ooops! Ok it works now!
-Pooya

On Apr 26, 2007, at 10:58 AM, Anand Avati wrote:

> Pooya,
>       please update the appropriate 'option auth.ip.<brick>.allow'
> lines in the protocol/server translator to allow the new volume names
> (locks*) from the preivous names (brick*)
>
> thanks,
> avati
>
>
> On Thu, Apr 26, 2007 at 10:27:46AM -0700, Pooya Woodcock wrote:
>> Avati,
>> Thanks, I confirmed that the bug is fixed!
>>
>> I loaded the posix-locks translater as follows on the server side.
>> Did I do this right? I couldn't get the volume to mount when I put
>> "subvolumes locks locks-afr" in the protocol/server translator (and
>> then also changing the client side to look for remote-subvolume locks
>> and remote-subvolume locks-afr). When I had the remote subvolumes set
>> to locks and locks-afr, I got this error:
>> Apr 26 10:05:13] [ERROR/tcp-client.c:123/do_handshake()] tcp/
>> client:SETVOLUME on remo
>> te server failed (Volume "locks-afr" is not attachable from host
>> 10.175.175.134)
>>
>> ## current server.vol ###
>> volume brick
>>         type storage/posix
>>         option directory /GlusterFS
>> end-volume
>>
>> volume locks
>>   type features/posix-locks
>>   subvolumes brick
>> end-volume
>>
>> volume brick-afr
>>         type storage/posix
>>         option directory /GlusterFS-AFR
>> end-volume
>>
>> volume locks-afr
>>   type features/posix-locks
>>   subvolumes brick-afr
>> end-volume
>>
>> volume server
>>         type protocol/server
>>         option transport-type tcp/server
>>         option bind-address 10.175.175.137
>>         option listen-port 6996
>>         subvolumes brick brick-afr
>>         option auth.ip.brick.allow 10.175.175.*
>>         option auth.ip.brick-afr.allow 10.175.175.*
>> end-volume
>>
>>
>>
>> On Apr 26, 2007, at 9:43 AM, Anand Avati wrote:
>>
>>> Pooya,
>>>  the bug was in posix-locks. posis-locks is supposed to be loaded on
>>> server side just above the storage/posix trnaslator. loading it  
>>> on the
>>> client side makes it lock only for processes within the same  
>>> machine.
>>>
>>> the bug has been fixed in posix locks, but nonetheless you need to
>>> load posix-locks on server side for it to be put to use at all.  
>>> please
>>> tla update to get the fix (or just loading posix-locks on serverside
>>> alone will fix it too)
>>>
>>> avati
>>>
>>> On Thu, Apr 26, 2007 at 08:31:27AM -0700, Pooya Woodcock wrote:
>>>>
>>>> On Apr 26, 2007, at 2:51 AM, Anand Avati wrote:
>>>>> Pooya,
>>>>> the svn was failing because of the same root cause as brent was
>>>>> facing for mtime failures. i have committed some changes to
>>>>> write-behind to use a more 'play safe' flushing policy. I am  
>>>>> able to
>>>>> successfully svn co now.
>>>>>
>>>>> I am still unable to reprodcue the cp issue. please give me your
>>>>> spec files and the commands you executed before you got the bug.
>>>>>
>>>>> regards,
>>>>> avati
>>>>
>>>>
>>>> Hi Avati,
>>>>   After upgrading to patch-123, all I did was start copying files
>>>> around and then noticed that I can't overwrite any file. My shell
>>>> thinks that the file is a directory when it is not. I just did
>>>> another test:
>>>>
>>>> [root at node01 www]# touch hi
>>>> [root at node01 www]# touch hi2
>>>> [root at node01 www]# cp hi2 hi
>>>> cp: overwrite `hi'? yes
>>>> cp: cannot create regular file `hi': No such file or directory
>>>>
>>>> [root at node01 www]# mv hi2 hi
>>>> mv: overwrite `hi'? yes
>>>>
>>>> [root at node01 www]# cp hi hi2
>>>> [root at node01 www]# ls -l hi*
>>>> -rw-r--r--  1 root root 0 Apr 26 08:22 hi
>>>> -rw-r--r--  1 root root 0 Apr 26 08:23 hi2
>>>>
>>>> [root at node01 www]# rm hi
>>>> rm: remove regular empty file `hi'? yes
>>>> [root at node01 www]# rm hi2
>>>> rm: remove regular empty file `hi2'? yes
>>>> [root at node01 www]#
>>>>
>>>>
>>>> So you see, it is just the cp/overwrite which fails. This  
>>>> happens on
>>>> nodes 2-4 also.
>>>> ***After reverting to patch 119, all the above commands work  
>>>> fine***
>>>>
>>>> Here are the conf files for glusterfs/glusterfsd I am using....
>>>>
>>>> There are 4 servers and 4 clients.
>>>> ## SERVER.VOL (same on each except for bind-address) ##
>>>>
>>>> volume brick
>>>>        type storage/posix
>>>>        option directory /GlusterFS
>>>> end-volume
>>>>
>>>> volume brick-afr
>>>>        type storage/posix
>>>>        option directory /GlusterFS-AFR
>>>> end-volume
>>>>
>>>> volume server
>>>>        type protocol/server
>>>>        option transport-type tcp/server
>>>>        option bind-address 10.175.175.134
>>>>        option listen-port 6996
>>>>        subvolumes brick brick-afr
>>>>        option auth.ip.brick.allow 10.175.175.*
>>>>        option auth.ip.brick-afr.allow 10.175.175.*
>>>> end-volume
>>>>
>>>>
>>>> ### CLIENT.VOL ###
>>>> volume brick1
>>>> type protocol/client
>>>> option transport-type tcp/client
>>>> option remote-host 10.175.175.134
>>>> option remote-port 6996
>>>> option remote-subvolume brick
>>>> end-volume
>>>>
>>>> volume brick1-afr
>>>> type protocol/client
>>>> option transport-type tcp/client
>>>> option remote-host 10.175.175.135
>>>> option remote-port 6996
>>>> option remote-subvolume brick-afr
>>>> end-volume
>>>>
>>>> volume brick2
>>>> type protocol/client
>>>> option transport-type tcp/client
>>>> option remote-host 10.175.175.135
>>>> option remote-port 6996
>>>> option remote-subvolume brick
>>>> end-volume
>>>>
>>>> volume brick2-afr
>>>> type protocol/client
>>>> option transport-type tcp/client
>>>> option remote-host 10.175.175.136
>>>> option remote-port 6996
>>>> option remote-subvolume brick-afr
>>>> end-volume
>>>>
>>>> volume brick3
>>>> type protocol/client
>>>> option transport-type tcp/client
>>>> option remote-host 10.175.175.136
>>>> option remote-port 6996
>>>> option remote-subvolume brick
>>>> end-volume
>>>>
>>>> volume brick3-afr
>>>> type protocol/client
>>>> option transport-type tcp/client
>>>> option remote-host 10.175.175.137
>>>> option remote-port 6996
>>>> option remote-subvolume brick-afr
>>>> end-volume
>>>>
>>>> volume brick4
>>>> type protocol/client
>>>> option transport-type tcp/client
>>>> option remote-host 10.175.175.137
>>>> option remote-port 6996
>>>> option remote-subvolume brick
>>>> end-volume
>>>>
>>>> volume brick4-afr
>>>> type protocol/client
>>>> option transport-type tcp/client
>>>> option remote-host 10.175.175.134
>>>> option remote-port 6996
>>>> option remote-subvolume brick-afr
>>>> end-volume
>>>>
>>>> volume afr1
>>>> type cluster/afr
>>>> subvolumes brick1 brick1-afr
>>>> option replicate *:2
>>>> end-volume
>>>>
>>>> volume afr2
>>>> type cluster/afr
>>>> subvolumes brick2 brick2-afr
>>>> option replicate *:2
>>>> end-volume
>>>>
>>>> volume afr3
>>>> type cluster/afr
>>>> subvolumes brick3 brick3-afr
>>>> option replicate *:2
>>>> end-volume
>>>>
>>>> volume afr4
>>>> type cluster/afr
>>>> subvolumes brick4 brick4-afr
>>>> option replicate *:2
>>>> end-volume
>>>>
>>>> volume unify1
>>>> type cluster/unify
>>>> subvolumes afr1 afr2 afr3 afr4
>>>>  option readdir-force-success on
>>>>  option scheduler rr
>>>>  option rr.limits.min-free-disk 10GB
>>>> end-volume
>>>>
>>>> volume locks
>>>>        type features/posix-locks
>>>>        subvolumes unify1
>>>> end-volume
>>>>
>>>
>>> -- 
>>> ultimate_answer_t
>>> deep_thought (void)
>>> {
>>>  sleep (years2secs (7500000));
>>>  return 42;
>>> }
>>
>>
>
> -- 
> ultimate_answer_t
> deep_thought (void)
> {
>   sleep (years2secs (7500000));
>   return 42;
> }






More information about the Gluster-devel mailing list