[Gluster-users] Fwd: GlusterFS running, but not syncing is done

Stas Oskin stas.oskin at gmail.com
Thu Mar 12 10:38:49 UTC 2009


Hi.

I still don't have the synchronizing working- can someone advice on the
matter?

Thanks in advance!

2009/3/9 Stas Oskin <stas.oskin at gmail.com>

> Hi.
> These are my new 2 vol files, one for client and one for server.
>
> Can you advice if they are correct?
>
> Thanks in advance.
>
> glusterfs.vol (client)
>
> ## Reference volume "home2" from remote server
> volume home2
>  type protocol/client
>  option transport-type tcp/client
>  option remote-host 192.168.253.41      # IP address of remote host
>  option remote-subvolume posix-locks-home1     # use home1 on remote host
>  option transport-timeout 10           # value in seconds; it should be set
> relatively low
> end-volume
>
> ### Create automatic file replication
> volume home
>  type cluster/afr
>  option metadata-self-heal on
>  option read-subvolume posix-locks-home1
> #  option favorite-child home2
>  subvolumes posix-locks-home1 home2
> end-volume
>
>
> glusterfsd.vol (server)
>
> volume home1
>  type storage/posix                   # POSIX FS translator
>  option directory /media/storage        # Export this directory
> end-volume
>
> volume posix-locks-home1
>  type features/posix-locks
>  option mandatory-locks on
>  subvolumes home1
> end-volume
>
> ### Add network serving capability to above home.
> volume server
>  type protocol/server
>  option transport-type tcp
>  subvolumes posix-locks-home1
>  option auth.addr.posix-locks-home1.allow 192.168.253.41,127.0.0.1 # Allow
> access to "home1" volume
> end-volume
>
> 2009/3/9 Krishna Srinivas <krishna at zresearch.com>
>
> Stats,
>>
>> I think there was nothing changed between rc2 and rc4 that could
>> affect this functionality.
>>
>> Your vol files look fine, i will look into why it is not working.
>>
>> Do not use single process as both server and client as we saw issues
>> related to locking. Can you see if using different processes for
>> server and client works fine w.r.t replication?
>>
>> Also subvolumes list of all AFRs should be in same order (in your case
>> its interchanged)
>>
>> Regards
>> Krishna
>>
>> On Mon, Mar 9, 2009 at 5:44 PM, Stas Oskin <stas.oskin at gmail.com> wrote:
>> > Actually, I see a new version came out, rc4.
>> > Any idea if anything related was fixed?
>> > Regards.
>> > 2009/3/9 Stas Oskin <stas.oskin at gmail.com>
>> >>
>> >> Hi.
>> >>>
>> >>> Was it working for your previously? Any other error logs on machine
>> >>> with afr? what version are you using? If it was working previously
>> >>> what changed in your setup recently? Can you paste your vol files
>> >>> (just to be sure)
>> >>
>> >>
>> >> Nope, it actually my first setup in lab. No errors - it just seems as
>> not
>> >> synchronizing anything. The version I'm using is the latest one - 2
>> rc2.
>> >> Perhaps I need to modify anything else in addition to GlusterFS
>> >> installation - like file-systems attributes or something?
>> >> The approach I'm using is the one that was recommended by Kieth over
>> >> direct emails (Keith, hope you don't mind me posting them :) ).
>> >> The idea is basically to have single vol file both for client and for
>> >> server, and to have one glusterfs process doing the job both as client
>> and
>> >> as server.
>> >> Thanks for the help.
>> >> Server 1:
>> >> volume home1
>> >>  type storage/posix                   # POSIX FS translator
>> >>  option directory /media/storage        # Export this directory
>> >> end-volume
>> >>
>> >> volume posix-locks-home1
>> >>  type features/posix-locks
>> >>  option mandatory-locks on
>> >>  subvolumes home1
>> >> end-volume
>> >>
>> >> ## Reference volume "home2" from remote server
>> >> volume home2
>> >>  type protocol/client
>> >>  option transport-type tcp/client
>> >>  option remote-host 192.168.253.42      # IP address of remote host
>> >>  option remote-subvolume posix-locks-home1     # use home1 on remote
>> host
>> >>  option transport-timeout 10           # value in seconds; it should be
>> >> set relatively low
>> >> end-volume
>> >>
>> >> ### Add network serving capability to above home.
>> >> volume server
>> >>  type protocol/server
>> >>  option transport-type tcp
>> >>  subvolumes posix-locks-home1
>> >>  option auth.addr.posix-locks-home1.allow 192.168.253.42,127.0.0.1 #
>> Allow
>> >> access to "home1" volume
>> >> end-volume
>> >>
>> >> ### Create automatic file replication
>> >> volume home
>> >>  type cluster/afr
>> >>  option metadata-self-heal on
>> >>  option read-subvolume posix-locks-home1
>> >> #  option favorite-child home2
>> >>  subvolumes home2 posix-locks-home1
>> >> end-volume
>> >>
>> >>
>> >> Server 2:
>> >>
>> >> volume home1
>> >>  type storage/posix                   # POSIX FS translator
>> >>  option directory /media/storage        # Export this directory
>> >> end-volume
>> >>
>> >> volume posix-locks-home1
>> >>  type features/posix-locks
>> >>  option mandatory-locks on
>> >>  subvolumes home1
>> >> end-volume
>> >>
>> >> ## Reference volume "home2" from remote server
>> >> volume home2
>> >>  type protocol/client
>> >>  option transport-type tcp/client
>> >>  option remote-host 192.168.253.41      # IP address of remote host
>> >>  option remote-subvolume posix-locks-home1     # use home1 on remote
>> host
>> >>  option transport-timeout 10           # value in seconds; it should be
>> >> set relatively low
>> >> end-volume
>> >>
>> >> ### Add network serving capability to above home.
>> >> volume server
>> >>  type protocol/server
>> >>  option transport-type tcp
>> >>  subvolumes posix-locks-home1
>> >>  option auth.addr.posix-locks-home1.allow 192.168.253.41,127.0.0.1 #
>> Allow
>> >> access to "home1" volume
>> >> end-volume
>> >>
>> >> ### Create automatic file replication
>> >> volume home
>> >>  type cluster/afr
>> >>  option metadata-self-heal on
>> >>  option read-subvolume posix-locks-home1
>> >> #  option favorite-child home2
>> >>  subvolumes home2 posix-locks-home1
>> >> end-volume
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20090312/165d641e/attachment.html>


More information about the Gluster-users mailing list