[Gluster-devel] AFR: machine crash hangs other mounts or transport endpoint not connected
Gerry Reno
greno at verizon.net
Fri May 2 03:40:11 UTC 2008
Gerry Reno wrote:
> Krishna and Anand:
>
> Hope these will help. Here are my configs:
>
> ================================================
> ### file: glusterfs-server.vol
> ### GRP: this files goes on all storage bricks
> ##############################################
> ### GlusterFS Server Volume Specification ##
> ##############################################
> #### CONFIG FILE RULES:
> ### "#" is comment character.
> ### - Config file is case sensitive
> ### - Options within a volume block can be in any order.
> ### - Spaces or tabs are used as delimitter within a line.
> ### - Multiple values to options will be : delimitted.
> ### - Each option should end within a line.
> ### - Missing or commented fields will assume default values.
> ### - Blank/commented lines are allowed.
> ### - Sub-volumes should already be defined above before referring.
>
> ### Export volume "brick" with the contents of "/home/export"
> directory.
> volume brick
> type storage/posix # POSIX FS translator
> option directory /home/vmail/mailbrick # Export this directory
> end-volume
>
> ### Add network serving capability to above brick.
> volume server
> type protocol/server
> option transport-type tcp/server # For TCP/IP transport
> # option ibv-send-work-request-size 131072
> # option ibv-send-work-request-count 64
> # option ibv-recv-work-request-size 131072
> # option ibv-recv-work-request-count 64
> # option transport-type ib-sdp/server # For Infiniband transport
> # option transport-type ib-verbs/server # For ib-verbs transport
> # option bind-address 127.0.0.1 # Default is to listen on all
> interfaces
> option listen-port 6996 # Default is 6996
> # option client-volume-filename /etc/glusterfs/glusterfs-client.vol
>
> subvolumes brick
> # NOTE: Access to any volume through protocol/server is denied by
> # default. You need to explicitly grant access through # "auth"
> # option.
> # option auth.ip.brick.allow * # Allow full access to
> "brick" volume
> # option auth.ip.brick.allow 192.168.* # Allow subnet access to
> "brick" volume
> option auth.ip.brick.allow 127.0.0.1,192.168.1.220,192.168.1.221 #
> Allow access to "brick" volume
> end-volume
> ================================================
>
> ================================================
> ### file: glusterfs-client.vol
> ### GRP: this file goes on every client node in cluster
> ##############################################
> ### GlusterFS Client Volume Specification ##
> ##############################################
> #### CONFIG FILE RULES:
> ### "#" is comment character.
> ### - Config file is case sensitive
> ### - Options within a volume block can be in any order.
> ### - Spaces or tabs are used as delimitter within a line.
> ### - Each option should end within a line.
> ### - Missing or commented fields will assume default values.
> ### - Blank/commented lines are allowed.
> ### - Sub-volumes should already be defined above before referring.
>
> ### Add client feature and declare local subvolume
> volume client-local
> type storage/posix
> option directory /home/vmail/mailbrick
> end-volume
>
> ### Add client feature and attach to remote subvolume
> volume client1
> type protocol/client
> option transport-type tcp/client # for TCP/IP transport
> option remote-host 192.168.1.220 # IP address of the remote
> brick
> option remote-port 6996 # default server port is 6996
> option remote-subvolume brick # name of the remote volume
> end-volume
>
> volume client2
> type protocol/client
> option transport-type tcp/client
> option remote-host 192.168.1.221
> option remote-port 6996
> option remote-subvolume brick
> end-volume
>
>
> #volume bricks
> # type cluster/unify
> # subvolumes *
> # option scheduler nufa
> # # does this brick name must be in local server.vol
> # option nufa.local-volume-name brick # note 'brick' is singular
> #end-volume
>
> ### Add automatic file replication (AFR) feature
> volume afr
> type cluster/afr
> subvolumes client1 client2
> # option replicate:*.html 2
> # option replicate:*.db 5
> ## ok, this would be RAID-1 on 2 nodes
> # option replicate:* 2
> ## so how would you say RAID-1 on all nodes? with * ?
> # option replicate *:2
> # option replicate client1,client2:2
> # option replicate is no longer supported:
> http://www.mail-archive.com/gluster-devel@nongnu.org/msg02201.html
> # pattern-matching translator will be provided later in 1.4
> end-volume
>
> ### Add writeback feature
> #volume writeback
> # type performance/write-behind
> # option aggregate-size 131072 # unit in bytes
> # subvolumes client #end-volume
>
> ### Add readahead feature
> #volume readahead
> # type performance/read-ahead
> # option page-size 65536 # unit in bytes
> # option page-count 16 # cache per file = (page-count x
> page-size)
> # subvolumes writeback
> #end-volume
> ================================================
>
>
> Regards,
> Gerry
>
And here is how I mount the client (from /etc/fstab):
/usr/local/etc/glusterfs/glusterfs-client.vol /home/vmail/mailstore
glusterfs defaults 0 0
Regards,
Gerry
More information about the Gluster-devel
mailing list