[Gluster-users] SPOF question

Bala.JA bala at gluster.com
Sun May 23 19:57:39 UTC 2010


Hi Roberto,

Gluster Storage Platform provides client volume spec file through server for 
created volumes.

You can mount using

mount -t glusterfs <server>:<volume>-<transport> <your-mount-point>

for example,
mount -t glusterfs node01:gluster01-tcp /mnt/gluster01

Its not required to write your own spec file for mounting it.

Thanks,

Regards,
Bala



Roberto Lucignani wrote:
> Hi all,
> 
> I installed two Gluster Storage Platorm 3.0.4 on two servers node01 e
> node02.
> 
> I created a volume called gluster01 than I mounted it on a Debian box in
> this way:
> 
>  
> 
> mount -t glusterfs /etc/glusterfs/gluster01-tcp.vol /mnt/gluster01/
> 
>  
> 
> the gluster01-tcp.vol is the following:
> 
>  
> 
>  
> 
> volume 192.168.0.200-1
> 
>     type protocol/client
> 
>     option transport-type tcp
> 
>     option remote-host 192.168.0.200
> 
>     option transport.socket.nodelay on
> 
>     option transport.remote-port 10012
> 
>     option remote-subvolume brick1
> 
> end-volume
> 
>  
> 
> volume 192.168.0.200-2
> 
>     type protocol/client
> 
>     option transport-type tcp
> 
>     option remote-host 192.168.0.200
> 
>     option transport.socket.nodelay on
> 
>     option transport.remote-port 10012
> 
>     option remote-subvolume brick2
> 
> end-volume
> 
>  
> 
> volume 192.168.0.201-1
> 
>     type protocol/client
> 
>     option transport-type tcp
> 
>     option remote-host 192.168.0.201
> 
>     option transport.socket.nodelay on
> 
>     option transport.remote-port 10012
> 
>     option remote-subvolume brick1
> 
> end-volume
> 
>  
> 
> volume 192.168.0.201-2
> 
>     type protocol/client
> 
>     option transport-type tcp
> 
>     option remote-host 192.168.0.201
> 
>     option transport.socket.nodelay on
> 
>     option transport.remote-port 10012
> 
>     option remote-subvolume brick2
> 
> end-volume
> 
>  
> 
> volume mirror-0
> 
>     type cluster/replicate
> 
>     subvolumes 192.168.0.201-1 192.168.0.200-1
> 
> end-volume
> 
>  
> 
> volume mirror-1
> 
>     type cluster/replicate
> 
>     subvolumes 192.168.0.201-2 192.168.0.200-2
> 
> end-volume
> 
>  
> 
> volume distribute
> 
>     type cluster/distribute
> 
>     subvolumes mirror-0 mirror-1
> 
> end-volume
> 
>  
> 
> volume readahead
> 
>     type performance/read-ahead
> 
>     option page-count 4
> 
>     subvolumes distribute
> 
> end-volume
> 
>  
> 
> volume iocache
> 
>     type performance/io-cache
> 
>     option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
> 's/[^0-9]//g') / 5120 ))`MB
> 
>     option cache-timeout 1
> 
>     subvolumes readahead
> 
> end-volume
> 
>  
> 
> volume quickread
> 
>     type performance/quick-read
> 
>     option cache-timeout 1
> 
>     option max-file-size 64kB
> 
>     subvolumes iocache
> 
> end-volume
> 
>  
> 
> volume writebehind
> 
>     type performance/write-behind
> 
>     option cache-size 4MB
> 
>     subvolumes quickread
> 
> end-volume
> 
>  
> 
> volume statprefetch
> 
>     type performance/stat-prefetch
> 
>     subvolumes writebehind
> 
> end-volume
> 
>  
> 
>  
> 
>  
> 
> all works fine and smooth, I can write and read on that volume without any
> problem.
> 
> The problem is when the node01 is unavailable, I can't access the volume via
> mount on the Debian box. This doesn't happen if it is the node02 to be
> unavailable.
> 
> I expected the same behavior in the two cases, in this way the node01
> represents an SPOF, am I wrong ? am I missing something the configuration ?
> 
>  
> 
> Tnx in advance
> 
> Rpberto
> 
> 
> 
> 
> ------------------------------------------------------------------------
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users




More information about the Gluster-users mailing list