[Gluster-users] R: SPOF question

Count Zero countz at gmail.com
Tue May 25 22:25:02 UTC 2010


Hi Roberto,

Try to reverse mirror-0 and mirror-1 such that mirror-1 is the first in the list.
Does this make a difference?


On May 26, 2010, at 2:21 AM, Roberto Lucignani wrote:

> Hi Bala,
> thank you very much for your reply. Yes I tried both methods, one building a
> vol file and one retrieving the configuration from the server and both work
> fine
> 
> However my problem do not regards the mount process but it regards the spof
> represented by node01. If it is unavailable I can't access my volume.
> 
> Why this ? and why it happens only with the first node ?
> 
> Regards
> M4dG
> 
> -----Messaggio originale-----
> Da: Bala.JA [mailto:bala at gluster.com] 
> Inviato: domenica 23 maggio 2010 20.58
> A: roberto.lucignani at caleidos.it
> Cc: gluster-users at gluster.org
> Oggetto: Re: [Gluster-users] SPOF question
> 
> 
> Hi Roberto,
> 
> Gluster Storage Platform provides client volume spec file through server for
> 
> created volumes.
> 
> You can mount using
> 
> mount -t glusterfs <server>:<volume>-<transport> <your-mount-point>
> 
> for example,
> mount -t glusterfs node01:gluster01-tcp /mnt/gluster01
> 
> Its not required to write your own spec file for mounting it.
> 
> Thanks,
> 
> Regards,
> Bala
> 
> 
> 
> Roberto Lucignani wrote:
>> Hi all,
>> 
>> I installed two Gluster Storage Platorm 3.0.4 on two servers node01 e
>> node02.
>> 
>> I created a volume called gluster01 than I mounted it on a Debian box in
>> this way:
>> 
>> 
>> 
>> mount -t glusterfs /etc/glusterfs/gluster01-tcp.vol /mnt/gluster01/
>> 
>> 
>> 
>> the gluster01-tcp.vol is the following:
>> 
>> 
>> 
>> 
>> 
>> volume 192.168.0.200-1
>> 
>>    type protocol/client
>> 
>>    option transport-type tcp
>> 
>>    option remote-host 192.168.0.200
>> 
>>    option transport.socket.nodelay on
>> 
>>    option transport.remote-port 10012
>> 
>>    option remote-subvolume brick1
>> 
>> end-volume
>> 
>> 
>> 
>> volume 192.168.0.200-2
>> 
>>    type protocol/client
>> 
>>    option transport-type tcp
>> 
>>    option remote-host 192.168.0.200
>> 
>>    option transport.socket.nodelay on
>> 
>>    option transport.remote-port 10012
>> 
>>    option remote-subvolume brick2
>> 
>> end-volume
>> 
>> 
>> 
>> volume 192.168.0.201-1
>> 
>>    type protocol/client
>> 
>>    option transport-type tcp
>> 
>>    option remote-host 192.168.0.201
>> 
>>    option transport.socket.nodelay on
>> 
>>    option transport.remote-port 10012
>> 
>>    option remote-subvolume brick1
>> 
>> end-volume
>> 
>> 
>> 
>> volume 192.168.0.201-2
>> 
>>    type protocol/client
>> 
>>    option transport-type tcp
>> 
>>    option remote-host 192.168.0.201
>> 
>>    option transport.socket.nodelay on
>> 
>>    option transport.remote-port 10012
>> 
>>    option remote-subvolume brick2
>> 
>> end-volume
>> 
>> 
>> 
>> volume mirror-0
>> 
>>    type cluster/replicate
>> 
>>    subvolumes 192.168.0.201-1 192.168.0.200-1
>> 
>> end-volume
>> 
>> 
>> 
>> volume mirror-1
>> 
>>    type cluster/replicate
>> 
>>    subvolumes 192.168.0.201-2 192.168.0.200-2
>> 
>> end-volume
>> 
>> 
>> 
>> volume distribute
>> 
>>    type cluster/distribute
>> 
>>    subvolumes mirror-0 mirror-1
>> 
>> end-volume
>> 
>> 
>> 
>> volume readahead
>> 
>>    type performance/read-ahead
>> 
>>    option page-count 4
>> 
>>    subvolumes distribute
>> 
>> end-volume
>> 
>> 
>> 
>> volume iocache
>> 
>>    type performance/io-cache
>> 
>>    option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
>> 's/[^0-9]//g') / 5120 ))`MB
>> 
>>    option cache-timeout 1
>> 
>>    subvolumes readahead
>> 
>> end-volume
>> 
>> 
>> 
>> volume quickread
>> 
>>    type performance/quick-read
>> 
>>    option cache-timeout 1
>> 
>>    option max-file-size 64kB
>> 
>>    subvolumes iocache
>> 
>> end-volume
>> 
>> 
>> 
>> volume writebehind
>> 
>>    type performance/write-behind
>> 
>>    option cache-size 4MB
>> 
>>    subvolumes quickread
>> 
>> end-volume
>> 
>> 
>> 
>> volume statprefetch
>> 
>>    type performance/stat-prefetch
>> 
>>    subvolumes writebehind
>> 
>> end-volume
>> 
>> 
>> 
>> 
>> 
>> 
>> 
>> all works fine and smooth, I can write and read on that volume without any
>> problem.
>> 
>> The problem is when the node01 is unavailable, I can't access the volume
> via
>> mount on the Debian box. This doesn't happen if it is the node02 to be
>> unavailable.
>> 
>> I expected the same behavior in the two cases, in this way the node01
>> represents an SPOF, am I wrong ? am I missing something the configuration
> ?
>> 
>> 
>> 
>> Tnx in advance
>> 
>> Rpberto
>> 
>> 
>> 
>> 
>> ------------------------------------------------------------------------
>> 
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
> 
> 
> 
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users




More information about the Gluster-users mailing list