[Gluster-users] Distributed volume going to Read only mode if any of the Brick is not available

Varadharajan S rajanvaradhu at gmail.com
Wed May 20 11:48:57 UTC 2015


Hi Team,
Anyone can suggest my below query, so that I can get clear idea

Regards,
Varad
On 19 May 2015 20:28, "Varadharajan S" <rajanvaradhu at gmail.com> wrote:

> FYI
> On 19 May 2015 20:25, "Varadharajan S" <rajanvaradhu at gmail.com> wrote:
>
>> Hi,
>> Replication means, I won't get space. Distribution is not like striping
>> right? If one brick is not available in the volume, other bricks can
>> distribute data in between. If I do any tuning will get solution?
>>  On 19 May 2015 20:02, "Atin Mukherjee" <atin.mukherjee83 at gmail.com>
>> wrote:
>>
>>>
>>> On 19 May 2015 17:10, "Varadharajan S" <rajanvaradhu at gmail.com> wrote:
>>> >
>>> > Hi,
>>> >
>>> > We are using Ubuntu 14.04 server and for storage purpose we configured
>>> gluster 3.5 as distributed volume and find the below details,
>>> >
>>> > 1).4 Servers - 14.04 Ubuntu Server and each server disks free spaces
>>> are  configured as ZFS raiddz2 volume
>>> >
>>> > 2). Each server has /pool/gluster zfs volume and capacity as - 5 TB,8
>>> TB,6 TB and 10 TB
>>> >
>>> > 3). Bricks are - rep1,rep2,rep3 and st1 and all the bricks are
>>> connected as Distributed Volume and mounted on each system as,
>>> >
>>> >   For E.x in rep1 -> mount -t glusterfs  rep1:/glustervol  /data.
>>> >                   rep2  -> mount  -t glusterfs  rep2:/glustervol  /data
>>> >                   rep3  -> mount  -t glusterfs  rep3:/glustervol  /data
>>> >                   st1    ->  mount  -t glusterfs  st1:/glustervol
>>> /data
>>> >
>>> > So we get /data is having  around 29 TB and all our applications
>>> data's are stored in /data mount point.
>>> >
>>> > Details about volume:
>>> >
>>> > volume glustervol-client-0
>>> >     type protocol/client
>>> >     option send-gids true
>>> >     option password b217da9d1d8b-bb55
>>> >     option username 9d76-4553-8c75
>>> >     option transport-type tcp
>>> >     option remote-subvolume /pool/gluster
>>> >     option remote-host rep1
>>> >     option ping-timeout 42
>>> > end-volume
>>> >
>>> > volume glustervol-client-1
>>> >     type protocol/client
>>> >     option send-gids true
>>> >     option password b217da9d1d8b-bb55
>>> >     option username jkd76-4553-5347
>>> >     option transport-type tcp
>>> >     option remote-subvolume /pool/gluster
>>> >     option remote-host rep2
>>> >     option ping-timeout 42
>>> > end-volume
>>> >
>>> > volume glustervol-client-2
>>> >     type protocol/client
>>> >     option send-gids true
>>> >     option password b217da9d1d8b-bb55
>>> >     option username 19d7-5a190c2
>>> >     option transport-type tcp
>>> >     option remote-subvolume /pool/gluster
>>> >     option remote-host rep3
>>> >     option ping-timeout 42
>>> > end-volume
>>> >
>>> > volume glustervol-client-3
>>> >     type protocol/client
>>> >     option send-gids true
>>> >     option password b217da9d1d8b-bb55
>>> >     option username c75-5436b5a168347
>>> >     option transport-type tcp
>>> >     option remote-subvolume /pool/gluster
>>> >     option remote-host st1
>>> >
>>> >     option ping-timeout 42
>>> > end-volume
>>> >
>>> > volume glustervol-dht
>>> >     type cluster/distribute
>>> >     subvolumes glustervol-client-0 glustervol-client-1
>>> glustervol-client-2 glustervol-client-3
>>> > end-volume
>>> >
>>> > volume glustervol-write-behind
>>> >     type performance/write-behind
>>> >     subvolumes glustervol-dht
>>> > end-volume
>>> >
>>> > volume glustervol-read-ahead
>>> >     type performance/read-ahead
>>> >     subvolumes glustervol-write-behind
>>> > end-volume
>>> >
>>> > volume glustervol-io-cache
>>> >     type performance/io-cache
>>> >     subvolumes glustervol-read-ahead
>>> > end-volume
>>> >
>>> > volume glustervol-quick-read
>>> >     type performance/quick-read
>>> >     subvolumes glustervol-io-cache
>>> > end-volume
>>> >
>>> > volume glustervol-open-behind
>>> >     type performance/open-behind
>>> >     subvolumes glustervol-quick-read
>>> > end-volume
>>> >
>>> > volume glustervol-md-cache
>>> >     type performance/md-cache
>>> >     subvolumes glustervol-open-behind
>>> > end-volume
>>> >
>>> > volume glustervol
>>> >     type debug/io-stats
>>> >     option count-fop-hits off
>>> >     option latency-measurement off
>>> >     subvolumes glustervol-md-cache
>>> > end-volume
>>> >
>>> >
>>> > ap at rep3:~$ sudo gluster volume info
>>> >
>>> > Volume Name: glustervol
>>> > Type: Distribute
>>> > Volume ID: 165b-XXXXX
>>> > Status: Started
>>> > Number of Bricks: 4
>>> > Transport-type: tcp
>>> > Bricks:
>>> > Brick1: rep1:/pool/gluster
>>> > Brick2: rep2:/pool/gluster
>>> > Brick3: rep3:/pool/gluster
>>> > Brick4: st1:/pool/gluster
>>> >
>>> > Problem:
>>> >
>>> > If we shutdown any of the bricks , the volume size is reduced (this is
>>> ok) but from the other servers , i can see my mount point /data but it's
>>> only listing contents and i can't write or edit any single files/folders.
>>> >
>>> > Solution Required:
>>> >
>>> > If anyone brick is not available, From other servers should allow for
>>> Write and edit functions
>>> This is expected since you are using distributed volume. You wouldn't be
>>> able to write/edit files belonging to the brick which is down. Solution
>>> would be to migrate to distributed replicate volume.
>>> >
>>> > Please let us know, what can i try further ?
>>> >
>>> > Regards,
>>> > Varad
>>> >
>>> >
>>> > _______________________________________________
>>> > Gluster-users mailing list
>>> > Gluster-users at gluster.org
>>> > http://www.gluster.org/mailman/listinfo/gluster-users
>>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150520/8cfef64d/attachment.html>


More information about the Gluster-users mailing list