[Gluster-users] Distributed volume going to Read only mode if any of the Brick is not available

Varadharajan S rajanvaradhu at gmail.com
Tue May 19 14:58:10 UTC 2015


FYI
On 19 May 2015 20:25, "Varadharajan S" <rajanvaradhu at gmail.com> wrote:

> Hi,
> Replication means, I won't get space. Distribution is not like striping
> right? If one brick is not available in the volume, other bricks can
> distribute data in between. If I do any tuning will get solution?
>  On 19 May 2015 20:02, "Atin Mukherjee" <atin.mukherjee83 at gmail.com>
> wrote:
>
>>
>> On 19 May 2015 17:10, "Varadharajan S" <rajanvaradhu at gmail.com> wrote:
>> >
>> > Hi,
>> >
>> > We are using Ubuntu 14.04 server and for storage purpose we configured
>> gluster 3.5 as distributed volume and find the below details,
>> >
>> > 1).4 Servers - 14.04 Ubuntu Server and each server disks free spaces
>> are  configured as ZFS raiddz2 volume
>> >
>> > 2). Each server has /pool/gluster zfs volume and capacity as - 5 TB,8
>> TB,6 TB and 10 TB
>> >
>> > 3). Bricks are - rep1,rep2,rep3 and st1 and all the bricks are
>> connected as Distributed Volume and mounted on each system as,
>> >
>> >   For E.x in rep1 -> mount -t glusterfs  rep1:/glustervol  /data.
>> >                   rep2  -> mount  -t glusterfs  rep2:/glustervol  /data
>> >                   rep3  -> mount  -t glusterfs  rep3:/glustervol  /data
>> >                   st1    ->  mount  -t glusterfs  st1:/glustervol  /data
>> >
>> > So we get /data is having  around 29 TB and all our applications data's
>> are stored in /data mount point.
>> >
>> > Details about volume:
>> >
>> > volume glustervol-client-0
>> >     type protocol/client
>> >     option send-gids true
>> >     option password b217da9d1d8b-bb55
>> >     option username 9d76-4553-8c75
>> >     option transport-type tcp
>> >     option remote-subvolume /pool/gluster
>> >     option remote-host rep1
>> >     option ping-timeout 42
>> > end-volume
>> >
>> > volume glustervol-client-1
>> >     type protocol/client
>> >     option send-gids true
>> >     option password b217da9d1d8b-bb55
>> >     option username jkd76-4553-5347
>> >     option transport-type tcp
>> >     option remote-subvolume /pool/gluster
>> >     option remote-host rep2
>> >     option ping-timeout 42
>> > end-volume
>> >
>> > volume glustervol-client-2
>> >     type protocol/client
>> >     option send-gids true
>> >     option password b217da9d1d8b-bb55
>> >     option username 19d7-5a190c2
>> >     option transport-type tcp
>> >     option remote-subvolume /pool/gluster
>> >     option remote-host rep3
>> >     option ping-timeout 42
>> > end-volume
>> >
>> > volume glustervol-client-3
>> >     type protocol/client
>> >     option send-gids true
>> >     option password b217da9d1d8b-bb55
>> >     option username c75-5436b5a168347
>> >     option transport-type tcp
>> >     option remote-subvolume /pool/gluster
>> >     option remote-host st1
>> >
>> >     option ping-timeout 42
>> > end-volume
>> >
>> > volume glustervol-dht
>> >     type cluster/distribute
>> >     subvolumes glustervol-client-0 glustervol-client-1
>> glustervol-client-2 glustervol-client-3
>> > end-volume
>> >
>> > volume glustervol-write-behind
>> >     type performance/write-behind
>> >     subvolumes glustervol-dht
>> > end-volume
>> >
>> > volume glustervol-read-ahead
>> >     type performance/read-ahead
>> >     subvolumes glustervol-write-behind
>> > end-volume
>> >
>> > volume glustervol-io-cache
>> >     type performance/io-cache
>> >     subvolumes glustervol-read-ahead
>> > end-volume
>> >
>> > volume glustervol-quick-read
>> >     type performance/quick-read
>> >     subvolumes glustervol-io-cache
>> > end-volume
>> >
>> > volume glustervol-open-behind
>> >     type performance/open-behind
>> >     subvolumes glustervol-quick-read
>> > end-volume
>> >
>> > volume glustervol-md-cache
>> >     type performance/md-cache
>> >     subvolumes glustervol-open-behind
>> > end-volume
>> >
>> > volume glustervol
>> >     type debug/io-stats
>> >     option count-fop-hits off
>> >     option latency-measurement off
>> >     subvolumes glustervol-md-cache
>> > end-volume
>> >
>> >
>> > ap at rep3:~$ sudo gluster volume info
>> >
>> > Volume Name: glustervol
>> > Type: Distribute
>> > Volume ID: 165b-XXXXX
>> > Status: Started
>> > Number of Bricks: 4
>> > Transport-type: tcp
>> > Bricks:
>> > Brick1: rep1:/pool/gluster
>> > Brick2: rep2:/pool/gluster
>> > Brick3: rep3:/pool/gluster
>> > Brick4: st1:/pool/gluster
>> >
>> > Problem:
>> >
>> > If we shutdown any of the bricks , the volume size is reduced (this is
>> ok) but from the other servers , i can see my mount point /data but it's
>> only listing contents and i can't write or edit any single files/folders.
>> >
>> > Solution Required:
>> >
>> > If anyone brick is not available, From other servers should allow for
>> Write and edit functions
>> This is expected since you are using distributed volume. You wouldn't be
>> able to write/edit files belonging to the brick which is down. Solution
>> would be to migrate to distributed replicate volume.
>> >
>> > Please let us know, what can i try further ?
>> >
>> > Regards,
>> > Varad
>> >
>> >
>> > _______________________________________________
>> > Gluster-users mailing list
>> > Gluster-users at gluster.org
>> > http://www.gluster.org/mailman/listinfo/gluster-users
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150519/367db631/attachment.html>


More information about the Gluster-users mailing list