[Gluster-users] Arbitier doesn't create
Ravishankar N
ravishankar at redhat.com
Wed Mar 23 15:39:21 UTC 2016
On 03/23/2016 06:56 PM, Ralf Simon wrote:
> Hello,
>
> I've installed ....
>
> # yum info glusterfs-server
> Loaded plugins: fastestmirror
> Loading mirror speeds from cached hostfile
> Installed Packages
> Name : glusterfs-server
> Arch : x86_64
> Version : 3.7.6
I think the patch http://review.gluster.org/#/c/12748/ to enhance the
volinfo output went in for 3.7.7. But I would advise you to use 3.7.9
that was recently released because a couple of arbiter related fixes
have happened since 3.7.6
Hope this helps,
Ravi
>
> Release : 1.el7
> Size : 4.3 M
> Repo : installed
> From repo : latest
> Summary : Clustered file-system server
> URL : http://www.gluster.org/docs/index.php/GlusterFS
> License : GPLv2 or LGPLv3+
> Description : GlusterFS is a distributed file-system capable of
> scaling to several
> : petabytes. It aggregates various storage bricks over Infiniband RDMA
> : or TCP/IP interconnect into one large parallel network file
> : system. GlusterFS is one of the most sophisticated file systems in
> : terms of features and extensibility. It borrows a powerful concept
> : called Translators from GNU Hurd kernel. Much of the code in GlusterFS
> : is in user space and easily manageable.
> :
> : This package provides the glusterfs server daemon.
>
> I wanted to build a ...
>
> # gluster volume create gv0 replica 3 arbiter 1 d90029:/data/brick0
> d90031:/data/brick0 d90034:/data/brick0
> volume create: gv0: success: please start the volume to access data
>
> ... but I got a ...
>
> # gluster volume info
>
> Volume Name: gv0
> Type: Replicate
> Volume ID: 329325fc-ceed-4dee-926f-038f44281678
> Status: Created
> Number of Bricks: *1 x 3 = 3*
> Transport-type: tcp
> Bricks:
> Brick1: d90029:/data/brick0
> Brick2: d90031:/data/brick0
> Brick3: d90034:/data/brick0
> Options Reconfigured:
> performance.readdir-ahead: on
>
> ... without the requested arbiter !
>
> The same situation with 6 bricks ...
>
> # gluster volume create gv0 replica 3 arbiter 1 d90029:/data/brick0
> d90031:/data/brick0 d90034:/data/brick0 d90029:/data/brick1
> d90031:/data/brick1 d90034:/data/brick1
> volume create: gv0: success: please start the volume to access data
> [root at d90029 ~]# gluster vol info
>
> Volume Name: gv0
> Type: Distributed-Replicate
> Volume ID: 2b8dbcc0-c4bb-41e3-a870-e164d8d10c49
> Status: Created
> Number of Bricks: *2 x 3 = 6*
> Transport-type: tcp
> Bricks:
> Brick1: d90029:/data/brick0
> Brick2: d90031:/data/brick0
> Brick3: d90034:/data/brick0
> Brick4: d90029:/data/brick1
> Brick5: d90031:/data/brick1
> Brick6: d90034:/data/brick1
> Options Reconfigured:
> performance.readdir-ahead: on
>
>
> In contrast the documentation tells ....
>
>
> *Arbiter configuration*
>
> The arbiter configuration a.k.a. the arbiter volume is the perfect
> sweet spot between a 2-way replica and 3-way replica to avoid files
> getting into split-brain, */without the 3x storage space/* as
> mentioned earlier. The syntax for creating the volume is:
>
> *gluster volume create replica 3 arbiter 1 host1:brick1 host2:brick2
> host3:brick3*
>
> For example:
>
> *gluster volume create testvol replica 3 arbiter 1
> 127.0.0.2:/bricks/brick{1..6} force*
>
> volume create: testvol: success: please start the volume to access data
>
> *gluster volume info*
>
> Volume Name: testvol
> Type: Distributed-Replicate
> Volume ID: ae6c4162-38c2-4368-ae5d-6bad141a4119
> Status: Created
> Number of Bricks: *2 x (2 + 1) = 6*
> Transport-type: tcp
> Bricks:
> Brick1: 127.0.0.2:/bricks/brick1
> Brick2: 127.0.0.2:/bricks/brick2
> Brick3: 127.0.0.2:/bricks/brick3 *(arbiter)*
> Brick4: 127.0.0.2:/bricks/brick4
> Brick5: 127.0.0.2:/bricks/brick5
> Brick6: 127.0.0.2:/bricks/brick6 *(arbiter)*
> Options Reconfigured : transport.address-family: inet
> performance.readdir-ahead: on `
>
>
>
> What's going wrong ? Can anybody help ?
>
> Kind Regards
> Ralf Simon
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160323/2b9012a5/attachment.html>
More information about the Gluster-users
mailing list