[Gluster-users] distributed replicated volume - added bricks

Herb Burnswell herbert.burnswell at gmail.com
Fri Oct 18 22:40:40 UTC 2019


All,

We recently added 4 new bricks to an establish distributed replicated
volume.  The original volume was created via gdeploy as:

[volume1]
action=create
volname=tank
replica_count=2
force=yes
key=performance.parallel-readdir,network.inode-lru-limit,performance.md-cache-timeout,performance.cache-invalidation,performance.stat-prefetch,features.cache-invalidation-timeout,features.cache-invalidation,performance.cache-samba-metadata
value=on,500000,600,on,on,600,on,on
brick_dirs=/gluster_bricks/data1,/gluster_bricks/data2
ignore_errors=no

This created the volume as:

# gluster vol status tank
Status of volume: tank
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick server1:/gluster_bricks/data1       49162     0          Y       20318
Brick server2:/gluster_bricks/data1       49166     0          Y       3432
Brick server1:/gluster_bricks/data2       49163     0          Y       20323
Brick server2:/gluster_bricks/data2       49167     0          Y       3435
Self-heal Daemon on localhost               N/A       N/A        Y
25874
Self-heal Daemon on server2               N/A       N/A        Y       12536

Task Status of Volume tank
------------------------------------------------------------------------------
There are no active volume tasks

I have read (
https://docs.gluster.org/en/latest/Administrator%20Guide/Setting%20Up%20Volumes)
that the way one creates distributed replicated volumes is sensitive to
replica-sets:

*Note*: The number of bricks should be a multiple of the replica count for
a distributed replicated volume. Also, the order in which bricks are
specified has a great effect on data protection. Each replica_count
consecutive bricks in the list you give will form a replica set, with all
replica sets combined into a volume-wide distribute set. To make sure that
replica-set members are not placed on the same node, list the first brick
on every server, then the second brick on every server in the same order,
and so on.

I just noticed that the way we added the new bricks we did not indicate a
replica:

gluster volume add-brick tank server1:/gluster_bricks/data3
server1:/gluster_bricks/data4 force
gluster volume add-brick tank server1:/gluster_bricks/data5
server1:/gluster_bricks/data6 force
gluster volume add-brick tank server2:/gluster_bricks/data3
server2:/gluster_bricks/data4 force
gluster volume add-brick tank server2:/gluster_bricks/data5
server2:/gluster_bricks/data6 force

Which modified the volume as:

# gluster vol status tank
Status of volume: tank
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick server1:/gluster_bricks/data1       49162     0          Y       20318
Brick server2:/gluster_bricks/data1       49166     0          Y       3432
Brick server1:/gluster_bricks/data2       49163     0          Y       20323
Brick server2:/gluster_bricks/data2       49167     0          Y       3435
Brick server1:/gluster_bricks/data3       49164     0          Y       4625
Brick server1:/gluster_bricks/data4       49165     0          Y       4644
Brick server1:/gluster_bricks/data5       49166     0          Y       5088
Brick server1:/gluster_bricks/data6       49167     0          Y       5128
Brick server2:/gluster_bricks/data3       49168     0          Y       22314
Brick server2:/gluster_bricks/data4       49169     0          Y       22345
Brick server2:/gluster_bricks/data5       49170     0          Y       22889
Brick server2:/gluster_bricks/data6       49171     0          Y       22932
Self-heal Daemon on localhost               N/A       N/A        Y
12366
Self-heal Daemon on server2               N/A       N/A        Y       21446

Task Status of Volume tank
------------------------------------------------------------------------------
Task                 : Rebalance
ID                   : ec958aee-edbd-4106-b896-97c688fde0e3
Status               : completed

As you can see the added 3,4,5,6 bricks appear differently:

Brick server1:/gluster_bricks/data3       49164     0          Y       4625
Brick server1:/gluster_bricks/data4       49165     0          Y       4644
Brick server1:/gluster_bricks/data5       49166     0          Y       5088
Brick server1:/gluster_bricks/data6       49167     0          Y       5128
Brick server2:/gluster_bricks/data3       49168     0          Y       22314
Brick server2:/gluster_bricks/data4       49169     0          Y       22345
Brick server2:/gluster_bricks/data5       49170     0          Y       22889
Brick server2:/gluster_bricks/data6       49171     0          Y
22932

My question is what does this mean for the volume?  Everything appears to
be running as expected, but:

- Is there a serious problem with the way the volume is now configured?
- Have we messed up the high availability of the 2 nodes?
- Is there a way to reconfigure the volume to get it to a more optimal
state?

Any help is greatly appreciated...

Thanks in advance,

HB
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20191018/88e3d504/attachment.html>


More information about the Gluster-users mailing list