[Gluster-users] Proper procedure to reduce an active volume

Diego Zuccato diego.zuccato at unibo.it
Fri Feb 5 06:57:27 UTC 2021

Il 04/02/21 19:28, Nag Pavan Chilakam ha scritto:

>     What is the proper procedure to reduce a "replica 3 arbiter 1" volume?
> Can you kindly elaborate the volume configuration.  Is this a plain
> arbiter volume or is it a distributed arbiter volume?
> Please share the volume info so that we can help you better
Sure. Here it is. Shortened a bit :)
# gluster v info

Volume Name: BigVol
Type: Distributed-Replicate
Volume ID: c51926bd-6715-46b2-8bb3-8c915ec47e28
Status: Started
Snapshot Count: 0
Number of Bricks: 28 x (2 + 1) = 84
Transport-type: tcp
Brick1: str957-biostor2:/srv/bricks/00/BigVol
Brick2: str957-biostor:/srv/bricks/00/BigVol
Brick3: str957-biostq:/srv/arbiters/00/BigVol (arbiter)
Brick4: str957-biostor2:/srv/bricks/01/BigVol
Brick5: str957-biostor:/srv/bricks/01/BigVol
Brick6: str957-biostq:/srv/arbiters/01/BigVol (arbiter)
Brick79: str957-biostor:/srv/bricks/26/BigVol
Brick80: str957-biostor2:/srv/bricks/26/BigVol
Brick81: str957-biostq:/srv/arbiters/26/BigVol (arbiter)
Brick82: str957-biostor:/srv/bricks/27/BigVol
Brick83: str957-biostor2:/srv/bricks/27/BigVol
Brick84: str957-biostq:/srv/arbiters/27/BigVol (arbiter)
Options Reconfigured:
features.scrub-throttle: aggressive
server.manage-gids: on
features.quota-deem-statfs: on
features.inode-quota: on
features.quota: on
cluster.self-heal-daemon: enable
ssl.certificate-depth: 1
auth.ssl-allow: str957-bio*
features.scrub-freq: biweekly
features.scrub: Active
features.bitrot: on
transport.address-family: inet
performance.readdir-ahead: on
nfs.disable: on
client.ssl: on
server.ssl: on
server.event-threads: 8
client.event-threads: 8
cluster.granular-entry-heal: enable

>     The procedure I've found is:
>     1) # gluster volume remove-brick VOLNAME BRICK start
>     (repeat for each brick to be removed, but being a r3a1 should I remove
>     both bricks and the arbiter in a single command or multiple ones?)
> No , you can mention bricks of a distributed subvolume in one command.
> If you are having a 1x(2+1a) volume , then you should mention only one
> brick. Start by removing the arbiter brick

>     2) # gluster volume remove-brick VOLNAME BRICK status
>     (to monitor migration)
>     3) # gluster volume remove-brick VOLNAME BRICK commit
>     (to finalize the removal)
>     4) umount and reformat the freed (now unused) bricks
>     Is this safe?
> What is the actual need to remove bricks?
I need to move a couple of disks to a new server, to keep it all well
balanced and increase the available space.

> If you feel this volume is not needed anymore , then just delete the
> volume, instead of going through each brick deletion
Nono, the volume is needed and is currently hosting data I cannot
lose... But I haven't space to copy it elsewhere...

Diego Zuccato
DIFA - Dip. di Fisica e Astronomia
Servizi Informatici
Alma Mater Studiorum - Università di Bologna
V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
tel.: +39 051 20 95786

More information about the Gluster-users mailing list