[Gluster-users] Removing subvolume from dist/rep volume

Nithya Balachandran nbalacha at redhat.com
Thu Jun 27 06:47:10 UTC 2019


Hi,


On Tue, 25 Jun 2019 at 15:26, Dave Sherohman <dave at sherohman.org> wrote:

> I have a 9-brick, replica 2+A cluster and plan to (permanently) remove
> one of the three subvolumes.  I think I've worked out how to do it, but
> want to verify first that I've got it right, since downtime or data loss
> would be Bad Things.
>
> The current configuration has six data bricks across six hosts (B
> through G), and all three arbiter bricks on the same host (A), such as
> one might create with
>
> # gluster volume create myvol replica 3 arbiter 1 B:/data C:/data A:/arb1
> D:/data E:/data A:/arb2 F:/data G:/data A:/arb3
>
>
> My objective is to remove nodes B and C entirely.
>
> First up is to pull their bricks from the volume:
>
> # gluster volume remove-brick myvol B:/data C:/data A:/arb1 start
> (wait for data to be migrated)
> # gluster volume remove-brick myvol B:/data C:/data A:/arb1 commit
>
>
There are some edge cases that may prevent a file from being migrated
during a remove-brick. Please do the following after this:

   1. Check the remove-brick status for any failures.  If there are any,
   check the rebalance log file for errors.
   2. Even if there are no failures, check the removed bricks to see if any
   files have not been migrated. If there are any, please check that they are
   valid files on the brick and copy them to the volume from the brick to the
   mount point.

The rest of the steps look good.

Regards,
Nithya

> And then remove the nodes with:
>
> # gluster peer detach B
> # gluster peer detach C
>
>
> Is this correct, or did I forget any steps and/or mangle the syntax on
> any commands?
>
> Also, for the remove-brick command, is there any way to throttle the
> amount of bandwidth which will be used for the data migration?
> Unfortunately, I was not able to provision a dedicated VLAN for the
> gluster servers to communicate among themselves, so I don't want it
> hogging all available capacity if that can be avoided.
>
>
> If it makes a difference, my gluster version is 3.12.15-1, running on
> Debian and installed from the debs at
>
> deb
> https://download.gluster.org/pub/gluster/glusterfs/3.12/LATEST/Debian/9/amd64/apt
> stretch main
>
> --
> Dave Sherohman
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190627/de1ff438/attachment.html>


More information about the Gluster-users mailing list