[Gluster-users] Removing subvolume from dist/rep volume
Nithya Balachandran
nbalacha at redhat.com
Fri Jun 28 09:56:00 UTC 2019
On Fri, 28 Jun 2019 at 14:34, Dave Sherohman <dave at sherohman.org> wrote:
> On Thu, Jun 27, 2019 at 12:17:10PM +0530, Nithya Balachandran wrote:
> > On Tue, 25 Jun 2019 at 15:26, Dave Sherohman <dave at sherohman.org> wrote:
> > > My objective is to remove nodes B and C entirely.
> > >
> > > First up is to pull their bricks from the volume:
> > >
> > > # gluster volume remove-brick myvol B:/data C:/data A:/arb1 start
> > > (wait for data to be migrated)
> > > # gluster volume remove-brick myvol B:/data C:/data A:/arb1 commit
> > >
> > >
> > There are some edge cases that may prevent a file from being migrated
> > during a remove-brick. Please do the following after this:
> >
> > 1. Check the remove-brick status for any failures. If there are any,
> > check the rebalance log file for errors.
> > 2. Even if there are no failures, check the removed bricks to see if
> any
> > files have not been migrated. If there are any, please check that
> they are
> > valid files on the brick and copy them to the volume from the brick
> to the
> > mount point.
> >
> > The rest of the steps look good.
>
> Apparently, they weren't quite right. I tried it and it just gives me
> the usage notes in return. Transcript of the commands and output is below.
>
> Any insight on how I got the syntax wrong?
>
> --- cut here ---
> root at merlin:/# gluster volume status
> Status of volume: palantir
> Gluster process TCP Port RDMA Port Online
> Pid
>
> ------------------------------------------------------------------------------
> Brick saruman:/var/local/brick0/data 49153 0 Y
> 17995
> Brick gandalf:/var/local/brick0/data 49153 0 Y
> 9415
> Brick merlin:/var/local/arbiter1/data 49170 0 Y
> 35034
> Brick azathoth:/var/local/brick0/data 49153 0 Y
> 25312
> Brick yog-sothoth:/var/local/brick0/data 49152 0 Y
> 10671
> Brick merlin:/var/local/arbiter2/data 49171 0 Y
> 35043
> Brick cthulhu:/var/local/brick0/data 49153 0 Y
> 21925
> Brick mordiggian:/var/local/brick0/data 49152 0 Y
> 12368
> Brick merlin:/var/local/arbiter3/data 49172 0 Y
> 35050
> Self-heal Daemon on localhost N/A N/A Y
> 1209
> Self-heal Daemon on saruman.lub.lu.se N/A N/A Y
> 23253
> Self-heal Daemon on gandalf.lub.lu.se N/A N/A Y
> 9542
> Self-heal Daemon on mordiggian.lub.lu.se N/A N/A Y
> 11016
> Self-heal Daemon on yog-sothoth.lub.lu.se N/A N/A Y
> 8126
> Self-heal Daemon on cthulhu.lub.lu.se N/A N/A Y
> 30998
> Self-heal Daemon on azathoth.lub.lu.se N/A N/A Y
> 34399
>
> Task Status of Volume palantir
>
> ------------------------------------------------------------------------------
> Task : Rebalance
> ID : e58bc091-5809-4364-af83-2b89bc5c7106
> Status : completed
>
> root at merlin:/# gluster volume remove-brick palantir
> saruman:/var/local/brick0/data gandalf:/var/local/brick0/data
> merlin:/var/local/arbiter1/data
>
>
You had it right in the first email.
gluster volume remove-brick palantir replica 3 arbiter 1
saruman:/var/local/brick0/data gandalf:/var/local/brick0/data
merlin:/var/local/arbiter1/data *start*
Usage:
> volume remove-brick <VOLNAME> [replica <COUNT>] <BRICK> ...
> <start|stop|status|commit|force>
>
> root at merlin:/# gluster volume remove-brick palantir replica 3 arbiter 1
> saruman:/var/local/brick0/data gandalf:/var/local/brick0/data
> merlin:/var/local/arbiter1/data
>
> Usage:
> volume remove-brick <VOLNAME> [replica <COUNT>] <BRICK> ...
> <start|stop|status|commit|force>
>
> root at merlin:/# gluster volume remove-brick palantir replica 3
> saruman:/var/local/brick0/data gandalf:/var/local/brick0/data
> merlin:/var/local/arbiter1/data
>
> Usage:
> volume remove-brick <VOLNAME> [replica <COUNT>] <BRICK> ...
> <start|stop|status|commit|force>
> --- cut here ---
>
> --
> Dave Sherohman
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190628/68888183/attachment.html>
More information about the Gluster-users
mailing list