[Gluster-users] Proper command for replace-brick on distribute–replicate?

Alan Orth alan.orth at gmail.com
Wed Jun 12 08:08:53 UTC 2019


Dear Ravi,

Thanks for the confirmation—I replaced a brick in a volume last night and
by the morning I see that Gluster has replicated data there, though I don't
have any indication of its progress. The `gluster v heal volume info` and
`gluster v heal volume info split-brain` are all looking good so I guess
that's enough of an indication.

One question, though. Immediately after I replaced the brick I checked
`gluster v status volume` and I saw the following:

Task Status of Volume volume
------------------------------------------------------------------------------
Task                 : Rebalance
ID                   : a890e99c-5715-4bc1-80ee-c28490612135
Status               : not started

I did not initiate a rebalance, so it's strange to see it there. Is Gluster
hinting that I should start a rebalance? If a rebalance is "not started"
shouldn't Gluster just not show it at all?

Regarding the patch to the documentation: absolutely! Let me just get my
Gluster back in order after my confusing upgrade last month. :P

Thanks,

On Tue, Jun 11, 2019 at 7:32 PM Ravishankar N <ravishankar at redhat.com>
wrote:

>
> On 11/06/19 9:11 PM, Alan Orth wrote:
>
> Dear list,
>
> In a recent discussion on this list Ravi suggested that the documentation
> for replace-brick¹ was out of date. For a distribute–replicate volume the
> documentation currently says that we need to kill the old brick's PID,
> create a temporary empty directory on the FUSE mount, check the xattrs,
> replace-brick with commit force.
>
> Is all this still necessary? I'm running Gluster 5.6 on CentOS 7 with a
> distribute–replicate volume.
>
> No,  all these very steps are 'codified' into the `replace brick commit
> force` command via https://review.gluster.org/#/c/glusterfs/+/10076/ and
> https://review.gluster.org/#/c/glusterfs/+/10448/.  You can see the
> commit messages of these 2 patches for more details.
>
> You can play around with most of these commands in a 1 node setup if you
> want to convince yourself that they work. There is no need to form a
> cluster.
> [root at tuxpad glusterfs]# gluster v create testvol replica 3 127.0.0.2:/home/ravi/bricks/brick{1..3}
> force
> [root at tuxpad glusterfs]# gluster v start testvol
> [root at tuxpad glusterfs]# mount -t glusterfs 127.0.0.2:testvol
> /mnt/fuse_mnt/
> [root at tuxpad glusterfs]# touch /mnt/fuse_mnt/FILE
> [root at tuxpad glusterfs]# ll /home/ravi/bricks/brick*/FILE
> -rw-r--r--. 2 root root 0 Jun 11 21:55 /home/ravi/bricks/brick1/FILE
> -rw-r--r--. 2 root root 0 Jun 11 21:55 /home/ravi/bricks/brick2/FILE
> -rw-r--r--. 2 root root 0 Jun 11 21:55 /home/ravi/bricks/brick3/FILE
>
> [root at tuxpad glusterfs]# gluster v replace-brick testvol 127.0.0.2:/home/ravi/bricks/brick3
> 127.0.0.2:/home/ravi/bricks/brick3_new commit force
> volume replace-brick: success: replace-brick commit force operation
> successful
> [root at tuxpad glusterfs]# ll /home/ravi/bricks/brick3_new/FILE
>
> -rw-r--r--. 2 root root 0 Jun 11 21:55 /home/ravi/bricks/brick3_new/FILE
> Why don't you send a patch to update the doc for replace-brick? I'd be
> happy to review it. ;-)
> HTH,
> Ravi
>
>
> Thank you,
>
> ¹ https://docs.gluster.org/en/latest/Administrator Guide/Managing Volumes/
> --
> Alan Orth
> alan.orth at gmail.com
> https://picturingjordan.com
> https://englishbulgaria.net
> https://mjanja.ch
> "In heaven all the interesting people are missing." ―Friedrich Nietzsche
>
> _______________________________________________
> Gluster-users mailing listGluster-users at gluster.orghttps://lists.gluster.org/mailman/listinfo/gluster-users
>
>

-- 
Alan Orth
alan.orth at gmail.com
https://picturingjordan.com
https://englishbulgaria.net
https://mjanja.ch
"In heaven all the interesting people are missing." ―Friedrich Nietzsche
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190612/0505d714/attachment.html>


More information about the Gluster-users mailing list