[Gluster-users] Fwd: Moving brick of replica volume to new mount on filesystem.
Joseph Jozwik
jjozwik at printsites.com
Tue Aug 26 21:29:29 UTC 2014
To add to this it appears that replace brick is in a broken state. I can't
abort it, or commit it. And I can run any other commands until it thinks
the replace-brick is complete.
Is there a way to manually remove the task since it failed?
root at pixel-glusterfs1:/# gluster volume status gdata2tb
Status of volume: gdata2tb
Gluster process Port Online Pid
------------------------------------------------------------------------------
Brick 10.0.1.31:/mnt/data2tb/gbrick3 49157 Y
14783
Brick 10.0.1.152:/mnt/raid10/gbrick3 49158 Y 2622
Brick 10.0.1.153:/mnt/raid10/gbrick3 49153 Y 3034
NFS Server on localhost 2049 Y
14790
Self-heal Daemon on localhost N/A Y
14794
NFS Server on 10.0.0.205 N/A N N/A
Self-heal Daemon on 10.0.0.205 N/A Y
10323
NFS Server on 10.0.1.153 2049 Y
12735
Self-heal Daemon on 10.0.1.153 N/A Y
12742
NFS Server on 10.0.1.152 2049 Y 2629
Self-heal Daemon on 10.0.1.152 N/A Y 2636
Task ID Status
---- -- ------
Replace brick 1dace9f0-ba98-4db9-9124-c962e74cce07 completed
---------- Forwarded message ----------
From: Joseph Jozwik <jjozwik at printsites.com>
Date: Tue, Aug 26, 2014 at 3:42 PM
Subject: Moving brick of replica volume to new mount on filesystem.
To: gluster-users at gluster.org
Hello,
I need to move a brick to another location on the filesystem.
My initial plan was to stop the gluster server with
1. service glusterfs-server stop
2. rsync -ap brick3 folder to new volume on server
3. umount old volume and bind mount the new to the same location.
However I stopped the glusterfs-server on the node and there was still
background processes running glusterd. So I was not sure how to safely stop
them.
I also attempted to replace-brick to a new location on the server but that
did not work with "volume replace-brick: failed: Commit failed on
localhost. Please check the log file for more details."
Then attempted remove brick with
"volume remove-brick gdata2tb replica 2 10.0.1.31:/mnt/data2tb/gbrick3
start"
gluster> volume remove-brick gdata2tb 10.0.1.31:/mnt/data2tb/gbrick3 status
volume remove-brick: failed: Volume gdata2tb is not a distribute volume or
contains only 1 brick.
Not performing rebalance
gluster>
Volume Name: gdata2tb
Type: Replicate
Volume ID: 6cbcb2fc-9fd7-467e-9561-bff1937e8492
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: 10.0.1.31:/mnt/data2tb/gbrick3
Brick2: 10.0.1.152:/mnt/raid10/gbrick3
Brick3: 10.0.1.153:/mnt/raid10/gbrick3
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140826/50f8da0d/attachment.html>
More information about the Gluster-users
mailing list