[Gluster-users] delete brick / format / add empty brick
Ravishankar N
ravishankar at redhat.com
Wed Jan 8 14:37:05 UTC 2014
On 01/08/2014 05:57 PM, Alun James wrote:
> I have this a go.
>
> /gluster volume add-brick myvol replica 2 server2:/brick1 /
> /gluster volume heal myvol full/
>
> It seems to be syncing the files but very slowly. Also the server load
> on server01 has risen to 200+ and the gluster clients are no longer
> able to access the mounts. Is there a way to do this that is not as
> impactful? Could I manually rsync the bricks before adding the second
> node back in?
>
The recommended way to heal is using the command mentioned. The gluster
self heal daemon takes appropriate file locks before healing. Since
clients are accessing the volume, I don't think bypassing that and
rsyncing the bricks is a good idea.
Regards,
Ravi
>
> Alun.
>
> ------------------------------------------------------------------------
> *From: *"Ravishankar N" <ravishankar at redhat.com>
> *To: *"Alun James" <ajames at tibus.com>
> *Cc: *gluster-users at gluster.org
> *Sent: *Wednesday, 8 January, 2014 4:00:44 AM
> *Subject: *Re: [Gluster-users] delete brick / format / add empty brick
>
> On 01/07/2014 09:40 PM, Alun James wrote:
>
> Hi folks,
>
> I had a 2 node (1 brick each) replica, some network meltdown
> issues seemed to cause problems with second node
> (server02). glusterfsd process reaching 200-300% and errors
> relating to split brain possibilities and self heal errors.
>
> Original volume info:
>
> /Volume Name: myvol/
> /Type: Replicate/
> /Status: Started/
> /Number of Bricks: 2/
> /Transport-type: tcp/
> /Bricks:/
> /Brick1: server01:/brick1/
> /Brick2: server02:/brick1/
>
> I removed the second brick (that was showing server problems).
>
> /gluster volume remove-brick myvol replica 1 server02:/brick1/
> /
> /
> Now the volume status is:
>
> /Volume Name: tsfsvol0/
> /Type: Distribute/
> /Status: Started/
> /Number of Bricks: 1/
> /Transport-type: tcp/
> /Bricks:/
> /Brick1: server01:/brick1/
> /
> /
> All is fine and the data on working server is sound.
>
> The xfs partition for /server02:/brick1/ has been formatted and
> therefore the data gone. All other gluster config data has
> remained untouched. Can I re-add the second server to the volume
> with an empty brick and the data will auto replicate over from the
> working server?
>
> /gluster volume add-brick myvol replica 2 server2:/brick1 ??/
>
>
> Yes this should work fine. You will need to run a `gluster volume
> heal myvol full` to manually trigger the replication.
>
>
>
>
> ALUN JAMES
> Senior Systems Engineer
> Tibus
>
> T: +44 (0)28 9033 1122
> E: ajames at tibus.com
> W: www.tibus.com <http://www.tibus.com>
>
> Follow us on Twitter @tibus
> <http://twitter.com/intent/user?screen_name=tibus>
>
> Tibus is a trading name of The Internet Business Ltd, a company
> limited by share capital and registered in Northern Ireland,
> NI31325. It is part of UTV Media Plc.
>
> This email and any attachment may contain confidential information
> for the sole use of the intended recipient. Any review, use,
> distribution or disclosure by others is strictly prohibited. If
> you are not the intended recipient (or authorised to receive for
> the recipient), please contact the sender by reply email and
> delete all copies of this message.
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140108/5445c58d/attachment.html>
More information about the Gluster-users
mailing list