[Gluster-users] delete brick / format / add empty brick

Ravishankar N ravishankar at redhat.com
Wed Jan 8 14:37:05 UTC 2014


On 01/08/2014 05:57 PM, Alun James wrote:
> I have this a go.
>
> /gluster volume add-brick myvol replica 2 server2:/brick1 /
> /gluster volume heal myvol full/
>
> It seems to be syncing the files but very slowly. Also the server load 
> on server01 has risen to 200+ and the gluster clients are no longer 
> able to access the mounts. Is there a way to do this that is not as 
> impactful? Could I manually rsync the bricks before adding the second 
> node back in?
>
The recommended way to heal is using the command mentioned. The gluster 
self heal daemon takes appropriate file locks before healing. Since 
clients are accessing the volume, I don't think bypassing that and 
rsyncing the bricks is a good idea.

Regards,
Ravi


>
> Alun.
>
> ------------------------------------------------------------------------
> *From: *"Ravishankar N" <ravishankar at redhat.com>
> *To: *"Alun James" <ajames at tibus.com>
> *Cc: *gluster-users at gluster.org
> *Sent: *Wednesday, 8 January, 2014 4:00:44 AM
> *Subject: *Re: [Gluster-users] delete brick / format / add empty brick
>
> On 01/07/2014 09:40 PM, Alun James wrote:
>
>     Hi folks,
>
>     I had a 2 node (1 brick each) replica, some network meltdown
>     issues seemed to cause problems with second node
>     (server02). glusterfsd process reaching 200-300% and errors
>     relating to split brain possibilities and self heal errors.
>
>     Original volume info:
>
>     /Volume Name: myvol/
>     /Type: Replicate/
>     /Status: Started/
>     /Number of Bricks: 2/
>     /Transport-type: tcp/
>     /Bricks:/
>     /Brick1: server01:/brick1/
>     /Brick2: server02:/brick1/
>
>     I removed the second brick (that was showing server problems).
>
>     /gluster volume remove-brick myvol replica 1 server02:/brick1/
>     /
>     /
>     Now the volume status is:
>
>     /Volume Name: tsfsvol0/
>     /Type: Distribute/
>     /Status: Started/
>     /Number of Bricks: 1/
>     /Transport-type: tcp/
>     /Bricks:/
>     /Brick1: server01:/brick1/
>     /
>     /
>     All is fine and the data on working server is sound.
>
>     The xfs partition for /server02:/brick1/ has been formatted and
>     therefore the data gone. All other gluster config data has
>     remained untouched. Can I re-add the second server to the volume
>     with an empty brick and the data will auto replicate over from the
>     working server?
>
>     /gluster volume add-brick myvol replica 2 server2:/brick1 ??/
>
>
> Yes this should work fine. You will need to run a  `gluster volume 
> heal myvol full` to manually trigger the replication.
>
>
>
>
>     ALUN JAMES
>     Senior Systems Engineer
>     Tibus
>
>     T: +44 (0)28 9033 1122
>     E: ajames at tibus.com
>     W: www.tibus.com <http://www.tibus.com>
>
>     Follow us on Twitter @tibus
>     <http://twitter.com/intent/user?screen_name=tibus>
>
>     Tibus is a trading name of The Internet Business Ltd, a company
>     limited by share capital and registered in Northern Ireland,
>     NI31325. It is part of UTV Media Plc.
>
>     This email and any attachment may contain confidential information
>     for the sole use of the intended recipient. Any review, use,
>     distribution or disclosure by others is strictly prohibited. If
>     you are not the intended recipient (or authorised to receive for
>     the recipient), please contact the sender by reply email and
>     delete all copies of this message.
>
>
>     _______________________________________________
>     Gluster-users mailing list
>     Gluster-users at gluster.org
>     http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140108/5445c58d/attachment.html>


More information about the Gluster-users mailing list