[Gluster-users] delete brick / format / add empty brick

Lalatendu Mohanty lmohanty at redhat.com
Wed Jan 8 14:59:25 UTC 2014


On 01/08/2014 08:10 PM, Alun James wrote:
> Is it normal to expect very high server load and clients being unable 
> to access the mounts during this process? It means the application 
> running on this will need to be offline for hours.
>

Nope. I never seen glusterfsd reaching 200% cpu . However it depends on 
the server hardware configuration also. what is the RAM size of your 
servers? Also I will suggest you to check logs and see anything unusual 
happening.


> ------------------------------------------------------------------------
> *From: *"Ravishankar N" <ravishankar at redhat.com>
> *To: *"Alun James" <ajames at tibus.com>
> *Cc: *gluster-users at gluster.org
> *Sent: *Wednesday, 8 January, 2014 2:37:05 PM
> *Subject: *Re: [Gluster-users] delete brick / format / add empty brick
>
> On 01/08/2014 05:57 PM, Alun James wrote:
>
>     I have this a go.
>
>     /gluster volume add-brick myvol replica 2 server2:/brick1 /
>     /gluster volume heal myvol full/
>
>     It seems to be syncing the files but very slowly. Also the server
>     load on server01 has risen to 200+ and the gluster clients are no
>     longer able to access the mounts. Is there a way to do this that
>     is not as impactful? Could I manually rsync the bricks before
>     adding the second node back in?
>
> The recommended way to heal is using the command mentioned. The 
> gluster self heal daemon takes appropriate file locks before healing. 
> Since clients are accessing the volume, I don't think bypassing that 
> and rsyncing the bricks is a good idea.
>
> Regards,
> Ravi
>
>
>
>     Alun.
>
>     ------------------------------------------------------------------------
>     *From: *"Ravishankar N" <ravishankar at redhat.com>
>     *To: *"Alun James" <ajames at tibus.com>
>     *Cc: *gluster-users at gluster.org
>     *Sent: *Wednesday, 8 January, 2014 4:00:44 AM
>     *Subject: *Re: [Gluster-users] delete brick / format / add empty brick
>
>     On 01/07/2014 09:40 PM, Alun James wrote:
>
>         Hi folks,
>
>         I had a 2 node (1 brick each) replica, some network meltdown
>         issues seemed to cause problems with second node
>         (server02). glusterfsd process reaching 200-300% and errors
>         relating to split brain possibilities and self heal errors.
>
>         Original volume info:
>
>         /Volume Name: myvol/
>         /Type: Replicate/
>         /Status: Started/
>         /Number of Bricks: 2/
>         /Transport-type: tcp/
>         /Bricks:/
>         /Brick1: server01:/brick1/
>         /Brick2: server02:/brick1/
>
>         I removed the second brick (that was showing server problems).
>
>         /gluster volume remove-brick myvol replica 1 server02:/brick1/
>         /
>         /
>         Now the volume status is:
>
>         /Volume Name: tsfsvol0/
>         /Type: Distribute/
>         /Status: Started/
>         /Number of Bricks: 1/
>         /Transport-type: tcp/
>         /Bricks:/
>         /Brick1: server01:/brick1/
>         /
>         /
>         All is fine and the data on working server is sound.
>
>         The xfs partition for /server02:/brick1/ has been formatted
>         and therefore the data gone. All other gluster config data has
>         remained untouched. Can I re-add the second server to the
>         volume with an empty brick and the data will auto replicate
>         over from the working server?
>
>         /gluster volume add-brick myvol replica 2 server2:/brick1 ??/
>
>
>     Yes this should work fine. You will need to run a  `gluster volume
>     heal myvol full` to manually trigger the replication.
>
>
>
>
>         ALUN JAMES
>         Senior Systems Engineer
>         Tibus
>
>         T: +44 (0)28 9033 1122
>         E: ajames at tibus.com
>         W: www.tibus.com <http://www.tibus.com>
>
>         Follow us on Twitter @tibus
>         <http://twitter.com/intent/user?screen_name=tibus>
>
>         Tibus is a trading name of The Internet Business Ltd, a
>         company limited by share capital and registered in Northern
>         Ireland, NI31325. It is part of UTV Media Plc.
>
>         This email and any attachment may contain confidential
>         information for the sole use of the intended recipient. Any
>         review, use, distribution or disclosure by others is strictly
>         prohibited. If you are not the intended recipient (or
>         authorised to receive for the recipient), please contact the
>         sender by reply email and delete all copies of this message.
>
>
>         _______________________________________________
>         Gluster-users mailing list
>         Gluster-users at gluster.org
>         http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140108/33b75c02/attachment.html>


More information about the Gluster-users mailing list