[Gluster-users] Re-provisioning a node and it's bricks

John Mark Walker johnmark at redhat.com
Thu Sep 6 07:01:37 UTC 2012


Eric - was good to see you in San Diego. Glad to see you on the list. 

I would recommend trying the IRC channel tomorrow morning. Should be someone there who can help you. 

-JM 

----- Original Message -----

> I've created a distributed replicated volume:

> > gluster> volume info
> >
> > Volume Name: Repositories
> > Type: Distributed-Replicate
> > Volume ID: 926262ae-2aa6-4bf7-b19e-cf674431b06c
> > Status: Started
> > Number of Bricks: 2 x 2 = 4
> > Transport-type: tcp
> > Bricks:
> > Brick1: 192.168.1.1:/srv/sda7
> > Brick2: 192.168.1.2:/srv/sda7
> > Brick3: 192.168.1.1:/srv/sdb7
> > Brick4: 192.168.1.2:/srv/sdb7

> ...by allocating physical partitions on each HDD of each node for the
> volumes' bricks: e.g.,

> > [eric at sn1 srv]$ mount | grep xfs
> > /dev/sda7 on /srv/sda7 type xfs (rw)
> > /dev/sdb7 on /srv/sdb7 type xfs (rw)
> > /dev/sda8 on /srv/sda8 type xfs (rw)
> > /dev/sdb8 on /srv/sdb8 type xfs (rw)
> > /dev/sda9 on /srv/sda9 type xfs (rw)
> > /dev/sdb9 on /srv/sdb9 type xfs (rw)
> > /dev/sda10 on /srv/sda10 type xfs (rw)
> > /dev/sdb10 on /srv/sdb10 type xfs (rw)

> I plan to re-provision both nodes (e.g., convert them from CentOS ->
> SLES) and need to preserve the data on the current bricks.

> It seems to me that the procedure for this endeavor would be to:
> detach the node that will be re-provisioned; re-provision the node;
> add the node back to the trusted storage pool, and then; add the
> bricks back to the volume - *but* this plan fails at Step #1. i.e.,

> * When attempting to detach the second node from the volume, the
> Console Manager
> complains "Brick(s) with the peer 192.168.1.2 exist in cluster".
> * When attempting to remove the second node's bricks from the volume,
> the Console
> Manager complains "Bricks not from same subvol for replica".

> Is this even feasible? I've already verified that bricks can be
> *added* to the volume (by adding two additional local partitions to
> the volume) but I'm not sure where to begin preparing the nodes for
> re-provisioning.

> Eric Pretorious
> Truckee, CA

> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20120906/0ecd3752/attachment.html>


More information about the Gluster-users mailing list