[Gluster-users] How to remove a dead node and re-balance volume?

Harshavardhana harsha at harshavardhana.net
Sun Sep 8 23:53:16 UTC 2013


Joe,

Perhaps a typo

"""So first we move server1:/data/brick2 to server3:/data/brick1""" -
http://joejulian.name/blog/how-to-expand-glusterfs-replicated-clusters-by-one-server/

Should be "server3:/data/brick2"


On Sun, Sep 8, 2013 at 12:34 PM, Joe Julian <joe at julianfamily.org> wrote:

>  On 09/05/2013 02:16 AM, Anup Nair wrote:
>
> On Thu, Sep 5, 2013 at 12:41 AM, Vijay Bellur <vbellur at redhat.com> wrote:
>
>>  On 09/03/2013 01:18 PM, Anup Nair wrote:
>>
>>>  Glusterfs version 3.2.2
>>>
>>> I have a Gluster volume in which one our of the 4 peers/nodes had
>>> crashed some time ago, prior to my joining service here.
>>>
>>> I see from volume info that the crashed (non-existing) node is still
>>> listed as one of the peers and the bricks are also listed. I would like
>>> to detach this node and its bricks and rebalance the volume with
>>> remaining 3 peers. But I am unable to do so. Here are my setps:
>>>
>>> 1. #gluster peer status
>>>    Number of Peers: 3 -- (note: excluding the one I run this command
>>> from)
>>>
>>>    Hostname: dbstore4r294 --- (note: node/peer that is down)
>>>    Uuid: 8bf13458-1222-452c-81d3-565a563d768a
>>>    State: Peer in Cluster (Disconnected)
>>>
>>>    Hostname: 172.16.1.90
>>>    Uuid: 77ebd7e4-7960-4442-a4a4-00c5b99a61b4
>>>    State: Peer in Cluster (Connected)
>>>
>>>    Hostname: dbstore3r294
>>>    Uuid: 23d7a18c-fe57-47a0-afbc-1e1a5305c0eb
>>>    State: Peer in Cluster (Connected)
>>>
>>> 2. #gluster peer detach dbstore4r294
>>>    Brick(s) with the peer dbstore4r294 exist in cluster
>>>
>>> 3. #gluster volume info
>>>
>>>    Volume Name: test-volume
>>>    Type: Distributed-Replicate
>>>    Status: Started
>>>    Number of Bricks: 4 x 2 = 8
>>>    Transport-type: tcp
>>>    Bricks:
>>>    Brick1: dbstore1r293:/datastore1
>>>    Brick2: dbstore2r293:/datastore1
>>>    Brick3: dbstore3r294:/datastore1
>>>    Brick4: dbstore4r294:/datastore1
>>>    Brick5: dbstore1r293:/datastore2
>>>    Brick6: dbstore2r293:/datastore2
>>>    Brick7: dbstore3r294:/datastore2
>>>    Brick8: dbstore4r294:/datastore2
>>>    Options Reconfigured:
>>>    network.ping-timeout: 42s
>>>    performance.cache-size: 64MB
>>>    performance.write-behind-window-size: 3MB
>>>    performance.io-thread-count: 8
>>>    performance.cache-refresh-timeout: 2
>>>
>>> Note that the non-existent node/peer is  -- dbstore4r294 (bricks are
>>> :/datastore1 & /datastore2  - i.e.  brick4 and brick8)
>>>
>>> 4. #gluster volume remove-brick test-volume dbstore4r294:/datastore1
>>>    Removing brick(s) can result in data loss. Do you want to Continue?
>>> (y/n) y
>>>    Remove brick incorrect brick count of 1 for replica 2
>>>
>>> 5. #gluster volume remove-brick test-volume dbstore4r294:/datastore1
>>> dbstore4r294:/datastore2
>>>    Removing brick(s) can result in data loss. Do you want to Continue?
>>> (y/n) y
>>>    Bricks not from same subvol for replica
>>>
>>> How do I remove the peer? What are the steps considering that the node
>>> is non-existent?
>>>  */
>>>
>>
>>
>> Do you plan to replace the dead server with a new server? If so, this
>> could be a possible sequence of steps:
>>
>>
>  No. We are not going to replace it. So, I need to resize it to a 3 node
> cluster.
>
>  I discovered the issue when one of the node hung and I had to reboot it.
> I expected Gluster volume to be available for one node failure. The volume
> was non-responsive.
>
> Surprised at that, I checked the details and found it was running with one
> node missing for many months now, perhaps an year!
>
>  I have no node to replace it with. So, I am looking for a method by
> which I can resize it.
>
>    The problem is that you want to do a replica 2 volume with an odd
> number of servers. This can be done but requires that you think of bricks
> individually rather than tying sets of bricks to servers. Your goal is to
> simply have each pair of replica bricks on two unique servers.
>
> See
> http://joejulian.name/blog/how-to-expand-glusterfs-replicated-clusters-by-one-server/for an example.
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>



-- 
*Religious confuse piety with mere ritual, the virtuous confuse regulation
with outcomes*
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130908/8b68d3ad/attachment.html>


More information about the Gluster-users mailing list