[Gluster-users] Recovering lost node in dispersed volume

Tony Schreiner anthony.schreiner at bc.edu
Thu Sep 22 14:16:52 UTC 2016


I set uo a dispersed volume with 1 x (3 + 1) nodes ( i do know that 3+1 is
not optimal).
Originally created in version 3.7 but recently upgraded without issue to
3.8.

# gluster vol info
Volume Name: rvol
Type: Disperse
Volume ID: e8f15248-d9de-458e-9896-f1a5782dcf74
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (3 + 1) = 4
Transport-type: tcp
Bricks:
Brick1: calliope:/brick/p1
Brick2: euterpe:/brick/p1
Brick3: lemans:/brick/p1
Brick4: thalia:/brick/p1
Options Reconfigured:
performance.readdir-ahead: on
nfs.disable: off

I inadvertently allowed one of the nodes (thalia) to be reinstalled; which
overwrote the system, but not the brick, and I need guidance in getting it
back into the volume.

(on lemans)
gluster peer status
Number of Peers: 3

Hostname: calliope
Uuid: 72373eb1-8047-405a-a094-891e559755da
State: Peer in Cluster (Connected)

Hostname: euterpe
Uuid: 9fafa5c4-1541-4aa0-9ea2-923a756cadbb
State: Peer in Cluster (Connected)

Hostname: thalia
Uuid: 843169fa-3937-42de-8fda-9819efc75fe8
State: Peer Rejected (Connected)

the thalia peer is rejected. If I try to peer probe thalia I am told it
already part of the pool. If from thalia, I try to peer probe one of the
others, I am told that they are already part of another pool.

I have tried removing the thalia brick with
gluster vol remove-brick rvol thalia:/brick/p1 start
but get the error
volume remove-brick start: failed: Remove brick incorrect brick count of 1
for disperse 4

I am not finding much guidance for this particular situation. I could use a
suggestion on how to recover. It's a lab situation so no biggie if I lose
it.
Cheers

Tony Schreiner
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160922/16e32a00/attachment.html>


More information about the Gluster-users mailing list