[Gluster-users] how to recover a accidentally delete brick directory?
shwetha
spandura at redhat.com
Thu Nov 28 09:47:16 UTC 2013
Since the "trusted.glusterfs.volume-id" is not set on the bricks , the
volume start will fail.
1) Execute : "getfattr -e hex -n trusted.glusterfs.volume-id
/opt/gluster_data/eccp_glance" from any of the node which the brick
process running.
You will get the hex value of the "trusted.glusterfs.volume-id" extended
attribute . Set this value on the bricks created (On the nodes where you
deleted the bricks)
2) setfattr -n trusted.glusterfs.volume-id -v
<hex_value_of_the_trusted_glusterfs_volume-id> /opt/gluster_data/eccp_glance
3) gluster volume start <volume_name> force : To restart the brick process
4) gluster volume status <volume_name> : Check all the brick process are
started.
5) gluster volume heal <volume_name> full : To trigger self-heal on to
the removed bricks.
Shwetha
On 11/28/2013 02:59 PM, weiyuanke wrote:
> hi Shwetha,
>
>
> command "gluster volume start eccp_glance force" on the other node
> gives following:
> Pasted Graphic.tiff
>
> with cli.log
> Pasted Graphic 1.tiff
>
> on the damaged node,gluster volume start eccp_glance force gives:
> Pasted Graphic 2.tiff
>
>
>
>
> ---------------------------
> 韦远科
> 010 5881 3749
> 中国科学院 计算机网络信息中心
>
>
>
>
>
> On 2013年11月28日, at 下午4:50, shwetha <spandura at redhat.com
> <mailto:spandura at redhat.com>> wrote:
>
>> 1) create the brick directory "/opt/gluster_data/eccp_glance" on the
>> nodes where you deleted the directories.
>>
>> 2) From any of the storage node execute :
>>
>> 1. gluster volume start <volume_name> force : To restart the brick
>> process
>> 2. gluster volume status <volume_name> : Check all the brick process
>> are started.
>> 3. gluster volume heal <volume_name> full : To trigger self-heal on
>> to the removed bricks.
>>
>> -Shwetha
>>
>> On 11/28/2013 02:09 PM, 韦远科 wrote:
>>> hi all,
>>>
>>> I accidentally removed the brick directory of a volume on one node,
>>> the replica for this volume is 2.
>>>
>>> now the situation is , there is no corresponding glusterfsd process
>>> on this node, and 'glusterfs volume status' shows that the brick is
>>> offline, like this:
>>> Brick 192.168.64.11:/opt/gluster_data/eccp_glance N/A Y 2513
>>> Brick 192.168.64.12:/opt/gluster_data/eccp_glance 49161 Y 2542
>>> Brick 192.168.64.17:/opt/gluster_data/eccp_glance 49164 Y 2537
>>> Brick 192.168.64.18:/opt/gluster_data/eccp_glance 49154 Y 4978
>>> Brick 192.168.64.29:/opt/gluster_data/eccp_glance N/A N N/A
>>> Brick 192.168.64.30:/opt/gluster_data/eccp_glance 49154 Y 4072
>>> Brick 192.168.64.25:/opt/gluster_data/eccp_glance 49155 Y 11975
>>> Brick 192.168.64.26:/opt/gluster_data/eccp_glance 49155 Y 17947
>>> Brick 192.168.64.13:/opt/gluster_data/eccp_glance 49154 Y 26045
>>> Brick 192.168.64.14:/opt/gluster_data/eccp_glance 49154 Y 22143
>>>
>>>
>>> so are there ways to bring this brick back to normal?
>>>
>>> thanks!
>>>
>>>
>>> -----------------------------------------------------------------
>>> 韦远科
>>> 中国科学院 计算机网络信息中心
>>>
>>>
>>>
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20131128/b54406e9/attachment.html>
More information about the Gluster-users
mailing list