[Gluster-users] Resync or how to force the replication

shwetha spandura at redhat.com
Wed Nov 27 08:16:17 UTC 2013


To start self-healing data execute : "gluster volume heal datastore1 
full" .

To monitor self-heal completion status exeucte : "gluster volume heal 
datastore1 info". The number of entries under each brick should be 0. 
When the number of entries count becomes 0 the self-heal is completed.

Other way to trigger self-heal is to create a new fuse mount and execute 
"find . | xargs stat" from the mount point.

-Shwetha
On 11/27/2013 01:07 PM, gandalf istari wrote:
> Thank you so much,
>
> It seems to be working the directories are now created but still 
> empty. I suppose this will take a while to sync 44GB
>
> The only change I have done was: leave out nas-02-data for your command.
>
> whats the best way now to monitor the sync process ?
>
>
>
> You couldn't force start the volume because the brick 
> "nas-02-data:/datastore" doesn't have the "trusted.glusterfs.volume-id" .
>
> From nas-02 node execute :
>
> 1. setfattr -n trusted.glusterfs.volume-id -v 
> 0xfdff519085ef4cba9056a6bbbd8d6863 nas-02-data:/datastore
>
> 2. gluster volume start datastore1 force.
>
>
> -Shwetha
> On 11/27/2013 12:23 PM, gandalf istari wrote:
>> Hi Shwetha,
>>
>> [root at nas-01 ~]# getfattr -d -e hex -m . /datastore
>>
>> getfattr: Removing leading '/' from absolute path names
>>
>> # file: datastore
>>
>> trusted.afr.datastore1-client-0=0x000000000000000000000000
>>
>> trusted.afr.datastore1-client-1=0x000000000000000000000000
>>
>> trusted.gfid=0x00000000000000000000000000000001
>>
>> trusted.glusterfs.dht=0x000000010000000000000000ffffffff
>>
>> trusted.glusterfs.volume-id=0xfdff519085ef4cba9056a6bbbd8d6863
>>
>>
>>
>> [root at nas-02 ~]# getfattr -d -e hex -m . /datastore
>>
>> getfattr: Removing leading '/' from absolute path names
>>
>> # file: datastore
>>
>> security.selinux=0x73797374656d5f753a6f626a6563745f723a6574635f72756e74696d655f743a733000
>>
>> I don't understand what happened .
>>
>>
>> gr
>>
>> Patrick
>>
>>
>> Hi Gandalf,
>>
>> can you run the following command on the brick path?
>>
>> "getfattr -d -e hex -m . /datastore" on both "nas-01-data" and 
>> "nas-02-data" nodes.
>>
>> This will let us know whether there is "trusted.glusterfs.volume-id" 
>> set.
>>
>> -Shwetha
>>
>> On 11/26/2013 07:36 PM, gandalf istari wrote:
>>> hi thanks for the quick answer.
>>>
>>> I'm running glusterfs 3.4.1
>>>
>>> [root at nas-02 datastore]# gluster volume start datastore1 force
>>>
>>> volume start: datastore1: failed: Failed to get extended attribute 
>>> trusted.glusterfs.volume-id for brick dir /datastore. Reason : No 
>>> data available
>>>
>>> It seems that the .gluster directory is missing for some reason.
>>>
>>>
>>> volume replace-brick datastore1 nas-01-data:/datastorenas-02-data:/datastore
>>> commit force
>>> To rebuild/replace the missing brick ?
>>> I'm quite new with glusterfs
>>> Thanks
>>>
>>>
>>>
>>>
>>>
>>>
>>> On 26/11/13 12:47, gandalf istari wrote:
>>>> Hi have setup a two node replication glusterfs. After the initial 
>>>> installation the "master" node was put into the datacenter and 
>>>> after two week we moved the second one also to the datacenter.
>>>>
>>>> But the sync has not started yet.
>>>>
>>>> On the "master"
>>>>
>>>> gluster> volume info all
>>>>
>>>> Volume Name: datastore1
>>>>
>>>> Type: Replicate
>>>>
>>>> Volume ID: fdff5190-85ef-4cba-9056-a6bbbd8d6863
>>>>
>>>> Status: Started
>>>>
>>>> Number of Bricks: 1 x 2 = 2
>>>>
>>>> Transport-type: tcp
>>>>
>>>> Bricks:
>>>>
>>>> Brick1: nas-01-data:/datastore
>>>>
>>>> Brick2: nas-02-data:/datastore
>>>>
>>>> gluster> peer status
>>>>
>>>> Number of Peers: 1
>>>>
>>>>
>>>> Hostname: nas-02-data
>>>>
>>>> Uuid: 71df9f86-a87b-481d-896c-c0d4ab679cfa
>>>>
>>>> State: Peer in Cluster (Connected)
>>>>
>>>>
>>>> On the "slave"
>>>>
>>>> gluster> peer status
>>>>
>>>> Number of Peers: 1
>>>>
>>>> Hostname: 192.168.70.6
>>>>
>>>> Uuid: 97ef0154-ad7b-402a-b0cb-22be09134a3c
>>>>
>>>> State: Peer in Cluster (Connected)
>>>>
>>>>
>>>> gluster> volume status all
>>>>
>>>> Status of volume: datastore1
>>>>
>>>> Gluster processPortOnlinePid
>>>>
>>>> ------------------------------------------------------------------------------
>>>>
>>>> Brick nas-01-data:/datastore49152Y2130
>>>>
>>>> Brick nas-02-data:/datastoreN/ANN/A
>>>>
>>>> NFS Server on localhost2049Y8064
>>>>
>>>> Self-heal Daemon on localhostN/AY8073
>>>>
>>>> NFS Server on 192.168.70.62049Y3379
>>>>
>>>> Self-heal Daemon on 192.168.70.6N/AY3384
>>>>
>>> Which version of glusterfs are you running?
>>>
>>> volume status suggests that the second brick 
>>> (nas-02-data:/datastore) is not running.
>>>
>>> Can you run "gluster volume start <volname> force" in any of these 
>>> two nodes and try again?
>>> Then you would also required to run `find . | xargs stat` on the 
>>> mountpoint of the volume. That should trigger the self heal.
>>>>
>>>> There are no active volume tasks
>>>>
>>>>
>>>> I would like to run on the "slave" gluster volume sync nas-01-data 
>>>> datastore1
>>>>
>>> BTW, There is no concept of "master" and "slave" in afr 
>>> (replication). However there is concept of "master volume" and 
>>> "slave volume" in gluster geo-replication.
>>>>
>>>> But then the virtual machines hosted will be unavailible is there 
>>>> another way to start the replication ?
>>>>
>>>>
>>>> Thanks
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org  <mailto:Gluster-users at gluster.org>
>>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>>
>>>
>>>
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org  <mailto:Gluster-users at gluster.org>
>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>
>>
>>
>
>


-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20131127/b374d2f4/attachment.html>


More information about the Gluster-users mailing list