[Gluster-users] Resync or how to force the replication

shwetha spandura at redhat.com
Wed Nov 27 06:13:39 UTC 2013


Hi Gandalf,

can you run the following command on the brick path?

"getfattr -d -e hex -m . /datastore" on both "nas-01-data" and 
"nas-02-data" nodes.

This will let us know whether there is "trusted.glusterfs.volume-id" set.

-Shwetha

On 11/26/2013 07:36 PM, gandalf istari wrote:
> hi thanks for the quick answer.
>
> I'm running glusterfs 3.4.1
>
> [root at nas-02 datastore]# gluster volume start datastore1 force
>
> volume start: datastore1: failed: Failed to get extended attribute 
> trusted.glusterfs.volume-id for brick dir /datastore. Reason : No data 
> available
>
> It seems that the .gluster directory is missing for some reason.
>
>
> volume replace-brick datastore1 nas-01-data:/datastorenas-02-data:/datastore
> commit force
> To rebuild/replace the missing brick ?
> I'm quite new with glusterfs
> Thanks
>
>
>
>
>
>
> On 26/11/13 12:47, gandalf istari wrote:
>> Hi have setup a two node replication glusterfs. After the initial 
>> installation the "master" node was put into the datacenter and after 
>> two week we moved the second one also to the datacenter.
>>
>> But the sync has not started yet.
>>
>> On the "master"
>>
>> gluster> volume info all
>>
>> Volume Name: datastore1
>>
>> Type: Replicate
>>
>> Volume ID: fdff5190-85ef-4cba-9056-a6bbbd8d6863
>>
>> Status: Started
>>
>> Number of Bricks: 1 x 2 = 2
>>
>> Transport-type: tcp
>>
>> Bricks:
>>
>> Brick1: nas-01-data:/datastore
>>
>> Brick2: nas-02-data:/datastore
>>
>> gluster> peer status
>>
>> Number of Peers: 1
>>
>>
>> Hostname: nas-02-data
>>
>> Uuid: 71df9f86-a87b-481d-896c-c0d4ab679cfa
>>
>> State: Peer in Cluster (Connected)
>>
>>
>> On the "slave"
>>
>> gluster> peer status
>>
>> Number of Peers: 1
>>
>> Hostname: 192.168.70.6
>>
>> Uuid: 97ef0154-ad7b-402a-b0cb-22be09134a3c
>>
>> State: Peer in Cluster (Connected)
>>
>>
>> gluster> volume status all
>>
>> Status of volume: datastore1
>>
>> Gluster processPortOnlinePid
>>
>> ------------------------------------------------------------------------------
>>
>> Brick nas-01-data:/datastore49152Y2130
>>
>> Brick nas-02-data:/datastoreN/ANN/A
>>
>> NFS Server on localhost2049Y8064
>>
>> Self-heal Daemon on localhostN/AY8073
>>
>> NFS Server on 192.168.70.62049Y3379
>>
>> Self-heal Daemon on 192.168.70.6N/AY3384
>>
> Which version of glusterfs are you running?
>
> volume status suggests that the second brick (nas-02-data:/datastore) 
> is not running.
>
> Can you run "gluster volume start <volname> force" in any of these two 
> nodes and try again?
> Then you would also required to run `find . | xargs stat` on the 
> mountpoint of the volume. That should trigger the self heal.
>>
>> There are no active volume tasks
>>
>>
>> I would like to run on the "slave" gluster volume sync nas-01-data 
>> datastore1
>>
> BTW, There is no concept of "master" and "slave" in afr (replication). 
> However there is concept of "master volume" and "slave volume" in 
> gluster geo-replication.
>>
>> But then the virtual machines hosted will be unavailible is there 
>> another way to start the replication ?
>>
>>
>> Thanks
>>
>>
>>
>>
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org  <mailto:Gluster-users at gluster.org>
>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users


-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20131127/6607f9ab/attachment.html>


More information about the Gluster-users mailing list