[Gluster-users] Entries in heal pending

Szilágyi Balázs szilagyi.balazs at pcvilag.hu
Thu Dec 19 09:21:27 UTC 2019


Hi Strahil,

Thanks for the input. It worked flawlesly!
I copy/paste the process here (maybe useful for someone).

### entering the "corrupt" gluster brick folder
[root at node1 data]# cd 
7ac28c32-947b-4ad5-8d69-213a205f06e8/images/d7c11f2e-58e4-4fe1-8236-6ded0f4dd757/
[root at node1 d7c11f2e-58e4-4fe1-8236-6ded0f4dd757]# ls -la
total 1080
drwxr-xr-x. 2 vdsm kvm    8192 Jan  1  1970 .
drwxr-xr-x. 4 vdsm kvm      94 Dec 18 11:11 ..
-rw-rw----. 2 vdsm kvm   30720 Dec 17 13:17 
e566f230-df72-4073-aecf-7e5a8d6b569b
-rw-rw----. 2 vdsm kvm 1048576 Dec  2 14:55 
e566f230-df72-4073-aecf-7e5a8d6b569b.lease
-rw-r--r--. 2 vdsm kvm     429 Dec 17 13:17 
e566f230-df72-4073-aecf-7e5a8d6b569b.meta
### making a backup of the files (did it on all the nodes)
[root at node1 d7c11f2e-58e4-4fe1-8236-6ded0f4dd757]# mkdir -p 
/root/save/gluster_bricks/data/data/7ac28c32-947b-4ad5-8d69-213a205f06e8/images/d7c11f2e-58e4-4fe1-8236-6ded0f4dd757
[root at node1 d7c11f2e-58e4-4fe1-8236-6ded0f4dd757]# cp * 
/root/save/gluster_bricks/data/data/7ac28c32-947b-4ad5-8d69-213a205f06e8/images/d7c11f2e-58e4-4fe1-8236-6ded0f4dd757/
### rsyncing the files from the selected source (i choosen node2 as 
soruce, and did this on node3 also)
[root at node1 d7c11f2e-58e4-4fe1-8236-6ded0f4dd757]# rsync -avh 
root at node2.ovirt.local:/gluster_bricks/data/data/7ac28c32-947b-4ad5-8d69-213a205f06e8/images/d7c11f2e-58e4-4fe1-8236-6ded0f4dd757/e566f230-df72-4073-aecf-7e5a8d6b569b* 
.
receiving incremental file list

sent 20 bytes  received 129 bytes  99.33 bytes/sec
total size is 1.08M  speedup is 7,246.48
### started the healing
[root at node1 d7c11f2e-58e4-4fe1-8236-6ded0f4dd757]# gluster volume heal data
Launching heal operation to perform index self heal on volume data has 
been successful
Use heal info commands to check status.
### checking result
[root at node1 d7c11f2e-58e4-4fe1-8236-6ded0f4dd757]# gluster volume heal 
data info
Brick node1storage.ovirt.local:/gluster_bricks/data/data
/7ac28c32-947b-4ad5-8d69-213a205f06e8/images/079904a4-71af-492c-bb2f-b45a918e8a2e/fce4b64d-2444-4f11-b226-db75bb2960c2.meta
Status: Connected
Number of entries: 1

Brick node2storage.ovirt.local:/gluster_bricks/data/data
/7ac28c32-947b-4ad5-8d69-213a205f06e8/images/079904a4-71af-492c-bb2f-b45a918e8a2e/fce4b64d-2444-4f11-b226-db75bb2960c2.meta
Status: Connected
Number of entries: 1

Brick node3storage.ovirt.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0

### As you can see the only one pending entry left. I did the above fix 
for the other files, and it also healed successfully.

Best Regards,
   Balazs Szilagyi

2019.12.18. 20:07 keltezéssel, Strahil Nikolov írta:
> You are the second person (excluding me) who observes this behaviour.
> The easiest way to resolve this is to:
> 1. Check which file is newest (there is a timestamp in the file) for :
> /gluster_bricks/data/data/7ac28c32-947b-4ad5-8d69-213a205f06e8/images/d7c11f2e-58e4-4fe1-8236-6ded0f4dd757/e566f230-df72-4073-aecf-7e5a8d6b569b.meta
> And for :
> /gluster_bricks/data/data/7ac28c32-947b-4ad5-8d69-213a205f06e8/images/079904a4-71af-492c-bb2f-b45a918e8a2e/fce4b64d-2444-4f11-b226-db75bb2960c2.meta
>
> Let's assume node3storage.ovirt.local has the newest data.
>
> 2.Then you can just backup (just in case you change your mind) the 
> files locally and rsync from node3storage.ovirt.local (actually 
> replace with the node with newest timestamp in the file) to the other 
> bricks.
>
> 3. Run a gluster heal just to notify gluster that the issue is resolved.
>
> In my case one of the nodes had a newer version of the file (I am 
> using replica 2 arbiter 1 volume) and the gfid was different and this 
> prevented Gluster from healing that.
>
> Usually , oVirt just updates the timestamp in the meta files - so even 
> an older version is not a problem.
>
> P.S.: What version of gluster are you using. I suppose v6.5 or v6.6 , 
> right ?
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
> В сряда, 18 декември 2019 г., 16:18:12 ч. Гринуич+2, Szilágyi Balázs 
> <szilagyi.balazs at pcvilag.hu> написа:
>
>
> Dear Gluster Users,
>
> I'm a newbie to gluster storage and during the stability testing I made
> a node reboot and I've got some heal issues after that, that i'm unable
> to fix.
> Anyway the vm's are running fine from the storage, and I did not
> discovered any data corruption.
> The system is oVirt version 4.3.7. I have 3 nodes with a replica 3.
> Please let me know what to do with the pending heals that are unable to
> finish.
> Also let me know if some more details are necessary.
>
> Thanks,
>   Balazs
>
> [root at node2 <mailto:root at node2> ~]# gluster volume status data
> Status of volume: data
> Gluster process TCP Port  RDMA Port Online  Pid
> ------------------------------------------------------------------------------
> Brick node1storage.ovirt.local:/gluster_bri
> cks/data/data 49152     0 Y       4187
> Brick node2storage.ovirt.local:/gluster_bri
> cks/data/data 49154     0 Y       6163
> Brick node3storage.ovirt.local:/gluster_bri
> cks/data/data 49154     0 Y       19439
> Self-heal Daemon on localhost N/A       N/A Y       3136
> Self-heal Daemon on node3storage.ovirt.loca
> l N/A       N/A Y       5876
> Self-heal Daemon on node1storage.ovirt.loca
> l N/A       N/A Y       15479
>
> Task Status of Volume data
> ------------------------------------------------------------------------------
> There are no active volume tasks
>
> [root at node2 <mailto:root at node2> ~]#
> [root at node2 <mailto:root at node2> ~]# gluster volume heal data info summary
> Brick node1storage.ovirt.local:/gluster_bricks/data/data
> Status: Connected
> Total Number of entries: 2
> Number of entries in heal pending: 2
> Number of entries in split-brain: 0
> Number of entries possibly healing: 0
>
> Brick node2storage.ovirt.local:/gluster_bricks/data/data
> Status: Connected
> Total Number of entries: 2
> Number of entries in heal pending: 2
> Number of entries in split-brain: 0
> Number of entries possibly healing: 0
>
> Brick node3storage.ovirt.local:/gluster_bricks/data/data
> Status: Connected
> Total Number of entries: 0
> Number of entries in heal pending: 0
> Number of entries in split-brain: 0
> Number of entries possibly healing: 0
>
> [root at node2 <mailto:root at node2> ~]# gluster volume heal data info
> Brick node1storage.ovirt.local:/gluster_bricks/data/data
> /7ac28c32-947b-4ad5-8d69-213a205f06e8/images/d7c11f2e-58e4-4fe1-8236-6ded0f4dd757/e566f230-df72-4073-aecf-7e5a8d6b569b.meta
> /7ac28c32-947b-4ad5-8d69-213a205f06e8/images/079904a4-71af-492c-bb2f-b45a918e8a2e/fce4b64d-2444-4f11-b226-db75bb2960c2.meta
> Status: Connected
> Number of entries: 2
>
> Brick node2storage.ovirt.local:/gluster_bricks/data/data
> /7ac28c32-947b-4ad5-8d69-213a205f06e8/images/d7c11f2e-58e4-4fe1-8236-6ded0f4dd757/e566f230-df72-4073-aecf-7e5a8d6b569b.meta
> /7ac28c32-947b-4ad5-8d69-213a205f06e8/images/079904a4-71af-492c-bb2f-b45a918e8a2e/fce4b64d-2444-4f11-b226-db75bb2960c2.meta
> Status: Connected
> Number of entries: 2
>
> Brick node3storage.ovirt.local:/gluster_bricks/data/data
> Status: Connected
> Number of entries: 0
>
> ________
>
> Community Meeting Calendar:
>
> APAC Schedule -
> Every 2nd and 4th Tuesday at 11:30 AM IST
> Bridge: https://bluejeans.com/441850968
>
> NA/EMEA Schedule -
> Every 1st and 3rd Tuesday at 01:00 PM EDT
> Bridge: https://bluejeans.com/441850968
>
> Gluster-users mailing list
> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
> https://lists.gluster.org/mailman/listinfo/gluster-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20191219/fbf4b805/attachment.html>


More information about the Gluster-users mailing list