[Gluster-users] Continual heals happening on sharded cluster, me too...
Nicolas Ecarnot
nicolas at ecarnot.net
Thu May 12 15:08:53 UTC 2016
Hello,
On a replica 3 sharded cluster in gluster 3.7.11 (CentOS 7.2), I've
witnessed a very sound behaviour for months.
Recently, I've upgraded from [something quite recent, like
3.7.something] to 3.7.11, and immediately, I'm seeing continual healing
appearing.
I'm using this cluster to host 3 oVirt VMs with very low I/O.
I read and tried to use informations here :
https://www.mail-archive.com/gluster-users%40gluster.org/msg24598.html
but I'm missing knowledge to know how to fix things.
Here what I see when running "gluster volume heal data-shard-03 info" :
* serv-vm-al01
Brick serv-vm-al01:/gluster/data/brick03
Status: Connected
Number of entries: 0
Brick serv-vm-al02:/gluster/data/brick03
Status: Connected
Number of entries: 0
Brick serv-vm-al03:/gluster/data/brick03
Status: Connected
Number of entries: 0
* serv-vm-al02
gluster volume heal data-shard-03 info
Brick serv-vm-al01:/gluster/data/brick03
<gfid:75ab0b1c-258f-4349-85bd-49ef80592919>
<gfid:39bdcf05-f5b1-4df2-9941-614838b50e18>
<gfid:7e0d84a0-0749-4a2d-9390-e95d489ec66a>
Status: Connected
Number of entries: 3
Brick serv-vm-al02:/gluster/data/brick03
/.shard/41573624-feb9-4ea6-bbd4-f0a912429b2f.1003
/.shard/26ec536a-8919-478c-834c-f6ac70882ee6.2351
/.shard/26ec536a-8919-478c-834c-f6ac70882ee6.1113
Status: Connected
Number of entries: 3
Brick serv-vm-al03:/gluster/data/brick03
/.shard/34948b18-5991-4761-b867-2bd4fe4879d4.2807
/.shard/26ec536a-8919-478c-834c-f6ac70882ee6.2351
/.shard/34948b18-5991-4761-b867-2bd4fe4879d4.2062
/.shard/26ec536a-8919-478c-834c-f6ac70882ee6.1113
/.shard/41573624-feb9-4ea6-bbd4-f0a912429b2f.1833
Status: Connected
Number of entries: 5
* serv-vm-al03
Brick serv-vm-al01:/gluster/data/brick03
/.shard/41573624-feb9-4ea6-bbd4-f0a912429b2f.882
<gfid:39bdcf05-f5b1-4df2-9941-614838b50e18>
<gfid:93c83d37-71e5-4f17-b757-ff00a4384f2f>
<gfid:75fcdab7-6d18-4371-bbb2-be156dd3da86>
<gfid:75ab0b1c-258f-4349-85bd-49ef80592919>
<gfid:f933a07a-4f89-4bf9-ab82-a05432a0162f>
/.shard/34948b18-5991-4761-b867-2bd4fe4879d4.2329
Status: Connected
Number of entries: 7
Brick serv-vm-al02:/gluster/data/brick03
/.shard/26ec536a-8919-478c-834c-f6ac70882ee6.1113
/.shard/34948b18-5991-4761-b867-2bd4fe4879d4.2807
/.shard/41573624-feb9-4ea6-bbd4-f0a912429b2f.190
/.shard/26ec536a-8919-478c-834c-f6ac70882ee6.2351
Status: Connected
Number of entries: 4
Brick serv-vm-al03:/gluster/data/brick03
/.shard/26ec536a-8919-478c-834c-f6ac70882ee6.1113
/.shard/26ec536a-8919-478c-834c-f6ac70882ee6.185
/.shard/26ec536a-8919-478c-834c-f6ac70882ee6.183
/.shard/26ec536a-8919-478c-834c-f6ac70882ee6.2351
Status: Connected
Number of entries: 4
We are monitoring healed entries in Nagios/centreon, and history graph
are showing that from the time I upgraded to 3.7.11, it went from around
0~1 files to heal every 5 minutes, to an average of 6 to 10 files to
heal every minutes.
The VMs are still doing the exact same nothing as previous.
Questions :
- how comes?
- how to fix?
--
Nicolas ECARNOT
More information about the Gluster-users
mailing list