[Gluster-users] Self-heal never finishes

Karthik Subrahmanya ksubrahm at redhat.com
Mon Mar 1 06:48:49 UTC 2021


I think [1] should help you.
If you can't find anything matching your situation or can't resolve it with
any of the methods listed there, please open an issue for this at [2], with
the following information.
- volume info, volume status, heal info and shd logs from node-1 & arbiter.
- Output of "getfattr -d -e hex -m. <path-to-file-on-brick>" of a few
entries from all the bricks which are listed in the heal info output.

[2] https://github.com/gluster/glusterfs/issues


On Mon, Mar 1, 2021 at 8:12 AM Ben <gravyfish at gmail.com> wrote:

> I'm having a problem where once one of my volumes requires healing, it
> never finishes the process. I use a 3-node replica cluster (2 node +
> arbiter) as oVirt storage for virtual machines. I'm using Gluster version
> 8.3.
> When I patch my Gluster nodes, I try to keep the system online by
> rebooting them one at a time. However, I've found that once I reboot node
> 2, when it comes back up, self-heal will begin on both node 1 and the
> arbiter and never finish. I have let it run for weeks and still have
> entries in gluster volume heal <volname> info. No heal entries are reported
> on the node that rebooted.
> I've set the volumes to the virt group (gluster volume set <volname> group
> virt) per the RHEV documentation, and the gluster nodes don't seem to be
> overly busy. I'm hoping someone can point me in the right direction --
> since the volumes never heal, I'm basically running on one node. Let me
> know what additional info will be helpful for troubleshooting, and thank
> you in advance.
> ________
> Community Meeting Calendar:
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://meet.google.com/cpu-eiue-hvk
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20210301/9c103eb0/attachment.html>

More information about the Gluster-users mailing list