[Gluster-users] How to make sure self-heal backlog is empty ?

Karthik Subrahmanya ksubrahm at redhat.com
Wed Dec 20 05:45:19 UTC 2017


Hi,

Can you provide the
- volume info
- shd log
- mount log
of the volumes which are showing pending entries, to debug the issue.

Thanks & Regards,
Karthik

On Wed, Dec 20, 2017 at 3:11 AM, Matt Waymack <mwaymack at nsgdv.com> wrote:

> Mine also has a list of files that seemingly never heal.  They are usually
> isolated on my arbiter bricks, but not always.  I would also like to find
> an answer for this behavior.
>
> -----Original Message-----
> From: gluster-users-bounces at gluster.org [mailto:gluster-users-bounces@
> gluster.org] On Behalf Of Hoggins!
> Sent: Tuesday, December 19, 2017 12:26 PM
> To: gluster-users <gluster-users at gluster.org>
> Subject: [Gluster-users] How to make sure self-heal backlog is empty ?
>
> Hello list,
>
> I'm not sure what to look for here, not sure if what I'm seeing is the
> actual "backlog" (that we need to make sure is empty while performing a
> rolling upgrade before going to the next node), how can I tell, while
> reading this, if it's okay to reboot / upgrade my next node in the pool ?
> Here is what I do for checking :
>
>     for i in `gluster volume list`; do gluster volume heal $i info; done
>
> And here is what I get :
>
>     Brick ngluster-1.network.hoggins.fr:/export/brick/clem
>     Status: Connected
>     Number of entries: 0
>
>     Brick ngluster-2.network.hoggins.fr:/export/brick/clem
>     Status: Connected
>     Number of entries: 0
>
>     Brick ngluster-3.network.hoggins.fr:/export/brick/clem
>     Status: Connected
>     Number of entries: 0
>
>     Brick ngluster-1.network.hoggins.fr:/export/brick/mailer
>     Status: Connected
>     Number of entries: 0
>
>     Brick ngluster-2.network.hoggins.fr:/export/brick/mailer
>     Status: Connected
>     Number of entries: 0
>
>     Brick ngluster-3.network.hoggins.fr:/export/brick/mailer
>     <gfid:98642fd6-f8a4-4966-9c30-32fedbecfc1a>
>     Status: Connected
>     Number of entries: 1
>
>     Brick ngluster-1.network.hoggins.fr:/export/brick/rom
>     Status: Connected
>     Number of entries: 0
>
>     Brick ngluster-2.network.hoggins.fr:/export/brick/rom
>     Status: Connected
>     Number of entries: 0
>
>     Brick ngluster-3.network.hoggins.fr:/export/brick/rom
>     <gfid:52b09fb6-78da-46db-af0e-e6a16194a977>
>     Status: Connected
>     Number of entries: 1
>
>     Brick ngluster-1.network.hoggins.fr:/export/brick/thedude
>     Status: Connected
>     Number of entries: 0
>
>     Brick ngluster-2.network.hoggins.fr:/export/brick/thedude
>     <gfid:4b1f4d9b-f2d8-4a50-83f7-3f014fe0b9f6>
>     Status: Connected
>     Number of entries: 1
>
>     Brick ngluster-3.network.hoggins.fr:/export/brick/thedude
>     Status: Connected
>     Number of entries: 0
>
>     Brick ngluster-1.network.hoggins.fr:/export/brick/web
>     Status: Connected
>     Number of entries: 0
>
>     Brick ngluster-2.network.hoggins.fr:/export/brick/web
>     <gfid:491c59f7-bf42-4d7c-be56-842317c55ac5>
>     <gfid:9deb7b0d-0459-4dd1-a93c-f4eab03df6d6>
>     <gfid:3803e1ec-9327-4e08-8f31-f3dc90aaa403>
>     Status: Connected
>     Number of entries: 3
>
>     Brick ngluster-3.network.hoggins.fr:/export/brick/web
>     <gfid:0f29326d-d273-4299-ba71-a5d8722a9149>
>     <gfid:b5f0dd49-00a1-4a1d-97c1-0be973b097d6>
>     <gfid:22d21ac4-8ad8-4390-a07b-26c8a75f2f5d>
>     <gfid:5b432df5-8e8d-4789-abea-c35e88490e41>
>     <gfid:b3621d26-3a60-4803-8039-a89933c306d8>
>     <gfid:5eb83bc2-c975-4182-a7c3-c8cc9b39a064>
>     <gfid:3803e1ec-9327-4e08-8f31-f3dc90aaa403>
>     <gfid:491c59f7-bf42-4d7c-be56-842317c55ac5>
>     <gfid:9deb7b0d-0459-4dd1-a93c-f4eab03df6d6>
>     <gfid:44a64b36-cdb4-4c0f-be18-419b72add380>
>     <gfid:e12d8e6e-56b0-4db4-9e89-e80bdee3a435>
>     Status: Connected
>     Number of entries: 11
>
>
> Should I be worrying with this never ending ?
>
>     Thank you,
>
>         Hoggins!
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20171220/9cd5ed71/attachment.html>


More information about the Gluster-users mailing list