[Gluster-users] v3.6.1 vs v3.5.2 self heal - help (Nagios related)

Nishanth Thomas nthomas at redhat.com
Wed Nov 19 18:14:05 UTC 2014


Hi Vince,

Are you referring the monitoring scripts mentioned in the blog( http://gopukrish.wordpress.com/2014/11/16/monitor-glusterfs-using-nagios-plugin/) or the scripts part of the gluster(http://gluster.org/pipermail/gluster-users.old/2014-June/017819.html)?
Please confirm?

Thanks,
Nishanth

----- Original Message -----
From: "Humble Devassy Chirammal" <humble.devassy at gmail.com>
To: "Vince Loschiavo" <vloschiavo at gmail.com>
Cc: "gluster-users at gluster.org" <Gluster-users at gluster.org>, "Sahina Bose" <sabose at redhat.com>, nthomas at redhat.com
Sent: Wednesday, November 19, 2014 11:22:18 PM
Subject: Re: [Gluster-users] v3.6.1 vs v3.5.2 self heal - help (Nagios related)

Hi Vince,
It could be a behavioural change in heal process output capture with latest
GlusterFS. If that is the case, we may tune the interval which  nagios
collect heal info output  or some other settings to avoid continuous
alerts. I am Ccing  gluster nagios devs.

--Humble

--Humble


On Wed, Nov 19, 2014 at 9:50 PM, Vince Loschiavo <vloschiavo at gmail.com>
wrote:

>
> Hello Gluster Community,
>
> I have been using the Nagios monitoring scripts, mentioned in the below
> thread, on 3.5.2 with great success. The most useful of these is the self
> heal.
>
> However, I've just upgraded to 3.6.1 on the lab and the self heal daemon
> has become quite aggressive.  I continually get alerts/warnings on 3.6.1
> that virt disk images need self heal, then they clear.  This is not the
> case on 3.5.2.  This
>
> Configuration:
> 2 node, 2 brick replicated volume with 2x1GB LAG network between the peers
> using this volume as a QEMU/KVM virt image store through the fuse mount on
> Centos 6.5.
>
> Example:
> on 3.5.2:
> *gluster volume heal volumename info:  *shows the bricks and number of
> entries to be healed: 0
>
> On v3.5.2 - During normal gluster operations, I can run this command over
> and over again, 2-4 times per second, and it will always show 0 entries to
> be healed.  I've used this as an indicator that the bricks are
> synchronized.
>
> Last night, I upgraded to 3.6.1 in lab and I'm seeing different behavior.
> Running *gluster volume heal volumename info*, during normal operations,
> will show a file out-of-sync, seemingly between every block written to disk
> then synced to the peer.  I can run the command over and over again, 2-4
> times per second, and it will almost always show something out of sync.
> The individual files change, meaning:
>
> Example:
> 1st Run: shows file1 out of sync
> 2nd run: shows file 2 and file 3 out of sync but file 1 is now in sync
> (not in the list)
> 3rd run: shows file 3 and file 4 out of sync but file 1 and 2 are in sync
> (not in the list).
> ...
> nth run: shows 0 files out of sync
> nth+1 run: shows file 3 and 12 out of sync.
>
> From looking at the virtual machines running off this gluster volume, it's
> obvious that gluster is working well.  However, this obviously plays havoc
> with Nagios and alerts.  Nagios will run the heal info and get different
> and non-useful results each time, and will send alerts.
>
> Is this behavior change (3.5.2 vs 3.6.1) expected?  Is there a way to tune
> the settings or change the monitoring method to get better results into
> Nagios.
>
> Thank you,
>
> --
> -Vince Loschiavo
>
>
> On Wed, Nov 19, 2014 at 4:35 AM, Humble Devassy Chirammal <
> humble.devassy at gmail.com> wrote:
>
>> Hi Gopu,
>>
>> Awesome !!
>>
>> We can  have a Gluster blog about this implementation.
>>
>> --Humble
>>
>>
>>
>> --Humble
>>
>>
>> On Wed, Nov 19, 2014 at 5:38 PM, Gopu Krishnan <gopukrishnantec at gmail.com
>> > wrote:
>>
>>> Thanks for all your help... I was able to configure nagios using the
>>> glusterfs plugin. Following link shows how I configured it. Hope it helps
>>> someone else.:
>>>
>>>
>>> http://gopukrish.wordpress.com/2014/11/16/monitor-glusterfs-using-nagios-plugin/
>>>
>>> On Sun, Nov 16, 2014 at 11:44 AM, Humble Devassy Chirammal <
>>> humble.devassy at gmail.com> wrote:
>>>
>>>> Hi,
>>>>
>>>> Please look at this thread
>>>> http://gluster.org/pipermail/gluster-users.old/2014-June/017819.html
>>>>
>>>> Btw,  if you are around, we have a talk on same topic in upcoming
>>>> GlusterFS India meetup.
>>>>
>>>> Details can be fetched from:
>>>>  http://www.meetup.com/glusterfs-India/
>>>>
>>>> --Humble
>>>>
>>>> --Humble
>>>>
>>>>
>>>> On Sun, Nov 16, 2014 at 11:23 AM, Gopu Krishnan <
>>>> gopukrishnantec at gmail.com> wrote:
>>>>
>>>>> How can we monitor the glusters and alert us if something happened
>>>>> wrong. I found some nagios plugins and didn't work until this time. I am
>>>>> still experimenting with those. Any suggestions would be much helpful
>>>>>
>>>>> _______________________________________________
>>>>> Gluster-users mailing list
>>>>> Gluster-users at gluster.org
>>>>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>>>>
>>>>
>>>>
>>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>
>
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>


More information about the Gluster-users mailing list