[Gluster-users] Gluster-Nagios
Ramesh Nachimuthu
rnachimu at redhat.com
Thu Oct 8 10:49:14 UTC 2015
Looks like you are hitting the NRPE Payload issue. Standard NRPE
packages from epel/fedora has 1024 bytes payload limit. We have to
increment this to 8192 to fix this. You can see more info at
http://serverfault.com/questions/613288/truncating-return-data-as-it-is-bigger-then-nrpe-allows.
Let me know if u need any more info.
Regards,
Ramesh
On 10/08/2015 02:48 PM, Punit Dambiwal wrote:
> Hi,
>
> I am getting the following error :-
>
> ----------------
> [root at monitor-001 yum.repos.d]#
> /usr/lib64/nagios/plugins/gluster/discovery.py -c ssd -H stor1
> Traceback (most recent call last):
> File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 510, in
> <module>
> clusterdata = discoverCluster(args.hostip, args.cluster, args.timeout)
> File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 88, in
> discoverCluster
> componentlist = discoverVolumes(hostip, timeout)
> File "/usr/lib64/nagios/plugins/gluster/discovery.py", line 56, in
> discoverVolumes
> timeout=timeout)
> File "/usr/lib64/nagios/plugins/gluster/server_utils.py", line 107,
> in execNRPECommand
> resultDict = json.loads(outputStr)
> File "/usr/lib64/python2.6/json/__init__.py", line 307, in loads
> return _default_decoder.decode(s)
> File "/usr/lib64/python2.6/json/decoder.py", line 319, in decode
> obj, end = self.raw_decode(s, idx=_w(s, 0).end())
> File "/usr/lib64/python2.6/json/decoder.py", line 336, in raw_decode
> obj, end = self._scanner.iterscan(s, **kw).next()
> File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
> rval, next_pos = action(m, context)
> File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
> value, end = iterscan(s, idx=end, context=context).next()
> File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
> rval, next_pos = action(m, context)
> File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
> value, end = iterscan(s, idx=end, context=context).next()
> File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
> rval, next_pos = action(m, context)
> File "/usr/lib64/python2.6/json/decoder.py", line 217, in JSONArray
> value, end = iterscan(s, idx=end, context=context).next()
> File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
> rval, next_pos = action(m, context)
> File "/usr/lib64/python2.6/json/decoder.py", line 183, in JSONObject
> value, end = iterscan(s, idx=end, context=context).next()
> File "/usr/lib64/python2.6/json/scanner.py", line 55, in iterscan
> rval, next_pos = action(m, context)
> File "/usr/lib64/python2.6/json/decoder.py", line 155, in JSONString
> return scanstring(match.string, match.end(), encoding, strict)
> ValueError: ('Invalid control character at: line 1 column 1023 (char
> 1023)', '{"ssd": {"name": "ssd", "disperseCount": "0", "bricks":
> [{"brickpath": "/bricks/b/vol1", "brickaddress": "stor1", "hostUuid":
> "5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath":
> "/bricks/b/vol1", "brickaddress": "stor2", "hostUuid":
> "b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath":
> "/bricks/b/vol1", "brickaddress": "stor3", "hostUuid":
> "40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath":
> "/bricks/b/vol1", "brickaddress": "stor4", "hostUuid":
> "5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath":
> "/bricks/c/vol1", "brickaddress": "stor1", "hostUuid":
> "5fcb5150-f0a5-4af8-b383-11fa5d3f82f0"}, {"brickpath":
> "/bricks/c/vol1", "brickaddress": "stor2", "hostUuid":
> "b78d42c1-6ad7-4044-b900-3ccfe915859f"}, {"brickpath":
> "/bricks/c/vol1", "brickaddress": "stor3", "hostUuid":
> "40500a9d-418d-4cc0-aec5-6efbfb3c24e5"}, {"brickpath":
> "/bricks/c/vol1", "brickaddress": "stor4", "hostUuid":
> "5886ef94-df5e-4845-a54c-0e01546d66ea"}, {"brickpath":
> "/bricks/d/vol1", "brickaddress": "stor1", "hostUuid":
> "5fcb5150-f0a5-4a\n')
> [root at monitor-001 yum.repos.d]#
> -------------------------
>
> --------------
> [root at monitor-001 yum.repos.d]# /usr/lib64/nagios/plugins/check_nrpe
> -H stor1 -c discover_volume_list
> {"ssd": {"type": "DISTRIBUTED_REPLICATE", "name": "ssd"}, "lockvol":
> {"type": "REPLICATE", "name": "lockvol"}}
> [root at monitor-001 yum.repos.d]#
> --------------
>
> Please help me to solve this issue...
>
> Thanks,
> Punit
>
> On Fri, Oct 2, 2015 at 12:15 AM, Sahina Bose <sabose at redhat.com
> <mailto:sabose at redhat.com>> wrote:
>
> The gluster-nagios packages have not been tested on Ubuntu
>
> Looking at the error below, it looks like the rpm has not updated
> the nrpe.cfg correctly. You may need to edit the spec file for the
> config file paths on Ubuntu and rebuild.
>
>
> On 10/01/2015 05:45 PM, Amudhan P wrote:
>> OSError: [Errno 2] No such file or directory is now sorted out by
>> by changing NRPE_PATH in "constants.py".
>>
>> now if i run discovery.py
>>
>> testusr at gfsovirt:/usr/local/lib/nagios/plugins/gluster$ sudo
>> python discovery.py -c vm-gfs -H 192.168.1.11
>> Failed to execute NRPE command 'discover_volume_list' in host
>> '192.168.1.11'
>> Error : NRPE: Command 'discover_volume_list' not defined
>> Make sure NRPE server in host '192.168.1.11' is configured to
>> accept requests from Nagios server
>>
>>
>> testusr at gfsovirt:/usr/local/lib/nagios/plugins/gluster$
>> /usr/lib/nagios/plugins/check_nrpe -H 192.168.1.11 -c
>> discover_volume_list
>> NRPE: Command 'discover_volume_list' not defined
>>
>>
>> My client is responding to other nrpe command.
>> testusr at gfsovirt:/usr/local/lib/nagios/plugins/gluster$
>> /usr/lib/nagios/plugins/check_nrpe -H 192.168.1.11 -c check_load
>> OK - load average: 0.01, 0.03, 0.10|load1=0.010;15.000;30.000;0;
>> load5=0.030;10.000;25.000;0; load15=0.100;5.000;20.000;0;
>>
>>
>>
>> On Thu, Oct 1, 2015 at 5:20 PM, Sahina Bose <sabose at redhat.com
>> <mailto:sabose at redhat.com>> wrote:
>>
>> Looks like a conflict in versions of python and python-cpopen.
>> Can you give us the version of these packages?
>>
>> Also, what's the output of
>> /usr/lib64/nagios/plugins/check_nrpe -H 192.168.1.11 -c
>> discover_volume_list
>>
>>
>>
>>
>> On 10/01/2015 04:10 PM, Amudhan P wrote:
>>> Hi,
>>>
>>> I am getting a error when i run discovery.py.
>>>
>>> discovery.py -c vm-gfs -H 192.168.1.11
>>>
>>> Traceback (most recent call last):
>>> File "discovery.py", line 541, in <module>
>>> clusterdata = discoverCluster(args.hostip, args.cluster,
>>> args.timeout)
>>> File "discovery.py", line 90, in discoverCluster
>>> componentlist = discoverVolumes(hostip, timeout)
>>> File "discovery.py", line 53, in discoverVolumes
>>> timeout=timeout)
>>> File
>>> "/usr/local/lib/nagios/plugins/gluster/server_utils.py",
>>> line 114, in execNRPECommand
>>> (returncode, outputStr, err) = utils.execCmd(nrpeCmd,
>>> raw=True)
>>> File
>>> "/usr/lib/python2.7/dist-packages/glusternagios/utils.py",
>>> line 403, in execCmd
>>> deathSignal=deathSignal, childUmask=childUmask)
>>> File
>>> "/usr/local/lib/python2.7/dist-packages/cpopen/__init__.py",
>>> line 63, in __init__
>>> **kw)
>>> File "/usr/lib/python2.7/subprocess.py", line 710, in __init__
>>> errread, errwrite)
>>> File
>>> "/usr/local/lib/python2.7/dist-packages/cpopen/__init__.py",
>>> line 82, in _execute_child_v276
>>> restore_sigpipe=restore_sigpipe
>>> File
>>> "/usr/local/lib/python2.7/dist-packages/cpopen/__init__.py",
>>> line 107, in _execute_child_v275
>>> restore_sigpipe
>>> OSError: [Errno 2] No such file or directory
>>>
>>> Gluster version : 3.7.4
>>> OS : Ubuntu 14.04
>>> Complied from source tar file.
>>>
>>>
>>> regards
>>> Amudhan
>>>
>>>
>>>
>>>
>>> On Wed, Sep 30, 2015 at 6:21 PM, Humble Devassy Chirammal
>>> <humble.devassy at gmail.com <mailto:humble.devassy at gmail.com>>
>>> wrote:
>>>
>>> The EL7 rpms of gluster-nagios are available @
>>> http://download.gluster.org/pub/gluster/glusterfs-nagios/1.1.0/
>>>
>>> Hope it helps!
>>>
>>> --Humble
>>>
>>>
>>> On Tue, Sep 29, 2015 at 10:56 AM, Sahina Bose
>>> <sabose at redhat.com <mailto:sabose at redhat.com>> wrote:
>>>
>>> We will publish the EL7 builds soon.
>>>
>>> The source tarballs are now available at -
>>> http://download.gluster.org/pub/gluster/glusterfs-nagios/
>>>
>>> thanks
>>> sahina
>>>
>>>
>>> On 09/25/2015 12:55 PM, Humble Devassy Chirammal wrote:
>>>> HI Michael,
>>>>
>>>> Yes, only el6 packages are available @
>>>> http://download.gluster.org/pub/gluster/glusterfs-nagios/
>>>> . I am looping nagios project team leads to this
>>>> thread. Lets wait for them to revert.
>>>>
>>>> --Humble
>>>>
>>>>
>>>> On Sun, Sep 20, 2015 at 2:32 PM, Prof. Dr. Michael
>>>> Schefczyk <michael at schefczyk.net
>>>> <mailto:michael at schefczyk.net>> wrote:
>>>>
>>>> Dear All,
>>>>
>>>> In June 2014, the gluster-nagios team (thanks!)
>>>> published the availability of
>>>> gluster-nagios-common and gluster-nagios-addons
>>>> on this list. As far as I can tell, this quite
>>>> extensive gluster nagios monitoring tool is
>>>> available for el6 only. Are there known plans
>>>> to make this available for el7 outside the
>>>> RHEL-repos
>>>> (http://ftp.redhat.de/pub/redhat/linux/enterprise/7Server/en/RHS/SRPMS/),
>>>> e.g. for use with oVirt / Centos 7 also? It
>>>> would be good to be able to monitor gluster
>>>> without playing around with scripts from
>>>> sources other than a rpm repo.
>>>>
>>>> Regards,
>>>>
>>>> Michael
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> <mailto:Gluster-users at gluster.org>
>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>
>>>>
>>>
>>>
>>>
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>
>>>
>>
>>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
> http://www.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20151008/bf5a86b1/attachment.html>
More information about the Gluster-users
mailing list