<div dir="ltr">Please do volume profile also for around 10 minutes when CPU% is high.<br></div><br><div class="gmail_quote"><div dir="ltr">On Fri, Aug 17, 2018 at 11:56 AM Pranith Kumar Karampuri <<a href="mailto:pkarampu@redhat.com">pkarampu@redhat.com</a>> wrote:<br></div><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><div dir="ltr">As per the output, all io-threads are using a lot of CPU. It is better to check what the volume profile is to see what is leading to so much work for io-threads. Please follow the documentation at <a href="https://gluster.readthedocs.io/en/latest/Administrator%20Guide/Monitoring%20Workload/" target="_blank">https://gluster.readthedocs.io/en/latest/Administrator%20Guide/Monitoring%20Workload/</a> section: "<h2 id="m_-6912089762367665109gmail-running-glusterfs-volume-profile-command"><font size="2">Running GlusterFS Volume Profile Command"</font></h2>and attach output of "<code>gluster volume profile info", <br></code></div><br><div class="gmail_quote"><div dir="ltr">On Fri, Aug 17, 2018 at 11:24 AM Hu Bert <<a href="mailto:revirii@googlemail.com" target="_blank">revirii@googlemail.com</a>> wrote:<br></div><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">Good morning,<br>
<br>
i ran the command during 100% CPU usage and attached the file.<br>
Hopefully it helps.<br>
<br>
2018-08-17 7:33 GMT+02:00 Pranith Kumar Karampuri <<a href="mailto:pkarampu@redhat.com" target="_blank">pkarampu@redhat.com</a>>:<br>
> Could you do the following on one of the nodes where you are observing high<br>
> CPU usage and attach that file to this thread? We can find what<br>
> threads/processes are leading to high usage. Do this for say 10 minutes when<br>
> you see the ~100% CPU.<br>
><br>
> top -bHd 5 > /tmp/top.${HOSTNAME}.txt<br>
><br>
> On Wed, Aug 15, 2018 at 2:37 PM Hu Bert <<a href="mailto:revirii@googlemail.com" target="_blank">revirii@googlemail.com</a>> wrote:<br>
>><br>
>> Hello again :-)<br>
>><br>
>> The self heal must have finished as there are no log entries in<br>
>> glustershd.log files anymore. According to munin disk latency (average<br>
>> io wait) has gone down to 100 ms, and disk utilization has gone down<br>
>> to ~60% - both on all servers and hard disks.<br>
>><br>
>> But now system load on 2 servers (which were in the good state)<br>
>> fluctuates between 60 and 100; the server with the formerly failed<br>
>> disk has a load of 20-30.I've uploaded some munin graphics of the cpu<br>
>> usage:<br>
>><br>
>> <a href="https://abload.de/img/gluster11_cpu31d3a.png" rel="noreferrer" target="_blank">https://abload.de/img/gluster11_cpu31d3a.png</a><br>
>> <a href="https://abload.de/img/gluster12_cpu8sem7.png" rel="noreferrer" target="_blank">https://abload.de/img/gluster12_cpu8sem7.png</a><br>
>> <a href="https://abload.de/img/gluster13_cpud7eni.png" rel="noreferrer" target="_blank">https://abload.de/img/gluster13_cpud7eni.png</a><br>
>><br>
>> This can't be normal. 2 of the servers under heavy load and one not<br>
>> that much. Does anyone have an explanation of this strange behaviour?<br>
>><br>
>><br>
>> Thx :-)<br>
>><br>
>> 2018-08-14 9:37 GMT+02:00 Hu Bert <<a href="mailto:revirii@googlemail.com" target="_blank">revirii@googlemail.com</a>>:<br>
>> > Hi there,<br>
>> ><br>
>> > well, it seems the heal has finally finished. Couldn't see/find any<br>
>> > related log message; is there such a message in a specific log file?<br>
>> ><br>
>> > But i see the same behaviour when the last heal finished: all CPU<br>
>> > cores are consumed by brick processes; not only by the formerly failed<br>
>> > bricksdd1, but by all 4 brick processes (and their threads). Load goes<br>
>> > up to > 100 on the 2 servers with the not-failed brick, and<br>
>> > glustershd.log gets filled with a lot of entries. Load on the server<br>
>> > with the then failed brick not that high, but still ~60.<br>
>> ><br>
>> > Is this behaviour normal? Is there some post-heal after a heal has<br>
>> > finished?<br>
>> ><br>
>> > thx in advance :-)<br>
><br>
><br>
><br>
> --<br>
> Pranith<br>
</blockquote></div><br clear="all"><br>-- <br><div dir="ltr" class="m_-6912089762367665109gmail_signature" data-smartmail="gmail_signature"><div dir="ltr">Pranith<br></div></div>
</blockquote></div><br clear="all"><br>-- <br><div dir="ltr" class="gmail_signature" data-smartmail="gmail_signature"><div dir="ltr">Pranith<br></div></div>