[Gluster-devel] [Gluster-users] 120k context switches on GlsuterFS nodes
Joe Julian
joe at julianfamily.org
Thu May 18 13:09:47 UTC 2017
On the other hand, tracking that stat between versions with a known test sequence may be valuable for watching for performance issues or improvements.
On May 17, 2017 10:03:28 PM PDT, Ravishankar N <ravishankar at redhat.com> wrote:
>On 05/17/2017 11:07 PM, Pranith Kumar Karampuri wrote:
>> + gluster-devel
>>
>> On Wed, May 17, 2017 at 10:50 PM, mabi <mabi at protonmail.ch
>> <mailto:mabi at protonmail.ch>> wrote:
>>
>> I don't know exactly what kind of context-switches it was but
>what
>> I know is that it is the "cs" number under "system" when you run
>> vmstat.
>>
>Okay, that could be due to the syscalls themselves or pre-emptive
>multitasking in case there aren't enough cpu cores. I think the spike
>in
>numbers is due to more users accessing the files at the same time like
>you observed, translating into more syscalls. You can try capturing
>the
>gluster volume profile info the next time it occurs and co-relate with
>the cs count. If you don't see any negative performance impact, I think
>
>you don't need to be bothered much by the numbers.
>
>HTH,
>Ravi
>>
>>
>> Also I use the percona linux monitoring template for cacti
>>
>(https://www.percona.com/doc/percona-monitoring-plugins/LATEST/cacti/linux-templates.html
>>
><https://www.percona.com/doc/percona-monitoring-plugins/LATEST/cacti/linux-templates.html>)
>> which monitors context switches too. If that's of any use
>> interrupts where also quite high during that time with peaks up
>to
>> 50k interrupts.
>>
>>
>>
>>> -------- Original Message --------
>>> Subject: Re: [Gluster-users] 120k context switches on GlsuterFS
>nodes
>>> Local Time: May 17, 2017 2:37 AM
>>> UTC Time: May 17, 2017 12:37 AM
>>> From: ravishankar at redhat.com <mailto:ravishankar at redhat.com>
>>> To: mabi <mabi at protonmail.ch <mailto:mabi at protonmail.ch>>,
>>> Gluster Users <gluster-users at gluster.org
>>> <mailto:gluster-users at gluster.org>>
>>>
>>>
>>> On 05/16/2017 11:13 PM, mabi wrote:
>>>> Today I even saw up to 400k context switches for around 30
>>>> minutes on my two nodes replica... Does anyone else have so
>high
>>>> context switches on their GlusterFS nodes?
>>>>
>>>> I am wondering what is "normal" and if I should be worried...
>>>>
>>>>
>>>>
>>>>
>>>>> -------- Original Message --------
>>>>> Subject: 120k context switches on GlsuterFS nodes
>>>>> Local Time: May 11, 2017 9:18 PM
>>>>> UTC Time: May 11, 2017 7:18 PM
>>>>> From: mabi at protonmail.ch <mailto:mabi at protonmail.ch>
>>>>> To: Gluster Users <gluster-users at gluster.org>
>>>>> <mailto:gluster-users at gluster.org>
>>>>>
>>>>> Hi,
>>>>>
>>>>> Today I noticed that for around 50 minutes my two GlusterFS
>>>>> 3.8.11 nodes had a very high amount of context switches,
>around
>>>>> 120k. Usually the average is more around 1k-2k. So I checked
>>>>> what was happening and there where just more users accessing
>>>>> (downloading) their files at the same time. These are
>>>>> directories with typical cloud files, which means files of any
>>>>> sizes ranging from a few kB to MB and a lot of course.
>>>>>
>>>>> Now I never saw such a high number in context switches in my
>>>>> entire life so I wanted to ask if this is normal or to be
>>>>> expected? I do not find any signs of errors or warnings in any
>>>>> log files.
>>>>>
>>>
>>> What context switch are you referring to (syscalls
>context-switch
>>> on the bricks?) ? How did you measure this?
>>> -Ravi
>>>
>>>>> My volume is a replicated volume on two nodes with ZFS as
>>>>> filesystem behind and the volume is mounted using FUSE on the
>>>>> client (the cloud server). On that cloud server the glusterfs
>>>>> process was using quite a lot of system CPU but that server
>>>>> (VM) only has 2 vCPUs so maybe I should increase the number of
>>>>> vCPUs...
>>>>>
>>>>> Any ideas or recommendations?
>>>>>
>>>>>
>>>>>
>>>>> Regards,
>>>>> M.
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>>> <http://lists.gluster.org/mailman/listinfo/gluster-users>
>>>
>> _______________________________________________ Gluster-users
>> mailing list Gluster-users at gluster.org
>> <mailto:Gluster-users at gluster.org>
>> http://lists.gluster.org/mailman/listinfo/gluster-users
>> <http://lists.gluster.org/mailman/listinfo/gluster-users>
>>
>> --
>> Pranith
--
Sent from my Android device with K-9 Mail. Please excuse my brevity.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-devel/attachments/20170518/8bea508b/attachment.html>
More information about the Gluster-devel
mailing list