[Gluster-users] Diagnosing Intermittent Performance Problems Possibly Caused by Gremlins

Matt matt at mattlantis.com
Tue Feb 3 16:38:41 UTC 2015


I’ve been trying for weeks to reproduce the performance problems in 
our preproduction environments but can’t. As a result, selling that 
just upgrading to 3.6.x and hoping it goes away might be tricky. 3.6 is 
perceived as a little too bleeding edge, and we’ve actually had some 
other not fully explained issues with this cluster recently that make 
us hesitate. I don’t think they’re related.

On Tue, Feb 3, 2015 at 4:58 AM, Justin Clift <justin at gluster.org> wrote:
> ----- Original Message -----
>>  Hello List,
>> 
>>  So I've been frustraded by intermittent performance problems 
>> throughout
>>  January. The problem occurs on a two node setup running 3.4.5, 16 
>> gigs
>>  of ram with a bunch of local disk. For sometimes an hour for 
>> sometimes
>>  weeks at a time (I have extensive graphs in OpenNMS) our Gluster 
>> boxes
>>  will get their CPUs pegged, and in vmstat they'll show extremely 
>> high
>>  numbers of context switches and interrupts. Eventually things calm
>>  down. During this time, memory usage actually drops. Overall usage 
>> on
>>  the box goes from between 6-10 gigs to right around 4 gigs, and 
>> stays
>>  there. That's what really puzzles me.
>> 
>>  When performance is problematic, sar shows one device, the device
>>  corresponding to the glusterfsd problem using all the CPU doing 
>> lots of
>>  little reads, Sometimes 70k/second, very small avg rq size, say 
>> 10-12.
>>  Afraid I don't have any saved output handy, but I can try to capture
>>  some next time it happens. I have tons of information frankly, but 
>> am
>>  trying to keep this reasonably brief.
>> 
>>  There are more than a dozen volumes on this two node setup. The CPU
>>  usage is pretty much entirely contained to one volume, a 1.5 TB 
>> volume
>>  that is just shy of 70% full. It stores uploaded files for a web 
>> app.
>>  What I hate about this app and so am always suspicious of, is that 
>> it
>>  stores a directory for every user in one level, so under the /data
>>  directory in the volume, there are 450,000 sub directories at this
>>  point.
>> 
>>  The only real mitigation step that's been taken so far was to turn 
>> off
>>  the self-heal daemon on the volume, as I thought maybe crawling that
>>  large directory was getting expensive. This doesn't seem to have 
>> done
>>  anything as the problem still occurs.
>> 
>>  At this point I figure there are one of two things sorts of things
>>  happening really broadly: one we're running into some sort of bug or
>>  performance problem with gluster we should either fix perhaps by
>>  upgrading or tuning around, or two, some process we're running but 
>> not
>>  aware of is hammering the file system causing problems.
>> 
>>  If it's the latter option, can anyone give me any tips on figuring 
>> out
>>  what might be hammering the system? I can use volume top to see 
>> what a
>>  brick is doing, but I can't figure out how to tell what clients are
>>  doing what.
>> 
>>  Apologies for the somewhat broad nature of the question, any input
>>  thoughts would be much appreciated. I can certainly provide more 
>> info
>>  about some things if it would help, but I've tried not to write a 
>> novel
>>  here.
> 
> Out of curiosity, are you able to test using GlusterFS 3.6.2?  We've
> had a bunch of pretty in-depth upstream testing at decent scale (100+
> nodes) from 3.5.x onwards, with lots of performance issues identified
> and fixed on the way through.
> 
> So, I'm kinda hopeful the problem you're describing is fixed in newer
> releases. :D
> 
> Regards and best wishes,
> 
> Justin Clift
> 
> --
> GlusterFS - http://www.gluster.org
> 
> An open source, distributed file system scaling to several
> petabytes, and handling thousands of clients.
> 
> My personal twitter: twitter.com/realjustinclift
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150203/8fd2cb2d/attachment.html>


More information about the Gluster-users mailing list