[Gluster-users] Gluster 3.6.3 performance.cache-size not working as expected in some cases

Raghavendra Bhat rabhat at redhat.com
Wed Sep 2 07:15:37 UTC 2015


Hi Christian,

I have been working on it since couple of days. I have not been able to 
recreate the issue. I will continue to recreate and get back to you in a 
day or two.

Regards,
Raghavendra Bhat


On 09/02/2015 12:45 AM, Christian Rice wrote:
> This is still an issue for me, I don’t need anyone to tear the code 
> apart, but I’d be grateful if someone would even chime in and say 
> “yeah, we’ve seen that too."
>
> From: Christian Rice <crice at pandora.com <mailto:crice at pandora.com>>
> Date: Sunday, August 30, 2015 at 11:18 PM
> To: "gluster-users at gluster.org <mailto:gluster-users at gluster.org>" 
> <gluster-users at gluster.org <mailto:gluster-users at gluster.org>>
> Subject: [Gluster-users] Gluster 3.6.3 performance.cache-size not 
> working as expected in some cases
>
> I am confused about my caching problem.  I’ll try to keep this as 
> straightforward as possible and include the basic details...
>
> I have a sixteen node distributed volume, one brick per node, XFS 
> isize=512, Debian 7/Wheezy, 32GB RAM minimally.  Every brick node is 
> also a gluster client, and also importantly an HTTP server.  We use a 
> back-end 1GbE network for gluster traffic (eth1).  There are a couple 
> dozen gluster client-only systems accessing this volume, as well.
>
> We had a really hot spot on one brick due to an oft-requested file, 
> and every time any httpd process on any gluster client was asked to 
> deliver the file, it was physically fetching it (we could see this 
> traffic using, say, ‘iftop -i eth1’,) so we thought to increase the 
> volume cache timeout and cache size.  We set the following values for 
> testing:
>
> performance.cache-size 16GB
> performance.cache-refresh-timeout: 30
>
> This test was run from a node that didn’t have the requested file on 
> the local brick:
>
> while(true); do cat /path/to/file > /dev/null; done
>
> and what had been very high traffic on the gluster backend network, 
> delivering the data repeatedly to my requesting node, dropped to 
> nothing visible.
>
> I thought good, problem fixed.  Caching works.  My colleague had run a 
> test early on to show this perf issue, so he ran it again to sign off.
>
> His testing used curl, because all the real front end traffic is HTTP, 
> and all the gluster nodes are web servers, which are of course using 
> the fuse mount to access the document root.  Even with our performance 
> tuning, the traffic on the gluster backend subnet was continuous and 
> undiminished.  I saw no evidence of cache (again using ‘iftop -i 
> eth1’, which showed a steady 75+% of line rate on a 1GbE link.
>
> Does that make sense at all?  We had theorized that we wouldn’t get to 
> use VFS/kernel page cache on any node except maybe the one which held 
> the data in the local brick.  That’s what drove us to setting the 
> gluster performance cache.  But it doesn’t seem to come into play with 
> http access.
>
>
> Volume info:
> Volume Name: DOCROOT
> Type: Distribute
> Volume ID: 3aecd277-4d26-44cd-879d-cffbb1fec6ba
> Status: Started
> Number of Bricks: 16
> Transport-type: tcp
> Bricks:
> <snipped list of bricks>
> Options Reconfigured:
> performance.cache-refresh-timeout: 30
> performance.cache-size: 16GB
>
> The net result of being overwhelmed by a hot spot is all the gluster 
> client nodes lose access to the gluster volume—it becomes so busy it 
> hangs.  When the traffic goes away (failing health checks by load 
> balancers causes requests to be redirected elsewhere), the volume 
> eventually unfreezes and life goes on.
>
> I wish I could type ALL that into a google query and get a lucid answer :)
>
> Regards,
> Christian
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150902/1da52d16/attachment.html>


More information about the Gluster-users mailing list