[Gluster-users] Performance gluster 3.2.5 + QLogic Infiniband

Michael Mayer michael at mayer.cx
Sat Apr 21 08:40:53 UTC 2012


Hi all,

thanks for your suggestions,

i think I have "solved" the performance issue now. I had a few too many 
kernel patches included. I am back to the stock RHEL 5.8 kernel with 
stock QLogic OFED and everything works ok (IPoIB). My original intent 
was to explore cachefs on RHEL5 by building a 2.6.32 kernel but while 
cachefs worked like a treat performance for gluster was as bad as 
reported previously - so will go without cachefs for now and reintroduce 
cachefs in an OS upgrade later on.

I even have a nicely working rdma setup now and - using that - 
performance is 900 MB/s + and that consistently so.

Since I have two volumes exported by the same bricks it seems I only can 
get one of them to use RDMA, the other will then refuse to mount and 
only mount if not using rdma on that one - but that is not a real 
problem for now as the second one is only used for backup purposes.

Michael,

On 04/12/2012 01:13 AM, Fabricio Cannini wrote:
> Hi there
>
> The only time i setup a gluster "distributed scratch" like Michael is 
> doing,
> ( 3.0.5 Debian packages ) i too choose IPoIB simply because i could 
> not get rdma working at all.
> Time was short and IPoIB "Just worked" well enough for our demand at 
> the time, so i didn't looked into this issue. Plus, pinging and 
> ssh'ing into a node through the IB interface comes handy when 
> diagnosing and fixing networking issues.
>
> Em quarta-feira, 11 de abril de 2012, Sabuj 
> Pattanayek<sabujp at gmail.com <mailto:sabujp at gmail.com>> escreveu:
> > I wonder if it's possible to have both rdma and ipoib served by a
> > single glusterfsd so I can test this? I guess so, since it's just a
> > tcp mount?
> >
> > On Wed, Apr 11, 2012 at 1:43 PM, Harry Mangalam 
> <harry.mangalam at uci.edu <mailto:harry.mangalam at uci.edu>> wrote:
> >> On Tuesday 10 April 2012 15:47:08 Bryan Whitehead wrote:
> >>
> >>> with my infiniband setup I found my performance was much better by
> >>
> >>> setting up a TCP network over infiniband and then using pure tcp as
> >>
> >>> the transport with my gluster volume. For the life of me I couldn't
> >>
> >>> get rdma to beat tcp.
> >>
> >> Thanks for that data point, Brian.
> >>
> >> Very interesting. Is this a common experience? The RDMA experience 
> has not
> >> been a very smooth one for me and doing everything with IPoIB would 
> save a
> >> lot of headaches, especially if it's also higher performance.
> >>
> >> hjm
> >>
> >> --
> >>
> >> Harry Mangalam - Research Computing, OIT, Rm 225 MSTB, UC Irvine
> >>
> >> [ZOT 2225] / 92697 Google Voice Multiplexer: (949) 478-4487
> >>
> >> 415 South Circle View Dr, Irvine, CA, 92697 [shipping]
> >>
> >> MSTB Lat/Long: (33.642025,-117.844414) (paste into Google Maps)
> >>
> >> --
> >>
> >>
> >> _______________________________________________
> >> Gluster-users mailing list
> >> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
> >> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
> >>
> > _______________________________________________
> > Gluster-users mailing list
> > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
> > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
> >
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20120421/b40503a3/attachment.html>


More information about the Gluster-users mailing list