[Gluster-users] Slow read performance

Thomas Wakefield twake at cola.iges.org
Wed Feb 27 17:31:37 UTC 2013

Help please-

I am running 3.3.1 on Centos using a 10GB network.  I get reasonable write speeds, although I think they could be faster.  But my read speeds are REALLY slow.

Executive summary:

On gluster client-
Writes average about 700-800MB/s
Reads average about 70-80MB/s

On server-
Writes average about 1-1.5GB/s
Reads average about 2-3GB/s

Any thoughts?

Here are some additional details:

Nothing interesting in any of the log files, everything is very quite.
All servers had no other load, and all clients are performing the same way.

Volume Name: shared
Type: Distribute
Volume ID: de11cc19-0085-41c3-881e-995cca244620
Status: Started
Number of Bricks: 26
Transport-type: tcp
Brick1: fs-disk2:/storage/disk2a
Brick2: fs-disk2:/storage/disk2b
Brick3: fs-disk2:/storage/disk2d
Brick4: fs-disk2:/storage/disk2e
Brick5: fs-disk2:/storage/disk2f
Brick6: fs-disk2:/storage/disk2g
Brick7: fs-disk2:/storage/disk2h
Brick8: fs-disk2:/storage/disk2i
Brick9: fs-disk2:/storage/disk2j
Brick10: fs-disk2:/storage/disk2k
Brick11: fs-disk2:/storage/disk2l
Brick12: fs-disk2:/storage/disk2m
Brick13: fs-disk2:/storage/disk2n
Brick14: fs-disk2:/storage/disk2o
Brick15: fs-disk2:/storage/disk2p
Brick16: fs-disk2:/storage/disk2q
Brick17: fs-disk2:/storage/disk2r
Brick18: fs-disk2:/storage/disk2s
Brick19: fs-disk2:/storage/disk2t
Brick20: fs-disk2:/storage/disk2u
Brick21: fs-disk2:/storage/disk2v
Brick22: fs-disk2:/storage/disk2w
Brick23: fs-disk2:/storage/disk2x
Brick24: fs-disk3:/storage/disk3a
Brick25: fs-disk3:/storage/disk3b
Brick26: fs-disk3:/storage/disk3c
Options Reconfigured:
performance.write-behind: on
performance.read-ahead: on
performance.io-cache: on
performance.stat-prefetch: on
performance.quick-read: on
cluster.min-free-disk: 500GB
nfs.disable: off

sysctl.conf settings for 10GBe
# increase TCP max buffer size settable using setsockopt()
net.core.rmem_max = 67108864 
net.core.wmem_max = 67108864 
# increase Linux autotuning TCP buffer limit
net.ipv4.tcp_rmem = 4096 87380 67108864
net.ipv4.tcp_wmem = 4096 65536 67108864
# increase the length of the processor input queue
net.core.netdev_max_backlog = 250000
# recommended default congestion control is htcp 
# recommended for hosts with jumbo frames enabled

Thomas W.
Sr.  Systems Administrator COLA/IGES
twake at cola.iges.org
Affiliate Computer Scientist GMU

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130227/2d08810b/attachment.html>

More information about the Gluster-users mailing list