[Gluster-users] performance in 3.3

Doug Schouten dschoute at sfu.ca
Fri Oct 19 15:01:12 UTC 2012


Hi again,

Addendum: I explicitly set direct-io to 'disable' and now the 
performance is at 110 MB/s ... which is what I would expect over a 1Gb link.

thanks! Doug

On 10/19/2012 07:29 AM, Doug Schouten wrote:
> Hi Fernando, Anand,
>
> Thanks for the suggestion. I removed the enable-direct-io option in
> /etc/fstab, remounted, and the performance is the same within
> statistical precision.
>
> cheers, Doug
>
>
> On 10/19/2012 02:13 AM, Fernando Frediani (Qube) wrote:
>> Hi Doug,
>>
>> Try to make the change suggested by Anand and let us know how you get
>> on. I am interested to hear about the performance on 3.3 because bad
>> performance has been subject of many emails for a while here.
>>
>> Regards,
>>
>> Fernando
>>
>> -----Original Message-----
>> From: gluster-users-bounces at gluster.org
>> [mailto:gluster-users-bounces at gluster.org] On Behalf Of Doug Schouten
>> Sent: 19 October 2012 02:45
>> To: gluster-users at gluster.org
>> Subject: [Gluster-users] performance in 3.3
>>
>> Hi,
>>
>> I am noticing a rather slow read performance using GlusterFS 3.3 with
>> the following configuration:
>>
>> Number of Bricks: 4
>> Transport-type: tcp
>> Bricks:
>> Brick1: server1:/srv/data
>> Brick2: server2:/srv/data
>> Brick3: server3:/srv/data
>> Brick4: server4:/srv/data
>> Options Reconfigured:
>> features.quota: off
>> features.quota-timeout: 1800
>> performance.flush-behind: on
>> performance.io-thread-count: 64
>> performance.quick-read: on
>> performance.stat-prefetch: on
>> performance.io-cache: on
>> performance.write-behind: on
>> performance.read-ahead: on
>> performance.write-behind-window-size: 4MB
>> performance.cache-refresh-timeout: 1
>> performance.cache-size: 4GB
>> nfs.rpc-auth-allow: none
>> network.frame-timeout: 60
>> nfs.disable: on
>> performance.cache-max-file-size: 1GB
>>
>>
>> The servers are connected with bonded 1Gb ethernet, and have LSI
>> MegaRAID arrays with 12x1 TB disks in RAID-6 array, using XFS file
>> system mounted like:
>>
>> xfs     logbufs=8,logbsize=32k,noatime,nodiratime  0    0
>>
>> and we use the FUSE client
>>
>> localhost:/global /global glusterfs
>> defaults,direct-io-mode=enable,log-level=WARNING,log-file=/var/log/gluster.log
>>
>> 0 0
>>
>> Our files are all >= 2MB. When rsync-ing we see about 50MB/s read
>> performance which improves to 250MB/s after the first copy. This
>> indicates to me that the disk caching is working as expected. However
>> I am rather surprised by the low 50MB/s read speed; this is too low to
>> be limited by network, and the native disk read performance is way
>> better.
>> Is there some configuration that can improve this situation?
>>
>> thanks,
>>
>>
>

-- 


  Doug Schouten
  Research Associate
  TRIUMF



More information about the Gluster-users mailing list