[Gluster-users] tuning glusterfs?
pkoelle
pkoelle at gmail.com
Wed Jan 20 22:49:28 UTC 2010
Am 20.01.2010 19:02, schrieb Jiann-Ming Su:
> On Wed, Jan 20, 2010 at 10:23 AM, pkoelle<pkoelle at gmail.com> wrote:
>> Am 19.01.2010 18:30, schrieb Jiann-Ming Su:
>>>
>>> I ran a simple bonnie++ benchmark on 3.0.0. The read performance
>>> through FUSE is horrible. Are there some parameters that can be tuned
>>> to improve read performance?
>>
>> Hi Jiann,
>>
>> You didn't provide information for anyone to help you. You should at least
>> post the exact bonnie++ command line you used and some information about
>> your setup (config files, transport, etc.).
>>
>
> Bonnie++ command: ./bonnie++ -u 99:99 -d /var/boot -s 2000
>
> The system I'm testing on has 1GB RAM. /var/boot is the glusterfs
> mount via fuse.
>
> The transport is tcp but all on localhost. Here's the glusterfsd.vol:
>
> volume posix1
> type storage/posix
> option directory /mnt/gluster/gfs2
This looks suspicious. Is this a gfs2 mount? what kind of performance do
you get wen running bonnie++ directly against /mnt/gluster/gfs2?
My setup is vastly different so I can't give direct advice. It should be
possible to watch bonnie++ and the glusterfs processes via strace to
find out whats happening (especially for the ~8sec read latency).
Can you leave out some of the performance/* translators and also try a
"real" network interface with lower MTU values to see of that helps? I
had terrible results with bigger MTU (6500) and bigger values for
net.core.(r/w)mem_max in /etc/sysctl.conf. IOPS where in the range of
5/sec.
hth
Paul
> end-volume
>
> volume locks1
> type features/locks
> subvolumes posix1
> end-volume
>
> volume brick1
> type performance/io-threads
> option thread-count 8
> subvolumes locks1
> end-volume
>
> volume server-tcp
> type protocol/server
> option transport-type tcp
> option auth.addr.brick1.allow 127.0.0.1
> option transport.socket.listen-port 6996
> option transport.socket.nodelay on
> subvolumes brick1
> end-volume
>
>
> Here's the client config:
>
> # TRANSPORT-TYPE tcp
> volume localhost-1
> type protocol/client
> option transport-type tcp
> option remote-host localhost
> option transport.socket.nodelay on
> option transport.remote-port 6996
> option remote-subvolume brick1
> end-volume
>
> volume writebehind
> type performance/write-behind
> option cache-size 4MB
> subvolumes localhost-1
> end-volume
>
> volume readahead
> type performance/read-ahead
> option page-count 4
> subvolumes writebehind
> end-volume
>
> volume iocache
> type performance/io-cache
> option page-size 1MB
> option cache-size 1GB
> option cache-timeout 1
> subvolumes readahead
> end-volume
>
> volume quickread
> type performance/quick-read
> option cache-timeout 1
> option max-file-size 64kB
> subvolumes iocache
> end-volume
>
> volume statprefetch
> type performance/stat-prefetch
> subvolumes quickread
> end-volume
>
>
More information about the Gluster-users
mailing list