[Gluster-devel] Performance question.

Kevan Benson kbenson at a-1networks.com
Wed Nov 21 17:29:35 UTC 2007


Have you tried running the tests with a single server exporting a 
filesystem without stripe or unify?  You could run it from the same NFS 
server to get a better comparison.  If glusterfs performs well as a 
simple share with one volume (maybe a few perf translators in there), 
you could add on translators and change the layout as you wanted to see 
what's causing your slowdown.

Chris Johnson wrote:
>      Hi, again,
> 
>      I asked about stack building philosophy.  Apparently there isn't
> one.  So I tried a few things.  The configs are down the end here.
> 
>      Two systems, CentOS5, both running fuse-devel-2.7.0-1 gluster
> enhanced, glusterfs-1.3.5-2.  Both have gigabit ethernet, server runs
> a SATABeast.  Currently I ge the following from from iozone.
> 
> iozone -aN -r 32k -s 131072k -f /mnt/glusterfs/sdm1/junknstuff
> 
>                                                             random  
> random    bkwd  record  stride
>               KB  reclen   write rewrite    read    reread    read   
> write    read rewrite    read   fwrite frewrite   fread  freread
>           131072      32     589     587      345      343     818     
> 621     757     624     845      592      591     346      366
> 
> Now, a similar test using NFS on a CentOS4.4 system running a 3ware RAID 
> card gives this
> 
> iozone -aN -r 32k -s 131072k -f /space/sake/5/admin/junknstuff
> 
>                                                             random  
> random    bkwd  record  stride
>               KB  reclen   write rewrite    read    reread    read   
> write    read rewrite    read   fwrite frewrite   fread  freread
>           131072      32      27      26      292       11      11      
> 24     542       9     539       30       28     295       11
> 
> And you can see that the NFS system is faster.  Is this because of the
> hardware 3ware RAID or is NFS really that much faster here?  Is there
> a better way to stack this that would improve things?  And I tried with
> and without striping.  No noticable difference in gluster performance.
> 
>      Help appreciated.
> 
> ============  server config
> 
> volume brick1
>   type storage/posix
>   option directory /home/sdm1
> end-volume
> 
> volume brick2
>   type storage/posix
>   option directory /home/sdl1
> end-volume
> 
> volume brick3
>   type storage/posix
>   option directory /home/sdk1
> end-volume
> 
> volume brick4
>   type storage/posix
>   option directory /home/sdk1
> end-volume
> 
> volume ns-brick
>   type storage/posix
>   option directory /home/sdk1
> end-volume
> 
> volume stripe1
>  type cluster/stripe
>  subvolumes brick1 brick2
> # option block-size *:10KB, end-volume
> 
> volume stripe2
>  type cluster/stripe
>  subvolumes brick3 brick4
> # option block-size *:10KB, end-volume
> 
> volume unify0
>  type cluster/unify
>  subvolumes stripe1 stripe2
>  option namespace ns-brick
>  option scheduler rr # option rr.limits.min-disk-free 5
> end-volume
> 
> volume iot
>  type performance/io-threads
>  subvolumes unify0
>  option thread-count 8
> end-volume
> 
> volume writebehind
>   type performance/write-behind
>   option aggregate-size 131072 # in bytes
>   subvolumes iot
> end-volume
> 
> volume readahead
>   type performance/read-ahead
> #  option page-size 65536 ### in bytes
>   option page-size 128kb ### in bytes
> #  option page-count 16 ### memory cache size is page-count x
> page-size per file
>   option page-count 2 ### memory cache size is page-count x page-size
> per file
>   subvolumes writebehind
> end-volume
> 
> volume server
>   type protocol/server
>   subvolumes readahead
>   option transport-type tcp/server     # For TCP/IP transport
> #  option client-volume-filename /etc/glusterfs/glusterfs-client.vol
>   option auth.ip.readahead.allow *
> end-volume
> 
> 
> ============  client config
> 
> volume client
>   type protocol/client
>   option transport-type tcp/client
>   option remote-host xxx.xxx.xxx.xxx
>   option remote-subvolume readahead
> end-volume
> 
> ------------------------------------------------------------------------------- 
> Chris Johnson               |Internet: johnson at nmr.mgh.harvard.edu
> Systems Administrator       |Web:      
> http://www.nmr.mgh.harvard.edu/~johnson
> NMR Center                  |Voice:    617.726.0949
> Mass. General Hospital      |FAX:      617.726.7422
> 149 (2301) 13th Street      |A compromise is a solution nobody is happy 
> with.
> Charlestown, MA., 02129 USA |     Observation, Unknown
> ------------------------------------------------------------------------------- 
> 
> 
> 
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at nongnu.org
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
> .
> 


-- 

-Kevan Benson
-A-1 Networks





More information about the Gluster-devel mailing list