[Gluster-devel] Performance

Raghavendra G raghavendra at gluster.com
Thu Mar 18 04:46:25 UTC 2010


Hi Roland,

What are the applications you are using on glusterfs? In particular, what is
the i/o pattern of applications? As a general guideline, you can try
enabling/disabling each of the performance translators and observe the
gain/loss of performance and tune the configuration accordingly.

regards,
On Wed, Mar 17, 2010 at 9:05 PM, Roland Fischer
<roland.fischer at xidras.com>wrote:

> Hi Community,
>
> i need your help. i have performance problems with glusterfs 3.0.0 and
> domUs (xen)
>
> i use 2 ident glusterfs-server (physikal HW) and two xen-server (physikal)
>
> currently i use client side replication - which is awful slow. i use a
> monitoring tool and see in domUs that there is a lot of cpu waiting. (before
> i switch to glusterfs there was no wait CPU)
>
> is server-side-replication faster and failsave. i mean if one glusterfs
> server goes down, does the other take over the domUs?
>
> is there anything in volfiles which i can tune?!? should i use
> server-side-replication?!?
>
> should i use the --disable-direct-io-mode? if yes on server side or
> client-side or both  -  and how to add in fstab (with
> --disable-direct-io-mode)?????
>
> Thank you for your help!!!
>
> servervolfile:
> cat /etc/glusterfs/export-domU-images-client_repl.vol
> #############
> volume posix
>  type storage/posix
>  option directory /GFS/domU-images
> end-volume
>
> volume locks
>  type features/locks
>  subvolumes posix
> end-volume
>
> volume domU-images
>  type performance/io-threads
>  option thread-count 8 # default is 16
>  subvolumes locks
> end-volume
>
> volume server
>  type protocol/server
>  option transport-type tcp
>  option auth.addr.domU-images.allow 192.*.*.*,127.0.0.1
>  option transport.socket.listen-port 6997
>  subvolumes domU-images
> end-volume
> ######################
>
> clientvolfiles:
>
> cat /etc/glusterfs/mount-domU-images-client_repl.vol
> volume gfs-01-01
>  type protocol/client
>  option transport-type tcp
>  option remote-host hostname
>  option transport.socket.nodelay on
>  option remote-port 6997
>  option remote-subvolume domU-images
>  option ping-timeout 5
> end-volume
>
> volume gfs-01-02
>  type protocol/client
>  option transport-type tcp
>  option remote-host hostname
>  option transport.socket.nodelay on
>  option remote-port 6997
>  option remote-subvolume domU-images
>  option ping-timeout 5
> end-volume
>
> volume gfs-replicate
>  type cluster/replicate
>  subvolumes gfs-01-01 gfs-01-02
> end-volume
>
> volume writebehind
>  type performance/write-behind
>  option cache-size 4MB   #default 16
>  subvolumes gfs-replicate
> end-volume
>
> volume readahead
>  type performance/read-ahead
>  option page-count 8              # cache per file  = (page-count x
> page-size)
>  subvolumes writebehind
> end-volume
>
> volume iocache
>  type performance/io-cache
>  option cache-size 1GB   #new 1GB supported
>  option cache-timeout 1
>  subvolumes readahead
> end-volume
>
> volume statprefetch
>    type performance/stat-prefetch
>    subvolumes iocache
> end-volume
>
> #################################################
>
>
> Best regards,
> Roland
>
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at nongnu.org
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>



-- 
Raghavendra G
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-devel/attachments/20100318/672bd9d5/attachment-0003.html>


More information about the Gluster-devel mailing list