[Gluster-devel] Performance
Roland Fischer
roland.fischer at xidras.com
Wed Mar 17 17:05:24 UTC 2010
Hi Community,
i need your help. i have performance problems with glusterfs 3.0.0 and
domUs (xen)
i use 2 ident glusterfs-server (physikal HW) and two xen-server (physikal)
currently i use client side replication - which is awful slow. i use a
monitoring tool and see in domUs that there is a lot of cpu waiting.
(before i switch to glusterfs there was no wait CPU)
is server-side-replication faster and failsave. i mean if one glusterfs
server goes down, does the other take over the domUs?
is there anything in volfiles which i can tune?!? should i use
server-side-replication?!?
should i use the --disable-direct-io-mode? if yes on server side or
client-side or both - and how to add in fstab (with
--disable-direct-io-mode)?????
Thank you for your help!!!
servervolfile:
cat /etc/glusterfs/export-domU-images-client_repl.vol
#############
volume posix
type storage/posix
option directory /GFS/domU-images
end-volume
volume locks
type features/locks
subvolumes posix
end-volume
volume domU-images
type performance/io-threads
option thread-count 8 # default is 16
subvolumes locks
end-volume
volume server
type protocol/server
option transport-type tcp
option auth.addr.domU-images.allow 192.*.*.*,127.0.0.1
option transport.socket.listen-port 6997
subvolumes domU-images
end-volume
######################
clientvolfiles:
cat /etc/glusterfs/mount-domU-images-client_repl.vol
volume gfs-01-01
type protocol/client
option transport-type tcp
option remote-host hostname
option transport.socket.nodelay on
option remote-port 6997
option remote-subvolume domU-images
option ping-timeout 5
end-volume
volume gfs-01-02
type protocol/client
option transport-type tcp
option remote-host hostname
option transport.socket.nodelay on
option remote-port 6997
option remote-subvolume domU-images
option ping-timeout 5
end-volume
volume gfs-replicate
type cluster/replicate
subvolumes gfs-01-01 gfs-01-02
end-volume
volume writebehind
type performance/write-behind
option cache-size 4MB #default 16
subvolumes gfs-replicate
end-volume
volume readahead
type performance/read-ahead
option page-count 8 # cache per file = (page-count x
page-size)
subvolumes writebehind
end-volume
volume iocache
type performance/io-cache
option cache-size 1GB #new 1GB supported
option cache-timeout 1
subvolumes readahead
end-volume
volume statprefetch
type performance/stat-prefetch
subvolumes iocache
end-volume
#################################################
Best regards,
Roland
More information about the Gluster-devel
mailing list