[Gluster-users] tuning
Papp Tamas
tompos at martos.bme.hu
Wed Jul 6 21:58:05 UTC 2011
hi!
I'm almost absolutely new to glusterfs.
Until now we used Storage Platform (3.0.5).
Today we installed Ubuntu 11.04 and glusterfs 3.2.1.
$ cat w-vol-fuse.vol
volume w-vol-client-0
type protocol/client
option remote-host gl0
option remote-subvolume /mnt/brick1
option transport-type tcp
end-volume
volume w-vol-client-1
type protocol/client
option remote-host gl1
option remote-subvolume /mnt/brick1
option transport-type tcp
end-volume
volume w-vol-client-2
type protocol/client
option remote-host gl2
option remote-subvolume /mnt/brick1
option transport-type tcp
end-volume
volume w-vol-client-3
type protocol/client
option remote-host gl3
option remote-subvolume /mnt/brick1
option transport-type tcp
end-volume
volume w-vol-client-4
type protocol/client
option remote-host gl4
option remote-subvolume /mnt/brick1
option transport-type tcp
end-volume
volume w-vol-dht
type cluster/distribute
subvolumes w-vol-client-0 w-vol-client-1 w-vol-client-2
w-vol-client-3 w-vol-client-4
end-volume
volume w-vol-write-behind
type performance/write-behind
option cache-size 4MB
subvolumes w-vol-dht
end-volume
volume w-vol-read-ahead
type performance/read-ahead
subvolumes w-vol-write-behind
end-volume
volume w-vol-io-cache
type performance/io-cache
option cache-size 128MB
subvolumes w-vol-read-ahead
end-volume
volume w-vol-quick-read
type performance/quick-read
option cache-size 128MB
subvolumes w-vol-io-cache
end-volume
volume w-vol-stat-prefetch
type performance/stat-prefetch
subvolumes w-vol-quick-read
end-volume
volume w-vol
type debug/io-stats
option latency-measurement off
option count-fop-hits off
subvolumes w-vol-stat-prefetch
end-volume
$ cat w-vol.gl0.mnt-brick1.vol
volume w-vol-posix
type storage/posix
option directory /mnt/brick1
end-volume
volume w-vol-access-control
type features/access-control
subvolumes w-vol-posix
end-volume
volume w-vol-locks
type features/locks
subvolumes w-vol-access-control
end-volume
volume w-vol-io-threads
type performance/io-threads
subvolumes w-vol-locks
end-volume
volume w-vol-marker
type features/marker
option volume-uuid ad362448-7ef0-49ae-b13c-74cb82ce9be5
option timestamp-file /etc/glusterd/vols/w-vol/marker.tstamp
option xtime off
option quota off
subvolumes w-vol-io-threads
end-volume
volume /mnt/brick1
type debug/io-stats
option latency-measurement off
option count-fop-hits off
subvolumes w-vol-marker
end-volume
volume w-vol-server
type protocol/server
option transport-type tcp
option auth.addr./mnt/brick1.allow *
subvolumes /mnt/brick1
end-volume
There is 5 nodes. 3 have 8 disks in RAID 6 (supermicro server, are
controller), 2 have 8 disks in raid5+spare (DL180).
Filesystem of datas was created via this command (of course a bit
different on HPs):
mkfs.xfs -b size=4096 -d sunit=256,swidth=1536 -L gluster /dev/sda4
The performance is far away that was before. I tried to modify
performance.write-behind-window-size 4MB
gluster volume set w-vol performance.cache-size 128MB
gluster volume set w-vol nfs.disable on
echo 512 > /sys/block/sda/queue/nr_requests
blockdev --setra 16384 /dev/sda
sysctl -w vm.swappiness=5
sysctl -w vm.dirty_background_ratio=3
sysctl -w vm.dirty_ratio=40
sysctl -w kernel.sysrq=0
Nothing really helped.
Can somebody give some instructions?
Thank you,
tamas
More information about the Gluster-users
mailing list