[Gluster-devel] Best performance glusterfs v. 3.0.0 and samba
Roland Fischer
roland.fischer at xidras.com
Tue Feb 16 09:27:27 UTC 2010
Hi Harshavardhana,
thank you for your reply. i will use the performance options for samba.
to your questions:
> volume web-data-replicate
> type cluster/replicate
> subvolumes gfs-01-01 gfs-01-02
> end-volume
>
> volume readahead
> type performance/read-ahead
> option page-count 16 # cache per file = (page-count x
> page-size)
> subvolumes web-data-replicate
> end-volume
>
>>what is the client side and server side TOTAL ram ?. How many servers and
>>clients do you have?. Coz having read-ahead count on 16 is no good for an
>>ethernet link, you might be choking up the bandwidth unnecessarily.
we have 2 physikal xen server and 2 physikal gfs-server (one gfsserver with 48TB space - 24 x 2TB SATAII)
we use on xenserver 14 domUs
alle domUs are located on glusterfs. And now we try to share in domU1 another gfs-lun (gfs-partition) via samba
each of the physikal xen server have 2x Quadcore with 12GB ram
each of the physikal glusterfs server have 2x Quardcore (hyperthreading) with 12GB ram
xen - domU1 has 1 CPU and 2GB ram and shares another gfs-lun via samba
>> Even with this we would need to know the backend disk performance with
>> o-direct to properly analyse the cost of using buffering on server side to
>> get better performance out of the system.
24 x 2TB - Western Digital WD RE4 - GP with 64MB Cache
Raid Controller = Areca - ARC-1280
Controller Name ARC-1280
Serial Number Y907CAAXAR800316
Main Processor 800MHz IOP341 C1
CPU ICache Size 32KBytes
CPU DCache Size 32KBytes/Write Back
CPU SCache Size 512KBytes/Write Back
System Memory 256MB/533MHz/ECC
Thank you
regards
Roland
More information about the Gluster-devel
mailing list