[Gluster-users] Gluster single file I/O
Nathan Stratton
nathan at robotics.net
Sun Feb 22 12:52:01 UTC 2009
On Sat, 21 Feb 2009, Keith Freedman wrote:
> is direct, a locally attached hard drive?
> a network filesystem will NEVER perform as well as a locally attached disk.
Odd, actually you should be able to make it preform BETTER, as you scale
Lustre the I/O for even a single file scales.
> I think you're numbers aren't too unreasonable.
> You could probably improve your performance by adding some performance
> translators.
I have them on the server, should I also put them on the client?
> Write behind would likely help you a bit.
Also, even DRBD is better and it is IPoIP, not raw infiniband:
[root at xen0 share]# dd if=/dev/zero of=/share/bar bs=1G count=8
8+0 records in
8+0 records out
8589934592 bytes (8.6 GB) copied, 60.8988 seconds, 141 MB/s
> At 06:59 PM 2/21/2009, Nathan Stratton wrote:
>
>> Direct:
>> [root at xen0 unify]# dd if=/dev/zero of=/sdb2/bar bs=1G count=8
>> 8+0 records in
>> 8+0 records out
>> 8589934592 bytes (8.6 GB) copied, 51.3145 seconds, 167 MB/s
>>
>> Gluster:
>> [root at xen0 unify]# dd if=/dev/zero of=/unify/foo bs=1G count=8
>> 8+0 records in
>> 8+0 records out
>> 8589934592 bytes (8.6 GB) copied, 87.7885 seconds, 97.8 MB/s
>>
>> Boxes are connected with 10 gig Infiniband so that should not be an issue.
>>
>> http://share.robotics.net/glusterfs.vol
>> http://share.robotics.net/glusterfsd.vol
>>
>>> <>
>> Nathan Stratton CTO, BlinkMind, Inc.
>> nathan at robotics.net nathan at blinkmind.com
>> http://www.robotics.net http://www.blinkmind.com
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users
>
More information about the Gluster-users
mailing list