[Gluster-users] Problem: Gluster performance too high..
Jeff White
jaw171 at pitt.edu
Thu Mar 29 19:40:23 UTC 2012
Maybe it's cheating by writing sparse files or something of the like
because it knows it's all zeros? Create some files locally from
/dev/urandom and copy them. I think you'll see much lower performance.
Better yet, use bonnie++.
Jeff White - Linux/Unix Systems Engineer
University of Pittsburgh - CSSD
On 03/29/2012 03:34 PM, Jeff Darcy wrote:
> On 03/29/2012 03:30 PM, Harry Mangalam wrote:
>> I'm doing some perf tests on a small gluster filesystem - 5 bricks on 4
>> servers, all single-homed on the private net.
>>
>> I've spawned up to 70 simultaneous jobs on our cluster nodes writing files of
>> various sizes from /dev/zero to the gluster fs to see what the effect on the
>> aggregate bandwith and the data is slightly unbelievable in that it seems to
>> exceed the theoretical max of the network. (I used /dev/zero instead of
>> /dev/urandom since /dev/urandom couldn't generate data fast enough.
>>
>> The 35,000 files of the right size do hit the filesystem (of course they're all
>> zero's) but the speed at which they transfer exceeds (by quite a bit) the
>> theoretical max of a 1 Gb network.
>>
>> Does gluster (or anything else) do transparent compression? What else would
>> explain this oddity?
> How do you define "theoretical max of a 1Gb network"? If it's a switched
> network, the actual maximum throughput depends on the capabilities of the
> switch but is likely to be far in excess of 1Gb/s. Could that be it? Could
> you give more detail about the actual traffic patterns and results?
>
> BTW, this is my favorite message title ever. Thanks for that. :)
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
More information about the Gluster-users
mailing list