[Gluster-devel] Question about compile performance over GlusterFS

Craig Tierney Craig.Tierney at noaa.gov
Mon Mar 17 15:11:03 UTC 2008


Amar S. Tumballi wrote:
> Hi Craig,
>  I will be looking into this issue. Btw, is there any reason you are not
> using ib-verbs? instead using ib-sdp?
>  Let me get back to you regarding this. Give me few days.
> 


For streaming bandwidth, verbs and ib-sdp did not provide any different
results.  However, I didn't really think about it after getting it setup.  I
will re-test with ib-verbs today.

Craig


> Regards,
> Amar
> 
> On Thu, Mar 13, 2008 at 8:41 AM, Craig Tierney <Craig.Tierney at noaa.gov>
> wrote:
> 
>> Amar S. Tumballi wrote:
>>> Hi Craig,
>>>  Thanks for a nice comparison  between GlusterFS and other network file
>>> systems. But sure, before concluding about the performance, I would
>> suggest
>>> few improvement to your GlusterFS setup.
>>>
>>> 1. Try with client/protocol instead of having unify with only one
>> subvolume.
>>> (Unify makes sense when you have more than one subvolume, but when there
>> is
>>> only one subvolume, its a extra layer which may count as overhead),
>> below
>>> io-thread volume.
>>>
>>> 2. in io-cache on client side, as during kernel compile lot of *.h files
>> are
>>> re-read, you can give preference to *h files only.
>>>
>>> volume ioc
>>>   type performance/io-cache
>>>   subvolumes wb
>>>   option priority *.h:100
>>> end-volume
>>>
>> I changed the io-cache settings to those above and eliminated the use of
>> the Unify
>> subvolume (my scripts generate server/client configs automatically, and in
>> most
>> cases multiple servers are used, in this case they weren't).    The
>> compile time
>> went down, but not by much.  The latest test finished in 1042 seconds.
>>
>> What I didn't test this time is the compile directly on the storage that
>> is exported
>> by Gluster.  The runtime there is 399 seconds, so the underlying
>> filesystem is fast.
>>
>> I am not making any conclusions about the performance based on these
>> numbers.
>> Things are going great so far, and this should be a solveable problem
>> based on the
>> other performance characteristics I have seen.
>>
>> Craig
>>
>>
>>
>>> Regards,
>>> Amar
>>>
>>> On Wed, Mar 12, 2008 at 3:55 PM, Craig Tierney <Craig.Tierney at noaa.gov>
>>> wrote:
>>>
>>>> I have been testing out my GlusterFS setup.  I have been
>>>> very happy with the streaming IO performance and scalability.
>>>> We have some users on the system now and they are seeing
>>>> very good performance (fast and consistent) as compared
>>>> to our other filesystem.
>>>>
>>>> I have a test that I created that tries to measure metadata
>>>> performance by building the linux kernel.  What I have
>>>> found is that GlusterFS is slower than local disk, NFS,
>>>> and Panasas.  The compile time on those three systems
>>>> is roughly 500 seconds.  For GlusterFS (1.3.7), the
>>>> compile time is roughly 1200 seconds.  My GlusterFS filesystem
>>>> is using ramdisks on the servers and communicating using
>>>> IB-Verbs.  My server and client configs are below.
>>>>
>>>> Note I did not implement both write-behind and not read-behind
>>>> based on some benchmarks I saw on the list on how it affects
>>>> re-write.
>>>>
>>>> So, is this just because mmap isn't (yet) supported in FUSE?
>>>> Or, is there something else I should be looking at.
>>>>
>>>> Thanks,
>>>> Craig
>>>>
>>>>
>>>> server.cfg
>>>> ----------
>>>>
>>>> volume brick
>>>>   type storage/posix                   # POSIX FS translator
>>>>   option directory /tmp/scratch/export        # Export this directory
>>>> end-volume
>>>>
>>>> volume server
>>>>   type protocol/server
>>>>   subvolumes brick
>>>>   option transport-type ib-sdp/server     # For TCP/IP transport
>>>>   option auth.ip.brick.allow *
>>>> end-volume
>>>>
>>>> client.cfgvolume client-ns
>>>>   type protocol/client
>>>>   option transport-type ib-sdp/client
>>>>   option remote-host w8-ib0
>>>>   option remote-subvolume brick-ns
>>>> end-volume
>>>>
>>>>
>>>>
>>>> volume client-w8
>>>>   type protocol/client
>>>>   option transport-type ib-sdp/client
>>>>   option remote-host w8-ib0
>>>>   option remote-subvolume brick
>>>> end-volume
>>>>
>>>> volume unify
>>>>         type cluster/unify
>>>>         subvolumes  client-w8
>>>>         option namespace client-ns
>>>>         option scheduler rr
>>>> end-volume
>>>>
>>>> volume iot
>>>>         type performance/io-threads
>>>>         subvolumes unify
>>>>         option thread-count 4
>>>> end-volume
>>>>
>>>> volume wb
>>>>         type performance/write-behind
>>>>         subvolumes iot
>>>> end-volume
>>>>
>>>> volume ioc
>>>>         type performance/io-cache
>>>>         subvolumes wb
>>>> end-volume
>>>>
>>>> ----------
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Craig Tierney (craig.tierney at noaa.gov)
>>>>
>>>>
>>>
>>
>> --
>> Craig Tierney (craig.tierney at noaa.gov)
>>
>>
> 
> 


-- 
Craig Tierney (craig.tierney at noaa.gov)





More information about the Gluster-devel mailing list