[Gluster-users] 3.0.5: server on tmpfs for small files

Jeremy Enos jenos at ncsa.uiuc.edu
Sat Nov 20 02:14:34 UTC 2010


btw- I'm using Gluster 3.0.5 and centos55 x86_64 in this case.  (updated 
subject line too)

     Jeremy

On 11/19/2010 7:27 PM, Craig Carl wrote:
> Jeremy -
>    We use ramfs to test with here, it should work. If you want to 
> troubleshoot that issue please let me know.
>
> Craig
>
>
> On 11/19/2010 05:12 PM, Jeremy Enos wrote:
>> Looks like RAMFS has the same issue TMPFS does...  I'm looking into 
>> the RNA networks.  Thanks-
>>
>>     Jeremy
>>
>> On 11/18/2010 6:55 PM, Craig Carl wrote:
>>> On 11/18/2010 04:33 PM, Jeremy Enos wrote:
>>>> Post is almost a year old...  ever any response here?  Is it 
>>>> possible to export tmpfs locations w/ gluster?
>>>> thx-
>>>>
>>>>     Jeremy
>>>>
>>>> On 12/1/2009 8:14 AM, Alexander Beregalov wrote:
>>>>> Hi
>>>>>
>>>>> Is it possible to start server on tmpfs ?
>>>>> It is announced that stripe can be used over tmpfs, but stripe is
>>>>> client plugin and server cannot start on tmpfs because lack of
>>>>> xattrs.
>>>>>
>>>>> I am trying to setup a small fast storage for small files 
>>>>> (compiling purpose).
>>>>> I made ext2 with xattr on ramdisk on 4 hosts, joined them with
>>>>> replicate plugin and mounted it on one client. Also io-cache,
>>>>> write-behind, quick-read and io threads  were used on client side.
>>>>> I compiled linux kernel, performance was 10 times worse than tmpfs
>>>>> exported by NFS on one node.
>>>>>
>>>>> Any ideas?
>>>>>
>>>>> Servers:
>>>>>
>>>>> volume posix
>>>>>    type storage/posix                   # POSIX FS translator
>>>>>    option directory /mnt/ost        # Export this directory
>>>>> end-volume
>>>>>
>>>>> volume locks
>>>>>    type features/locks
>>>>>    option mandatory-locks on
>>>>>    subvolumes posix
>>>>> end-volume
>>>>>
>>>>> volume brick
>>>>>    type performance/io-threads
>>>>>    option thread-count 4                # Four CPUs
>>>>>    subvolumes locks
>>>>> end-volume
>>>>>
>>>>> volume server
>>>>>    type protocol/server
>>>>>    option transport-type tcp
>>>>>    option transport.socket.nodelay on
>>>>>    subvolumes brick
>>>>>    option auth.addr.brick.allow * # Allow access to "brick" volume
>>>>> end-volume
>>>>>
>>>>>
>>>>> Client:
>>>>>
>>>>> volume server1
>>>>>    type protocol/client
>>>>>    option transport-type tcp
>>>>>    option remote-host<IP>
>>>>>    option transport.socket.nodelay on
>>>>>    option remote-subvolume brick        # name of the remote volume
>>>>> end-volume
>>>>>
>>>>> <the same for server[2-4]>
>>>>>
>>>>> volume replicated
>>>>>    type cluster/replicate
>>>>>    subvolumes server1 server2 server3 server4
>>>>> end-volume
>>>>>
>>>>> volume iocache
>>>>>    type performance/io-cache
>>>>>    option cache-size 1000MB        # default is 32MB
>>>>>    option priority *.h:3,*.o:2,*:1     # default is '*:0'
>>>>>    option cache-timeout 1                 # default is 1 second
>>>>>    subvolumes replicated
>>>>> end-volume
>>>>>
>>>>> volume writeback
>>>>>    type performance/write-behind
>>>>>    option cache-size 500MB     # default is equal to aggregate-size
>>>>>    option flush-behind off      # default is 'off'
>>>>>    subvolumes iocache
>>>>> end-volume
>>>>>
>>>>> volume quickread
>>>>>    type performance/quick-read
>>>>>    option cache-timeout 1         # default 1 second
>>>>>    option max-file-size 256KB        # default 64Kb
>>>>>    subvolumes iocache
>>>>> end-volume
>>>>>
>>>>> volume iothreads
>>>>>    type performance/io-threads
>>>>>    option thread-count 16 # default is 16
>>>>>    subvolumes quickread
>>>>> end-volume
>>>>> _______________________________________________
>>>>> Gluster-users mailing list
>>>>> Gluster-users at gluster.org
>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>
>>> Jeremy -
>>>    xattrs are required for any Gluster setup regardless of the 
>>> volume design, tmpfs doesn't support user xattrs. RamFS works well, 
>>> if you have the budget FisionIO cards are very fast and work well 
>>> with Gluster, so does the solution from RNA Networks. 
>>> (http://www.rnanetworks.com/)
>>>
>>> Thanks,
>>>
>>> Craig
>>>
>>> -->
>>> Craig Carl
>>> Senior Systems Engineer
>>> Gluster
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>
>



More information about the Gluster-users mailing list