[Gluster-users] 2.0.8: server on tmpfs for small files

Jeremy Enos jenos at ncsa.uiuc.edu
Sat Nov 20 02:13:44 UTC 2010


I'm definitely interested in getting ramfs to work directly (instead of 
the hacky loopback on top of it tmpfs).  I'll check again to be sure, 
but I'm pretty sure I got the same errors I did w/ tmpfs about xattrs 
when I tried to export ramfs.
thx-

     Jeremy

On 11/19/2010 7:27 PM, Craig Carl wrote:
> Jeremy -
>    We use ramfs to test with here, it should work. If you want to 
> troubleshoot that issue please let me know.
>
> Craig
>
>
> On 11/19/2010 05:12 PM, Jeremy Enos wrote:
>> Looks like RAMFS has the same issue TMPFS does...  I'm looking into 
>> the RNA networks.  Thanks-
>>
>>     Jeremy
>>
>> On 11/18/2010 6:55 PM, Craig Carl wrote:
>>> On 11/18/2010 04:33 PM, Jeremy Enos wrote:
>>>> Post is almost a year old...  ever any response here?  Is it 
>>>> possible to export tmpfs locations w/ gluster?
>>>> thx-
>>>>
>>>>     Jeremy
>>>>
>>>> On 12/1/2009 8:14 AM, Alexander Beregalov wrote:
>>>>> Hi
>>>>>
>>>>> Is it possible to start server on tmpfs ?
>>>>> It is announced that stripe can be used over tmpfs, but stripe is
>>>>> client plugin and server cannot start on tmpfs because lack of
>>>>> xattrs.
>>>>>
>>>>> I am trying to setup a small fast storage for small files 
>>>>> (compiling purpose).
>>>>> I made ext2 with xattr on ramdisk on 4 hosts, joined them with
>>>>> replicate plugin and mounted it on one client. Also io-cache,
>>>>> write-behind, quick-read and io threads  were used on client side.
>>>>> I compiled linux kernel, performance was 10 times worse than tmpfs
>>>>> exported by NFS on one node.
>>>>>
>>>>> Any ideas?
>>>>>
>>>>> Servers:
>>>>>
>>>>> volume posix
>>>>>    type storage/posix                   # POSIX FS translator
>>>>>    option directory /mnt/ost        # Export this directory
>>>>> end-volume
>>>>>
>>>>> volume locks
>>>>>    type features/locks
>>>>>    option mandatory-locks on
>>>>>    subvolumes posix
>>>>> end-volume
>>>>>
>>>>> volume brick
>>>>>    type performance/io-threads
>>>>>    option thread-count 4                # Four CPUs
>>>>>    subvolumes locks
>>>>> end-volume
>>>>>
>>>>> volume server
>>>>>    type protocol/server
>>>>>    option transport-type tcp
>>>>>    option transport.socket.nodelay on
>>>>>    subvolumes brick
>>>>>    option auth.addr.brick.allow * # Allow access to "brick" volume
>>>>> end-volume
>>>>>
>>>>>
>>>>> Client:
>>>>>
>>>>> volume server1
>>>>>    type protocol/client
>>>>>    option transport-type tcp
>>>>>    option remote-host<IP>
>>>>>    option transport.socket.nodelay on
>>>>>    option remote-subvolume brick        # name of the remote volume
>>>>> end-volume
>>>>>
>>>>> <the same for server[2-4]>
>>>>>
>>>>> volume replicated
>>>>>    type cluster/replicate
>>>>>    subvolumes server1 server2 server3 server4
>>>>> end-volume
>>>>>
>>>>> volume iocache
>>>>>    type performance/io-cache
>>>>>    option cache-size 1000MB        # default is 32MB
>>>>>    option priority *.h:3,*.o:2,*:1     # default is '*:0'
>>>>>    option cache-timeout 1                 # default is 1 second
>>>>>    subvolumes replicated
>>>>> end-volume
>>>>>
>>>>> volume writeback
>>>>>    type performance/write-behind
>>>>>    option cache-size 500MB     # default is equal to aggregate-size
>>>>>    option flush-behind off      # default is 'off'
>>>>>    subvolumes iocache
>>>>> end-volume
>>>>>
>>>>> volume quickread
>>>>>    type performance/quick-read
>>>>>    option cache-timeout 1         # default 1 second
>>>>>    option max-file-size 256KB        # default 64Kb
>>>>>    subvolumes iocache
>>>>> end-volume
>>>>>
>>>>> volume iothreads
>>>>>    type performance/io-threads
>>>>>    option thread-count 16 # default is 16
>>>>>    subvolumes quickread
>>>>> end-volume
>>>>> _______________________________________________
>>>>> Gluster-users mailing list
>>>>> Gluster-users at gluster.org
>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>
>>> Jeremy -
>>>    xattrs are required for any Gluster setup regardless of the 
>>> volume design, tmpfs doesn't support user xattrs. RamFS works well, 
>>> if you have the budget FisionIO cards are very fast and work well 
>>> with Gluster, so does the solution from RNA Networks. 
>>> (http://www.rnanetworks.com/)
>>>
>>> Thanks,
>>>
>>> Craig
>>>
>>> -->
>>> Craig Carl
>>> Senior Systems Engineer
>>> Gluster
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>
>



More information about the Gluster-users mailing list