[Gluster-users] 3.0.5: server on tmpfs for small files
Jeremy Enos
jenos at ncsa.uiuc.edu
Sat Nov 20 02:14:34 UTC 2010
btw- I'm using Gluster 3.0.5 and centos55 x86_64 in this case. (updated
subject line too)
Jeremy
On 11/19/2010 7:27 PM, Craig Carl wrote:
> Jeremy -
> We use ramfs to test with here, it should work. If you want to
> troubleshoot that issue please let me know.
>
> Craig
>
>
> On 11/19/2010 05:12 PM, Jeremy Enos wrote:
>> Looks like RAMFS has the same issue TMPFS does... I'm looking into
>> the RNA networks. Thanks-
>>
>> Jeremy
>>
>> On 11/18/2010 6:55 PM, Craig Carl wrote:
>>> On 11/18/2010 04:33 PM, Jeremy Enos wrote:
>>>> Post is almost a year old... ever any response here? Is it
>>>> possible to export tmpfs locations w/ gluster?
>>>> thx-
>>>>
>>>> Jeremy
>>>>
>>>> On 12/1/2009 8:14 AM, Alexander Beregalov wrote:
>>>>> Hi
>>>>>
>>>>> Is it possible to start server on tmpfs ?
>>>>> It is announced that stripe can be used over tmpfs, but stripe is
>>>>> client plugin and server cannot start on tmpfs because lack of
>>>>> xattrs.
>>>>>
>>>>> I am trying to setup a small fast storage for small files
>>>>> (compiling purpose).
>>>>> I made ext2 with xattr on ramdisk on 4 hosts, joined them with
>>>>> replicate plugin and mounted it on one client. Also io-cache,
>>>>> write-behind, quick-read and io threads were used on client side.
>>>>> I compiled linux kernel, performance was 10 times worse than tmpfs
>>>>> exported by NFS on one node.
>>>>>
>>>>> Any ideas?
>>>>>
>>>>> Servers:
>>>>>
>>>>> volume posix
>>>>> type storage/posix # POSIX FS translator
>>>>> option directory /mnt/ost # Export this directory
>>>>> end-volume
>>>>>
>>>>> volume locks
>>>>> type features/locks
>>>>> option mandatory-locks on
>>>>> subvolumes posix
>>>>> end-volume
>>>>>
>>>>> volume brick
>>>>> type performance/io-threads
>>>>> option thread-count 4 # Four CPUs
>>>>> subvolumes locks
>>>>> end-volume
>>>>>
>>>>> volume server
>>>>> type protocol/server
>>>>> option transport-type tcp
>>>>> option transport.socket.nodelay on
>>>>> subvolumes brick
>>>>> option auth.addr.brick.allow * # Allow access to "brick" volume
>>>>> end-volume
>>>>>
>>>>>
>>>>> Client:
>>>>>
>>>>> volume server1
>>>>> type protocol/client
>>>>> option transport-type tcp
>>>>> option remote-host<IP>
>>>>> option transport.socket.nodelay on
>>>>> option remote-subvolume brick # name of the remote volume
>>>>> end-volume
>>>>>
>>>>> <the same for server[2-4]>
>>>>>
>>>>> volume replicated
>>>>> type cluster/replicate
>>>>> subvolumes server1 server2 server3 server4
>>>>> end-volume
>>>>>
>>>>> volume iocache
>>>>> type performance/io-cache
>>>>> option cache-size 1000MB # default is 32MB
>>>>> option priority *.h:3,*.o:2,*:1 # default is '*:0'
>>>>> option cache-timeout 1 # default is 1 second
>>>>> subvolumes replicated
>>>>> end-volume
>>>>>
>>>>> volume writeback
>>>>> type performance/write-behind
>>>>> option cache-size 500MB # default is equal to aggregate-size
>>>>> option flush-behind off # default is 'off'
>>>>> subvolumes iocache
>>>>> end-volume
>>>>>
>>>>> volume quickread
>>>>> type performance/quick-read
>>>>> option cache-timeout 1 # default 1 second
>>>>> option max-file-size 256KB # default 64Kb
>>>>> subvolumes iocache
>>>>> end-volume
>>>>>
>>>>> volume iothreads
>>>>> type performance/io-threads
>>>>> option thread-count 16 # default is 16
>>>>> subvolumes quickread
>>>>> end-volume
>>>>> _______________________________________________
>>>>> Gluster-users mailing list
>>>>> Gluster-users at gluster.org
>>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>
>>> Jeremy -
>>> xattrs are required for any Gluster setup regardless of the
>>> volume design, tmpfs doesn't support user xattrs. RamFS works well,
>>> if you have the budget FisionIO cards are very fast and work well
>>> with Gluster, so does the solution from RNA Networks.
>>> (http://www.rnanetworks.com/)
>>>
>>> Thanks,
>>>
>>> Craig
>>>
>>> -->
>>> Craig Carl
>>> Senior Systems Engineer
>>> Gluster
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>>
>
More information about the Gluster-users
mailing list