[Gluster-users] Running Gluster client/server on single process

Tejas N. Bhise tejas at gluster.com
Wed May 19 13:06:49 UTC 2010


Roberto,

We recently made some code changes we think will considerably help small file performance -

selective readdirp - http://patches.gluster.com/patch/3203/
dht lookup revalidation optimization - http://patches.gluster.com/patch/3204/
updated write-behind default values - http://patches.gluster.com/patch/3223/

These are tentatively scheduled to go into 3.0.5.
If its possible for you, I would suggest you test them in a non-production environment
and see if it  helps with distribute config itself.

Please do not use in production, for that wait for the release which these patches go in.

Do let me know if you have any questions about this.

Regards,
Tejas.


----- Original Message -----
From: "Roberto Franchini" <ro.franchini at gmail.com>
To: "gluster-users" <gluster-users at gluster.org>
Sent: Wednesday, May 19, 2010 5:29:47 PM
Subject: Re: [Gluster-users] Running Gluster client/server on single process

On Sat, May 15, 2010 at 10:06 PM, Craig Carl <craig at gluster.com> wrote:
> Robert -
>       NUFA has been deprecated and doesn't apply to any recent version of
> Gluster. What version are you running? ('glusterfs --version')

We run 3.0.4 on ubuntu 9.10 and 10.04 server.
Is there a way to mimic NUFA behaviour?

We are using gluster to store Lucene indexes. Indexes are created
locally from milions of small files and then copied to the storage.
I tried read this little files from gluster but was too slow.
So maybe a NUFA way, e.g. prefer local disk for read, could improve performance.
Let me know :)

At the moment we use dht/replicate:


#CLIENT

volume remote1
 type protocol/client
 option transport-type tcp
 option remote-host zeus
 option remote-subvolume brick
end-volume

volume remote2
 type protocol/client
 option transport-type tcp
 option remote-host hera
 option remote-subvolume brick
end-volume

volume remote3
 type protocol/client
 option transport-type tcp
 option remote-host apollo
 option remote-subvolume brick
end-volume

volume remote4
 type protocol/client
 option transport-type tcp
 option remote-host demetra
 option remote-subvolume brick
end-volume

volume remote5
 type protocol/client
 option transport-type tcp
 option remote-host ade
 option remote-subvolume brick
end-volume

volume remote6
 type protocol/client
 option transport-type tcp
 option remote-host athena
 option remote-subvolume brick
end-volume

volume replicate1
 type cluster/replicate
 subvolumes remote1 remote2
end-volume

volume replicate2
 type cluster/replicate
 subvolumes remote3 remote4
end-volume

volume replicate3
 type cluster/replicate
 subvolumes remote5 remote6
end-volume

volume distribute
 type cluster/distribute
 subvolumes replicate1 replicate2 replicate3
end-volume

volume writebehind
 type performance/write-behind
 option window-size 1MB
 subvolumes distribute
end-volume

volume quickread
 type performance/quick-read
 option cache-timeout 1         # default 1 second
#  option max-file-size 256KB        # default 64Kb
 subvolumes writebehind
end-volume

### Add io-threads for parallel requisitions
volume iothreads
 type performance/io-threads
 option thread-count 16 # default is 16
 subvolumes quickread
end-volume


#SERVER

volume posix
 type storage/posix
 option directory /data/export
end-volume

volume locks
 type features/locks
 subvolumes posix
end-volume

volume brick
 type performance/io-threads
 option thread-count 8
 subvolumes locks
end-volume

volume server
 type protocol/server
 option transport-type tcp
 option auth.addr.brick.allow *
 subvolumes brick
end-volume
-- 
Roberto Franchini
http://www.celi.it
http://www.blogmeter.it
http://www.memesphere.it
Tel +39.011.562.71.15
jabber:ro.franchini at gmail.com skype:ro.franchini
_______________________________________________
Gluster-users mailing list
Gluster-users at gluster.org
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users



More information about the Gluster-users mailing list