[Gluster-devel] clustered afr

Tibor Veres tibor.veres at gmail.com
Mon Mar 12 14:15:41 UTC 2007


i'm trying to build a 3-node storage cluster which should be able to
withstand 1 node going down.
first I tried glusterfs 1.3.0-pre2.2, but had some memory leakage
which seems to be fixed in the source checked from the repository

i'm exporting 3 bricks with this configs like this:
volume brick[1-3]
  type storage/posix
  option directory /mnt/export/shared/[1-3]
end-volume
volume server
  type protocol/server
  option transport-type tcp/server     # For TCP/IP transport
 option listen-port 699[6-8]              # Default is 6996
  subvolumes brick[1-3]
  option auth.ip.brick[1-3].allow * # Allow access to "brick" volume
end-volume

my client config looks like this:
volume b[1-3]
  type protocol/client
  option transport-type tcp/client     # for TCP/IP transport
  option remote-host 127.0.0.1         # IP address of the remote brick
 option remote-port 699[6-8]              # default server port is 6996
  option remote-subvolume brick[1-3]        # name of the remote volume
end-volume
volume afr
    type cluster/afr
    subvolumes b1 b2 b3
    option replicate *:2
    option scheduler rr
    option rr.limits.min-free-disk 512MB
    option rr.refresh-interval 10
end-volume

i didnt activate any performance-enhance translators.

This setup sort of works, except that i saw files created only on
bricks1 and 2, brick3 got only the directories and symlinks created on
it. After killing the brick2 glusterfsd, the filesystem stayed up,
which is promising, but still no files are created on brick3.

is this setup supposed to work? can i get comparable functionality set
up with current glusterfs? preferably in a way that can be extended to
5 nodes, withstanding 2 going down. is there any plan for some
raid6-like functionality, or this would kill performance alltogether?


-- 
Tibor Veres
  tibor.veres at gmail.com





More information about the Gluster-devel mailing list