[Gluster-devel] GlusterFS Example

DeeDee Park deedee6905 at hotmail.com
Tue Jul 31 02:57:34 UTC 2007


>Client unifies S1 and S2 (call him U1)
>Client unifies S3 and S4 (call him U2)
>Client then AFR's together U1 and U2 and S5
>
>In this example, if I save a file and its *:3, it will appear....
>* on U1 (either S1 or S2 depending on scheduler)
>* on U2 (either S3 or S4 depending on sched.)
>* on S5

Is that supported? I was trying to do something similiar earlier, but was 
told that
AFR is under(is a subvolume of) unify since I was having too many problems. 
Can I create
a couple unifies, and then have both U1 and U2 as subvolumes of AFR? If so, 
let me
know, and I can test it.

Can I do something like the following?
BrickS1-1
BrickS1-2
Unify US1
   subvolumes S1-1,S1-2
BrickS2-1
BrickS2-2
Unify US2
   subvolumes S2-1,S2-2
AFR
  subvolumes US1, US2
  replicate *:2

>From: Matt Paine <matt at mattsoftware.com>
>To: gluster-devel at nongnu.org
>Subject: Re: [Gluster-devel] GlusterFS Example
>Date: Tue, 31 Jul 2007 09:15:32 +1000
>
>
>>>
>>>option replicate *.tmp:1,*.html:4,*.php:4,*.db:1,*:2
>>>
>>>In this case any .tmp or .db files will not get replicated (will sit on
>>>the first afr brick only), all html and php files will be mirrored
>>>across the first four bricks of afr, and all other files will be
>>>mirrorred across the first two bricks. (note all on one line, note comma
>>>delimited, note also no spaces before or after the comma).
>>
>>
>>Right, but is it possible to just tell the config you want to store 2
>>copies of everything without telling each system where the redundant
>>copies should be stored? I.E. I am looking for something like RAID 6,
>>where I have say 8 bricks with two copies of every file. If I just enter
>>*:2 then the spare copies will just be on the first two bricks and since
>>each 8 brick has the same amount of storage I will run out of space on the
>>first two before the others vs the redundant copies distributed over the 8
>>bricks.
>
>
>Absolutly, except its there is a bit more work to do. AFR works by starting 
>at the first brick and any replicating goes to the second, third etc as 
>described above. But these are normal gluster bricks, so there is nothing 
>stopping you from creating an AFR for the first brick, and AFR for the 
>second, and just a posix brick for the rest (if thats what you want of 
>course). Let me try to explain.....
>
>
>Server one has one brick: S1
>Server two has one brick: S2
>Server three has one brick: S3
>Server four has one brick: S4
>Server five has one brick: S5
>
>Client unifies S1 and S2 (call him U1)
>Client unifies S3 and S4 (call him U2)
>Client then AFR's together U1 and U2 and S5
>
>In this example, if I save a file and its *:3, it will appear....
>* on U1 (either S1 or S2 depending on scheduler)
>* on U2 (either S3 or S4 depending on sched.)
>* on S5
>
>If I save a file as only *:1, it will appear only on U1 (either S1 or S2 
>depending on scheduler).
>
>Ad nausium.
>
>
>Of coures there is nothing stopping you from unifying three bricks, or even 
>unifying an afr to afr.
>
>
>i.e. (might need a mono spaced font to see correctly...)
>
>                         A5
>             +-----------------------+
>             |                       |
>             U1                      U2
>      +-------------+           +---------+
>      |             |           |         |
>      A1            A2          A3        A4
>  +---+---+   +---+---+---+   +---+   +---+---+
>  |   |   |   |   |   |   |   |   |   |   |   |
>S01 S02 S03 S04 S05 S06 S07 S08 S09 S10 S11 S12
>
>
>Where Sxx = Server bricks
>        Ax = AFR brick
>        Ux = Unify brick
>
>
>
>So in this configuration (which you have already worked out i'm sure) that 
>if you save something *:2, then it will appear in both U1 and U2, which 
>means (depending on the spec from a[1-4], assume *:2) it will appear in 
>either A1 or A2 (because of the unify), AND it will also appear in either 
>A3 or A4. etc etc etc.....
>
>I think i've laboured the point far enough :)
>
>
>
>
>>
>>
>>>>4) In my application the server is also the client, in my current config
>>>>when one out of 3 servers is done I can no longer write, even tho I am
>>>>telling my config to write to the local brick. Is there any way to fix
>>>>this? If a note is down I understand how I would not have access to the
>>>>files, but I don't want to take down the full cluster.
>>>
>>>I'm note sure sorry. Maybe one of the devs can clarify that. I know
>>>there was an issue with an earlier relase to do with AFR. When the first
>>>brick was down then afr was no longer available.
>>
>>
>>Ah... That could be it.
>
>Which version of glusterfs are you using? tla, pre ?
>
>(that issue has been fixed for a little while now, so if your using pre6 
>you shoulnd't have come across it)
>
>
>
>
>
>Matt.
>
>
>
>_______________________________________________
>Gluster-devel mailing list
>Gluster-devel at nongnu.org
>http://lists.nongnu.org/mailman/listinfo/gluster-devel

_________________________________________________________________
http://newlivehotmail.com






More information about the Gluster-devel mailing list