[Gluster-devel] Choice of Translator question
Krishna Srinivas
krishna at zresearch.com
Fri Dec 28 19:34:33 UTC 2007
On Dec 29, 2007 12:08 AM, Gareth Bult <gareth at encryptec.net> wrote:
> Ok, a little more testing .. wait, it gets better!
>
> I now have a x10 stripe.
>
> 10 stripes on node a - 10 AFR's - 1 x10 Stripe
> 10 stripes on node b - (for self heal) (to divide the heal chunk size)
>
> Sample;
>
> a. create 500M file
> b. take down one glusterfsd process
> c. append 2M to file
> d. bring glusterfsd back up
> e. head -c1 on file
>
> Problem #1;
>
> On a self-heal, it does a self-heal on every stripe, regardless of the fact I only appended 2M.
It is a bug then. Only 2 afrs should have done self heal, the ones where
two chunks of 1MB of 2MBs were written to. How did you confirm all the
10 afrs did selfheal?
>
> Problem #2;
>
> Self-heal ignored the fact the file is sparse and copies the entire sripe ..
> so for a 500M file, the healing process actually copies 5Gb!!
I need to check how it behaves on holes as no one had complained on this bug
before.
Thanks
Krishna
>
> Help! Bug!
>
> Here's a client config summary; (server condig is fairly obvious)
> ...
> volume stripes-stripe
> type cluster/stripe
> subvolumes afr-1 afr-2 afr-3 afr-4 afr-5 afr-6 afr-7 afr-8 afr-9 afr-10
> option block-size *:1MB
> end-volume
> ...
> volume afr-1
> type cluster/afr
> subvolumes node1A node1B
> option replicate *:2
> option scheduler rr
> end-volume
> ...
> volume node1A
> type protocol/client
> option transport-type tcp/client
> option remote-host nodea
> option remote-subvolume stripes-1A
> end-volume
> volume node1B
> type protocol/client
> option transport-type tcp/client
> option remote-host nodez
> option remote-subvolume stripes-1B
> end-volume
>
>
>
>
>
> ----- Original Message -----
> From: "Csibra Gergo" <gergo at csibra.hu>
> To: "Gareth Bult" <gareth at encryptec.net>
> Sent: Friday, December 28, 2007 3:47:03 PM (GMT) Europe/London
> Subject: Re: [Gluster-devel] Choice of Translator question
>
> Friday, December 28, 2007, 3:57:52 PM, Gareth Bult wrote:
>
> >>Oh. I don't understand this, can you explain why need to change configs regularly?
>
> > To add new systems,
>
> Add new systems to...? To the server? Adding new clients?
>
> > install newer versions of fuse and glusterfs,
>
> This and the upper is not a reason to rebuild/remirror afr-ed files.
> If a new version of glusterfs or fuse comes out or need to add new
> systems to anywhere in glusterfs system, you make a shutdown (unmount
> , kill glusterfsd (this make actually a regular sgutdown)) install new
> version and restart the glusterfs. The xattrs (they contains version
> informations) of mirrored files will be the same before and after
> version changes, so they will not remirrored.
>
> > to recover from fuse kernel lockups ..
>
> Yes. If this happens in write, yes. In this situation AFR-ed files
> need to be healed.
>
> > .. I've not yet seen raid/afr comments from anyone who actually
> > understands the problem, so I'm not likely to see a fix (?!)
>
> I understand this problem, but this because of bugs. Peoples in Z
> research working hard to fix them, and you gives lot information if
> you gives bugreports :)
>
>
> --
> Best regards,
> Csibra Gergo mailto:gergo at csibra.hu
>
>
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at nongnu.org
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>
More information about the Gluster-devel
mailing list