[Gluster-users] RAID-1 over network scenario - incredible problems
stas.oskin at gmail.com
Wed Apr 1 08:55:59 UTC 2009
No need to do rsync - running ls -lR on the mount *should* synchronize all
That said, it doesn't always work - I'm currently battling with several
issues on the subject.
2009/4/1 Ondrej Jombik <nepto at platon.sk>
> Than you so much for this reply. We used to have client-side replication
> and failover worked very nice, but we were unable to do auto-restore
> (servers do not know about each other), thus I thought that moving
> tovards server-side replication would be a good idea.
> We will probably move back to client-size and do the rsync-like
> auto-restore after storage node failure.
> Again thanks, that helped me a lot, since now I have at least startpoint
> what to expect and whan not (regarding GlusterFS).
> PS: I like the GlusterFS architecture concept, and config files as well.
> On Wed, 1 Apr 2009, Stas Oskin wrote:
>> 2009/4/1 Ondrej Jombik <nepto at platon.sk>
>> I'm trying to configure GlusterFS setup with two replicating servers.
>> For now just without any client. Worked well so far, however after
>> I rebooted the second server I started to have difficult times...
>> (note: first server remains unrebooted)
>> Advice - use client side replication, it's considered more reliable and
>> supports fail-over.
>> 1. are all changes made on non-rebooted server during the second
>> reboot lost? they are not replicated after rebooted server is
>> again... is there OFFICIAL way how to acheive this? Does it have
>> binary log of non-performed write operations?
>> You only need to run ls -lR when the server comes up to sync between the
>> files. Now, only if it always worked...
>> 6. option favorite-child remote
>> This does nothing as well, but at least print some warning into the
>> log files. However what is written in the warning actually does not
>> happen. I tried to access all files on remote/local
>> device/mountpoint (4 ways), no change at all.
>> Happens to me to.
>> 7. if I define "subvolumes remote" (so kicking local from subvolumes)
>> than I finally get the right file contents, but only at mountpoint,
>> not in actual device; I need to get files into the actual device
>> (local disk) of rebooted server
>> Happen to me to.
>> 8. and finally I deleted all the files from device of rebooted server
>> and I was hopping for the replication to do the rest; and viola,
>> I have them replicated, so all files created during reboot are
>> but they are all filed with zeros!
>> (and no this is not that known XFS bug, it is actually on EXT3)
>> Welcome to the club :).
>> I know this all is pretty incredible and looks like a horror story,
>> I have read tons of documentation and still I'm not able to figure
>> out. I wish that it is problem between keyboard and chair and not in
>> software itself.
>> I'm only trying to have RAID-1 over network with automatic recovery
>> after reboot/outage. Is this that complicated??
>> (I need to metion that I did not started with clients yet, there I'm
>> expecting even bigger troubles like this)
>> As I said, perhaps you should try client-side AFR first?
> /\ Ondrej Jombik - nepto at platon.sk - http://nepto.sk - ICQ #122428216
> //\\ Platon Group - open source software development - http://platon.sk
> //\\ 10 types of people: those who understand binary & those who do not
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the Gluster-users