[Gluster-users] mysql replication between two nodes
Richard de Vries
rdevries1000 at gmail.com
Fri Oct 15 12:26:41 UTC 2010
The mysql database only runs on one node at a time.
I still find it hard to understand why you need to restart the service
if a brick goes down and comes back again.
this is the volume file I'm using
glusterfsd.vol server file:
volume posix1
type storage/posix
option directory /export/database
end-volume
volume locks1
type features/locks
subvolumes posix1
end-volume
volume database
type performance/io-threads
option thread-count 8
subvolumes locks1
end-volume
volume server
type protocol/server
option transport-type tcp/server
option auth.addr.database.allow *
option transport.socket.listen-port 6996
option transport.socket.nodelay on
subvolumes database
end-volume
database.vol file:
volume databasenode1
type protocol/client
option transport-type tcp
option transport.socket.nodelay on
option remote-port 6996
option ping-timeout 2
option remote-host node1
option remote-subvolume database
end-volume
volume databasenode2
type protocol/client
option transport-type tcp
option transport.socket.nodelay on
option remote-port 6996
option ping-timeout 2
option remote-host node2
option remote-subvolume database
end-volume
volume replicate
type cluster/replicate
subvolumes databasenode1 databasenode2
end-volume
volume stat-performance
type performance/stat-prefetch
subvolumes replicate
end-volume
Maybe the stat-performance translator has influence on the this stat output.
I'll try to disable this and test again.
Regards,
Richard
On Fri, Oct 15, 2010 at 12:44 PM, Deadpan110 <deadpan110 at gmail.com> wrote:
> I am very new to this list, but here is my 2 cents...
>
> In the past I used DRBD between 2 nodes to provide a master/slave
> setup with mySQL data stored on the filesystem.
>
> Upon a failover situation, mySQL would start on the remaining server
> and pick up where things left off.
>
> DRBD (8.0 +) now supports master/master but it would be unwise to run
> mySQL on such a setup live on 2 servers.
>
> mySQL has also advanced and replication is not restricted to master/slave.
>
> I use (and am loving) glusterfs in various guises on my 3 node cluster
> for my client filesystems.
>
> For mySQL I use master/master/master circular replication without
> depending on any type of clustered filesystem (only local on each
> node) - there have been people frowning on such a setup, but things
> have advanced with the latest stable mySQL versions and as such, I
> have been successfully using it in a clustered environment.
>
> Martin
>
> On 15 October 2010 20:33, Richard de Vries <rdevries1000 at gmail.com> wrote:
>> Hello Beat,
>>
>> This is a pitty. Because a stop of the service only to resync te
>> standby node is not so nice...
>>
>> The stat of the database file in: /opt/test/database after a reboot of
>> the node 2 shows different output,
>> one time from the node 1 and another time from node 2.
>>
>> What is the role of self heal in this? It is noticed that the files
>> are not equal (via stat).
>>
>> Would you see the same behaviour for example with qemu-kvm that keeps
>> also files open?
>>
>> Regards,
>> Richard
More information about the Gluster-users
mailing list