[Gluster-users] Expected behaviour of hypervisor on Gluster node loss
Vijay Bellur
vbellur at redhat.com
Wed Feb 1 04:35:15 UTC 2017
On Mon, Jan 30, 2017 at 6:25 AM, Niklaus Hofer <
niklaus.hofer at stepping-stone.ch> wrote:
> Hi
>
> I have a question concerning the 'correct' behaviour of GlusterFS:
>
> We a nice Gluster setup up and running. Most things are working nicely.
> Our setup is as follows:
> - Storage is a 2+1 Gluster setup (2 replicating hosts + 1 arbiter) with a
> volume for virtual machines.
> - Two virtualisation hosts running libvirt / qemu / kvm.
>
>
Are you using something like oVirt or proxmox for managing your
virtualization cluster?
> Now the question is, what is supposed to happen when we unplug one of the
> storage nodes (aka power outage in one of our data centers)?
> Initially we were hoping that the virtualisation hosts would automatically
> switch over to the second storage node and keep all VMs running.
>
> However, during our tests, we have found that this is not the case.
> Instead, when we unplug one of the storage nodes, the virtual machines run
> into all sorts of problems; being unable to read/write, crashing
> applications and even corrupting the filesystem. That is of course not
> acceptable.
>
> Reading the documentation again, we now think that we have misunderstood
> what we're supposed to be doing. To our understanding, what should happen
> is this:
> - If the virtualisation host is connected to the storage node which is
> still running:
> - everything is fine and the VM keeps running
> - If the virtualisation host was connected to the storage node which is
> now absent:
> - qemu is supposed to 'pause' / 'freeze' the VM
> - Virtualisation host waits for ping timeout
> - Virtualisation host switches over to the other storage node
> - qemu 'unpauses' the VMs
> - The VM is fully operational again
>
> Does my description match the 'optimal' GlusterFS behaviour?
>
Can you provide more details about your gluster volume configuration and
the options enabled on the volume?
Regards,
Vijay
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170131/13d7c069/attachment.html>
More information about the Gluster-users
mailing list