[Gluster-users] gluster and kvm livemigration
Samuli Heinonen
samppah at neutraali.net
Fri Jan 17 15:45:20 UTC 2014
Hello Bernhard,
Can you test if setting option network.remote-dio to enable allows you to use cache=none?
-samuli
Bernhard Glomm <bernhard.glomm at ecologic.eu> kirjoitti 17.1.2014 kello 16.41:
> Pranith,
> I stopped the volume
> started it again,
> mounted it on both hosts
> started the VM
> did the livemigration
> and collected the logs:
> - etc-glusterfs-glusterd.vol.log
> - glustershd.log
> - srv-vm_infrastructure-vm-atom01.log
> - cli.log
> from the beginning of the gluster volume start.
> You can found them here (part 1 to 3):
> http://pastebin.com/mnATm2BE
> http://pastebin.com/RYZFP3E9
> http://pastebin.com/HAXEGd54
>
> further more:
> gluster --version: glusterfs 3.4.2 built on Jan 11 2014 03:21:47
> ubuntu: raring
> filesystem on the gluster bricks: zfs-0.6.2
>
> gluster volume info fs_vm_atom01
> Volume Name: fs_vm_atom01
> Type: Replicate
> Volume ID: fea9bdcf-783e-442a-831d-f564f8dbe551
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: 172.24.1.11:/zp_ping_1/fs_vm_atom01
> Brick2: 172.24.1.13:/zp_pong_1/fs_vm_atom01
> Options Reconfigured:
> diagnostics.client-log-level: DEBUG
> server.allow-insecure: on
>
> disk part of VM configuration
>
> <emulator>/usr/bin/kvm-spice</emulator>
> <disk type='file' device='disk'>
> <driver name='qemu' type='raw' cache='writethrough'/>
> <source file='/srv/vm_infrastructure/vm/atom01/atom01.img'/>
> <target dev='vda' bus='virtio'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x04' function='0x0'/>
> </disk>
>
> can't use <source protocol='gluster' ...
> as josh suggested because couldn't get
> my qemu recompiled with gluster enabled yet.
>
> Are there other special tuning parameter for kvm/qemu/ to set on gluster?
> as mentioned: all works except the livemigration (disk image file becomes read only)
> and I have to use something different than cache=none...
>
> TIA
>
> Bernhard
>
>
> On 17.01.2014 05:04:52, Pranith Kumar Karampuri wrote:
> Bernhard,
> Configuration seems ok. Could you please give the log files of the bricks and mount please. If you think it is not a big procedure to do this live migration, could you set client-log-level to DEBUG and provide the log files of that run.
>
> Pranith
>
> ----- Original Message -----
> From: "Bernhard Glomm" <bernhard.glomm at ecologic.eu>
> To: pkarampu at redhat.com
> Cc: gluster-users at gluster.org
> Sent: Thursday, January 16, 2014 5:58:17 PM
> Subject: Re: [Gluster-users] gluster and kvm livemigration
>
>
> hi Pranith
>
> # gluster volume info fs_vm_atom01
>
> Volume Name: fs_vm_atom01
> Type: Replicate
> Volume ID: fea9bdcf-783e-442a-831d-f564f8dbe551
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: 172.24.1.11:/zp_ping_1/fs_vm_atom01
> Brick2: 172.24.1.13:/zp_pong_1/fs_vm_atom01
> Options Reconfigured:
> diagnostics.client-log-level: ERROR
>
>
> TIA
> Bernhard
>
>
> On 16.01.2014 13:05:12, Pranith Kumar Karampuri wrote:
> hi Bernhard,
> Could you give gluster volume info output?
>
> Pranith
>
> ----- Original Message -----
> From: "Bernhard Glomm" <> > bernhard.glomm at ecologic.eu> > >
> To: > > gluster-users at gluster.org
> Sent: Thursday, January 16, 2014 4:22:36 PM
> Subject: [Gluster-users] gluster and kvm livemigration
>
> I experienced a strange behavior of glusterfs during livemigration
> of a qemu-kvm guest
> using a 10GB file on a mirrored gluster 3.4.2 volume
> (both on ubuntu 13.04)
> I run
> virsh migrate --verbose --live --unsafe --p2p --domain atom01 --desturi
> qemu+ssh://<target_ip>/system
> and the migration works,
> the running machine is pingable and keeps sending pings.
> nevertheless, when I let the machine touch a file during migration
> it stops, complaining that it's filesystem is read only (from that moment
> that
> migration finished)
> A reboot from inside the machine failes,
> machine goes down and comes up with an error
> unable to write to sector xxxxxx on hd0
> (than falling into the initrd).
> a
> virsh destroy VM && virsh start VM
> leads to a perfect running VM again,
> no matter on which of the two hosts I start the machine
> anybody better experience with livemigration?
> any hint on a procedure how to debug that?
> TIA
> Bernhard
>
> --
>
>
> Bernhard Glomm
> IT Administration
>
> Phone: +49 (30) 86880 134
> Fax: +49 (30) 86880 100
> Skype: bernhard.glomm.ecologic
>
> Ecologic Institut gemeinnützige GmbH | Pfalzburger Str. 43/44 | 10717
> Berlin
> | Germany
> GF: R. Andreas Kraemer | AG: Charlottenburg HRB 57947 | USt/VAT-IdNr.:
> DE811963464
> Ecologic™ is a Trade Mark (TM) of Ecologic Institut gemeinnützige GmbH
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
> --
>
>
>
>
>
>
>
>
>
> Bernhard Glomm
>
> IT Administration
>
>
>
>
> Phone:
>
>
> +49 (30) 86880 134
>
>
> Fax:
>
>
> +49 (30) 86880 100
>
>
> Skype:
>
>
> bernhard.glomm.ecologic
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> Ecologic Institut gemeinnützige GmbH | Pfalzburger Str. 43/44 |
> 10717 Berlin | Germany
>
> GF: R. Andreas Kraemer | AG: Charlottenburg HRB 57947 |
> USt/VAT-IdNr.: DE811963464
>
> Ecologic™ is a Trade Mark (TM) of Ecologic Institut gemeinnützige
> GmbH
>
>
>
>
>
>
>
>
>
> --
> Bernhard Glomm
> IT Administration
>
> Phone: +49 (30) 86880 134
> Fax: +49 (30) 86880 100
> Skype: bernhard.glomm.ecologic
>
> Ecologic Institut gemeinnützige GmbH | Pfalzburger Str. 43/44 | 10717 Berlin | Germany
> GF: R. Andreas Kraemer | AG: Charlottenburg HRB 57947 | USt/VAT-IdNr.: DE811963464
> Ecologic™ is a Trade Mark (TM) of Ecologic Institut gemeinnützige GmbH
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20140117/73f16f94/attachment.html>
More information about the Gluster-users
mailing list