[Gluster-users] KVM guest I/O errors with xfs backed gluster volumes

Jacob Yundt jyundt at gmail.com
Tue Jul 16 11:30:11 UTC 2013


> I'm using gluster 3.3.0 and 3.3.1 with xfs bricks and kvm based VM's
> using qcow2 files on gluster volume fuse mounts. CentOS6.2 through 6.4
> w/CloudStack 3.0.2 - 4.1.0.
>
> I've not had any problems. Here is 1 host in a small 3 host cluster
> (using the cloudstack terminology). about 30 VM's are running across
> these 3 hosts - which all contribute to the volume with 2 bricks each.
> I'll also attach a virsh dumpxml for you to take a look at.
>
> [root ~]# w
>  06:21:53 up 320 days,  7:23,  1 user,  load average: 1.41, 1.07, 0.79
> USER     TTY      FROM              LOGIN@   IDLE   JCPU   PCPU WHAT
> root     pts/9    10.100.0.100     06:21    0.00s  0.00s  0.00s w
> [root ~]# cat /etc/redhat-release
> CentOS release 6.3 (Final)
> [root ~]# rpm -qa | grep gluster
> glusterfs-server-3.3.0-1.el6.x86_64
> glusterfs-fuse-3.3.0-1.el6.x86_64
> glusterfs-3.3.0-1.el6.x86_64
> [root@ ~]# cat /etc/fstab | grep glust
> /dev/storage/glust0    /gluster/0        xfs    defaults,inode64 0 0
> /dev/storage/glust1    /gluster/1        xfs    defaults,inode64 0 0
> 172.16.0.11:qcow2-share /gluster/qcow2        glusterfs    defaults,_netdev 0 0
> [root at cs0.la.vorstack.net ~]# df -h
> [cut.....]
> /dev/mapper/storage-glust0
>                       2.0T  217G  1.8T  11% /gluster/0
> /dev/mapper/storage-glust1
>                       2.0T  148G  1.9T   8% /gluster/1
> 172.16.0.11:qcow2-share
>                       6.0T  472G  5.6T   8% /gluster/qcow2
> [root@ ~]# virsh list
>  Id    Name                           State
> ----------------------------------------------------
>  10    i-2-19-VM                      running
>  21    i-3-44-VM                      running
>  22    i-2-12-VM                      running
>  28    i-4-58-VM                      running
>  37    s-5-VM                         running
>  38    v-2-VM                         running
>  39    i-2-56-VM                      running
>  41    i-7-59-VM                      running
>  46    i-4-87-VM                      running
> [root@ ~]# gluster volume info
>
> Volume Name: qcow2-share
> Type: Distributed-Replicate
> Volume ID: 22fcbaa9-4b2d-4d84-9353-eb77abcaf0db
> Status: Started
> Number of Bricks: 3 x 2 = 6
> Transport-type: tcp
> Bricks:
> Brick1: 172.16.0.10:/gluster/0
> Brick2: 172.16.0.11:/gluster/0
> Brick3: 172.16.0.12:/gluster/0
> Brick4: 172.16.0.10:/gluster/1
> Brick5: 172.16.0.11:/gluster/1
> Brick6: 172.16.0.12:/gluster/1
> [root@ ~]# gluster volume status
> Status of volume: qcow2-share
> Gluster process                        Port    Online    Pid
> ------------------------------------------------------------------------------
> Brick 172.16.0.10:/gluster/0                24009    Y    1873
> Brick 172.16.0.11:/gluster/0                24009    Y    1831
> Brick 172.16.0.12:/gluster/0                24009    Y    1938
> Brick 172.16.0.10:/gluster/1                24010    Y    1878
> Brick 172.16.0.11:/gluster/1                24010    Y    1837
> Brick 172.16.0.12:/gluster/1                24010    Y    1953
> NFS Server on localhost                    38467    Y    1899
> Self-heal Daemon on localhost                N/A    Y    1909
> NFS Server on 172.16.0.12                38467    Y    1959
> Self-heal Daemon on 172.16.0.12                N/A    Y    1964
> NFS Server on 172.16.0.11                38467    Y    1843
> Self-heal Daemon on 172.16.0.11                N/A    Y    1848
>
>

This information (including the attached xml) is very helpful, thank
you!  Can you provide the xml of your (KVM) gluster storage pool:
"virsh pool-dumpxml <pool>"

Have you ever tried using "raw" virtio disk images?  When I try to use
qcow2, I get errors when trying to start my VM:

qemu-kvm: -drive
file=/var/lib/libvirt/images/xfs/kvm2.img,if=none,id=drive-virtio-disk1,format=qcow2,cache=none:
could not open disk image /var/lib/libvirt/images/xfs/kvm2.img:
Invalid argument

-Jacob



More information about the Gluster-users mailing list