[Gluster-users] File Corruption with shards - 100% reproducable
Lindsay Mathieson
lindsay.mathieson at gmail.com
Fri Nov 13 06:27:15 UTC 2015
On 12 November 2015 at 15:46, Krutika Dhananjay <kdhananj at redhat.com> wrote:
> OK. What do the client logs say?
>
Dumb question - Which logs are those?
Could you share the exact steps to recreate this, and I will try it locally
> on my setup?
>
I'm running this on a 3 node proxmox cluster, which makes the vm creation &
migration easy to test.
Steps:
- Create 3 node gluster datastore using proxmox vm host nodes
- Add gluster datastore as a storage dvice to proxmox
* qemu vms use the gfapi to access the datastore
* proxmox also adds a fuse mount for easy acces
- create a VM on the gluster storage, QCOW2 format. I just created a simple
debain Mate vm
- start the vm, open a console to it.
- live migrate the VM to a another node
- It will rapdily barf itself with disk errors
- stop the VM
- qemu will show file corruption (many many errors)
* qemu-img check <vm disk image>
* qemu-img info <vm disk image>
Repeating the process with sharding off has no errors.
>
> Also, want to see the output of 'gluster volume info'.
>
I've trimmed settings down to a bare minimum. This is a test gluster
cluster so I can do with it as I wish.
gluster volume info
Volume Name: datastore1
Type: Replicate
Volume ID: 238fddd0-a88c-4edb-8ac5-ef87c58682bf
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: vnb.proxmox.softlog:/mnt/ext4
Brick2: vng.proxmox.softlog:/mnt/ext4
Brick3: vna.proxmox.softlog:/mnt/ext4
Options Reconfigured:
performance.strict-write-ordering: on
performance.readdir-ahead: off
cluster.quorum-type: auto
features.shard: on
--
Lindsay
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20151113/d0c9fcbf/attachment.html>
More information about the Gluster-users
mailing list