[Gluster-users] File Corruption when adding bricks to live replica volumes

Lindsay Mathieson lindsay.mathieson at gmail.com
Tue Jan 19 11:24:07 UTC 2016


gluster 3.7.6

I seem to be able to reliably reproduce this. I have a replica 2 volume 
with 1 test VM image. While the VM is  running with heavy disk 
read/writes  (disk benchmark) I add a 3rd brick for replica 3:

gluster volume add-brick datastore1 replica 3 
vng.proxmox.softlog:/vmdata/datastore1

I pretty much immediately get this:

    gluster volume heal datastore1 info
    Brick vna.proxmox.softlog:/vmdata/datastore1
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.20
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.22
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.55 - Possibly
    undergoing heal

    /images/301/vm-301-disk-1.qcow2 - Possibly undergoing heal

    Number of entries: 4

    Brick vnb.proxmox.softlog:/vmdata/datastore1
    /images/301/vm-301-disk-1.qcow2 - Possibly undergoing heal

    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.55 - Possibly
    undergoing heal

    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.20
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.22
    Number of entries: 4

    Brick vng.proxmox.softlog:/vmdata/datastore1
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.16
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.28
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.1
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.22
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.77
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.9
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.5
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.2
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.26
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.15
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.13
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.3
    /.shard/d6aad699-d71d-4b35-b021-d35e5ff297c4.18
    Number of entries: 13


The brick on vng is the new empty brick, but it has 13 shards being 
healed back to vna & vnb. That can't be right and if I leave it the VM 
becomes hopelessly corrupted. Also there are 81 shards in the files, 
they should all be queued for healing.

Additionally I get read errors when I run a qemu-img check on the VM 
image. If I remove the vng brick the problems are resolved.


If I do the same process while the VM is not running - i.e no files are 
being access, every proceeds as expect. All shard on vn & vnb are healed 
to vng,

-- 
Lindsay Mathieson

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160119/37126601/attachment.html>


More information about the Gluster-users mailing list