[Gluster-users] Continual heals happening on cluster

Lindsay Mathieson lindsay.mathieson at gmail.com
Sun Apr 17 04:12:19 UTC 2016


gluster 3.7.10
Proxmox (debian jessie)

I'm finding the following more than a little concerning. I've created a 
datastore with the following settings:

    Volume Name: datastore4
    Type: Replicate
    Volume ID: 0ba131ef-311d-4bb1-be46-596e83b2f6ce
    Status: Started
    Number of Bricks: 1 x 3 = 3
    Transport-type: tcp
    Bricks:
    Brick1: vnb.proxmox.softlog:/tank/vmdata/datastore4
    Brick2: vng.proxmox.softlog:/tank/vmdata/datastore4
    Brick3: vna.proxmox.softlog:/tank/vmdata/datastore4
    Options Reconfigured:
    features.shard-block-size: 64MB
    network.remote-dio: enable
    cluster.eager-lock: enable
    performance.io-cache: off
    performance.read-ahead: off
    performance.quick-read: off
    performance.stat-prefetch: on
    performance.strict-write-ordering: off
    nfs.enable-ino32: off
    nfs.addr-namelookup: off
    nfs.disable: on
    cluster.server-quorum-type: server
    cluster.quorum-type: auto
    features.shard: on
    cluster.data-self-heal: on
    cluster.self-heal-window-size: 1024
    transport.address-family: inet
    performance.readdir-ahead: on



I've transferred 12 Windows VM's to it (gfapi) and am running them all, 
spread across three nodes.

"gluster volume heal datastore3 statistics heal-count" shows zero heals 
on all nodes.

but "gluster volume heal datastore4 info" shows heals occurring on 
mutliple shards on all nodes, different shards each time its called.

    gluster volume heal datastore4 info
    Brick vnb.proxmox.softlog:/tank/vmdata/datastore4
    /.shard/d297f8d6-e263-4af3-9384-6492614dc115.221
    /.shard/744c5059-303d-4e82-b5be-0a5f53b1aeff.1362
    /.shard/bbdff876-290a-4e5e-93ef-a95276d57220.942
    /.shard/eaeb41ec-9c0d-4fed-984f-cf832d8d33e0.1032
    /.shard/f8ce4b49-14d0-46ef-9a95-456884f34fd4.623
    /.shard/e9a39d2e-a1b7-4ea0-9d8c-b55370048d03.483
    /.shard/f8ce4b49-14d0-46ef-9a95-456884f34fd4.47
    /.shard/eaeb41ec-9c0d-4fed-984f-cf832d8d33e0.160
    Status: Connected
    Number of entries: 8

    Brick vng.proxmox.softlog:/tank/vmdata/datastore4
    /.shard/bd493985-2ee6-43f1-b8d5-5f0d5d3aa6f5.33
    /.shard/d297f8d6-e263-4af3-9384-6492614dc115.48
    /.shard/744c5059-303d-4e82-b5be-0a5f53b1aeff.1304
    /.shard/d297f8d6-e263-4af3-9384-6492614dc115.47
    /.shard/719041d0-d755-4bc6-a5fc-6b59071fac17.142
    Status: Connected
    Number of entries: 5

    Brick vna.proxmox.softlog:/tank/vmdata/datastore4
    /.shard/d297f8d6-e263-4af3-9384-6492614dc115.357
    /.shard/bbdff876-290a-4e5e-93ef-a95276d57220.996
    /.shard/d297f8d6-e263-4af3-9384-6492614dc115.679
    /.shard/d297f8d6-e263-4af3-9384-6492614dc115.496
    /.shard/eaeb41ec-9c0d-4fed-984f-cf832d8d33e0.160
    /.shard/719041d0-d755-4bc6-a5fc-6b59071fac17.954
    /.shard/d297f8d6-e263-4af3-9384-6492614dc115.678
    /.shard/719041d0-d755-4bc6-a5fc-6b59071fac17.852
    /.shard/bbdff876-290a-4e5e-93ef-a95276d57220.1544
    Status: Connected
    Number of entries: 9




-- 
Lindsay Mathieson

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160417/203cdefb/attachment.html>


More information about the Gluster-users mailing list