[Gluster-users] heal info OK but statistics not working

lejeczek peljasz at yahoo.co.uk
Mon Sep 4 13:54:22 UTC 2017

1) one peer, out of four, got separated from the network, 
from the rest of the cluster.
2) that unavailable(while it was unavailable) peer got 
detached with "gluster peer detach" command which succeeded, 
so now cluster comprise of three peers
3) Self-heal daemon (for some reason) does not start(with an 
attempt to restart glusted) on the peer which probed that 
fourth peer.
4) fourth unavailable peer is still up & running but is 
inaccessible to other peers for network is disconnected, 
segmented. That peer's gluster status show peer is still in 
the cluster.
5) So, fourth peer's gluster(nor other processes) stack did 
not fail nor crushed, just network got, is disconnected.
6) peer status show ok & connected for current three peers.

This is third time when it happens to me, very same way: 
each time net-disjointed peer was brought back online then 
statistics & details worked again.

can you not reproduce it?

$ gluster vol info QEMU-VMs

Volume Name: QEMU-VMs
Type: Replicate
Volume ID: 8709782a-daa5-4434-a816-c4e0aef8fef2
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Options Reconfigured:
transport.address-family: inet
nfs.disable: on
storage.owner-gid: 107
storage.owner-uid: 107
performance.readdir-ahead: on
geo-replication.indexing: on
geo-replication.ignore-pid-check: on
changelog.changelog: on

$ gluster vol status QEMU-VMs
Status of volume: QEMU-VMs
Gluster process                             TCP Port  RDMA 
Port Online  Pid
TERs/0GLUSTER-QEMU-VMs                      49156     0 
Y       9302
TERs/0GLUSTER-QEMU-VMs                      49156     0 
Y       7610
STERs/0GLUSTER-QEMU-VMs                     49156     0 
Y       11013
Self-heal Daemon on localhost               N/A       N/A 
Y       3069276
Self-heal Daemon on               N/A       N/A 
Y       3315870
Self-heal Daemon on               N/A       N/A 
N       N/A  <--- HERE
Self-heal Daemon on               N/A       N/A 
Y       5163

Task Status of Volume QEMU-VMs
There are no active volume tasks

$ gluster vol heal QEMU-VMs statistics heal-count
Gathering count of entries to be healed on volume QEMU-VMs 
has been unsuccessful on bricks that are down. Please check 
if all brick processes are running.

On 04/09/17 11:47, Atin Mukherjee wrote:
> Please provide the output of gluster volume info, gluster 
> volume status and gluster peer status.
> On Mon, Sep 4, 2017 at 4:07 PM, lejeczek 
> <peljasz at yahoo.co.uk <mailto:peljasz at yahoo.co.uk>> wrote:
>     hi all
>     this:
>     $ vol heal $_vol info
>     outputs ok and exit code is 0
>     But if I want to see statistics:
>     $ gluster vol heal $_vol statistics
>     Gathering crawl statistics on volume GROUP-WORK has
>     been unsuccessful on bricks that are down. Please
>     check if all brick processes are running.
>     I suspect - gluster inability to cope with a situation
>     where one peer(which is not even a brick for a single
>     vol on the cluster!) is inaccessible to the rest of
>     cluster.
>     I have not played with any other variations of this
>     case, eg. more than one peer goes down, etc.
>     But I hope someone could try to replicate this simple
>     test case.
>     Cluster and vols, when something like this happens,
>     seem accessible and as such "all" works, except when
>     you want more details.
>     This also fails:
>     $ gluster vol status $_vol detail
>     Error : Request timed out
>     My gluster(3.10.5-1.el7.x86_64) exhibits these
>     symptoms every time one(at least) peers goes out of
>     the rest reach.
>     maybe @devel can comment?
>     many thanks, L.
>     _______________________________________________
>     Gluster-users mailing list
>     Gluster-users at gluster.org
>     <mailto:Gluster-users at gluster.org>
>     http://lists.gluster.org/mailman/listinfo/gluster-users
>     <http://lists.gluster.org/mailman/listinfo/gluster-users>

More information about the Gluster-users mailing list