[Gluster-users] Can't heal a volume: "Please check if all brick processes are running."

Laura Bailey lbailey at redhat.com
Tue Mar 13 23:03:42 UTC 2018


Can we add a smarter error message for this situation by checking volume
type first?

Cheers,
Laura B

On Wednesday, March 14, 2018, Karthik Subrahmanya <ksubrahm at redhat.com>
wrote:

> Hi Anatoliy,
>
> The heal command is basically used to heal any mismatching contents
> between replica copies of the files.
> For the command "gluster volume heal <volname>" to succeed, you should
> have the self-heal-daemon running,
> which is true only if your volume is of type replicate/disperse.
> In your case you have a plain distribute volume where you do not store the
> replica of any files.
> So the volume heal will return you the error.
>
> Regards,
> Karthik
>
> On Tue, Mar 13, 2018 at 7:53 PM, Anatoliy Dmytriyev <tolid at tolid.eu.org>
> wrote:
>
>> Hi,
>>
>>
>> Maybe someone can point me to a documentation or explain this? I can't
>> find it myself.
>> Do we have any other useful resources except doc.gluster.org? As I see
>> many gluster options are not described there or there are no explanation
>> what is doing...
>>
>>
>>
>> On 2018-03-12 15:58, Anatoliy Dmytriyev wrote:
>>
>>> Hello,
>>>
>>> We have a very fresh gluster 3.10.10 installation.
>>> Our volume is created as distributed volume, 9 bricks 96TB in total
>>> (87TB after 10% of gluster disk space reservation)
>>>
>>> For some reasons I can’t “heal” the volume:
>>> # gluster volume heal gv0
>>> Launching heal operation to perform index self heal on volume gv0 has
>>> been unsuccessful on bricks that are down. Please check if all brick
>>> processes are running.
>>>
>>> Which processes should be run on every brick for heal operation?
>>>
>>> # gluster volume status
>>> Status of volume: gv0
>>> Gluster process                             TCP Port  RDMA Port  Online
>>> Pid
>>> ------------------------------------------------------------
>>> ------------------
>>> Brick cn01-ib:/gfs/gv0/brick1/brick         0         49152      Y
>>>  70850
>>> Brick cn02-ib:/gfs/gv0/brick1/brick         0         49152      Y
>>>  102951
>>> Brick cn03-ib:/gfs/gv0/brick1/brick         0         49152      Y
>>>  57535
>>> Brick cn04-ib:/gfs/gv0/brick1/brick         0         49152      Y
>>>  56676
>>> Brick cn05-ib:/gfs/gv0/brick1/brick         0         49152      Y
>>>  56880
>>> Brick cn06-ib:/gfs/gv0/brick1/brick         0         49152      Y
>>>  56889
>>> Brick cn07-ib:/gfs/gv0/brick1/brick         0         49152      Y
>>>  56902
>>> Brick cn08-ib:/gfs/gv0/brick1/brick         0         49152      Y
>>>  94920
>>> Brick cn09-ib:/gfs/gv0/brick1/brick         0         49152      Y
>>>  56542
>>>
>>> Task Status of Volume gv0
>>> ------------------------------------------------------------
>>> ------------------
>>> There are no active volume tasks
>>>
>>>
>>> # gluster volume info gv0
>>> Volume Name: gv0
>>> Type: Distribute
>>> Volume ID: 8becaf78-cf2d-4991-93bf-f2446688154f
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 9
>>> Transport-type: rdma
>>> Bricks:
>>> Brick1: cn01-ib:/gfs/gv0/brick1/brick
>>> Brick2: cn02-ib:/gfs/gv0/brick1/brick
>>> Brick3: cn03-ib:/gfs/gv0/brick1/brick
>>> Brick4: cn04-ib:/gfs/gv0/brick1/brick
>>> Brick5: cn05-ib:/gfs/gv0/brick1/brick
>>> Brick6: cn06-ib:/gfs/gv0/brick1/brick
>>> Brick7: cn07-ib:/gfs/gv0/brick1/brick
>>> Brick8: cn08-ib:/gfs/gv0/brick1/brick
>>> Brick9: cn09-ib:/gfs/gv0/brick1/brick
>>> Options Reconfigured:
>>> client.event-threads: 8
>>> performance.parallel-readdir: on
>>> performance.readdir-ahead: on
>>> cluster.nufa: on
>>> nfs.disable: on
>>>
>>
>> --
>> Best regards,
>> Anatoliy
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>
>
>

-- 
Laura Bailey
Senior Technical Writer
Customer Content Services BNE
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20180314/dfe3ed76/attachment.html>


More information about the Gluster-users mailing list