[Gluster-users] Unable To Heal On 3.7 Branch With Arbiter

Kyle Harris kyle.harris98 at gmail.com
Thu Jan 7 22:38:34 UTC 2016


Hello,



I have a rather odd situation I’m hoping someone can help me out with.  I
have a 2 node gluster replica with an arbiter running on the 3.7 branch.  I
don’t appear to have a split-brain yet I am unable able to heal the cluster
after a power failure.  Perhaps someone can tell me how to fix this?  I
don't see much in the logs that may help.  ’Here is some command output
that might be helpful:



gluster volume info gv0:

Volume Name: gv0

Type: Replicate

Volume ID: 14e7bb9c-aa5e-4386-8dd2-83a88d93dc54

Status: Started

Number of Bricks: 1 x 3 = 3

Transport-type: tcp

Bricks:

Brick1: server1:/export/brick1

Brick2: server2:/export/brick1

Brick3: kvm:/export/brick1

Options Reconfigured:

nfs.acl: off

performance.readdir-ahead: on

performance.quick-read: off

performance.read-ahead: off

performance.io-cache: off

performance.stat-prefetch: off

cluster.eager-lock: enable

network.remote-dio: enable


---

gluster volume status gv0 detail:

Status of volume: gv0

------------------------------------------------------------------------------

Brick                : Brick server1:/export/brick1

TCP Port             : 49152

RDMA Port            : 0

Online               : Y

Pid                  : 4409

File System          : ext3

Device               : /dev/sdb1

Mount Options        : rw

Inode Size           : 128

Disk Space Free      : 1.7TB

Total Disk Space     : 1.8TB

Inode Count          : 244203520

Free Inodes          : 244203413

------------------------------------------------------------------------------

Brick                : Brick server2:/export/brick1

TCP Port             : 49152

RDMA Port            : 0

Online               : Y

Pid                  : 4535

File System          : ext3

Device               : /dev/sdb1

Mount Options        : rw

Inode Size           : 128

Disk Space Free      : 1.7TB

Total Disk Space     : 1.8TB

Inode Count          : 244203520

Free Inodes          : 244203405


---

Why doesn’t this accomplish anything?

gluster volume heal gv0:

Launching heal operation to perform index self heal on volume gv0 has been
successful

Use heal info commands to check status


---

Or this?

gluster volume heal gv0 full:

Launching heal operation to perform full self heal on volume gv0 has been
successful

Use heal info commands to check status


---

gluster volume heal gv0 info split-brain:

Brick server1:/export/brick1

Number of entries in split-brain: 0



Brick server2:/export/brick1

Number of entries in split-brain: 0



Brick kvm:/export/brick1

Status: Transport endpoint is not connected



---

I can't seem to get these to heal?

gluster volume heal gv0 info:

Brick server1:/export/brick1

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/8c524ed9-e382-40cd-9361-60c23a2c1ae2.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/0b16f938-e859-41e3-bb33-fefba749a578.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/715ddb6c-67af-4047-9fa0-728019b49d63.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/b0cdf43c-7e6b-44bf-ab2d-efb14e9d2156.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/d2873b74-f6be-43a9-bdf1-276761e3e228.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/asdf

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/940ee016-8288-4369-9fb8-9c64cb3af256.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/72a33878-59f7-4f6e-b3e1-e137aeb19ced.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/03070877-9cf4-4d55-a66c-fbd3538eedb9.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/c2645723-efd9-474b-8cce-fe07ac9fbba9.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/930196aa-0b85-4482-97ab-3d05e9928884.vhd

Number of entries: 12



Brick server2:/export/brick1

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/03070877-9cf4-4d55-a66c-fbd3538eedb9.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/d2873b74-f6be-43a9-bdf1-276761e3e228.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/715ddb6c-67af-4047-9fa0-728019b49d63.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/930196aa-0b85-4482-97ab-3d05e9928884.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/c2645723-efd9-474b-8cce-fe07ac9fbba9.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/asdf

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/940ee016-8288-4369-9fb8-9c64cb3af256.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/0b16f938-e859-41e3-bb33-fefba749a578.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/8c524ed9-e382-40cd-9361-60c23a2c1ae2.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/72a33878-59f7-4f6e-b3e1-e137aeb19ced.vhd

/4b37411d-97cd-0d4c-f898-a3b93cfe1b34/b0cdf43c-7e6b-44bf-ab2d-efb14e9d2156.vhd

Number of entries: 12



Brick kvm:/export/brick1

Status: Transport endpoint is not connected

---

Thank you.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160107/9383f9c2/attachment.html>


More information about the Gluster-users mailing list