[Gluster-users] broken gluster config

Thing thing.thing at gmail.com
Thu May 10 01:22:14 UTC 2018


Whatever repair happened has now finished but I still have this,

I cant find anything so far telling me how to fix it.  Looking at

http://staged-gluster-docs.readthedocs.io/en/release3.7.0beta1/Features/heal-info-and-split-brain-resolution/

I cant determine what file? dir gvo? is actually the issue.

[root at glusterp1 gv0]# gluster volume heal gv0 info split-brain
Brick glusterp1:/bricks/brick1/gv0
<gfid:eafb8799-4e7a-4264-9213-26997c5a4693>
Status: Connected
Number of entries in split-brain: 1

Brick glusterp2:/bricks/brick1/gv0
<gfid:eafb8799-4e7a-4264-9213-26997c5a4693>
Status: Connected
Number of entries in split-brain: 1

Brick glusterp3:/bricks/brick1/gv0
<gfid:eafb8799-4e7a-4264-9213-26997c5a4693>
Status: Connected
Number of entries in split-brain: 1

[root at glusterp1 gv0]#


On 10 May 2018 at 12:22, Thing <thing.thing at gmail.com> wrote:

> also I have this "split brain"?
>
> [root at glusterp1 gv0]# gluster volume heal gv0 info
> Brick glusterp1:/bricks/brick1/gv0
> <gfid:eafb8799-4e7a-4264-9213-26997c5a4693> - Is in split-brain
>
> Status: Connected
> Number of entries: 1
>
> Brick glusterp2:/bricks/brick1/gv0
> <gfid:eafb8799-4e7a-4264-9213-26997c5a4693> - Is in split-brain
>
> /glusterp1/images/centos-server-001.qcow2
> /glusterp1/images/kubernetes-template.qcow2
> /glusterp1/images/kworker01.qcow2
> /glusterp1/images/kworker02.qcow2
> Status: Connected
> Number of entries: 5
>
> Brick glusterp3:/bricks/brick1/gv0
> <gfid:eafb8799-4e7a-4264-9213-26997c5a4693> - Is in split-brain
>
> /glusterp1/images/centos-server-001.qcow2
> /glusterp1/images/kubernetes-template.qcow2
> /glusterp1/images/kworker01.qcow2
> /glusterp1/images/kworker02.qcow2
> Status: Connected
> Number of entries: 5
>
> [root at glusterp1 gv0]#
>
> On 10 May 2018 at 12:20, Thing <thing.thing at gmail.com> wrote:
>
>> [root at glusterp1 gv0]# !737
>> gluster v status
>> Status of volume: gv0
>> Gluster process                             TCP Port  RDMA Port  Online
>> Pid
>> ------------------------------------------------------------
>> ------------------
>> Brick glusterp1:/bricks/brick1/gv0          49152     0          Y
>>  5229
>> Brick glusterp2:/bricks/brick1/gv0          49152     0          Y
>>  2054
>> Brick glusterp3:/bricks/brick1/gv0          49152     0          Y
>>  2110
>> Self-heal Daemon on localhost               N/A       N/A        Y
>>  5219
>> Self-heal Daemon on glusterp2               N/A       N/A        Y
>>  1943
>> Self-heal Daemon on glusterp3               N/A       N/A        Y
>>  2067
>>
>> Task Status of Volume gv0
>> ------------------------------------------------------------
>> ------------------
>> There are no active volume tasks
>>
>> [root at glusterp1 gv0]# ls -l glusterp1/images/
>> total 2877064
>> -rw-------. 2 root root 107390828544 May 10 12:18 centos-server-001.qcow2
>> -rw-r--r--. 2 root root            0 May  8 14:32 file1
>> -rw-r--r--. 2 root root            0 May  9 14:41 file1-1
>> -rw-------. 2 root root  85912715264 May 10 12:18
>> kubernetes-template.qcow2
>> -rw-------. 2 root root            0 May 10 12:08 kworker01.qcow2
>> -rw-------. 2 root root            0 May 10 12:08 kworker02.qcow2
>> [root at glusterp1 gv0]#
>>
>>
>> while,
>>
>> [root at glusterp2 gv0]# ls -l glusterp1/images/
>> total 11209084
>> -rw-------. 2 root root 107390828544 May  9 14:45 centos-server-001.qcow2
>> -rw-r--r--. 2 root root            0 May  8 14:32 file1
>> -rw-r--r--. 2 root root            0 May  9 14:41 file1-1
>> -rw-------. 2 root root  85912715264 May  9 15:59
>> kubernetes-template.qcow2
>> -rw-------. 2 root root   3792371712 May  9 16:15 kworker01.qcow2
>> -rw-------. 2 root root   3792371712 May 10 11:20 kworker02.qcow2
>> [root at glusterp2 gv0]#
>>
>> So some files have re-synced but not the kworker machines   network
>> activity has stopped.
>>
>>
>>
>> On 10 May 2018 at 12:05, Diego Remolina <dijuremo at gmail.com> wrote:
>>
>>> Show us output from: gluster v status
>>>
>>> It should be easy to fix. Stop gluster daemon on that node, mount the
>>> brick, start gluster daemon again.
>>>
>>> Check: gluster v status
>>>
>>> Does it show the brick up?
>>>
>>> HTH,
>>>
>>> Diego
>>>
>>>
>>> On Wed, May 9, 2018, 20:01 Thing <thing.thing at gmail.com> wrote:
>>>
>>>> Hi,
>>>>
>>>> I have 3 Centos7.4 machines setup as a 3 way raid 1.
>>>>
>>>> Due to an oopsie on my part for glusterp1 /bricks/brick1/gv0 didnt
>>>> mount on boot and as a result its empty.
>>>>
>>>> Meanwhile I have data on glusterp2 /bricks/brick1/gv0  and glusterp3
>>>> /bricks/brick1/gv0 as expected.
>>>>
>>>> Is there a way to get glusterp1's gv0 to sync off the other 2? there
>>>> must be but,
>>>>
>>>> I have looked at the gluster docs and I cant find anything about
>>>> repairing  resyncing?
>>>>
>>>> Where am I meant to look for such info?
>>>>
>>>> thanks
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>>
>>>
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>>
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20180510/4083ba1d/attachment.html>


More information about the Gluster-users mailing list