[Gluster-users] Gluster 3.7.6 add new node state Peer Rejected (Connected)
Mohammed Rafi K C
rkavunga at redhat.com
Thu Feb 25 20:23:04 UTC 2016
On 02/26/2016 01:32 AM, Steve Dainard wrote:
> I haven't done anything more than peer thus far, so I'm a bit confused
> as to how the volume info fits in, can you expand on this a bit?
>
> Failed commits? Is this split brain on the replica volumes? I don't
> get any return from 'gluster volume heal <volname> info' on all the
> replica volumes, but if I try a gluster volume heal <volname> full I
> get: 'Launching heal operation to perform full self heal on volume
> <volname> has been unsuccessful'.
forget about this. it is not for metadata selfheal .
>
> I have 5 volumes total.
>
> 'Replica 3' volumes running on gluster01/02/03:
> vm-storage
> iso-storage
> export-domain-storage
> env-modules
>
> And one distributed only volume 'storage' info shown below:
>
> *From existing host gluster01/02:*
> type=0
> count=4
> status=1
> sub_count=0
> stripe_count=1
> replica_count=1
> disperse_count=0
> redundancy_count=0
> version=25
> transport-type=0
> volume-id=26d355cb-c486-481f-ac16-e25390e73775
> username=eb9e2063-6ba8-4d16-a54f-2c7cf7740c4c
> password=
> op-version=3
> client-op-version=3
> quota-version=1
> parent_volname=N/A
> restored_from_snap=00000000-0000-0000-0000-000000000000
> snap-max-hard-limit=256
> features.quota-deem-statfs=on
> features.inode-quota=on
> diagnostics.brick-log-level=WARNING
> features.quota=on
> performance.readdir-ahead=on
> performance.cache-size=1GB
> performance.stat-prefetch=on
> brick-0=10.0.231.50:-mnt-raid6-storage-storage
> brick-1=10.0.231.51:-mnt-raid6-storage-storage
> brick-2=10.0.231.52:-mnt-raid6-storage-storage
> brick-3=10.0.231.53:-mnt-raid6-storage-storage
>
> *From existing host gluster03/04:*
> type=0
> count=4
> status=1
> sub_count=0
> stripe_count=1
> replica_count=1
> disperse_count=0
> redundancy_count=0
> version=25
> transport-type=0
> volume-id=26d355cb-c486-481f-ac16-e25390e73775
> username=eb9e2063-6ba8-4d16-a54f-2c7cf7740c4c
> password=
> op-version=3
> client-op-version=3
> quota-version=1
> parent_volname=N/A
> restored_from_snap=00000000-0000-0000-0000-000000000000
> snap-max-hard-limit=256
> features.quota-deem-statfs=on
> features.inode-quota=on
> performance.stat-prefetch=on
> performance.cache-size=1GB
> performance.readdir-ahead=on
> features.quota=on
> diagnostics.brick-log-level=WARNING
> brick-0=10.0.231.50:-mnt-raid6-storage-storage
> brick-1=10.0.231.51:-mnt-raid6-storage-storage
> brick-2=10.0.231.52:-mnt-raid6-storage-storage
> brick-3=10.0.231.53:-mnt-raid6-storage-storage
>
> So far between gluster01/02 and gluster03/04 the configs are the same,
> although the ordering is different for some of the features.
>
> On gluster05/06 the ordering is different again, and the
> quota-version=0 instead of 1.
This is why the peer shows as rejected. Can you check the op-version of
all the glusterd including the one which is in reject state. you can
find out the op-version here in /var/lib/glusterd/glusterd.info
Rafi KC
>
> *From new hosts gluster05/gluster06:*
> type=0
> count=4
> status=1
> sub_count=0
> stripe_count=1
> replica_count=1
> disperse_count=0
> redundancy_count=0
> version=25
> transport-type=0
> volume-id=26d355cb-c486-481f-ac16-e25390e73775
> username=eb9e2063-6ba8-4d16-a54f-2c7cf7740c4c
> password=
> op-version=3
> client-op-version=3
> quota-version=0
> parent_volname=N/A
> restored_from_snap=00000000-0000-0000-0000-000000000000
> snap-max-hard-limit=256
> performance.stat-prefetch=on
> performance.cache-size=1GB
> performance.readdir-ahead=on
> features.quota=on
> diagnostics.brick-log-level=WARNING
> features.inode-quota=on
> features.quota-deem-statfs=on
> brick-0=10.0.231.50:-mnt-raid6-storage-storage
> brick-1=10.0.231.51:-mnt-raid6-storage-storage
> brick-2=10.0.231.52:-mnt-raid6-storage-storage
> brick-3=10.0.231.53:-mnt-raid6-storage-storage
>
> Also, I forgot to mention that when I initially peer'd the two new
> hosts, glusterd crashed on gluster03 and had to be restarted (log
> attached) but has been fine since.
>
> Thanks,
> Steve
>
> On Thu, Feb 25, 2016 at 11:27 AM, Mohammed Rafi K C
> <rkavunga at redhat.com <mailto:rkavunga at redhat.com>> wrote:
>
>
>
> On 02/25/2016 11:45 PM, Steve Dainard wrote:
>> Hello,
>>
>> I upgraded from 3.6.6 to 3.7.6 a couple weeks ago. I just peered
>> 2 new nodes to a 4 node cluster and gluster peer status is:
>>
>> # gluster peer status *<-- from node gluster01*
>> Number of Peers: 5
>>
>> Hostname: 10.0.231.51
>> Uuid: b01de59a-4428-486b-af49-cb486ab44a07
>> State: Peer in Cluster (Connected)
>>
>> Hostname: 10.0.231.52
>> Uuid: 75143760-52a3-4583-82bb-a9920b283dac
>> State: Peer in Cluster (Connected)
>>
>> Hostname: 10.0.231.53
>> Uuid: 2c0b8bb6-825a-4ddd-9958-d8b46e9a2411
>> State: Peer in Cluster (Connected)
>>
>> Hostname: 10.0.231.54 *<-- new node gluster05*
>> Uuid: 408d88d6-0448-41e8-94a3-bf9f98255d9c
>> *State: Peer Rejected (Connected)*
>>
>> Hostname: 10.0.231.55 *<-- new node gluster06*
>> Uuid: 9c155c8e-2cd1-4cfc-83af-47129b582fd3
>> *State: Peer Rejected (Connected)*
>
> Looks like your configuration files are mismatching, ie the
> checksum calculation differs on this two node than the others,
>
> Did you had any failed commit ?
>
> Compare your /var/lib/glusterd/<volname>/info of the failed node
> against good one, mostly you could see some difference.
>
> can you paste the /var/lib/glusterd/<volname>/info ?
>
> Regards
> Rafi KC
>
>
>> *
>> *
>> I followed the write-up
>> here: http://www.gluster.org/community/documentation/index.php/Resolving_Peer_Rejected
>> and the two new nodes peer'd properly but after a reboot of the
>> two new nodes I'm seeing the same Peer Rejected (Connected) State.
>>
>> I've attached logs from an existing node, and the two new nodes.
>>
>> Thanks for any suggestions,
>> Steve
>>
>>
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>> http://www.gluster.org/mailman/listinfo/gluster-users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160226/112aad19/attachment.html>
More information about the Gluster-users
mailing list