[Gluster-users] problem with Peer Rejected

Jiří Sléžka jiri.slezka at slu.cz
Fri Feb 4 15:39:08 UTC 2022


well, I tried to downgrade to 8.6 on node gluster07. It didn't help.

Fortunately I remember old post from Strahil in ovirt list which 
suggests to switch

gluster volume set  <VOLUME NAME> cluster.lookup-optimize off

when expanding cluster. As nodes were rejected due cksum mismatch on 
only one volume I tried to switch lookup-optimize on that volume, then 
restart glusterd on both nodes and it helped.

Problem seems to be solved...

Cheers,

Jiri


On 2/4/22 15:45, Jiří Sléžka wrote:
> Hello,
> 
> I have a glusterfs cluster in version 8.6, 6 nodes, 1 arbiter node, 
> distributed-replicated setup with arbiter (Number of Bricks: 3 x (2 + 1) 
> = 9).
> 
> Yesterday I added two new nodes. Because I plan to upgrade to gluster 9 
> I have installed them with Rocky Linux 8 and glusterfs 9 (from CentOS 
> stream repo). Then I added these two nodes and got this setup
> 
> Volume Name: samba
> Type: Distributed-Replicate
> Volume ID: a96ea622-7abb-4213-a39b-8a23a3035a5d
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 4 x (2 + 1) = 12
> Transport-type: tcp
> Bricks:
> Brick1: 10.10.102.91:/gluster/samba
> Brick2: 10.10.100.92:/gluster/samba
> Brick3: 10.10.100.90:/gluster/samba/brick1 (arbiter)
> Brick4: 10.10.100.93:/gluster/samba
> Brick5: 10.10.100.94:/gluster/samba
> Brick6: 10.10.100.90:/gluster/samba/brick2 (arbiter)
> Brick7: 10.10.100.95:/gluster/samba
> Brick8: 10.10.100.96:/gluster/samba
> Brick9: 10.10.100.90:/gluster/samba/brick3 (arbiter)
> Brick10: 10.10.100.97:/gluster/samba
> Brick11: 10.10.100.98:/gluster/samba
> Brick12: 10.10.100.90:/gluster/samba/brick4 (arbiter)
> Options Reconfigured:
> auth.allow: xxxxxxxxxxxxxxxxxxxxxxx
> cluster.self-heal-daemon: on
> cluster.entry-self-heal: on
> cluster.metadata-self-heal: on
> cluster.data-self-heal: on
> performance.client-io-threads: off
> nfs.disable: on
> transport.address-family: inet
> performance.readdir-ahead: on
> features.shard: on
> features.shard-block-size: 512MB
> cluster.quorum-type: auto
> cluster.server-quorum-type: server
> cluster.lookup-optimize: off
> 
> 
> op-version is still 80000
> 
> It worked well, I reballanced one of volumes but today I mentioned that 
> two new nodes are in Peer Rejected state (from gluster02 view)
> 
> gluster peer status
> Number of Peers: 8
> 
> Hostname: 10.10.100.91
> Uuid: 6d9e6170-2386-4b40-8fb5-7aeaef3d3122
> State: Peer in Cluster (Connected)
> 
> Hostname: 10.224.102.93
> Uuid: 4f74741e-7fee-41d0-a8db-916458f7280e
> State: Peer in Cluster (Connected)
> 
> Hostname: 10.10.100.94
> Uuid: cda31067-5bd9-44ea-816d-7c9dd947d78a
> State: Peer in Cluster (Connected)
> 
> Hostname: 10.10.100.95
> Uuid: 3c904f48-1ff3-4669-891b-27d4296ccf0e
> State: Peer in Cluster (Connected)
> 
> Hostname: 10.10.100.96
> Uuid: 0105494d-d5b4-40fb-ad31-c531efd818bb
> State: Peer in Cluster (Connected)
> 
> Hostname: 10.10.100.90
> Uuid: 291b7afd-3090-4733-a97f-20f8585adad2
> State: Peer in Cluster (Connected)
> 
> Hostname: 10.10.100.97
> Uuid: 82ac9abf-1678-43c9-a92f-94d0d472b2fe
> State: Peer Rejected (Disconnected)
> 
> Hostname: 10.10.100.98
> Uuid: 0f9e4891-250a-45b5-bdd3-e6a61aa49a29
> State: Peer Rejected (Connected)
> 
> from new node (gluster08) are Peer Rejected all nodes
> 
> there are log line in /var/log/glusterfs/glusterd.log like this
> 
> [2022-02-04 14:36:49.805753 +0000] E [MSGID: 106010] 
> [glusterd-utils.c:3851:glusterd_compare_friend_volume] 0-management: 
> Version of Cksums samba differ. local cksum = 3146523269, remote cksum = 
> 2206743689 on peer 10.10.100.97
> 
> there is a documentation for this particular problem...
> 
> https://docs.gluster.org/en/latest/Troubleshooting/troubleshooting-glusterd/#common-issues-and-how-to-resolve-them 
> 
> 
> ..but
> 
> gluster volume get all cluster.max-op-version
> 
> is still 80000
> 
> and I cannot set it lower or equal
> 
> gluster volume set all cluster.op-version 80000
> volume set: failed: Required op-version (80000) should not be equal or 
> lower than current cluster op-version (80000).
> 
> Unfortunately cluster seems broken on client's side. Any hints how can I 
> recover?
> 
> Thanks in advance,
> 
> Jiri
> 
> 
> ________
> 
> 
> 
> Community Meeting Calendar:
> 
> Schedule -
> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> Bridge: https://meet.google.com/cpu-eiue-hvk
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users

-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/pkcs7-signature
Size: 4269 bytes
Desc: S/MIME Cryptographic Signature
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20220204/8af1ce5e/attachment.p7s>


More information about the Gluster-users mailing list