[Bugs] [Bug 1540249] Gluster is trying to use a port outside documentation and firewalld' s glusterfs.xml
bugzilla at redhat.com
bugzilla at redhat.com
Mon Feb 12 09:40:33 UTC 2018
https://bugzilla.redhat.com/show_bug.cgi?id=1540249
--- Comment #21 from devianca at gmail.com ---
Both *were already* set, check
https://bugzilla.redhat.com/show_bug.cgi?id=1540249#c1
Now, also disabled both clients, and still getting Disconnected,
node1:
[root at ProdigyX ~]# uptime
10:37:35 up 11 min, 1 user, load average: 0,00, 0,01, 0,03
[root at ProdigyX ~]# systemctl status gluster.mount
● gluster.mount - Mount Gluster
Loaded: loaded (/etc/systemd/system/gluster.mount; disabled; vendor preset:
disabled)
Active: inactive (dead)
Where: /gluster
What: 127.0.0.1:/replica1
[root at ProdigyX ~]# gluster pool list
UUID Hostname State
2f6697f4-2529-4072-910c-8862fdc43562 10.250.1.1 Disconnected
f7976943-b81a-4bb4-a1fb-06253bf064c4 localhost Connected
[root at ProdigyX ~]# gluster peer status
Number of Peers: 1
Hostname: 10.250.1.1
Uuid: 2f6697f4-2529-4072-910c-8862fdc43562
State: Peer in Cluster (Disconnected)
[root at ProdigyX ~]# gluster volume status
Status of volume: replica1
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick 10.250.1.2:/array0/brick1 49152 0 Y 1390
Self-heal Daemon on localhost N/A N/A Y 1373
Task Status of Volume replica1
------------------------------------------------------------------------------
There are no active volume tasks
[root at ProdigyX ~]# gluster volume info
Volume Name: replica1
Type: Replicate
Volume ID: 5331fac2-42b6-4530-bf79-1ec0236efbc4
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 10.250.1.2:/array0/brick1
Brick2: 10.250.1.1:/raid0array1/brick2
Options Reconfigured:
client.bind-insecure: off
performance.client-io-threads: off
auth.allow: 10.250.1.1,10.250.1.2
transport.address-family: inet
nfs.disable: on
server.event-threads: 8
performance.io-thread-count: 64
performance.cache-size: 32MB
performance.write-behind-window-size: 64MB
server.allow-insecure: off
[root at ProdigyX ~]# netstat -tap | grep gluster
tcp 0 0 0.0.0.0:49152 0.0.0.0:* LISTEN
1390/glusterfsd
tcp 0 0 0.0.0.0:24007 0.0.0.0:* LISTEN
1361/glusterd
tcp 0 0 ProdigyX:1019 10.250.1.1:49152 ESTABLISHED
1373/glusterfs
tcp 0 0 ProdigyX:1020 ProdigyX:49152 ESTABLISHED
1373/glusterfs
tcp 0 0 ProdigyX:49152 10.250.1.1:exp2 ESTABLISHED
1390/glusterfsd
tcp 0 0 ProdigyX:24007 10.250.1.1:49151 ESTABLISHED
1361/glusterd
tcp 0 1 ProdigyX:49151 10.250.1.1:24007 SYN_SENT
1361/glusterd
tcp 0 0 ProdigyX:24007 ProdigyX:49149 ESTABLISHED
1361/glusterd
tcp 0 0 ProdigyX:49152 ProdigyX:1020 ESTABLISHED
1390/glusterfsd
tcp 0 0 ProdigyX:49149 ProdigyX:24007 ESTABLISHED
1390/glusterfsd
tcp 0 0 localhost:49150 localhost:24007 ESTABLISHED
1373/glusterfs
tcp 0 0 localhost:24007 localhost:49150 ESTABLISHED
1361/glusterd
node2:
[root at BUNKER ~]# uptime
10:37:34 up 26 min, 1 user, load average: 0,00, 0,01, 0,05
[root at BUNKER ~]# systemctl status gluster.mount
● gluster.mount - Mount Gluster
Loaded: loaded (/etc/systemd/system/gluster.mount; disabled; vendor preset:
disabled)
Active: inactive (dead)
Where: /gluster
What: 127.0.0.1:/replica1
[root at BUNKER ~]# gluster pool list
UUID Hostname State
f7976943-b81a-4bb4-a1fb-06253bf064c4 10.250.1.2 Connected
2f6697f4-2529-4072-910c-8862fdc43562 localhost Connected
[root at BUNKER ~]# gluster peer status
Number of Peers: 1
Hostname: 10.250.1.2
Uuid: f7976943-b81a-4bb4-a1fb-06253bf064c4
State: Peer in Cluster (Connected)
[root at BUNKER ~]# gluster volume status
Status of volume: replica1
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick 10.250.1.2:/array0/brick1 49152 0 Y 1390
Brick 10.250.1.1:/raid0array1/brick2 49152 0 Y 1334
Self-heal Daemon on localhost N/A N/A Y 1149
Self-heal Daemon on 10.250.1.2 N/A N/A Y 1373
Task Status of Volume replica1
------------------------------------------------------------------------------
There are no active volume tasks
[root at BUNKER ~]# gluster volume info
Volume Name: replica1
Type: Replicate
Volume ID: 5331fac2-42b6-4530-bf79-1ec0236efbc4
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 10.250.1.2:/array0/brick1
Brick2: 10.250.1.1:/raid0array1/brick2
Options Reconfigured:
client.bind-insecure: off
performance.client-io-threads: off
auth.allow: 10.250.1.1,10.250.1.2
transport.address-family: inet
nfs.disable: on
server.event-threads: 8
performance.io-thread-count: 64
performance.cache-size: 32MB
performance.write-behind-window-size: 64MB
server.allow-insecure: off
[root at BUNKER ~]# netstat -tap | grep gluster
tcp 0 0 0.0.0.0:49152 0.0.0.0:* LISTEN
1334/glusterfsd
tcp 0 0 0.0.0.0:24007 0.0.0.0:* LISTEN
866/glusterd
tcp 0 0 BUNKER:49152 10.250.1.2:1019 ESTABLISHED
1334/glusterfsd
tcp 0 0 BUNKER:exp2 10.250.1.2:49152 ESTABLISHED
1149/glusterfs
tcp 0 0 BUNKER:24007 BUNKER:49149 ESTABLISHED
866/glusterd
tcp 0 0 BUNKER:1020 BUNKER:49152 ESTABLISHED
1149/glusterfs
tcp 0 0 BUNKER:49151 10.250.1.2:24007 ESTABLISHED
866/glusterd
tcp 0 0 BUNKER:49152 BUNKER:1020 ESTABLISHED
1334/glusterfsd
tcp 0 0 localhost:49150 localhost:24007 ESTABLISHED
1149/glusterfs
tcp 0 0 localhost:24007 localhost:49150 ESTABLISHED
866/glusterd
tcp 0 0 BUNKER:49149 BUNKER:24007 ESTABLISHED
1334/glusterfsd
--
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.
More information about the Bugs
mailing list