[Bugs] [Bug 1540249] Gluster is trying to use a port outside documentation and firewalld' s glusterfs.xml

bugzilla at redhat.com bugzilla at redhat.com
Mon Feb 12 09:40:33 UTC 2018


https://bugzilla.redhat.com/show_bug.cgi?id=1540249



--- Comment #21 from devianca at gmail.com ---
Both *were already* set, check
https://bugzilla.redhat.com/show_bug.cgi?id=1540249#c1

Now, also disabled both clients, and still getting Disconnected,

node1:

[root at ProdigyX ~]# uptime
 10:37:35 up 11 min,  1 user,  load average: 0,00, 0,01, 0,03
[root at ProdigyX ~]# systemctl status gluster.mount
● gluster.mount - Mount Gluster
   Loaded: loaded (/etc/systemd/system/gluster.mount; disabled; vendor preset:
disabled)
   Active: inactive (dead)
    Where: /gluster
     What: 127.0.0.1:/replica1
[root at ProdigyX ~]# gluster pool list
UUID                                    Hostname        State
2f6697f4-2529-4072-910c-8862fdc43562    10.250.1.1      Disconnected
f7976943-b81a-4bb4-a1fb-06253bf064c4    localhost       Connected
[root at ProdigyX ~]# gluster peer status
Number of Peers: 1

Hostname: 10.250.1.1
Uuid: 2f6697f4-2529-4072-910c-8862fdc43562
State: Peer in Cluster (Disconnected)
[root at ProdigyX ~]# gluster volume status
Status of volume: replica1
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick 10.250.1.2:/array0/brick1             49152     0          Y       1390
Self-heal Daemon on localhost               N/A       N/A        Y       1373

Task Status of Volume replica1
------------------------------------------------------------------------------
There are no active volume tasks

[root at ProdigyX ~]# gluster volume info

Volume Name: replica1
Type: Replicate
Volume ID: 5331fac2-42b6-4530-bf79-1ec0236efbc4
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 10.250.1.2:/array0/brick1
Brick2: 10.250.1.1:/raid0array1/brick2
Options Reconfigured:
client.bind-insecure: off
performance.client-io-threads: off
auth.allow: 10.250.1.1,10.250.1.2
transport.address-family: inet
nfs.disable: on
server.event-threads: 8
performance.io-thread-count: 64
performance.cache-size: 32MB
performance.write-behind-window-size: 64MB
server.allow-insecure: off
[root at ProdigyX ~]# netstat -tap | grep gluster
tcp        0      0 0.0.0.0:49152           0.0.0.0:*               LISTEN     
1390/glusterfsd
tcp        0      0 0.0.0.0:24007           0.0.0.0:*               LISTEN     
1361/glusterd
tcp        0      0 ProdigyX:1019           10.250.1.1:49152        ESTABLISHED
1373/glusterfs
tcp        0      0 ProdigyX:1020           ProdigyX:49152          ESTABLISHED
1373/glusterfs
tcp        0      0 ProdigyX:49152          10.250.1.1:exp2         ESTABLISHED
1390/glusterfsd
tcp        0      0 ProdigyX:24007          10.250.1.1:49151        ESTABLISHED
1361/glusterd
tcp        0      1 ProdigyX:49151          10.250.1.1:24007        SYN_SENT   
1361/glusterd
tcp        0      0 ProdigyX:24007          ProdigyX:49149          ESTABLISHED
1361/glusterd
tcp        0      0 ProdigyX:49152          ProdigyX:1020           ESTABLISHED
1390/glusterfsd
tcp        0      0 ProdigyX:49149          ProdigyX:24007          ESTABLISHED
1390/glusterfsd
tcp        0      0 localhost:49150         localhost:24007         ESTABLISHED
1373/glusterfs
tcp        0      0 localhost:24007         localhost:49150         ESTABLISHED
1361/glusterd

node2:

[root at BUNKER ~]# uptime
 10:37:34 up 26 min,  1 user,  load average: 0,00, 0,01, 0,05
[root at BUNKER ~]# systemctl status gluster.mount
● gluster.mount - Mount Gluster
   Loaded: loaded (/etc/systemd/system/gluster.mount; disabled; vendor preset:
disabled)
   Active: inactive (dead)
    Where: /gluster
     What: 127.0.0.1:/replica1
[root at BUNKER ~]# gluster pool list
UUID                                    Hostname        State
f7976943-b81a-4bb4-a1fb-06253bf064c4    10.250.1.2      Connected
2f6697f4-2529-4072-910c-8862fdc43562    localhost       Connected
[root at BUNKER ~]# gluster peer status
Number of Peers: 1

Hostname: 10.250.1.2
Uuid: f7976943-b81a-4bb4-a1fb-06253bf064c4
State: Peer in Cluster (Connected)
[root at BUNKER ~]# gluster volume status
Status of volume: replica1
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick 10.250.1.2:/array0/brick1             49152     0          Y       1390
Brick 10.250.1.1:/raid0array1/brick2        49152     0          Y       1334
Self-heal Daemon on localhost               N/A       N/A        Y       1149
Self-heal Daemon on 10.250.1.2              N/A       N/A        Y       1373

Task Status of Volume replica1
------------------------------------------------------------------------------
There are no active volume tasks

[root at BUNKER ~]# gluster volume info

Volume Name: replica1
Type: Replicate
Volume ID: 5331fac2-42b6-4530-bf79-1ec0236efbc4
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 10.250.1.2:/array0/brick1
Brick2: 10.250.1.1:/raid0array1/brick2
Options Reconfigured:
client.bind-insecure: off
performance.client-io-threads: off
auth.allow: 10.250.1.1,10.250.1.2
transport.address-family: inet
nfs.disable: on
server.event-threads: 8
performance.io-thread-count: 64
performance.cache-size: 32MB
performance.write-behind-window-size: 64MB
server.allow-insecure: off
[root at BUNKER ~]# netstat -tap | grep gluster
tcp        0      0 0.0.0.0:49152           0.0.0.0:*               LISTEN     
1334/glusterfsd
tcp        0      0 0.0.0.0:24007           0.0.0.0:*               LISTEN     
866/glusterd
tcp        0      0 BUNKER:49152            10.250.1.2:1019         ESTABLISHED
1334/glusterfsd
tcp        0      0 BUNKER:exp2             10.250.1.2:49152        ESTABLISHED
1149/glusterfs
tcp        0      0 BUNKER:24007            BUNKER:49149            ESTABLISHED
866/glusterd
tcp        0      0 BUNKER:1020             BUNKER:49152            ESTABLISHED
1149/glusterfs
tcp        0      0 BUNKER:49151            10.250.1.2:24007        ESTABLISHED
866/glusterd
tcp        0      0 BUNKER:49152            BUNKER:1020             ESTABLISHED
1334/glusterfsd
tcp        0      0 localhost:49150         localhost:24007         ESTABLISHED
1149/glusterfs
tcp        0      0 localhost:24007         localhost:49150         ESTABLISHED
866/glusterd
tcp        0      0 BUNKER:49149            BUNKER:24007            ESTABLISHED
1334/glusterfsd

-- 
You are receiving this mail because:
You are on the CC list for the bug.
You are the assignee for the bug.


More information about the Bugs mailing list