[Gluster-users] Cannot connect to the cluster after customizing the port
sky
x_hsky at 163.com
Tue Dec 8 07:19:56 UTC 2020
There should be one port for communication and transport, both of which are defined by this option.
Judging from the current log, I suspect that the port is hard-coded.
Is there any other solution?
在 2020-12-08 14:50:13,"Strahil Nikolov" <hunter86_bg at yahoo.com> 写道:
>This option is only for the transport:
>option transport.socket.listen-port 24017
>
>So it still needs 24007 for communication.
>
>Maybe there is an option for changing the glusterd port , but I can't find it.
>
>Best Regards,
>Strahil Nikolov
>
>
>
>
>
>
>В вторник, 8 декември 2020 г., 07:52:38 Гринуич+2, sky <x_hsky at 163.com> написа:
>
>
>
>
>
>Hi:
> The firewall has been confirmed to be closed.
> The following is the startup log,and the ports of both nodes are 24017:
> Final graph:
> +------------------------------------------------------------------------------+
> 1: volume management
> 2: type mgmt/glusterd
> 3: option rpc-auth.auth-glusterfs on
> 4: option rpc-auth.auth-unix on
> 5: option rpc-auth.auth-null on
> 6: option rpc-auth-allow-insecure on
> 7: option transport.listen-backlog 1024
> 8: option max-port 60999
> 9: option base-port 49252
> 10: option event-threads 1
> 11: option ping-timeout 0
> 12: option transport.rdma.listen-port 24008
> 13: option transport.socket.listen-port 24017
> 14: option transport.socket.read-fail-log off
> 15: option transport.socket.keepalive-interval 2
> 16: option transport.socket.keepalive-time 10
> 17: option transport-type rdma
> 18: option working-directory /var/lib/glusterd
> 19: end-volume
> 20:
> +------------------------------------------------------------------------------+
> [2020-12-08 05:44:37.300602] I [MSGID: 101190] [event-epoll.c:682:event_dispatch_epoll_worker] 0-epoll: Started thread with index 0
>
> And the following is the log displayed after executing the command “gluster peer probe dev2”. It shows that the connected port is 24007. Is this port hard-coded?
>
> [2020-12-08 05:45:23.136309] I [MSGID: 106487] [glusterd-handler.c:1082:__glusterd_handle_cli_probe] 0-glusterd: Received CLI probe req dev2 24007
>[2020-12-08 05:45:23.138990] I [MSGID: 106128] [glusterd-handler.c:3541:glusterd_probe_begin] 0-glusterd: Unable to find peerinfo for host: dev2 (24007)
>[2020-12-08 05:45:23.208257] W [MSGID: 106061] [glusterd-handler.c:3315:glusterd_transport_inet_options_build] 0-glusterd: Failed to get tcp-user-timeout
>[2020-12-08 05:45:23.208387] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-management: setting frame-timeout to 600
>[2020-12-08 05:45:23.255375] I [MSGID: 106498] [glusterd-handler.c:3470:glusterd_friend_add] 0-management: connect returned 0
>[2020-12-08 05:45:23.260180] I [MSGID: 106004] [glusterd-handler.c:6204:__glusterd_peer_rpc_notify] 0-management: Peer <dev2> (<00000000-0000-0000-0000-000000000000>), in state <Establishing Connection>, has disconnected from glusterd.
>[2020-12-08 05:45:23.260997] I [MSGID: 106599] [glusterd-nfs-svc.c:161:glusterd_nfssvc_reconfigure] 0-management: nfs/server.so xlator is not installed
>[2020-12-08 05:45:23.261412] I [MSGID: 106544] [glusterd.c:152:glusterd_uuid_init] 0-management: retrieved UUID: 769fa9d7-a204-4c44-a15e-b5eac367e322
>[2020-12-08 05:45:23.261480] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-quotad: setting frame-timeout to 600
>[2020-12-08 05:45:23.261737] I [MSGID: 106131] [glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: quotad already stopped
>[2020-12-08 05:45:23.261780] I [MSGID: 106568] [glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: quotad service is stopped
>[2020-12-08 05:45:23.261915] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-bitd: setting frame-timeout to 600
>[2020-12-08 05:45:23.262039] I [MSGID: 106131] [glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: bitd already stopped
>[2020-12-08 05:45:23.262064] I [MSGID: 106568] [glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: bitd service is stopped
>[2020-12-08 05:45:23.262110] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-scrub: setting frame-timeout to 600
>[2020-12-08 05:45:23.262289] I [MSGID: 106131] [glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: scrub already stopped
>[2020-12-08 05:45:23.262331] I [MSGID: 106568] [glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: scrub service is stopped
>
>
>
>
>
>
>在 2020-12-08 05:01:12,"Strahil Nikolov" <hunter86_bg at yahoo.com> 写道:
>> Are you sure that the firewall is open. Usually , there could be a firewall inbetween the nodes.
>>
>> Also, you can run a tcpdump on nodeA and issue a peer probe from NodeB and see if there is anything hitting it.
>>
>> What about enabling trace logs and checking for any clues. You can find details about the log level at : https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3/html/administration_guide/configuring_the_log_level Best Regards,
>> Strahil Nikolov
>>
>> В 10:07 +0800 на 07.12.2020 (пн), sky написа:
>>> Hi:
>>> 1. The ports of both nodes are changed to 24017
>>> 2. After installation, change the port and start directly (non-restart)
>>> 3. I tried, but the nodes in the pool cannot connect after restarting. The log shows an attempt to connect to the default port
>>>
>>> Steps(both node1 and node2)
>>> # yum install glusterfs glusterfs-server glusterfs-fuse
>>> # vim /etc/glusterfs/glusterd.vol
>>> ....
>>> option transport.socket.listen-port 24017
>>> ....
>>> # systemctl start glusterd
>>>
>>> # gluster peer probe node2
>>> Probe returned with Transport endpoint is not connected
>>>
>>> After changing the port back to 24007 and restarting, the connection is normal
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> 在 2020-12-06 19:27:01,"Strahil Nikolov" <hunter86_bg at yahoo.com> 写道:>Did you change the port on both nodes ?>Did you restart glusterd on both nodes (the one you do peer probe and the other one that is being probed) ?>>Have you tried to first peer probe and then change the port on all nodes in the pool ?>>>Best Regards,>Strahil Nikolov>>>>>>В петък, 4 декември 2020 г., 12:18:42 Гринуич+2, sky <x_hsky at 163.com> написа: >>>>>>linux version: cenots 7.5 >gluster verison: 7.5.1>/etc/glusterfs/glusterd.vol: >volume management> type mgmt/glusterd> option working-directory /var/lib/glusterd> option transport-type socket,rdma> option transport.socket.keepalive-time 10> option transport.socket.keepalive-interval 2> option transport.socket.read-fail-log off> option transport.socket.listen-port 24017> option transport.rdma.listen-port 24008> option ping-timeout 0> option event-threads 1># option lock-timer 180># option transport.address-family inet6> option base-port 49252> option max-port 60999>end-volume>>I started normally after changing the ports on both nodes(port from 24007 to 24017), but I cannot add nodes through the ‘gluster peer probe node2‘ command,>It always prompts me: > Probe returned with Transport endpoint is not connected>>>>>>>>>>>> >>> >________>>>>Community Meeting Calendar:>>Schedule ->Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC>Bridge: https://meet.google.com/cpu-eiue-hvk>Gluster-users mailing list>Gluster-users at gluster.org>https://lists.gluster.org/mailman/listinfo/gluster-users
>>>
>>>
>>>
>>>
>>
>
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20201208/df4e3f2e/attachment.html>
More information about the Gluster-users
mailing list