There should be one port for communication and transport, both of which are defined by this option.
Judging from the current log, I suspect that the port is hard-coded.
Is there any other solution?
在 2020-12-08 14:50:13,"Strahil Nikolov" <hunter86_bg@xxxxxxxxx> 写道: >This option is only for the transport: >option transport.socket.listen-port 24017 > >So it still needs 24007 for communication. > >Maybe there is an option for changing the glusterd port , but I can't find it. > >Best Regards, >Strahil Nikolov > > > > > > >В вторник, 8 декември 2020 г., 07:52:38 Гринуич+2, sky <x_hsky@xxxxxxx> написа: > > > > > >Hi: > The firewall has been confirmed to be closed. > The following is the startup log,and the ports of both nodes are 24017: > Final graph: > +------------------------------------------------------------------------------+ > 1: volume management > 2: type mgmt/glusterd > 3: option rpc-auth.auth-glusterfs on > 4: option rpc-auth.auth-unix on > 5: option rpc-auth.auth-null on > 6: option rpc-auth-allow-insecure on > 7: option transport.listen-backlog 1024 > 8: option max-port 60999 > 9: option base-port 49252 > 10: option event-threads 1 > 11: option ping-timeout 0 > 12: option transport.rdma.listen-port 24008 > 13: option transport.socket.listen-port 24017 > 14: option transport.socket.read-fail-log off > 15: option transport.socket.keepalive-interval 2 > 16: option transport.socket.keepalive-time 10 > 17: option transport-type rdma > 18: option working-directory /var/lib/glusterd > 19: end-volume > 20: > +------------------------------------------------------------------------------+ > [2020-12-08 05:44:37.300602] I [MSGID: 101190] [event-epoll.c:682:event_dispatch_epoll_worker] 0-epoll: Started thread with index 0 > > And the following is the log displayed after executing the command “gluster peer probe dev2”. It shows that the connected port is 24007. Is this port hard-coded? > > [2020-12-08 05:45:23.136309] I [MSGID: 106487] [glusterd-handler.c:1082:__glusterd_handle_cli_probe] 0-glusterd: Received CLI probe req dev2 24007 >[2020-12-08 05:45:23.138990] I [MSGID: 106128] [glusterd-handler.c:3541:glusterd_probe_begin] 0-glusterd: Unable to find peerinfo for host: dev2 (24007) >[2020-12-08 05:45:23.208257] W [MSGID: 106061] [glusterd-handler.c:3315:glusterd_transport_inet_options_build] 0-glusterd: Failed to get tcp-user-timeout >[2020-12-08 05:45:23.208387] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-management: setting frame-timeout to 600 >[2020-12-08 05:45:23.255375] I [MSGID: 106498] [glusterd-handler.c:3470:glusterd_friend_add] 0-management: connect returned 0 >[2020-12-08 05:45:23.260180] I [MSGID: 106004] [glusterd-handler.c:6204:__glusterd_peer_rpc_notify] 0-management: Peer <dev2> (<00000000-0000-0000-0000-000000000000>), in state <Establishing Connection>, has disconnected from glusterd. >[2020-12-08 05:45:23.260997] I [MSGID: 106599] [glusterd-nfs-svc.c:161:glusterd_nfssvc_reconfigure] 0-management: nfs/server.so xlator is not installed >[2020-12-08 05:45:23.261412] I [MSGID: 106544] [glusterd.c:152:glusterd_uuid_init] 0-management: retrieved UUID: 769fa9d7-a204-4c44-a15e-b5eac367e322 >[2020-12-08 05:45:23.261480] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-quotad: setting frame-timeout to 600 >[2020-12-08 05:45:23.261737] I [MSGID: 106131] [glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: quotad already stopped >[2020-12-08 05:45:23.261780] I [MSGID: 106568] [glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: quotad service is stopped >[2020-12-08 05:45:23.261915] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-bitd: setting frame-timeout to 600 >[2020-12-08 05:45:23.262039] I [MSGID: 106131] [glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: bitd already stopped >[2020-12-08 05:45:23.262064] I [MSGID: 106568] [glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: bitd service is stopped >[2020-12-08 05:45:23.262110] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-scrub: setting frame-timeout to 600 >[2020-12-08 05:45:23.262289] I [MSGID: 106131] [glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: scrub already stopped >[2020-12-08 05:45:23.262331] I [MSGID: 106568] [glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: scrub service is stopped > > > > > > >在 2020-12-08 05:01:12,"Strahil Nikolov" <hunter86_bg@xxxxxxxxx> 写道: >> Are you sure that the firewall is open. Usually , there could be a firewall inbetween the nodes. >> >> Also, you can run a tcpdump on nodeA and issue a peer probe from NodeB and see if there is anything hitting it. >> >> What about enabling trace logs and checking for any clues. You can find details about the log level at : https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3/html/administration_guide/configuring_the_log_level Best Regards, >> Strahil Nikolov >> >> В 10:07 +0800 на 07.12.2020 (пн), sky написа: >>> Hi: >>> 1. The ports of both nodes are changed to 24017 >>> 2. After installation, change the port and start directly (non-restart) >>> 3. I tried, but the nodes in the pool cannot connect after restarting. The log shows an attempt to connect to the default port >>> >>> Steps(both node1 and node2) >>> # yum install glusterfs glusterfs-server glusterfs-fuse >>> # vim /etc/glusterfs/glusterd.vol >>> .... >>> option transport.socket.listen-port 24017 >>> .... >>> # systemctl start glusterd >>> >>> # gluster peer probe node2 >>> Probe returned with Transport endpoint is not connected >>> >>> After changing the port back to 24007 and restarting, the connection is normal >>> >>> >>> >>> >>> >>> >>> >>> 在 2020-12-06 19:27:01,"Strahil Nikolov" <hunter86_bg@xxxxxxxxx> 写道:>Did you change the port on both nodes ?>Did you restart glusterd on both nodes (the one you do peer probe and the other one that is being probed) ?>>Have you tried to first peer probe and then change the port on all nodes in the pool ?>>>Best Regards,>Strahil Nikolov>>>>>>В петък, 4 декември 2020 г., 12:18:42 Гринуич+2, sky <x_hsky@xxxxxxx> написа: >>>>>>linux version: cenots 7.5 >gluster verison: 7.5.1>/etc/glusterfs/glusterd.vol: >volume management> type mgmt/glusterd> option working-directory /var/lib/glusterd> option transport-type socket,rdma> option transport.socket.keepalive-time 10> option transport.socket.keepalive-interval 2> option transport.socket.read-fail-log off> option transport.socket.listen-port 24017> option transport.rdma.listen-port 24008> option ping-timeout 0> option event-threads 1># option lock-timer 180># option transport.address-family inet6> option base-port 49252> option max-port 60999>end-volume>>I started normally after changing the ports on both nodes(port from 24007 to 24017), but I cannot add nodes through the ‘gluster peer probe node2‘ command,>It always prompts me: > Probe returned with Transport endpoint is not connected>>>>>>>>>>>> >>> >________>>>>Community Meeting Calendar:>>Schedule ->Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC>Bridge: https://meet.google.com/cpu-eiue-hvk>Gluster-users mailing list>Gluster-users@xxxxxxxxxxx>https://lists.gluster.org/mailman/listinfo/gluster-users >>> >>> >>> >>> >> > > > >
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://meet.google.com/cpu-eiue-hvk Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users