Re: Cannot connect to the cluster after customizing the port

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



There should be one port for communication and transport, both of which are defined by this option.

Judging from the current log, I suspect that the port is hard-coded.

Is there any other solution?







在 2020-12-08 14:50:13,"Strahil Nikolov" <hunter86_bg@xxxxxxxxx> 写道: >This option is only for the transport: >option transport.socket.listen-port 24017 > >So it still needs 24007 for communication. > >Maybe there is an option for changing the glusterd port , but I can't find it. > >Best Regards, >Strahil Nikolov > > > > > > >В вторник, 8 декември 2020 г., 07:52:38 Гринуич+2, sky <x_hsky@xxxxxxx> написа: > > > > > >Hi: >     The firewall has been confirmed to be closed.  >     The following is the startup log,and the ports of both nodes are 24017: >          Final graph:                                                                                                                                                                   >         +------------------------------------------------------------------------------+                                                                                               >       1: volume management                                                                                                                                                         >       2:     type mgmt/glusterd                                                                                                                                                    >       3:     option rpc-auth.auth-glusterfs on                                                                                                                                     >       4:     option rpc-auth.auth-unix on                                                                                                                                          >       5:     option rpc-auth.auth-null on >       6:     option rpc-auth-allow-insecure on >       7:     option transport.listen-backlog 1024 >       8:     option max-port 60999 >       9:     option base-port 49252 >      10:     option event-threads 1 >     11:     option ping-timeout 0 >     12:     option transport.rdma.listen-port 24008 >     13:     option transport.socket.listen-port 24017 >     14:     option transport.socket.read-fail-log off >     15:     option transport.socket.keepalive-interval 2 >     16:     option transport.socket.keepalive-time 10 >     17:     option transport-type rdma >     18:     option working-directory /var/lib/glusterd >     19: end-volume                             >     20:                                        >     +------------------------------------------------------------------------------+ >     [2020-12-08 05:44:37.300602] I [MSGID: 101190] [event-epoll.c:682:event_dispatch_epoll_worker] 0-epoll: Started thread with index 0 > >    And the following is the log displayed after executing the command “gluster peer probe dev2”. It shows that the connected port is 24007. Is this port hard-coded? > > [2020-12-08 05:45:23.136309] I [MSGID: 106487] [glusterd-handler.c:1082:__glusterd_handle_cli_probe] 0-glusterd: Received CLI probe req dev2 24007  >[2020-12-08 05:45:23.138990] I [MSGID: 106128] [glusterd-handler.c:3541:glusterd_probe_begin] 0-glusterd: Unable to find peerinfo for host: dev2 (24007)  >[2020-12-08 05:45:23.208257] W [MSGID: 106061] [glusterd-handler.c:3315:glusterd_transport_inet_options_build] 0-glusterd: Failed to get tcp-user-timeout  >[2020-12-08 05:45:23.208387] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-management: setting frame-timeout to 600 >[2020-12-08 05:45:23.255375] I [MSGID: 106498] [glusterd-handler.c:3470:glusterd_friend_add] 0-management: connect returned 0  >[2020-12-08 05:45:23.260180] I [MSGID: 106004] [glusterd-handler.c:6204:__glusterd_peer_rpc_notify] 0-management: Peer <dev2> (<00000000-0000-0000-0000-000000000000>), in state <Establishing Connection>, has disconnected from glusterd.  >[2020-12-08 05:45:23.260997] I [MSGID: 106599] [glusterd-nfs-svc.c:161:glusterd_nfssvc_reconfigure] 0-management: nfs/server.so xlator is not installed  >[2020-12-08 05:45:23.261412] I [MSGID: 106544] [glusterd.c:152:glusterd_uuid_init] 0-management: retrieved UUID: 769fa9d7-a204-4c44-a15e-b5eac367e322  >[2020-12-08 05:45:23.261480] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-quotad: setting frame-timeout to 600 >[2020-12-08 05:45:23.261737] I [MSGID: 106131] [glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: quotad already stopped  >[2020-12-08 05:45:23.261780] I [MSGID: 106568] [glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: quotad service is stopped  >[2020-12-08 05:45:23.261915] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-bitd: setting frame-timeout to 600 >[2020-12-08 05:45:23.262039] I [MSGID: 106131] [glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: bitd already stopped  >[2020-12-08 05:45:23.262064] I [MSGID: 106568] [glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: bitd service is stopped  >[2020-12-08 05:45:23.262110] I [rpc-clnt.c:1014:rpc_clnt_connection_init] 0-scrub: setting frame-timeout to 600 >[2020-12-08 05:45:23.262289] I [MSGID: 106131] [glusterd-proc-mgmt.c:86:glusterd_proc_stop] 0-management: scrub already stopped  >[2020-12-08 05:45:23.262331] I [MSGID: 106568] [glusterd-svc-mgmt.c:265:glusterd_svc_stop] 0-management: scrub service is stopped > > > > > > >在 2020-12-08 05:01:12,"Strahil Nikolov" <hunter86_bg@xxxxxxxxx> 写道: >> Are you sure that the firewall is open. Usually , there could be a firewall inbetween the nodes. >> >> Also, you can run a tcpdump on nodeA and issue a peer probe from NodeB and see if there is anything hitting it. >> >> What about enabling trace logs and checking for any clues. You can find details about the log level at : https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3/html/administration_guide/configuring_the_log_level Best Regards, >> Strahil Nikolov >> >> В 10:07 +0800 на 07.12.2020 (пн), sky написа: >>> Hi:  >>> 1. The ports of both nodes are changed to 24017 >>> 2. After installation, change the port and start directly (non-restart) >>> 3. I tried, but the nodes in the pool cannot connect after restarting. The log shows an attempt to connect to the default port >>> >>> Steps(both node1 and node2) >>>    # yum install glusterfs glusterfs-server glusterfs-fuse >>>    # vim /etc/glusterfs/glusterd.vol >>>  .... >>>  option transport.socket.listen-port 24017 >>>  .... >>>  # systemctl start glusterd >>> >>>  # gluster peer probe node2 >>>  Probe returned with Transport endpoint is not connected >>> >>> After changing the port back to 24007 and restarting, the connection is normal >>> >>> >>> >>> >>> >>> >>> >>> 在 2020-12-06 19:27:01,"Strahil Nikolov" <hunter86_bg@xxxxxxxxx> 写道:>Did you change the port on both nodes ?>Did you restart glusterd on both nodes (the one you do peer probe and the other one that is being probed) ?>>Have you tried to first peer probe and then change the port on all nodes in the pool ?>>>Best Regards,>Strahil Nikolov>>>>>>В петък, 4 декември 2020 г., 12:18:42 Гринуич+2, sky <x_hsky@xxxxxxx> написа: >>>>>>linux version: cenots 7.5 >gluster verison: 7.5.1>/etc/glusterfs/glusterd.vol: >volume management>    type mgmt/glusterd>    option working-directory /var/lib/glusterd>    option transport-type socket,rdma>    option transport.socket.keepalive-time 10>    option transport.socket.keepalive-interval 2>    option transport.socket.read-fail-log off>    option transport.socket.listen-port 24017>    option transport.rdma.listen-port 24008>    option ping-timeout 0>    option event-threads 1>#   option lock-timer 180>#   option transport.address-family inet6>    option base-port 49252>    option max-port  60999>end-volume>>I started normally after changing the ports on both nodes(port from 24007 to 24017), but I cannot add nodes through the ‘gluster peer probe node2‘ command,>It always prompts me: >   Probe returned with Transport endpoint is not connected>>>>>>>>>>>> >>> >________>>>>Community Meeting Calendar:>>Schedule ->Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC>Bridge: https://meet.google.com/cpu-eiue-hvk>Gluster-users mailing list>Gluster-users@xxxxxxxxxxx>https://lists.gluster.org/mailman/listinfo/gluster-users >>> >>> >>> >>>   >> > > > > 


 

________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux