Try adding routes so it can connect to all.
Also curious if fuse mount does need access to all nodes. Supposedly it does wright to all at the same time unless you have halo feature enabled.
v
On Sun, Oct 28, 2018 at 1:07 AM Oğuz Yarımtepe <oguzyarimtepe@xxxxxxxxx> wrote:
My two nodes are at another vlan. Should my client have connection to all nodes, at replicated mod?Regards.On Fri, Oct 26, 2018 at 4:44 AM Poornima Gurusiddaiah <pgurusid@xxxxxxxxxx> wrote:Is this a new volume? Has it never been mounted successfully? If so try changing firewall settings to allow gluster ports, also check for selinux settings.
Regards,PoornimaOn Fri, Oct 26, 2018, 1:26 AM Oğuz Yarımtepe <oguzyarimtepe@xxxxxxxxx> wrote:One more addition:# gluster volume infoVolume Name: vol0Type: ReplicateVolume ID: 28384e2b-ea7e-407e-83ae-4d4e69a2cc7eStatus: StartedSnapshot Count: 0Number of Bricks: 1 x 4 = 4Transport-type: tcpBricks:Brick1: aslrplpgls01:/bricks/brick1/vol0Brick2: aslrplpgls02:/bricks/brick2/vol0Brick3: bslrplpgls01:/bricks/brick3/vol0Brick4: bslrplpgls02:/bricks/brick4/vol0Options Reconfigured:cluster.self-heal-daemon: enablecluster.halo-enabled: Truetransport.address-family: inetnfs.disable: onperformance.client-io-threads: offOn Thu, Oct 25, 2018 at 10:39 PM Oğuz Yarımtepe <oguzyarimtepe@xxxxxxxxx> wrote:I have 4 node GlusterFS cluster. Used Centos SIG 4.1 repo.# gluster peer statusNumber of Peers: 3Hostname: aslrplpgls02Uuid: 0876151a-058e-42ec-91f2-f25f353a0207State: Peer in Cluster (Connected)Hostname: bslrplpgls01Uuid: 6d73ed2a-2287-4872-9a8f-64d6e833181fState: Peer in Cluster (Connected)Hostname: bslrplpgls02Uuid: 8ab6b61f-f502-44c7-8966-2ab03a6b9f7eState: Peer in Cluster (Connected)# gluster volume status vol0Status of volume: vol0Gluster process TCP Port RDMA Port Online Pid------------------------------------------------------------------------------Brick aslrplpgls01:/bricks/brick1/vol0 49152 0 Y 12991Brick aslrplpgls02:/bricks/brick2/vol0 49152 0 Y 9344Brick bslrplpgls01:/bricks/brick3/vol0 49152 0 Y 61662Brick bslrplpgls02:/bricks/brick4/vol0 49152 0 Y 61843Self-heal Daemon on localhost N/A N/A Y 13014Self-heal Daemon on bslrplpgls02 N/A N/A Y 61866Self-heal Daemon on bslrplpgls01 N/A N/A Y 61685Self-heal Daemon on aslrplpgls02 N/A N/A Y 9367Task Status of Volume vol0------------------------------------------------------------------------------There are no active volume tasksThis is how volume area is mounted:/dev/gluster_vg/gluster_lv /bricks/brick1 xfs defaults 1 2When i try to mount vol0 on a remote machine below is what i got:[2018-10-25 19:37:23.033302] D [MSGID: 0] [write-behind.c:2396:wb_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-write-behind returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.033329] D [MSGID: 0] [io-cache.c:268:ioc_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-io-cache returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.033356] D [MSGID: 0] [quick-read.c:473:qr_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-quick-read returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.033373] D [MSGID: 0] [md-cache.c:1130:mdc_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-md-cache returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.033389] D [MSGID: 0] [io-stats.c:2278:io_stats_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0 returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.033408] W [fuse-resolve.c:132:fuse_resolve_gfid_cbk] 0-fuse: 00000000-0000-0000-0000-000000000001: failed to resolve (Transport endpoint is not connected)[2018-10-25 19:37:23.033426] E [fuse-bridge.c:928:fuse_getattr_resume] 0-glusterfs-fuse: 2: GETATTR 1 (00000000-0000-0000-0000-000000000001) resolution failed[2018-10-25 19:37:23.036511] D [MSGID: 0] [dht-common.c:3468:dht_lookup] 0-vol0-dht: Calling fresh lookup for / on vol0-replicate-0[2018-10-25 19:37:23.037347] D [MSGID: 0] [afr-common.c:3241:afr_discover_do] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-replicate-0 returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.037375] D [MSGID: 0] [dht-common.c:3020:dht_lookup_cbk] 0-vol0-dht: fresh_lookup returned for / with op_ret -1 [Transport endpoint is not connected][2018-10-25 19:37:23.037940] D [MSGID: 0] [afr-common.c:3241:afr_discover_do] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-replicate-0 returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.037963] D [MSGID: 0] [dht-common.c:1378:dht_lookup_dir_cbk] 0-vol0-dht: lookup of / on vol0-replicate-0 returned error [Transport endpoint is not connected][2018-10-25 19:37:23.037979] E [MSGID: 101046] [dht-common.c:1502:dht_lookup_dir_cbk] 0-vol0-dht: dict is null[2018-10-25 19:37:23.037994] D [MSGID: 0] [dht-common.c:1505:dht_lookup_dir_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-dht returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.038010] D [MSGID: 0] [write-behind.c:2396:wb_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-write-behind returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.038028] D [MSGID: 0] [io-cache.c:268:ioc_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-io-cache returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.038045] D [MSGID: 0] [quick-read.c:473:qr_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-quick-read returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.038061] D [MSGID: 0] [md-cache.c:1130:mdc_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-md-cache returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.038078] D [MSGID: 0] [io-stats.c:2278:io_stats_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0 returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected][2018-10-25 19:37:23.038096] W [fuse-resolve.c:132:fuse_resolve_gfid_cbk] 0-fuse: 00000000-0000-0000-0000-000000000001: failed to resolve (Transport endpoint is not connected)[2018-10-25 19:37:23.038110] E [fuse-bridge.c:928:fuse_getattr_resume] 0-glusterfs-fuse: 3: GETATTR 1 (00000000-0000-0000-0000-000000000001) resolution failed[2018-10-25 19:37:23.041169] D [fuse-bridge.c:5087:fuse_thread_proc] 0-glusterfs-fuse: terminating upon getting ENODEV when reading /dev/fuse[2018-10-25 19:37:23.041196] I [fuse-bridge.c:5199:fuse_thread_proc] 0-fuse: initating unmount of /mnt/gluster[2018-10-25 19:37:23.041306] D [logging.c:1795:gf_log_flush_extra_msgs] 0-logging-infra: Log buffer size reduced. About to flush 5 extra log messages[2018-10-25 19:37:23.041331] D [logging.c:1798:gf_log_flush_extra_msgs] 0-logging-infra: Just flushed 5 extra log messages[2018-10-25 19:37:23.041398] W [glusterfsd.c:1514:cleanup_and_exit] (-->/lib64/libpthread.so.0(+0x7e25) [0x7f0d24e0ae25] -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x5594b73edd65] -->/usr/sbin/glusterfs(cleanup_and_exit+0x6b) [0x5594b73edb8b] ) 0-: received signum (15), shutting down[2018-10-25 19:37:23.041417] D [mgmt-pmap.c:79:rpc_clnt_mgmt_pmap_signout] 0-fsd-mgmt: portmapper signout arguments not given[2018-10-25 19:37:23.041428] I [fuse-bridge.c:5981:fini] 0-fuse: Unmounting '/mnt/gluster'.[2018-10-25 19:37:23.041441] I [fuse-bridge.c:5986:fini] 0-fuse: Closing fuse connection to '/mnt/gluster'.This is how i added mount point to fstab10.35.72.138:/vol0 /mnt/gluster glusterfs defaults,_netdev,log-level=DEBUG 0 0Any idea what the problem is? I found some bug entries, not sure whether this situation is a bug.--Oğuz Yarımtepe
http://about.me/oguzy--Oğuz Yarımtepe_______________________________________________
http://about.me/oguzy
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users--Oğuz Yarımtepe_______________________________________________
http://about.me/oguzy
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users