Re: client glusterfs connection problem

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



It was about halo enabled. When i disabled mounting was succesfull. It seems just enabling is not enough.

So far i am happy with currwnt situation just need to figured out whether all data is replicated. With big and even small files, replication is fast. I tested with GB files. 
Just need to know the replication latency with multiple volumes usage, one volume and GB files, didnt realized a latency.

31 Eki 2018 Çar, saat 06:46 tarihinde Vlad Kopylov <vladkopy@xxxxxxxxx> şunu yazdı:
Try adding routes so it can connect to all.
Also curious if fuse mount does need access to all nodes. Supposedly it does wright to all at the same time unless you have halo feature enabled.

v




On Sun, Oct 28, 2018 at 1:07 AM Oğuz Yarımtepe <oguzyarimtepe@xxxxxxxxx> wrote:
My two nodes are at another vlan. Should my client have connection to all nodes, at replicated mod? 

Regards.

On Fri, Oct 26, 2018 at 4:44 AM Poornima Gurusiddaiah <pgurusid@xxxxxxxxxx> wrote:
Is this a new volume? Has it never been mounted successfully? If so try changing firewall settings to allow gluster ports, also check for selinux settings.

Regards,
Poornima

On Fri, Oct 26, 2018, 1:26 AM Oğuz Yarımtepe <oguzyarimtepe@xxxxxxxxx> wrote:
One more addition:

# gluster volume info

 
Volume Name: vol0
Type: Replicate
Volume ID: 28384e2b-ea7e-407e-83ae-4d4e69a2cc7e
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 4 = 4
Transport-type: tcp
Bricks:
Brick1: aslrplpgls01:/bricks/brick1/vol0
Brick2: aslrplpgls02:/bricks/brick2/vol0
Brick3: bslrplpgls01:/bricks/brick3/vol0
Brick4: bslrplpgls02:/bricks/brick4/vol0
Options Reconfigured:
cluster.self-heal-daemon: enable
cluster.halo-enabled: True
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off

On Thu, Oct 25, 2018 at 10:39 PM Oğuz Yarımtepe <oguzyarimtepe@xxxxxxxxx> wrote:
I have 4 node GlusterFS cluster. Used Centos SIG 4.1 repo. 

# gluster peer status
Number of Peers: 3

Hostname: aslrplpgls02
Uuid: 0876151a-058e-42ec-91f2-f25f353a0207
State: Peer in Cluster (Connected)

Hostname: bslrplpgls01
Uuid: 6d73ed2a-2287-4872-9a8f-64d6e833181f
State: Peer in Cluster (Connected)

Hostname: bslrplpgls02
Uuid: 8ab6b61f-f502-44c7-8966-2ab03a6b9f7e
State: Peer in Cluster (Connected)

# gluster volume status vol0
Status of volume: vol0
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick aslrplpgls01:/bricks/brick1/vol0      49152     0          Y       12991
Brick aslrplpgls02:/bricks/brick2/vol0      49152     0          Y       9344 
Brick bslrplpgls01:/bricks/brick3/vol0      49152     0          Y       61662
Brick bslrplpgls02:/bricks/brick4/vol0      49152     0          Y       61843
Self-heal Daemon on localhost               N/A       N/A        Y       13014
Self-heal Daemon on bslrplpgls02            N/A       N/A        Y       61866
Self-heal Daemon on bslrplpgls01            N/A       N/A        Y       61685
Self-heal Daemon on aslrplpgls02            N/A       N/A        Y       9367 
 
Task Status of Volume vol0
------------------------------------------------------------------------------
There are no active volume tasks
 
This is how volume area is mounted:

/dev/gluster_vg/gluster_lv /bricks/brick1 xfs defaults 1 2

When i try to mount vol0 on a remote machine below is what i got:

[2018-10-25 19:37:23.033302] D [MSGID: 0] [write-behind.c:2396:wb_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-write-behind returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.033329] D [MSGID: 0] [io-cache.c:268:ioc_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-io-cache returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.033356] D [MSGID: 0] [quick-read.c:473:qr_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-quick-read returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.033373] D [MSGID: 0] [md-cache.c:1130:mdc_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-md-cache returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.033389] D [MSGID: 0] [io-stats.c:2278:io_stats_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0 returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.033408] W [fuse-resolve.c:132:fuse_resolve_gfid_cbk] 0-fuse: 00000000-0000-0000-0000-000000000001: failed to resolve (Transport endpoint is not connected)
[2018-10-25 19:37:23.033426] E [fuse-bridge.c:928:fuse_getattr_resume] 0-glusterfs-fuse: 2: GETATTR 1 (00000000-0000-0000-0000-000000000001) resolution failed
[2018-10-25 19:37:23.036511] D [MSGID: 0] [dht-common.c:3468:dht_lookup] 0-vol0-dht: Calling fresh lookup for / on vol0-replicate-0
[2018-10-25 19:37:23.037347] D [MSGID: 0] [afr-common.c:3241:afr_discover_do] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-replicate-0 returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.037375] D [MSGID: 0] [dht-common.c:3020:dht_lookup_cbk] 0-vol0-dht: fresh_lookup returned for / with op_ret -1 [Transport endpoint is not connected]
[2018-10-25 19:37:23.037940] D [MSGID: 0] [afr-common.c:3241:afr_discover_do] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-replicate-0 returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.037963] D [MSGID: 0] [dht-common.c:1378:dht_lookup_dir_cbk] 0-vol0-dht: lookup of / on vol0-replicate-0 returned error [Transport endpoint is not connected]
[2018-10-25 19:37:23.037979] E [MSGID: 101046] [dht-common.c:1502:dht_lookup_dir_cbk] 0-vol0-dht: dict is null
[2018-10-25 19:37:23.037994] D [MSGID: 0] [dht-common.c:1505:dht_lookup_dir_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-dht returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.038010] D [MSGID: 0] [write-behind.c:2396:wb_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-write-behind returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.038028] D [MSGID: 0] [io-cache.c:268:ioc_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-io-cache returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.038045] D [MSGID: 0] [quick-read.c:473:qr_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-quick-read returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.038061] D [MSGID: 0] [md-cache.c:1130:mdc_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0-md-cache returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.038078] D [MSGID: 0] [io-stats.c:2278:io_stats_lookup_cbk] 0-stack-trace: stack-address: 0x7f0d04001038, vol0 returned -1 error: Transport endpoint is not connected [Transport endpoint is not connected]
[2018-10-25 19:37:23.038096] W [fuse-resolve.c:132:fuse_resolve_gfid_cbk] 0-fuse: 00000000-0000-0000-0000-000000000001: failed to resolve (Transport endpoint is not connected)
[2018-10-25 19:37:23.038110] E [fuse-bridge.c:928:fuse_getattr_resume] 0-glusterfs-fuse: 3: GETATTR 1 (00000000-0000-0000-0000-000000000001) resolution failed
[2018-10-25 19:37:23.041169] D [fuse-bridge.c:5087:fuse_thread_proc] 0-glusterfs-fuse: terminating upon getting ENODEV when reading /dev/fuse
[2018-10-25 19:37:23.041196] I [fuse-bridge.c:5199:fuse_thread_proc] 0-fuse: initating unmount of /mnt/gluster
[2018-10-25 19:37:23.041306] D [logging.c:1795:gf_log_flush_extra_msgs] 0-logging-infra: Log buffer size reduced. About to flush 5 extra log messages
[2018-10-25 19:37:23.041331] D [logging.c:1798:gf_log_flush_extra_msgs] 0-logging-infra: Just flushed 5 extra log messages
[2018-10-25 19:37:23.041398] W [glusterfsd.c:1514:cleanup_and_exit] (-->/lib64/libpthread.so.0(+0x7e25) [0x7f0d24e0ae25] -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x5594b73edd65] -->/usr/sbin/glusterfs(cleanup_and_exit+0x6b) [0x5594b73edb8b] ) 0-: received signum (15), shutting down
[2018-10-25 19:37:23.041417] D [mgmt-pmap.c:79:rpc_clnt_mgmt_pmap_signout] 0-fsd-mgmt: portmapper signout arguments not given
[2018-10-25 19:37:23.041428] I [fuse-bridge.c:5981:fini] 0-fuse: Unmounting '/mnt/gluster'.
[2018-10-25 19:37:23.041441] I [fuse-bridge.c:5986:fini] 0-fuse: Closing fuse connection to '/mnt/gluster'.

This is how i added mount point to fstab

10.35.72.138:/vol0 /mnt/gluster glusterfs defaults,_netdev,log-level=DEBUG 0 0

Any idea what the problem is? I found some bug entries, not sure whether this situation is a bug.

 

--
Oğuz Yarımtepe
http://about.me/oguzy


--
Oğuz Yarımtepe
http://about.me/oguzy
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users


--
Oğuz Yarımtepe
http://about.me/oguzy
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users
--
Oğuz Yarımtepe
http://about.me/oguzy
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux