Can't mount using volfile

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all!
I'm rather new to glusterfs, trying it out for redundant storage for my very small company.
I have a minimal setup of glusterfs, 2 servers (storage1 & storage2) with one brick each, both added to volume "testvol1". I then  mount "testvol1" on a third server (app1). This is fine as long as I use any one servers ip and filesystem glusterfs, but when trying to mount the volume using a file "/owndata/conf/glusterfs/storage.vol", it does not work. Also I can't use nfs - don't need it but maybe it's related.

So:
"mount -t glusterfs  192.168.160.21:/testvol1 /mnt/tmp/" works. I write a file to /mnt/tmp/filename, then mount 192.168.12.210:/testvol1 and the newly created file is there.

Trying to mount using config file storage.vol:
mount -t glusterfs /owndata/conf/glusterfs/storage.vol /mnt/tmp
Mount failed. Please check the log file for more details.
The main error in the log is
 E [client-handshake.c:1778:client_query_portmap_cbk] 0-remote1: failed to get the port number for remote subvolume. Please run 'gluster volume status' on server to see if brick process is running.
There are lots & lots of pages on the 'net about this error message, none of the solutions I've found has worked.

CentOS 6.5 on all servers, they are all kvm hosts under oVirt (this is just the testing stage, will be on real iron in production).
No firewall anywhere, selinux is permissive.

File storage.vol:
volume remote1
  type protocol/client
  option transport-type tcp
  option remote-host 192.168.12.210
  option remote-subvolume testvol1
end-volume
 
volume remote2
  type protocol/client
  option transport-type tcp
  option remote-host 192.168.160.21
  option remote-subvolume testvol1
end-volume
 
volume replicate
  type cluster/replicate
  subvolumes remote1 remote2
end-volume
 
volume writebehind
  type performance/write-behind
  option window-size 1MB
  subvolumes replicate
end-volume
 
volume cache
  type performance/io-cache
  option cache-size 256MB
  subvolumes writebehind
end-volume
# gluster volume info
Volume Name: testvol1
Type: Replicate
Volume ID: bcca4aa2-46c0-44a2-8175-1305faa8b4f9
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 192.168.12.210:/export/vdb/brick1
Brick2: 192.168.160.21:/export/vdb/brick1
# gluster volume status
Status of volume: testvol1
Gluster process                                         Port    Online  Pid
-------------------------------------------------------------------------
Brick 192.168.12.210:/export/vdb/brick1                 49152   Y       1656
Brick 192.168.160.21:/export/vdb/brick1                 49152   Y       139090
NFS Server on localhost                                 2049    Y       1670
Self-heal Daemon on localhost                           N/A     Y       1674
NFS Server on 192.168.160.21                            2049    Y       1481
Self-heal Daemon on 192.168.160.21                      N/A     Y       139105
 
Task Status of Volume testvol1
------------------------------------------------------------------------------
There are no active volume tasks
Complete log after fail:
[2014-10-02 14:38:22.252235] I [glusterfsd.c:2026:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.4.0.57rhs (/usr/sbin/glusterfs --fuse-mountopts=allow_other,default_permissions,max_read=131072 --volfile=/owndata/conf/glusterfs/storage.vol --fuse-mountopts=allow_other,default_permissions,max_read=131072 /mnt/glust)
[2014-10-02 14:38:22.284438] W [options.c:848:xl_opt_validate] 0-writebehind: option 'window-size' is deprecated, preferred is 'cache-size', continuing with correction
[2014-10-02 14:38:22.284476] W [io-cache.c:1672:init] 0-cache: dangling volume. check volfile
[2014-10-02 14:38:22.294306] I [socket.c:3505:socket_init] 0-remote2: SSL support is NOT enabled
[2014-10-02 14:38:22.294339] I [socket.c:3520:socket_init] 0-remote2: using system polling thread
[2014-10-02 14:38:22.294832] I [socket.c:3505:socket_init] 0-remote1: SSL support is NOT enabled
[2014-10-02 14:38:22.294848] I [socket.c:3520:socket_init] 0-remote1: using system polling thread
[2014-10-02 14:38:22.294870] I [client.c:2171:notify] 0-remote1: parent translators are ready, attempting connect on transport
[2014-10-02 14:38:22.306697] I [client.c:2171:notify] 0-remote2: parent translators are ready, attempting connect on transport
Final graph:
+------------------------------------------------------------------------------+
  1: volume remote1
  2:     type protocol/client
  3:     option remote-subvolume testvol1
  4:     option remote-host 192.168.12.210
  5:     option transport-type socket
  6: end-volume
  7:
  8: volume remote2
  9:     type protocol/client
 10:     option remote-subvolume testvol1
 11:     option remote-host 192.168.160.21
 12:     option transport-type socket
 13: end-volume
 14:
 15: volume replicate
 16:     type cluster/replicate
 17:     subvolumes remote1 remote2
 18: end-volume
 19:
 20: volume writebehind
 21:     type performance/write-behind
 22:     option cache-size 1MB
 23:     subvolumes replicate
 24: end-volume
 25:
 26: volume cache
 27:     type performance/io-cache
 28:     option cache-size 256MB
 29:     subvolumes writebehind
 30: end-volume
 31:
+------------------------------------------------------------------------------+
[2014-10-02 14:38:22.310830] E [client-handshake.c:1778:client_query_portmap_cbk] 0-remote1: failed to get the port number for remote subvolume. Please run 'gluster volume status' on server to see if brick process is running.
[2014-10-02 14:38:22.310887] I [client.c:2103:client_rpc_notify] 0-remote1: disconnected from 192.168.12.210:24007. Client process will keep trying to connect to glusterd until brick's port is available.
[2014-10-02 14:38:22.311031] E [client-handshake.c:1778:client_query_portmap_cbk] 0-remote2: failed to get the port number for remote subvolume. Please run 'gluster volume status' on server to see if brick process is running.
[2014-10-02 14:38:22.311059] I [client.c:2103:client_rpc_notify] 0-remote2: disconnected from 192.168.160.21:24007. Client process will keep trying to connect to glusterd until brick's port is available.
[2014-10-02 14:38:22.311070] E [afr-common.c:4025:afr_notify] 0-replicate: All subvolumes are down. Going offline until atleast one of them comes back up.
[2014-10-02 14:38:22.314827] I [fuse-bridge.c:5874:fuse_graph_setup] 0-fuse: switched to graph 0
[2014-10-02 14:38:22.316140] I [fuse-bridge.c:4811:fuse_init] 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.13 kernel 7.13
[2014-10-02 14:38:22.321404] W [fuse-bridge.c:1134:fuse_attr_cbk] 0-glusterfs-fuse: 2: LOOKUP() / => -1 (Transport endpoint is not connected)
[2014-10-02 14:38:22.324731] I [fuse-bridge.c:5715:fuse_thread_proc] 0-fuse: unmounting /mnt/glust
[2014-10-02 14:38:22.324931] W [glusterfsd.c:1099:cleanup_and_exit] (-->/lib64/libc.so.6(clone+0x6d) [0x7f6e2ec5e86d] (-->/lib64/libpthread.so.0(+0x79d1) [0x7f6e2f2f19d1] (-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xcd) [0x4052ad]))) 0-: received signum (15), shutting down
[2014-10-02 14:38:22.324946] I [fuse-bridge.c:6412:fini] 0-fuse: Unmounting '/mnt/glust'.

Regards
Peter H

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux