On Fri, Oct 03, 2014 at 03:26:04PM +0200, Peter Haraldson wrote: > Hi all! Hi Peter! > I'm rather new to glusterfs, trying it out for redundant storage for my very > small company. > I have a minimal setup of glusterfs, 2 servers (storage1 & storage2) with > one brick each, both added to volume "testvol1". I then mount "testvol1" on > a third server (app1). This is fine as long as I use any one servers ip and > filesystem glusterfs, but when trying to mount the volume using a file > "/owndata/conf/glusterfs/storage.vol", it does not work. Also I can't use > nfs - don't need it but maybe it's related. It is not recommended to use the volume-file for mounting. Current versions of Gluster manage the volume-file for you, and there is no need to make changes there, or use it directly. Mounting with "-t glusterfs ..." is the recommended way. If you would like to fall-back on the 2nd server while mounting, you can use the "backupvolfile-server=storage2" mount option. I can not say why mounting over NFS fails to work. The output of "gluster volume status" below shows that the NFS-server is running and listening on port 2049. You can find the logs for the nfs-server in /var/log/gluster/nfs.log, combine that with the output of # mount -vvv -t nfs storage1:/testvol1 /mnt/tmp to get some ideas on what might go wrong. HTH, Niels > > So: > "mount -t glusterfs 192.168.160.21:/testvol1 /mnt/tmp/" works. I write a > file to /mnt/tmp/filename, then mount 192.168.12.210:/testvol1 and the newly > created file is there. > > Trying to mount using config file storage.vol: > > mount -t glusterfs /owndata/conf/glusterfs/storage.vol /mnt/tmp > Mount failed. Please check the log file for more details. > > The main error in the log is > > /E [client-handshake.c:1778:client_query_portmap_cbk] 0-remote1: > failed to get the port number for remote subvolume. Please run > 'gluster volume status' on server to see if brick process is running/. > > There are lots & lots of pages on the 'net about this error message, none of > the solutions I've found has worked. > > CentOS 6.5 on all servers, they are all kvm hosts under oVirt (this is just > the testing stage, will be on real iron in production). > No firewall anywhere, selinux is permissive. > > *File storage.vol:* > > volume remote1 > type protocol/client > option transport-type tcp > option remote-host 192.168.12.210 > option remote-subvolume testvol1 > end-volume > > volume remote2 > type protocol/client > option transport-type tcp > option remote-host 192.168.160.21 > option remote-subvolume testvol1 > end-volume > > volume replicate > type cluster/replicate > subvolumes remote1 remote2 > end-volume > > volume writebehind > type performance/write-behind > option window-size 1MB > subvolumes replicate > end-volume > > volume cache > type performance/io-cache > option cache-size 256MB > subvolumes writebehind > end-volume > > *# gluster volume info* > > Volume Name: testvol1 > Type: Replicate > Volume ID: bcca4aa2-46c0-44a2-8175-1305faa8b4f9 > Status: Started > Number of Bricks: 1 x 2 = 2 > Transport-type: tcp > Bricks: > Brick1: 192.168.12.210:/export/vdb/brick1 > Brick2: 192.168.160.21:/export/vdb/brick1 > > *# gluster volume status* > > Status of volume: testvol1 > Gluster process Port Online Pid > ------------------------------------------------------------------------- > Brick 192.168.12.210:/export/vdb/brick1 49152 > Y 1656 > Brick 192.168.160.21:/export/vdb/brick1 49152 > Y 139090 > NFS Server on localhost 2049 Y > 1670 > Self-heal Daemon on localhost N/A Y > 1674 > NFS Server on 192.168.160.21 2049 Y > 1481 > Self-heal Daemon on 192.168.160.21 N/A Y > 139105 > > Task Status of Volume testvol1 > ------------------------------------------------------------------------------ > There are no active volume tasks > > *Complete log after fail: > * > > [2014-10-02 14:38:22.252235] I [glusterfsd.c:2026:main] > 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version > 3.4.0.57rhs (/usr/sbin/glusterfs > --fuse-mountopts=allow_other,default_permissions,max_read=131072 > --volfile=/owndata/conf/glusterfs/storage.vol > --fuse-mountopts=allow_other,default_permissions,max_read=131072 > /mnt/glust) > [2014-10-02 14:38:22.284438] W [options.c:848:xl_opt_validate] > 0-writebehind: option 'window-size' is deprecated, preferred is > 'cache-size', continuing with correction > [2014-10-02 14:38:22.284476] W [io-cache.c:1672:init] 0-cache: > dangling volume. check volfile > [2014-10-02 14:38:22.294306] I [socket.c:3505:socket_init] > 0-remote2: SSL support is NOT enabled > [2014-10-02 14:38:22.294339] I [socket.c:3520:socket_init] > 0-remote2: using system polling thread > [2014-10-02 14:38:22.294832] I [socket.c:3505:socket_init] > 0-remote1: SSL support is NOT enabled > [2014-10-02 14:38:22.294848] I [socket.c:3520:socket_init] > 0-remote1: using system polling thread > [2014-10-02 14:38:22.294870] I [client.c:2171:notify] 0-remote1: > parent translators are ready, attempting connect on transport > [2014-10-02 14:38:22.306697] I [client.c:2171:notify] 0-remote2: > parent translators are ready, attempting connect on transport > Final graph: > +------------------------------------------------------------------------------+ > 1: volume remote1 > 2: type protocol/client > 3: option remote-subvolume testvol1 > 4: option remote-host 192.168.12.210 > 5: option transport-type socket > 6: end-volume > 7: > 8: volume remote2 > 9: type protocol/client > 10: option remote-subvolume testvol1 > 11: option remote-host 192.168.160.21 > 12: option transport-type socket > 13: end-volume > 14: > 15: volume replicate > 16: type cluster/replicate > 17: subvolumes remote1 remote2 > 18: end-volume > 19: > 20: volume writebehind > 21: type performance/write-behind > 22: option cache-size 1MB > 23: subvolumes replicate > 24: end-volume > 25: > 26: volume cache > 27: type performance/io-cache > 28: option cache-size 256MB > 29: subvolumes writebehind > 30: end-volume > 31: > +------------------------------------------------------------------------------+ > [2014-10-02 14:38:22.310830] E > [client-handshake.c:1778:client_query_portmap_cbk] 0-remote1: failed > to get the port number for remote subvolume. Please run 'gluster > volume status' on server to see if brick process is running. > [2014-10-02 14:38:22.310887] I [client.c:2103:client_rpc_notify] > 0-remote1: disconnected from 192.168.12.210:24007. Client process > will keep trying to connect to glusterd until brick's port is > available. > [2014-10-02 14:38:22.311031] E > [client-handshake.c:1778:client_query_portmap_cbk] 0-remote2: failed > to get the port number for remote subvolume. Please run 'gluster > volume status' on server to see if brick process is running. > [2014-10-02 14:38:22.311059] I [client.c:2103:client_rpc_notify] > 0-remote2: disconnected from 192.168.160.21:24007. Client process > will keep trying to connect to glusterd until brick's port is > available. > [2014-10-02 14:38:22.311070] E [afr-common.c:4025:afr_notify] > 0-replicate: All subvolumes are down. Going offline until atleast > one of them comes back up. > [2014-10-02 14:38:22.314827] I [fuse-bridge.c:5874:fuse_graph_setup] > 0-fuse: switched to graph 0 > [2014-10-02 14:38:22.316140] I [fuse-bridge.c:4811:fuse_init] > 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.13 > kernel 7.13 > [2014-10-02 14:38:22.321404] W [fuse-bridge.c:1134:fuse_attr_cbk] > 0-glusterfs-fuse: 2: LOOKUP() / => -1 (Transport endpoint is not > connected) > [2014-10-02 14:38:22.324731] I [fuse-bridge.c:5715:fuse_thread_proc] > 0-fuse: unmounting /mnt/glust > [2014-10-02 14:38:22.324931] W [glusterfsd.c:1099:cleanup_and_exit] > (-->/lib64/libc.so.6(clone+0x6d) [0x7f6e2ec5e86d] > (-->/lib64/libpthread.so.0(+0x79d1) [0x7f6e2f2f19d1] > (-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xcd) [0x4052ad]))) 0-: > received signum (15), shutting down > [2014-10-02 14:38:22.324946] I [fuse-bridge.c:6412:fini] 0-fuse: > Unmounting '/mnt/glust'. > > > Regards > Peter H > > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://supercolony.gluster.org/mailman/listinfo/gluster-users _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://supercolony.gluster.org/mailman/listinfo/gluster-users