Hi all!
I'm rather new to glusterfs, trying it out for redundant storage
for my very small company.
I have a minimal setup of glusterfs, 2 servers (storage1 &
storage2) with one brick each, both added to volume "testvol1". I
then mount "testvol1" on a third server (app1). This is fine as
long as I use any one servers ip and filesystem glusterfs, but
when trying to mount the volume using a file
"/owndata/conf/glusterfs/storage.vol", it does not work. Also I
can't use nfs - don't need it but maybe it's related.
So:
"mount -t glusterfs 192.168.160.21:/testvol1 /mnt/tmp/" works. I
write a file to /mnt/tmp/filename, then mount
192.168.12.210:/testvol1 and the newly created file is there.
Trying to mount using config file storage.vol:
mount -t glusterfs
/owndata/conf/glusterfs/storage.vol /mnt/tmp
Mount failed. Please check the log file for more
details.
The main error in the log is
E
[client-handshake.c:1778:client_query_portmap_cbk] 0-remote1:
failed to get the port number for remote subvolume. Please run
'gluster volume status' on server to see if brick process is
running.
There are lots & lots of pages on the 'net about
this error message, none of the solutions I've found has worked.
CentOS 6.5 on all
servers, they are all kvm hosts under oVirt (this is just the
testing stage, will be on real iron in production).
No firewall anywhere, selinux is permissive.
File storage.vol:
volume remote1
type protocol/client
option transport-type tcp
option remote-host 192.168.12.210
option remote-subvolume testvol1
end-volume
volume remote2
type protocol/client
option transport-type tcp
option remote-host 192.168.160.21
option remote-subvolume testvol1
end-volume
volume replicate
type cluster/replicate
subvolumes remote1 remote2
end-volume
volume writebehind
type performance/write-behind
option window-size 1MB
subvolumes replicate
end-volume
volume cache
type performance/io-cache
option cache-size 256MB
subvolumes writebehind
end-volume
# gluster volume info
Volume Name: testvol1
Type: Replicate
Volume ID:
bcca4aa2-46c0-44a2-8175-1305faa8b4f9
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 192.168.12.210:/export/vdb/brick1
Brick2: 192.168.160.21:/export/vdb/brick1
# gluster volume status
Status of volume: testvol1
Gluster
process Port
Online Pid
-------------------------------------------------------------------------
Brick
192.168.12.210:/export/vdb/brick1 49152
Y 1656
Brick
192.168.160.21:/export/vdb/brick1 49152
Y 139090
NFS Server on
localhost 2049 Y 1670
Self-heal Daemon on
localhost N/A Y 1674
NFS Server on
192.168.160.21 2049 Y 1481
Self-heal Daemon on
192.168.160.21 N/A Y 139105
Task Status of Volume testvol1
------------------------------------------------------------------------------
There are no active volume tasks
Complete log after fail:
[2014-10-02 14:38:22.252235] I
[glusterfsd.c:2026:main] 0-/usr/sbin/glusterfs: Started running
/usr/sbin/glusterfs version 3.4.0.57rhs (/usr/sbin/glusterfs
--fuse-mountopts=allow_other,default_permissions,max_read=131072
--volfile=/owndata/conf/glusterfs/storage.vol
--fuse-mountopts=allow_other,default_permissions,max_read=131072
/mnt/glust)
[2014-10-02 14:38:22.284438] W
[options.c:848:xl_opt_validate] 0-writebehind: option
'window-size' is deprecated, preferred is 'cache-size',
continuing with correction
[2014-10-02 14:38:22.284476] W [io-cache.c:1672:init]
0-cache: dangling volume. check volfile
[2014-10-02 14:38:22.294306] I
[socket.c:3505:socket_init] 0-remote2: SSL support is NOT
enabled
[2014-10-02 14:38:22.294339] I
[socket.c:3520:socket_init] 0-remote2: using system polling
thread
[2014-10-02 14:38:22.294832] I
[socket.c:3505:socket_init] 0-remote1: SSL support is NOT
enabled
[2014-10-02 14:38:22.294848] I
[socket.c:3520:socket_init] 0-remote1: using system polling
thread
[2014-10-02 14:38:22.294870] I [client.c:2171:notify]
0-remote1: parent translators are ready, attempting connect on
transport
[2014-10-02 14:38:22.306697] I [client.c:2171:notify]
0-remote2: parent translators are ready, attempting connect on
transport
Final graph:
+------------------------------------------------------------------------------+
1: volume remote1
2: type protocol/client
3: option remote-subvolume testvol1
4: option remote-host 192.168.12.210
5: option transport-type socket
6: end-volume
7:
8: volume remote2
9: type protocol/client
10: option remote-subvolume testvol1
11: option remote-host 192.168.160.21
12: option transport-type socket
13: end-volume
14:
15: volume replicate
16: type cluster/replicate
17: subvolumes remote1 remote2
18: end-volume
19:
20: volume writebehind
21: type performance/write-behind
22: option cache-size 1MB
23: subvolumes replicate
24: end-volume
25:
26: volume cache
27: type performance/io-cache
28: option cache-size 256MB
29: subvolumes writebehind
30: end-volume
31:
+------------------------------------------------------------------------------+
[2014-10-02 14:38:22.310830] E
[client-handshake.c:1778:client_query_portmap_cbk] 0-remote1:
failed to get the port number for remote subvolume. Please run
'gluster volume status' on server to see if brick process is
running.
[2014-10-02 14:38:22.310887] I
[client.c:2103:client_rpc_notify] 0-remote1: disconnected from
192.168.12.210:24007. Client process will keep trying to connect
to glusterd until brick's port is available.
[2014-10-02 14:38:22.311031] E
[client-handshake.c:1778:client_query_portmap_cbk] 0-remote2:
failed to get the port number for remote subvolume. Please run
'gluster volume status' on server to see if brick process is
running.
[2014-10-02 14:38:22.311059] I
[client.c:2103:client_rpc_notify] 0-remote2: disconnected from
192.168.160.21:24007. Client process will keep trying to connect
to glusterd until brick's port is available.
[2014-10-02 14:38:22.311070] E
[afr-common.c:4025:afr_notify] 0-replicate: All subvolumes are
down. Going offline until atleast one of them comes back up.
[2014-10-02 14:38:22.314827] I
[fuse-bridge.c:5874:fuse_graph_setup] 0-fuse: switched to graph
0
[2014-10-02 14:38:22.316140] I
[fuse-bridge.c:4811:fuse_init] 0-glusterfs-fuse: FUSE inited
with protocol versions: glusterfs 7.13 kernel 7.13
[2014-10-02 14:38:22.321404] W
[fuse-bridge.c:1134:fuse_attr_cbk] 0-glusterfs-fuse: 2: LOOKUP()
/ => -1 (Transport endpoint is not connected)
[2014-10-02 14:38:22.324731] I
[fuse-bridge.c:5715:fuse_thread_proc] 0-fuse: unmounting
/mnt/glust
[2014-10-02 14:38:22.324931] W
[glusterfsd.c:1099:cleanup_and_exit]
(-->/lib64/libc.so.6(clone+0x6d) [0x7f6e2ec5e86d]
(-->/lib64/libpthread.so.0(+0x79d1) [0x7f6e2f2f19d1]
(-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xcd)
[0x4052ad]))) 0-: received signum (15), shutting down
[2014-10-02 14:38:22.324946] I [fuse-bridge.c:6412:fini]
0-fuse: Unmounting '/mnt/glust'.
Regards
Peter H
|
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://supercolony.gluster.org/mailman/listinfo/gluster-users