2) using gluster-client (same version as server 3.6.2 ) to
mount my gluster volume , im getting the following messages in
the gluster log.
1: volume gfsvolume-client-0
2: type protocol/client
3: option ping-timeout 30
4: option remote-host gfs1
5: option remote-subvolume /export/sda/brick
6: option transport-type socket
7: option frame-timeout 90
8: option send-gids true
9: end-volume
10:
11: volume gfsvolume-client-1
12: type protocol/client
13: option ping-timeout 30
14: option remote-host gfs2
15: option remote-subvolume /export/sda/brick
16: option transport-type socket
17: option frame-timeout 90
18: option send-gids true
19: end-volume
20:
21: volume gfsvolume-replicate-0
22: type cluster/replicate
23: option data-self-heal-algorithm diff
24: option quorum-type fixed
25: option quorum-count 1
26: subvolumes gfsvolume-client-0 gfsvolume-client-1
27: end-volume
28:
29: volume gfsvolume-dht
30: type cluster/distribute
31: subvolumes gfsvolume-replicate-0
32: end-volume
33:
34: volume gfsvolume-write-behind
35: type performance/write-behind
36: option cache-size 4MB
37: subvolumes gfsvolume-dht
38: end-volume
39:
40: volume gfsvolume-read-ahead
41: type performance/read-ahead
42: subvolumes gfsvolume-write-behind
43: end-volume
44:
45: volume gfsvolume-io-cache
46: type performance/io-cache
47: option max-file-size 2MB
48: option cache-timeout 60
49: option cache-size 6442450944
50: subvolumes gfsvolume-read-ahead
51: end-volume
52:
53: volume gfsvolume-open-behind
54: type performance/open-behind
55: subvolumes gfsvolume-io-cache
56: end-volume
57:
58: volume gfsvolume-md-cache
59: type performance/md-cache
60: subvolumes gfsvolume-open-behind
61: end-volume
62:
63: volume gfsvolume
[2015-07-16 03:31:27.231127] E [MSGID: 108006]
[afr-common.c:3591:afr_notify] 0-gfsvolume-replicate-0:
All subvolumes are down. Going offline until atleast one
of them comes back up.
[2015-07-16 03:31:27.232201] W [MSGID: 108001]
[afr-common.c:3635:afr_notify] 0-gfsvolume-replicate-0:
Client-quorum is not met
64: type debug/io-stats
65: option latency-measurement on
66: option count-fop-hits on
67: subvolumes gfsvolume-md-cache
68: end-volume
69:
70: volume meta-autoload
71: type meta
72: subvolumes gfsvolume
73: end-volume
74:
+------------------------------------------------------------------------------+
[2015-07-16 03:31:27.254509] I
[fuse-bridge.c:5080:fuse_graph_setup] 0-fuse: switched to
graph 0
[2015-07-16 03:31:27.255262] I
[fuse-bridge.c:4009:fuse_init] 0-glusterfs-fuse: FUSE
inited with protocol versions: glusterfs 7.22 kernel 7.17
[2015-07-16 03:31:27.256340] I
[afr-common.c:3722:afr_local_init]
0-gfsvolume-replicate-0: no subvolumes up
[2015-07-16 03:31:27.256722] I
[afr-common.c:3722:afr_local_init]
0-gfsvolume-replicate-0: no subvolumes up
[2015-07-16 03:31:27.256840] W
[fuse-bridge.c:779:fuse_attr_cbk] 0-glusterfs-fuse: 2:
LOOKUP() / => -1 (Transport endpoint is not connected)
[2015-07-16 03:31:27.284927] I
[fuse-bridge.c:4921:fuse_thread_proc] 0-fuse: unmounting
/export
[2015-07-16 03:31:27.285919] W
[glusterfsd.c:1194:cleanup_and_exit] (--> 0-: received
signum (15), shutting down
[2015-07-16 03:31:27.286052] I [fuse-bridge.c:5599:fini]
0-fuse: Unmounting '/export'.
I have deployed gluster successfully on my physical machines and
my network is fine. I used the exact same steps to install the
gluster client on my VM and my physical machines. Some info you
may find helpful in narrowing down the error is as below:
-My physical machines are fine and show no error messages
in log.
-I am able to ssh into the machine so it is not a
network problem.
-there are no firewall rules on the gluster server and
client .
-I have done modprobe fuse successfully.
- NFS is installed on all machines.
- gluster versions are same on all machines
- gluster volume status shows all processes are online
- gluster status client does not show my client machine
-server can ping to client and client can ping the server
-i tried restarting rpcbind but it didnt fix my problem
If there is anything else you would need to help me fix
this , please do let me know and ill pass the info on to you .
Eid Mubarak to all those who are celebrating and thank you
in advance for your assistance.
Thanks,
Kaamesh