That worked for me. Thanks! On Thu, Nov 4, 2010 at 12:46 AM, Lakshmipathi <lakshmipathi at gluster.com> wrote: > Hi Jeremy- > Did you start the volume - like > n128:~ # gluster volume start test > before mounting it ? If not,Please refer - http://www.gluster.com/community/documentation/index.php/Gluster_3.1:_Configuring_Distributed_Volumes > > > -- > ---- > Cheers, > Lakshmipathi.G > FOSS Programmer. > > > ----- Original Message ----- > From: "Jeremy Stout" <stout.jeremy at gmail.com> > To: gluster-users at gluster.org > Sent: Thursday, November 4, 2010 12:11:38 AM > Subject: Problems Mounting GlusterFS 3.1 Volumes > > Hello. I'm having problems with GlusterFS 3.1. Whenever I mount a > GlusterFS volume and perform a 'df' command, the command hangs. The > same thing also happens if I mount the volume and try to perform a > directory listing on it. This happens if I try to mount the volume > locally or over the Infiniband network. > > Here is a breakdown of my setup: > Platform: x86_64 > Distro: openSUSE 11.3 > Kernel: 2.6.32.22 > OFED: 1.5.1 > GlusterFS: 3.1 > > Assuming I've just performed a fresh install and started the daemon, > here is what I do: > n128:~ # gluster volume create test transport rdma n128:/scratch/ > Creation of volume test has been successful > n128:~ # mount -t glusterfs n128:/test /mnt/glusterfs/ > n128:~ # df > Filesystem ? ? ? ? ? 1K-blocks ? ? ?Used Available Use% Mounted on > /dev/sda1 ? ? ? ? ? ?103210272 ? 6269264 ?91698232 ? 7% / > devtmpfs ? ? ? ? ? ? ?12358632 ? ? ? 176 ?12358456 ? 1% /dev > tmpfs ? ? ? ? ? ? ? ? 12367684 ? ? ? ? 4 ?12367680 ? 1% /dev/shm > /dev/sda3 ? ? ? ? ? ?104804356 ? 2004256 102800100 ? 2% /scratch > /dev/sda4 ? ? ? ? ? ? ?9174072 ? ?151580 ? 8556472 ? 2% /spool > > It hangs while trying to list /mnt/glusterfs. If I manually kill the > glusterfs mount process in a second terminal, the terminal running the > df command will return. > > Here are the log files: > n128:/var/log/glusterfs # cat /var/log/glusterfs/etc-glusterfs-glusterd.vol.log > [2010-11-03 14:12:17.261374] I [glusterd.c:274:init] management: Using > /etc/glusterd as working directory > [2010-11-03 14:12:17.284874] I [glusterd.c:86:glusterd_uuid_init] > glusterd: retrieved UUID: 869f621b-622c-4ab9-9f61-3ee602c8ddf6 > Given volfile: > +------------------------------------------------------------------------------+ > ?1: volume management > ?2: ? ? type mgmt/glusterd > ?3: ? ? option working-directory /etc/glusterd > ?4: ? ? option transport-type socket,rdma > ?5: ? ? option transport.socket.keepalive-time 10 > ?6: ? ? option transport.socket.keepalive-interval 2 > ?7: end-volume > ?8: > > +------------------------------------------------------------------------------+ > [2010-11-03 14:20:43.874738] I > [glusterd-handler.c:775:glusterd_handle_create_volume] glusterd: > Received create volume req > [2010-11-03 14:20:43.876305] I [glusterd-utils.c:223:glusterd_lock] > glusterd: Cluster lock held by 869f621b-622c-4ab9-9f61-3ee602c8ddf6 > [2010-11-03 14:20:43.876326] I > [glusterd-handler.c:2653:glusterd_op_txn_begin] glusterd: Acquired > local lock > [2010-11-03 14:20:43.876337] I > [glusterd-op-sm.c:5061:glusterd_op_sm_inject_event] glusterd: > Enqueuing event: 'GD_OP_EVENT_START_LOCK' > [2010-11-03 14:20:43.876357] I [glusterd-op-sm.c:5109:glusterd_op_sm] > : Dequeued event of type: 'GD_OP_EVENT_START_LOCK' > [2010-11-03 14:20:43.876369] I > [glusterd3_1-mops.c:1105:glusterd3_1_cluster_lock] glusterd: Sent lock > req to 0 peers > [2010-11-03 14:20:43.876378] I > [glusterd-op-sm.c:5061:glusterd_op_sm_inject_event] glusterd: > Enqueuing event: 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.876386] I > [glusterd-op-sm.c:4740:glusterd_op_sm_transition_state] : > Transitioning from 'Default' to 'Lock sent' due to event > 'GD_OP_EVENT_START_LOCK' > [2010-11-03 14:20:43.876395] I [glusterd-op-sm.c:5109:glusterd_op_sm] > : Dequeued event of type: 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.876552] I > [glusterd3_1-mops.c:1247:glusterd3_1_stage_op] glusterd: Sent op req > to 0 peers > [2010-11-03 14:20:43.876568] I > [glusterd-op-sm.c:5061:glusterd_op_sm_inject_event] glusterd: > Enqueuing event: 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.876576] I > [glusterd-op-sm.c:4740:glusterd_op_sm_transition_state] : > Transitioning from 'Lock sent' to 'Stage op sent' due to event > 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.876584] I [glusterd-op-sm.c:5109:glusterd_op_sm] > : Dequeued event of type: 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.876592] I > [glusterd-op-sm.c:5061:glusterd_op_sm_inject_event] glusterd: > Enqueuing event: 'GD_OP_EVENT_STAGE_ACC' > [2010-11-03 14:20:43.876600] I > [glusterd-op-sm.c:5061:glusterd_op_sm_inject_event] glusterd: > Enqueuing event: 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.876608] I > [glusterd-op-sm.c:4740:glusterd_op_sm_transition_state] : > Transitioning from 'Stage op sent' to 'Stage op sent' due to event > 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.876615] I [glusterd-op-sm.c:5109:glusterd_op_sm] > : Dequeued event of type: 'GD_OP_EVENT_STAGE_ACC' > [2010-11-03 14:20:43.880391] I > [glusterd3_1-mops.c:1337:glusterd3_1_commit_op] glusterd: Sent op req > to 0 peers > [2010-11-03 14:20:43.880426] I > [glusterd-op-sm.c:5061:glusterd_op_sm_inject_event] glusterd: > Enqueuing event: 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.880436] I > [glusterd-op-sm.c:4740:glusterd_op_sm_transition_state] : > Transitioning from 'Stage op sent' to 'Commit op sent' due to event > 'GD_OP_EVENT_STAGE_ACC' > [2010-11-03 14:20:43.880445] I [glusterd-op-sm.c:5109:glusterd_op_sm] > : Dequeued event of type: 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.880455] I > [glusterd3_1-mops.c:1159:glusterd3_1_cluster_unlock] glusterd: Sent > unlock req to 0 peers > [2010-11-03 14:20:43.880463] I > [glusterd-op-sm.c:5061:glusterd_op_sm_inject_event] glusterd: > Enqueuing event: 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.880470] I > [glusterd-op-sm.c:4740:glusterd_op_sm_transition_state] : > Transitioning from 'Commit op sent' to 'Unlock sent' due to event > 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.880478] I [glusterd-op-sm.c:5109:glusterd_op_sm] > : Dequeued event of type: 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.880491] I > [glusterd-op-sm.c:4561:glusterd_op_txn_complete] glusterd: Cleared > local lock > [2010-11-03 14:20:43.880562] I > [glusterd-op-sm.c:4740:glusterd_op_sm_transition_state] : > Transitioning from 'Unlock sent' to 'Default' due to event > 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.880576] I [glusterd-op-sm.c:5109:glusterd_op_sm] > : Dequeued event of type: 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:43.880584] I > [glusterd-op-sm.c:4740:glusterd_op_sm_transition_state] : > Transitioning from 'Default' to 'Default' due to event > 'GD_OP_EVENT_ALL_ACC' > [2010-11-03 14:20:57.327303] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: rdma.management: pollin received on tcp socket > (peer: 192.168.40.128:1021) after handshake is complete > [2010-11-03 14:21:01.166661] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: rdma.management: pollin received on tcp socket > (peer: 192.168.40.128:1023) after handshake is complete > [2010-11-03 14:21:05.169694] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: rdma.management: pollin received on tcp socket > (peer: 192.168.40.128:1020) after handshake is complete > [2010-11-03 14:21:09.172853] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: rdma.management: pollin received on tcp socket > (peer: 192.168.40.128:1019) after handshake is complete > [2010-11-03 14:21:13.176023] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: rdma.management: pollin received on tcp socket > (peer: 192.168.40.128:1018) after handshake is complete > [2010-11-03 14:21:17.179219] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: rdma.management: pollin received on tcp socket > (peer: 192.168.40.128:1017) after handshake is complete > [2010-11-03 14:21:21.182370] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: rdma.management: pollin received on tcp socket > (peer: 192.168.40.128:1016) after handshake is complete > [2010-11-03 14:21:25.185607] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: rdma.management: pollin received on tcp socket > (peer: 192.168.40.128:1015) after handshake is complete > [2010-11-03 14:21:29.188771] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: rdma.management: pollin received on tcp socket > (peer: 192.168.40.128:1014) after handshake is complete > > > n128:/var/log/glusterfs # cat /var/log/glusterfs/mnt-glusterfs-.log > [2010-11-03 14:20:57.163632] W [io-stats.c:1637:init] test: dangling > volume. check volfile > [2010-11-03 14:20:57.163670] W [dict.c:1204:data_to_str] dict: @data=(nil) > [2010-11-03 14:20:57.163681] W [dict.c:1204:data_to_str] dict: @data=(nil) > Given volfile: > +------------------------------------------------------------------------------+ > ?1: volume test-client-0 > ?2: ? ? type protocol/client > ?3: ? ? option remote-host n128 > ?4: ? ? option remote-subvolume /scratch > ?5: ? ? option transport-type rdma > ?6: end-volume > ?7: > ?8: volume test-write-behind > ?9: ? ? type performance/write-behind > ?10: ? ? subvolumes test-client-0 > ?11: end-volume > ?12: > ?13: volume test-read-ahead > ?14: ? ? type performance/read-ahead > ?15: ? ? subvolumes test-write-behind > ?16: end-volume > ?17: > ?18: volume test-io-cache > ?19: ? ? type performance/io-cache > ?20: ? ? subvolumes test-read-ahead > ?21: end-volume > ?22: > ?23: volume test-quick-read > ?24: ? ? type performance/quick-read > ?25: ? ? subvolumes test-io-cache > ?26: end-volume > ?27: > ?28: volume test > ?29: ? ? type debug/io-stats > ?30: ? ? subvolumes test-quick-read > ?31: end-volume > > +------------------------------------------------------------------------------+ > [2010-11-03 14:20:57.327245] E > [client-handshake.c:773:client_query_portmap_cbk] test-client-0: > failed to get the port number for remote subvolume > [2010-11-03 14:20:57.327308] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: test-client-0: pollin received on tcp socket > (peer: 192.168.40.128:24008) after handshake is complete > [2010-11-03 14:21:01.166622] E > [client-handshake.c:773:client_query_portmap_cbk] test-client-0: > failed to get the port number for remote subvolume > [2010-11-03 14:21:01.166667] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: test-client-0: pollin received on tcp socket > (peer: 192.168.40.128:24008) after handshake is complete > [2010-11-03 14:21:05.169655] E > [client-handshake.c:773:client_query_portmap_cbk] test-client-0: > failed to get the port number for remote subvolume > [2010-11-03 14:21:05.169699] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: test-client-0: pollin received on tcp socket > (peer: 192.168.40.128:24008) after handshake is complete > [2010-11-03 14:21:09.172815] E > [client-handshake.c:773:client_query_portmap_cbk] test-client-0: > failed to get the port number for remote subvolume > [2010-11-03 14:21:09.172858] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: test-client-0: pollin received on tcp socket > (peer: 192.168.40.128:24008) after handshake is complete > [2010-11-03 14:21:13.175981] E > [client-handshake.c:773:client_query_portmap_cbk] test-client-0: > failed to get the port number for remote subvolume > [2010-11-03 14:21:13.176028] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: test-client-0: pollin received on tcp socket > (peer: 192.168.40.128:24008) after handshake is complete > [2010-11-03 14:21:17.179181] E > [client-handshake.c:773:client_query_portmap_cbk] test-client-0: > failed to get the port number for remote subvolume > [2010-11-03 14:21:17.179226] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: test-client-0: pollin received on tcp socket > (peer: 192.168.40.128:24008) after handshake is complete > [2010-11-03 14:21:21.182325] E > [client-handshake.c:773:client_query_portmap_cbk] test-client-0: > failed to get the port number for remote subvolume > [2010-11-03 14:21:21.182375] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: test-client-0: pollin received on tcp socket > (peer: 192.168.40.128:24008) after handshake is complete > [2010-11-03 14:21:25.185567] E > [client-handshake.c:773:client_query_portmap_cbk] test-client-0: > failed to get the port number for remote subvolume > [2010-11-03 14:21:25.185611] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: test-client-0: pollin received on tcp socket > (peer: 192.168.40.128:24008) after handshake is complete > [2010-11-03 14:21:29.188732] E > [client-handshake.c:773:client_query_portmap_cbk] test-client-0: > failed to get the port number for remote subvolume > [2010-11-03 14:21:29.188776] E [rdma.c:4370:rdma_event_handler] > rpc-transport/rdma: test-client-0: pollin received on tcp socket > (peer: 192.168.40.128:24008) after handshake is complete > > I've tried different versions of the mount command, but receive the > same results. > > Any help would be appreciated. > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >