BTW, is there someone have configured gluster like HDFS (unify with automatic replication). Could someone share the volfile here? I think I'm not the only fan waiting here ;-) Gnep On Wed, Jul 1, 2009 at 6:39 PM, Peng Zhao <blackass at gmail.com> wrote: > OK, my stupid. There was no fuse module. I built one and modprobe fuse. The > previous error is gone, with some new one: > Here are the DEBUG-level msg: > [2009-07-01 18:36:25] D [xlator.c:634:xlator_set_type] xlator: > dlsym(notify) on /usr/lib64/glusterfs/2.0.2/xlator/features/locks.so: > undefined symbol: notify -- neglecting > [2009-07-01 18:36:25] D [xlator.c:634:xlator_set_type] xlator: > dlsym(notify) on > /usr/lib64/glusterfs/2.0.2/xlator/performance/io-threads.so: undefined > symbol: notify -- neglecting > [2009-07-01 18:36:25] D [xlator.c:634:xlator_set_type] xlator: > dlsym(notify) on > /usr/lib64/glusterfs/2.0.2/xlator/performance/write-behind.so: undefined > symbol: notify -- neglecting > [2009-07-01 18:36:25] D [xlator.c:634:xlator_set_type] xlator: > dlsym(notify) on /usr/lib64/glusterfs/2.0.2/xlator/performance/io-cache.so: > undefined symbol: notify -- neglecting > [2009-07-01 18:36:25] D [glusterfsd.c:1179:main] glusterfs: running in pid > 6874 > [2009-07-01 18:36:25] D [client-protocol.c:5948:init] compute-5-0: > defaulting frame-timeout to 30mins > [2009-07-01 18:36:25] D [client-protocol.c:5959:init] compute-5-0: > defaulting ping-timeout to 10 > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] D [client-protocol.c:5948:init] compute-5-1: > defaulting frame-timeout to 30mins > [2009-07-01 18:36:25] D [client-protocol.c:5959:init] compute-5-1: > defaulting ping-timeout to 10 > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] D [client-protocol.c:5948:init] compute-5-2: > defaulting frame-timeout to 30mins > [2009-07-01 18:36:25] D [client-protocol.c:5959:init] compute-5-2: > defaulting ping-timeout to 10 > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] D [client-protocol.c:5948:init] compute-5-3: > defaulting frame-timeout to 30mins > [2009-07-01 18:36:25] D [client-protocol.c:5959:init] compute-5-3: > defaulting ping-timeout to 10 > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] D [unify.c:4288:init] unified: namespace node > specified as compute-5-4 > [2009-07-01 18:36:25] D [scheduler.c:48:get_scheduler] scheduler: attempt > to load file rr.so > [2009-07-01 18:36:25] D [unify.c:4320:init] unified: Child node count is 2 > [2009-07-01 18:36:25] D [rr-options.c:188:rr_options_validate] rr: using > scheduler.limits.min-free-disk = 15 [default] > [2009-07-01 18:36:25] D [rr-options.c:216:rr_options_validate] rr: using > scheduler.refresh-interval = 10 [default] > [2009-07-01 18:36:25] D [client-protocol.c:5948:init] compute-5-4: > defaulting frame-timeout to 30mins > [2009-07-01 18:36:25] D [client-protocol.c:5959:init] compute-5-4: > defaulting ping-timeout to 10 > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-4: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-4: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-4: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-4: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [write-behind.c:1859:init] writebehind: disabling > write-behind for first 1 bytes > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-0: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-0: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-1: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-1: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-2: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-2: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-3: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [client-protocol.c:6276:notify] compute-5-3: got > GF_EVENT_PARENT_UP, attempting connect on transport > [2009-07-01 18:36:25] D [io-threads.c:2280:init] brick: io-threads: > Autoscaling: off, min_threads: 16, max_threads: 16 > [2009-07-01 18:36:25] D [transport.c:141:transport_load] transport: attempt > to load file /usr/lib64/glusterfs/2.0.2/transport/socket.so > [2009-07-01 18:36:25] E [socket.c:206:__socket_server_bind] server: binding > to failed: Address already in use > [2009-07-01 18:36:25] E [socket.c:209:__socket_server_bind] server: Port is > already in use > [2009-07-01 18:36:25] E [server-protocol.c:7631:init] server: failed to > bind/listen on socket > [2009-07-01 18:36:25] E [xlator.c:736:xlator_init_rec] xlator: > Initialization of volume 'server' failed, review your volfile again > [2009-07-01 18:36:25] E [glusterfsd.c:498:_xlator_graph_init] glusterfs: > initializing translator failed > [2009-07-01 18:36:25] E [glusterfsd.c:1191:main] glusterfs: translator > initialization failed. exiting > [root at compute-5-0 gluster]# ps aux | grep gluster > > I think my volfile is wrong, though don't know where is the problem. > BR, > Gnep > > On Wed, Jul 1, 2009 at 2:41 PM, Shehjar Tikoo <shehjart at gluster.com>wrote: > >> Peng Zhao wrote: >> >>> Hi, all, >>> I'm new to gluster, but found it interesting. I want to setup gluster in >>> a way to be similar with HDFS. >>> There is my sample vol-file: >>> volume posix >>> type storage/posix >>> option directory /data1/gluster >>> end-volume >>> >>> volume locks >>> type features/locks >>> subvolumes posix >>> end-volume >>> >>> volume brick >>> type performance/io-threads >>> subvolumes locks >>> end-volume >>> >>> volume server >>> type protocol/server >>> option transport-type tcp >>> option auth.addr.brick.allow * >>> subvolumes brick >>> end-volume >>> >>> volume compute-5-0 >>> type protocol/client >>> option transport-type tcp >>> option remote-host compute-5-0 >>> option remote-subvolume brick >>> end-volume >>> >>> volume compute-5-1 >>> type protocol/client >>> option transport-type tcp >>> option remote-host compute-5-1 >>> option remote-subvolume brick >>> end-volume >>> >>> volume compute-5-2 >>> type protocol/client >>> option transport-type tcp >>> option remote-host compute-5-2 >>> option remote-subvolume brick >>> end-volume >>> >>> volume compute-5-3 >>> type protocol/client >>> option transport-type tcp >>> option remote-host compute-5-3 >>> option remote-subvolume brick >>> end-volume >>> >>> volume compute-5-4 >>> type protocol/client >>> option transport-type tcp >>> option remote-host compute-5-4 >>> option remote-subvolume brick-ns >>> end-volume >>> >>> volume primary >>> type cluster/replicate >>> option local-volume-name primary >>> subvolumes compute-5-0 compute-5-1 >>> end-volume >>> >>> volume secondary >>> type cluster/replicate >>> option local-volume-name secondary >>> subvolumes compute-5-2 compute-5-3 >>> end-volume >>> >>> volume unified >>> type cluster/unify >>> option scheduler rr >>> option local-volume-name unified # do I need this? >>> option namespace compute-5-4 # do I need this? >>> subvolumes primary secondary >>> end-volume >>> >>> volume writebehind >>> type performance/write-behind >>> option cache-size 1MB >>> subvolumes unified >>> end-volume >>> >>> volume cache >>> type performance/io-cache >>> option cache-size 512MB >>> subvolumes writebehind >>> end-volume >>> >>> The glusterd is up & running and no error msg in the logs. However, it >>> reports some error when I tried to mount it: >>> [2009-07-01 09:37:36] E [xlator.c:736:xlator_init_rec] xlator: >>> Initialization of volume 'fuse' failed, review your volfile again >>> [2009-07-01 09:37:36] E [glusterfsd.c:498:_xlator_graph_init] glusterfs: >>> initializing translator failed >>> [2009-07-01 09:37:36] E [glusterfsd.c:1191:main] glusterfs: translator >>> initialization failed. exiting >>> >>> I guess it is a very common question. Anyone has any idea? >>> BR, >>> >> >> Try generating the log file with log-level set to bug. You >> can do so by using the "-L DEBUG" command line parameter. >> >> The debug log level will give us a better idea of what >> exactly is failing. >> >> -Shehjar >> >>> Gnep >>> >>> >>> ------------------------------------------------------------------------ >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users >>> >> >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://zresearch.com/pipermail/gluster-users/attachments/20090701/ffd67097/attachment-0001.htm>