If you can see fuse version on client 2 is different than fuse version of client 1. Install the fuse -2.7.x release and recompile glusterfs on client 2 machine. It should work. Regards -- Harshavardhana Z Research Inc http://www.zresearch.com/ 2009/7/6 eagleeyes <eagleeyes at 126.com> > How could i remove one of fuse ? > > > client 1 > 2.6.30 > gfs1:~ # dmesg | grep -i fuse > fuse init (API version 7.11) > gfs1:~ # grep -i FUSE_MINOR_VERSION /usr/local/include/fuse/fuse_common.h > #define FUSE_MINOR_VERSION 7 > > #define FUSE_VERSION FUSE_MAKE_VERSION(FUSE_MAJOR_VERSION, FUSE_MINOR_VERSION) > # undef FUSE_MINOR_VERSION > # define FUSE_MINOR_VERSION 5 > # define FUSE_MINOR_VERSION 4 > # define FUSE_MINOR_VERSION 1 > # define FUSE_MINOR_VERSION 1 > > > client 2 > 2.6.16.21-0.8-smp > linux-2ca1:~ # dmesg | grep -i fuse > fuse: module not supported by Novell, setting U taint flag. > fuse init (API version 7.6) > > linux-2ca1:~ # grep -i FUSE_MINOR_VERSION /usr/local/include/fuse/fuse_common.h > #define FUSE_MINOR_VERSION 8 > > #define FUSE_VERSION FUSE_MAKE_VERSION(FUSE_MAJOR_VERSION, FUSE_MINOR_VERSION) > # undef FUSE_MINOR_VERSION > # define FUSE_MINOR_VERSION 5 > # define FUSE_MINOR_VERSION 4 > # define FUSE_MINOR_VERSION 1 > # define FUSE_MINOR_VERSION 1 > > > > > > > 2009-07-06 > ------------------------------ > eagleeyes > ------------------------------ > *????* Harshavardhana > *?????* 2009-07-06 17:18:34 > *????* eagleeyes > *???* Anand Avati; gluster-users > *???* Re: Error : gluster2.0.3rc1 with fuse2.8 > inkernel2.6.30 ,help !!!!! > Eagleyes, > > I think you are using glusterfs with two different versions of fuse API > versions. API versions for 2.6.30 kernel are not compatible with 2.6.16-21 > version. I would suggest you to use same fuse API versions for glusterfs. > Can i have a few details > > 1. dmesg | grep -i fuse (on each clients) > 2. grep -i FUSE_MINOR_VERSION /usr/include/fuse/fuse_common.h (on each > clients) > > Regards > -- > Harshavardhana > Z Research Inc http://www.zresearch.com/ > > > On Mon, Jul 6, 2009 at 12:31 PM, eagleeyes <eagleeyes at 126.com> wrote: > >> HI >> >> 1. I use gluster2.0.3rc2 with fuse init (API version 7.11) in SUSE >> sp10 ,kernel 2.6.30. >> There were some error log : >> pending frames: >> frame : type(1) op(WRITE) >> frame : type(1) op(READ) >> frame : type(1) op(READ) >> frame : type(1) op(READ) >> patchset: 65524f58b29f0b813549412ba6422711a505f5d8 >> signal received: 11 >> configuration details:argp 1 >> backtrace 1 >> dlfcn 1 >> fdatasync 1 >> libpthread 1 >> llistxattr 1 >> setfsid 1 >> spinlock 1 >> epoll.h 1 >> xattr.h 1 >> st_atim.tv_nsec 1 >> package-string: glusterfs 2.0.3rc2 >> [0xffffe400] >> /usr/local/lib/libfuse.so.2(fuse_session_process+0x26)[0xb752fb56] >> /lib/glusterfs/2.0.3rc2/xlator/mount/fuse.so[0xb755de25] >> /lib/libpthread.so.0[0xb7f0d2ab] >> /lib/libc.so.6(__clone+0x5e)[0xb7ea4a4e] >> --------- >> 2. Use glusterfs 2.0.3rc2 with fuse init (API version 7.6) in suse >> sp10, kernel 2.6.16.21-0.8-smp , >> when i expanded dht volumes from four to six ,then i "rm *" in gluster >> directory , there were some error : >> >> [2009-07-06 22:56:23] W [fuse-bridge.c:921:fuse_unlink_cbk] glusterfs-fuse: 1636: RMDIR() /scheduler => -1 (No such file or directory) >> >> [2009-07-06 22:56:23] W [fuse-bridge.c:921:fuse_unlink_cbk] glusterfs-fuse: 1643: RMDIR() /transport => -1 (No such file or directory) >> >> [2009-07-06 22:56:23] W [fuse-bridge.c:921:fuse_unlink_cbk] glusterfs-fuse: 1655: RMDIR() /xlators/cluster => -1 (No such file or directory) >> >> [2009-07-06 22:56:23] W [fuse-bridge.c:921:fuse_unlink_cbk] glusterfs-fuse: 1666: RMDIR() /xlators/debug => -1 (No such file or directory) >> [2009-07-06 22:56:23] W [fuse-bridge.c:921:fuse_unlink_cbk] glusterfs-fuse: 1677: RMDIR() /xlators/mount => -1 (No such file or directory) >> >> >> and new files didn't write into the new volumes after expansion . >> >> >> >> >> 2009-07-06 >> ------------------------------ >> eagleeyes >> ------------------------------ >> *????* Anand Avati >> *?????* 2009-07-06 12:09:13 >> *????* eagleeyes >> *???* gluster-users >> *???* Re: Error : gluster2.0.3rc1 with fuse2.8 in >> kernel2.6.30 ,help !!!!! >> Please use 2.0.3 stable, or upgrade to the next rc2 until then. This >> has been fixed in rc2. >> Avati >> On Mon, Jul 6, 2009 at 8:31 AM, eagleeyes<eagleeyes at 126.com> wrote: >> > HI >> > I use gluster2.0.3rc1 with fuse 2.8 in kernel >> >> > 2.6.30(SUSE Linux Enterprise Server 10 SP1 with kernel 2.6.30 ) . the mount >> > message was : >> > >> > /dev/hda4 on /data type reiserfs (rw,user_xattr) >> >> > glusterfs-client.vol.dht on /home type fuse.glusterfs (rw,allow_other,default_permissions,max_read=131072) >> > >> > >> > >> >> > There was some error when i "touce 111" in gluster directory ,the error was >> > : >> > /home: Transport endpoint is not connected >> > >> > pending frames: >> > patchset: e0db4ff890b591a58332994e37ce6db2bf430213 >> > signal received: 11 >> > configuration details:argp 1 >> > backtrace 1 >> > dlfcn 1 >> > fdatasync 1 >> > libpthread 1 >> > llistxattr 1 >> > setfsid 1 >> > spinlock 1 >> > epoll.h 1 >> > xattr.h 1 >> > st_atim.tv_nsec 1 >> > package-string: glusterfs 2.0.3rc1 >> > [0xffffe400] >> > /lib/glusterfs/2.0.3rc1/xlator/mount/fuse.so[0xb75c6288] >> >> > /lib/glusterfs/2.0.3rc1/xlator/performance/write-behind.so(wb_create_cbk+0xa7)[0xb75ccad7] >> >> > /lib/glusterfs/2.0.3rc1/xlator/performance/io-cache.so(ioc_create_cbk+0xde)[0xb7fbe8ae] >> >> > /lib/glusterfs/2.0.3rc1/xlator/performance/read-ahead.so(ra_create_cbk+0x167)[0xb7fc78b7] >> >> > /lib/glusterfs/2.0.3rc1/xlator/cluster/dht.so(dht_create_cbk+0xf7)[0xb75e25b7] >> >> > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(client_create_cbk+0x2ad)[0xb76004ad] >> >> > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(protocol_client_interpret+0x1ef)[0xb75ef8ff] >> >> > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(protocol_client_pollin+0xcf)[0xb75efaef] >> >> > /lib/glusterfs/2.0.3rc1/xlator/protocol/client.so(notify+0x1ec)[0xb75f6ddc] >> >> > /lib/glusterfs/2.0.3rc1/transport/socket.so(socket_event_poll_in+0x3b)[0xb75b775b] >> >> > /lib/glusterfs/2.0.3rc1/transport/socket.so(socket_event_handler+0xae)[0xb75b7b8e] >> > /lib/libglusterfs.so.0[0xb7facbda] >> > /lib/libglusterfs.so.0(event_dispatch+0x21)[0xb7fabac1] >> > glusterfs(main+0xc2e)[0x804b6ae] >> > /lib/libc.so.6(__libc_start_main+0xdc)[0xb7e6087c] >> > glusterfs[0x8049c11] >> > --------- >> > >> > the server configuration >> > >> > gfs1:/ # cat /etc/glusterfs/glusterfsd-sever.vol >> > volume posix1 >> > type storage/posix # POSIX FS translator >> > option directory /data/data1 # Export this directory >> > end-volume >> > volume posix2 >> > type storage/posix # POSIX FS translator >> > option directory /data/data2 # Export this directory >> > end-volume >> > volume posix3 >> > type storage/posix # POSIX FS translator >> > option directory /data/data3 # Export this directory >> > end-volume >> > volume posix4 >> > type storage/posix # POSIX FS translator >> > option directory /data/data4 # Export this directory >> > end-volume >> > volume posix5 >> > type storage/posix # POSIX FS translator >> > option directory /data/data5 # Export this directory >> > end-volume >> > volume posix6 >> > type storage/posix # POSIX FS translator >> > option directory /data/data6 # Export this directory >> > end-volume >> > volume posix7 >> > type storage/posix # POSIX FS translator >> > option directory /data/data7 # Export this directory >> > end-volume >> > volume posix8 >> > type storage/posix # POSIX FS translator >> > option directory /data/data8 # Export this directory >> > end-volume >> > volume brick1 >> > type features/posix-locks >> >> > option mandatory-locks on # enables mandatory locking on all files >> > subvolumes posix1 >> > end-volume >> > volume brick2 >> > type features/posix-locks >> >> > option mandatory-locks on # enables mandatory locking on all files >> > subvolumes posix2 >> > end-volume >> > volume brick3 >> > type features/posix-locks >> >> > option mandatory-locks on # enables mandatory locking on all files >> > subvolumes posix3 >> > end-volume >> > volume brick4 >> > type features/posix-locks >> >> > option mandatory-locks on # enables mandatory locking on all files >> > subvolumes posix4 >> > end-volume >> > volume brick5 >> > type features/posix-locks >> >> > option mandatory-locks on # enables mandatory locking on all files >> > subvolumes posix5 >> > end-volume >> > volume brick6 >> > type features/posix-locks >> >> > option mandatory-locks on # enables mandatory locking on all files >> > subvolumes posix6 >> > end-volume >> > volume brick7 >> > type features/posix-locks >> >> > option mandatory-locks on # enables mandatory locking on all files >> > subvolumes posix7 >> > end-volume >> > volume brick8 >> > type features/posix-locks >> >> > option mandatory-locks on # enables mandatory locking on all files >> > subvolumes posix8 >> > end-volume >> > ### Add network serving capability to above brick. >> > volume server >> > type protocol/server >> > option transport-type tcp >> >> > option transport.socket.bind-address 172.20.92.240 # Default is to listen on all interfaces >> >> > option transport.socket.listen-port 6996 # Default is 6996 >> > subvolumes brick1 brick2 brick3 brick4 >> > option auth.addr.brick1.allow * # Allow access to "brick" volume >> > option auth.addr.brick2.allow * # Allow access to "brick" volume >> > option auth.addr.brick3.allow * # Allow access to "brick" volume >> > option auth.addr.brick4.allow * # Allow access to "brick" volume >> > option auth.addr.brick5.allow * # Allow access to "brick" volume >> > option auth.addr.brick6.allow * # Allow access to "brick" volume >> > option auth.addr.brick7.allow * # Allow access to "brick" volume >> > option auth.addr.brick8.allow * # Allow access to "brick" volume >> > end-volume >> > >> > the client configuration: >> > >> > gfs1:/ # cat /etc/glusterfs/glusterfs-client.vol.dht >> > volume client1 >> > type protocol/client >> > option transport-type tcp >> >> > option remote-host 172.20.92.240 # IP address of the remote brick2 >> > option remote-port 6996 >> > option remote-subvolume brick1 # name of the remote volume >> > end-volume >> > volume client2 >> > type protocol/client >> > option transport-type tcp >> >> > option remote-host 172.20.92.240 # IP address of the remote brick2 >> > option remote-port 6996 >> > #option transport-timeout 10 # seconds to wait for a reply >> > option remote-subvolume brick2 # name of the remote volume >> > end-volume >> > volume client3 >> > type protocol/client >> > option transport-type tcp >> >> > option remote-host 172.20.92.240 # IP address of the remote brick2 >> > option remote-port 6996 >> > #option transport-timeout 10 # seconds to wait for a reply >> > option remote-subvolume brick3 # name of the remote volume >> > end-volume >> > volume client4 >> > type protocol/client >> > option transport-type tcp >> >> > option remote-host 172.20.92.240 # IP address of the remote brick2 >> > option remote-port 6996 >> > #option transport-timeout 10 # seconds to wait for a reply >> > option remote-subvolume brick4 # name of the remote volume >> > end-volume >> > volume client5 >> > type protocol/client >> > option transport-type tcp >> >> > option remote-host 172.20.92.240 # IP address of the remote brick2 >> > option remote-port 6996 >> > #option transport-timeout 10 # seconds to wait for a reply >> > option remote-subvolume brick1 # name of the remote volume >> > end-volume >> > volume client6 >> > type protocol/client >> > option transport-type tcp >> >> > option remote-host 172.20.92.240 # IP address of the remote brick2 >> > option remote-port 6996 >> > #option transport-timeout 10 # seconds to wait for a reply >> > option remote-subvolume brick2 # name of the remote volume >> > end-volume >> > volume client7 >> > type protocol/client >> > option transport-type tcp >> >> > option remote-host 172.20.92.240 # IP address of the remote brick2 >> > option remote-port 6996 >> > #option transport-timeout 10 # seconds to wait for a reply >> > option remote-subvolume brick3 # name of the remote volume >> > end-volume >> > volume client8 >> > type protocol/client >> > option transport-type tcp >> >> > option remote-host 172.20.92.240 # IP address of the remote brick2 >> > option remote-port 6996 >> > #option transport-timeout 10 # seconds to wait for a reply >> > option remote-subvolume brick4 # name of the remote volume >> > end-volume >> > #volume afr3 >> > # type cluster/afr >> > # subvolumes client3 client6 >> > #end-volume >> > volume dht >> > type cluster/dht >> > option lookup-unhashed yes >> > subvolumes client1 client2 client3 client4 >> > end-volume >> > >> > Could you help me ? >> > >> > >> > >> > 2009-07-06 >> > ________________________________ >> > eagleeyes >> > ________________________________ >> > ???? Sachidananda >> > ????? 2009-07-04 11:39:03 >> > ???? eagleeyes >> > ??? gluster-users >> > ??? Re: HELP : Files lost after DHT expansion >> > Hi, >> > eagleeyes wrote: >> >> > > When i update to gluster2.0.3 ,after dht expansion ,double directorys >> > > appear in the gluster directory ?why ? >> > > >> > > client configure >> > > volume dht >> > > type cluster/dht >> > > option lookup-unhashed yes >> > > option min-free-disk 10% >> > > subvolumes client1 client2 client3 client4 client5 client6 client7 >> > client8 >> > > #subvolumes client1 client2 client3 client4 >> > > end-volume >> > > >> > > >> > Can you please send us your server/client volume files? >> > -- >> > Sachidananda. >> > _______________________________________________ >> > Gluster-users mailing list >> > Gluster-users at gluster.org >> > http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users >> > >> > >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users >> >> > -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://zresearch.com/pipermail/gluster-users/attachments/20090706/21eec691/attachment-0001.htm>