KwangErn, Can you give us the setup details? Thanks Krishna On Sat, Sep 13, 2008 at 2:12 PM, KwangErn Liew <ke.liew@xxxxxxxxx> wrote: > Hm, it seems to be 'normal'. I have just ran dd across the network and this > is what I have... > > $ dd if=/dev/zero of=/home/storage/testfile bs=16384k count=100 > 100+0 records in > 100+0 records out > 1677721600 bytes (1.7 GB) copied, 82.5314 seconds, 20.3 MB/s > > $ ls storage -l > total 1638408 > drwxr-xr-x 2 bonnie nogroup 21 2008-09-13 08:34 . > drwxr-xr-x 7 root root 4096 2008-09-11 21:28 .. > -rw-r--r-- 1 root root 1677721600 2008-09-13 08:35 testfile > > ==== Client ==== > 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] glusterfs-fuse: > 458332: WRITE => 13 > 1072/131072,1677459456/1677590528 > 2008-09-13 08:35:29 D [fuse-bridge.c:1548:fuse_write] glusterfs-fuse: > 458333: WRITE (0x516760, > size=131072, offset=1677590528) > 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] glusterfs-fuse: > 458333: WRITE => 131072/131072,1677590528/1677721600 > 2008-09-13 08:35:29 D [fuse-bridge.c:1572:fuse_flush] glusterfs-fuse: > 458334: FLUSH 0x516760 > 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: > 458334: FLUSH() ERR => 0 > 2008-09-13 08:35:29 D [fuse-bridge.c:1599:fuse_release] glusterfs-fuse: > 458335: CLOSE 0x516760 > 2008-09-13 08:35:29 E [afr.c:3012:afr_close_cbk] gfs-afr: (path=/testfile > child=gfs-client2) op_ret=-1 op_errno=77(File descriptor in bad state) > 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: > 458335: CLOSE() ERR => 0 > 2008-09-13 08:35:29 D [inode.c:287:__inode_passivate] fuse/inode: > passivating inode(2013354535) lru=1/0 active=1 purge=30 > 2008-09-13 08:35:31 D [client-protocol.c:4126:client_protocol_reconnect] > gfs-client2: attempting reconnect > 2008-09-13 08:35:31 D [socket.c:924:socket_connect] gfs-client2: connect () > called on transport already connected > 2008-09-13 08:35:39 D [client-protocol.c:4065:client_setvolume_cbk] > gfs-client2: SETVOLUME on remote-host succeeded > 2008-09-13 08:35:39 D [afr.c:6309:notify] gfs-afr: GF_EVENT_CHILD_UP from > gfs-client2 > 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] glusterfs-fuse: > 458332: WRITE => 13 > 1072/131072,1677459456/1677590528 > 2008-09-13 08:35:29 D [fuse-bridge.c:1548:fuse_write] glusterfs-fuse: > 458333: WRITE (0x516760, > size=131072, offset=1677590528) > 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] glusterfs-fuse: > 458333: WRITE => 131072/131072,1677590528/1677721600 > 2008-09-13 08:35:29 D [fuse-bridge.c:1572:fuse_flush] glusterfs-fuse: > 458334: FLUSH 0x516760 > 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: > 458334: FLUSH() ERR => 0 > 2008-09-13 08:35:29 D [fuse-bridge.c:1599:fuse_release] glusterfs-fuse: > 458335: CLOSE 0x516760 > 2008-09-13 08:35:29 E [afr.c:3012:afr_close_cbk] gfs-afr: (path=/testfile > child=gfs-client2) op_ret=-1 op_errno=77(File descriptor in bad state) > 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: > 458335: CLOSE() ERR => 0 > 2008-09-13 08:35:29 D [inode.c:287:__inode_passivate] fuse/inode: > passivating inode(2013354535) lru=1/0 active=1 purge=30 > 2008-09-13 08:35:31 D [client-protocol.c:4126:client_protocol_reconnect] > gfs-client2: attempting reconnect > 2008-09-13 08:35:31 D [socket.c:924:socket_connect] gfs-client2: connect () > called on transport already connected > 2008-09-13 08:35:39 D [client-protocol.c:4065:client_setvolume_cbk] > gfs-client2: SETVOLUME on remote-host succeeded > 2008-09-13 08:35:39 D [afr.c:6309:notify] gfs-afr: GF_EVENT_CHILD_UP from > gfs-client2 > > > ==== Server ==== > > 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine] gfs-server2: > partial header re > ad on NB socket. continue later > 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine] gfs-server2: > partial header re > ad on NB socket. continue later > 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine] gfs-server2: > partial header re > ad on NB socket. continue later > 2008-09-13 08:35:39 W [socket.c:93:__socket_rwv] gfs-server2: EOF from peer > 62.212.91.118:1022 > 2008-09-13 08:35:39 D [socket.c:568:socket_proto_state_machine] gfs-server2: > socket read faile > d (Transport endpoint is not connected) in state 1 (62.212.91.118:1022) > 2008-09-13 08:35:39 D [inode.c:287:__inode_passivate] gfs-storage2/inode: > passivating inode(1879077150) lru=3/1024 active=1 purge=32 > 2008-09-13 08:35:39 D [server-protocol.c:6182:server_protocol_cleanup] > gfs-server2: cleaned up transport state for client 62.212.91.118:1022 > 2008-09-13 08:35:39 D [socket.c:1287:fini] gfs-server2: transport 0x50d980 > destroyed > 2008-09-13 08:35:39 D [addr.c:166:gf_auth] gfs-storage2: allowed = "*", > received addr = "62.212.91.118" > 2008-09-13 08:35:39 D [server-protocol.c:5544:mop_setvolume] gfs-server2: > accepted client from 62.212.91.118:1021 > 2008-09-13 08:38:58 D [inode.c:268:__inode_activate] gfs-storage2/inode: > activating inode(1879076204), lru=2/1024 active=2 purge=32 > 2008-09-13 08:38:58 D [inode.c:287:__inode_passivate] gfs-storage2/inode: > passivating inode(1879076204) lru=3/1024 active=1 purge=32 > 2008-09-13 08:38:58 D [inode.c:268:__inode_activate] gfs-storage2/inode: > activating inode(1879076204), lru=2/1024 active=2 purge=32 > 2008-09-13 08:38:58 D [inode.c:287:__inode_passivate] gfs-storage2/inode: > passivating inode(1879076204) lru=3/1024 active=1 purge=32 > > > Does that mean GlusterFS will always destroy the connection once the process > is complete? > > > KwangErn > _______________________________________________ > Gluster-devel mailing list > Gluster-devel@xxxxxxxxxx > http://lists.nongnu.org/mailman/listinfo/gluster-devel >