Hm, it seems to be 'normal'. I have just ran dd across the network and this is what I have... $ dd if=/dev/zero of=/home/storage/testfile bs=16384k count=100 100+0 records in 100+0 records out 1677721600 bytes (1.7 GB) copied, 82.5314 seconds, 20.3 MB/s $ ls storage -l total 1638408 drwxr-xr-x 2 bonnie nogroup 21 2008-09-13 08:34 . drwxr-xr-x 7 root root 4096 2008-09-11 21:28 .. -rw-r--r-- 1 root root 1677721600 2008-09-13 08:35 testfile ==== Client ==== 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] glusterfs-fuse: 458332: WRITE => 13 1072/131072,1677459456/1677590528 2008-09-13 08:35:29 D [fuse-bridge.c:1548:fuse_write] glusterfs-fuse: 458333: WRITE (0x516760, size=131072, offset=1677590528) 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] glusterfs-fuse: 458333: WRITE => 131072/131072,1677590528/1677721600 2008-09-13 08:35:29 D [fuse-bridge.c:1572:fuse_flush] glusterfs-fuse: 458334: FLUSH 0x516760 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: 458334: FLUSH() ERR => 0 2008-09-13 08:35:29 D [fuse-bridge.c:1599:fuse_release] glusterfs-fuse: 458335: CLOSE 0x516760 2008-09-13 08:35:29 E [afr.c:3012:afr_close_cbk] gfs-afr: (path=/testfile child=gfs-client2) op_ret=-1 op_errno=77(File descriptor in bad state) 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: 458335: CLOSE() ERR => 0 2008-09-13 08:35:29 D [inode.c:287:__inode_passivate] fuse/inode: passivating inode(2013354535) lru=1/0 active=1 purge=30 2008-09-13 08:35:31 D [client-protocol.c:4126:client_protocol_reconnect] gfs-client2: attempting reconnect 2008-09-13 08:35:31 D [socket.c:924:socket_connect] gfs-client2: connect () called on transport already connected 2008-09-13 08:35:39 D [client-protocol.c:4065:client_setvolume_cbk] gfs-client2: SETVOLUME on remote-host succeeded 2008-09-13 08:35:39 D [afr.c:6309:notify] gfs-afr: GF_EVENT_CHILD_UP from gfs-client2 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] glusterfs-fuse: 458332: WRITE => 13 1072/131072,1677459456/1677590528 2008-09-13 08:35:29 D [fuse-bridge.c:1548:fuse_write] glusterfs-fuse: 458333: WRITE (0x516760, size=131072, offset=1677590528) 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] glusterfs-fuse: 458333: WRITE => 131072/131072,1677590528/1677721600 2008-09-13 08:35:29 D [fuse-bridge.c:1572:fuse_flush] glusterfs-fuse: 458334: FLUSH 0x516760 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: 458334: FLUSH() ERR => 0 2008-09-13 08:35:29 D [fuse-bridge.c:1599:fuse_release] glusterfs-fuse: 458335: CLOSE 0x516760 2008-09-13 08:35:29 E [afr.c:3012:afr_close_cbk] gfs-afr: (path=/testfile child=gfs-client2) op_ret=-1 op_errno=77(File descriptor in bad state) 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: 458335: CLOSE() ERR => 0 2008-09-13 08:35:29 D [inode.c:287:__inode_passivate] fuse/inode: passivating inode(2013354535) lru=1/0 active=1 purge=30 2008-09-13 08:35:31 D [client-protocol.c:4126:client_protocol_reconnect] gfs-client2: attempting reconnect 2008-09-13 08:35:31 D [socket.c:924:socket_connect] gfs-client2: connect () called on transport already connected 2008-09-13 08:35:39 D [client-protocol.c:4065:client_setvolume_cbk] gfs-client2: SETVOLUME on remote-host succeeded 2008-09-13 08:35:39 D [afr.c:6309:notify] gfs-afr: GF_EVENT_CHILD_UP from gfs-client2 ==== Server ==== 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine] gfs-server2: partial header re ad on NB socket. continue later 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine] gfs-server2: partial header re ad on NB socket. continue later 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine] gfs-server2: partial header re ad on NB socket. continue later 2008-09-13 08:35:39 W [socket.c:93:__socket_rwv] gfs-server2: EOF from peer 62.212.91.118:1022 2008-09-13 08:35:39 D [socket.c:568:socket_proto_state_machine] gfs-server2: socket read faile d (Transport endpoint is not connected) in state 1 (62.212.91.118:1022) 2008-09-13 08:35:39 D [inode.c:287:__inode_passivate] gfs-storage2/inode: passivating inode(1879077150) lru=3/1024 active=1 purge=32 2008-09-13 08:35:39 D [server-protocol.c:6182:server_protocol_cleanup] gfs-server2: cleaned up transport state for client 62.212.91.118:1022 2008-09-13 08:35:39 D [socket.c:1287:fini] gfs-server2: transport 0x50d980 destroyed 2008-09-13 08:35:39 D [addr.c:166:gf_auth] gfs-storage2: allowed = "*", received addr = "62.212.91.118" 2008-09-13 08:35:39 D [server-protocol.c:5544:mop_setvolume] gfs-server2: accepted client from 62.212.91.118:1021 2008-09-13 08:38:58 D [inode.c:268:__inode_activate] gfs-storage2/inode: activating inode(1879076204), lru=2/1024 active=2 purge=32 2008-09-13 08:38:58 D [inode.c:287:__inode_passivate] gfs-storage2/inode: passivating inode(1879076204) lru=3/1024 active=1 purge=32 2008-09-13 08:38:58 D [inode.c:268:__inode_activate] gfs-storage2/inode: activating inode(1879076204), lru=2/1024 active=2 purge=32 2008-09-13 08:38:58 D [inode.c:287:__inode_passivate] gfs-storage2/inode: passivating inode(1879076204) lru=3/1024 active=1 purge=32 Does that mean GlusterFS will always destroy the connection once the process is complete? KwangErn