Hi, Can you try out with an higher transport-timeout? Currently you are having transport-timeout configured to 30s. Can you progressively increase the transport timeout and observe the behaviour? I think 180s can be kept as the ceiling. regards, On Tue, Sep 16, 2008 at 1:01 PM, KwangErn Liew <ke.liew@xxxxxxxxx> wrote: > Hm, a new error occurred when running Bonnie++ at one pass instead of 10. > > ==================== > $ bonnie++ -u bonnie -d /home/storage -s 7904m:256k -n 1024:-2:5m:30 > Using uid:100, gid:65534. > Writing with putc()...done > Writing intelligently...done > Rewriting...done > Reading with getc()...Can't getc(3) > ==================== > > The error logs are attached! > > Strangely, the server side logs doesn't quite correspond with the > client...! In terms of the time and its execution. > > > KwangErn > > On Sat, Sep 13, 2008 at 1:18 PM, Krishna Srinivas <krishna@xxxxxxxxxxxxx> > wrote: > > KwangErn, > > Can you give us the setup details? > > Thanks > > Krishna > > > > On Sat, Sep 13, 2008 at 2:12 PM, KwangErn Liew <ke.liew@xxxxxxxxx> > wrote: > >> Hm, it seems to be 'normal'. I have just ran dd across the network and > this > >> is what I have... > >> > >> $ dd if=/dev/zero of=/home/storage/testfile bs=16384k count=100 > >> 100+0 records in > >> 100+0 records out > >> 1677721600 bytes (1.7 GB) copied, 82.5314 seconds, 20.3 MB/s > >> > >> $ ls storage -l > >> total 1638408 > >> drwxr-xr-x 2 bonnie nogroup 21 2008-09-13 08:34 . > >> drwxr-xr-x 7 root root 4096 2008-09-11 21:28 .. > >> -rw-r--r-- 1 root root 1677721600 2008-09-13 08:35 testfile > >> > >> ==== Client ==== > >> 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] > glusterfs-fuse: > >> 458332: WRITE => 13 > >> 1072/131072,1677459456/1677590528 > >> 2008-09-13 08:35:29 D [fuse-bridge.c:1548:fuse_write] glusterfs-fuse: > >> 458333: WRITE (0x516760, > >> size=131072, offset=1677590528) > >> 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] > glusterfs-fuse: > >> 458333: WRITE => 131072/131072,1677590528/1677721600 > >> 2008-09-13 08:35:29 D [fuse-bridge.c:1572:fuse_flush] glusterfs-fuse: > >> 458334: FLUSH 0x516760 > >> 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: > >> 458334: FLUSH() ERR => 0 > >> 2008-09-13 08:35:29 D [fuse-bridge.c:1599:fuse_release] glusterfs-fuse: > >> 458335: CLOSE 0x516760 > >> 2008-09-13 08:35:29 E [afr.c:3012:afr_close_cbk] gfs-afr: > (path=/testfile > >> child=gfs-client2) op_ret=-1 op_errno=77(File descriptor in bad state) > >> 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: > >> 458335: CLOSE() ERR => 0 > >> 2008-09-13 08:35:29 D [inode.c:287:__inode_passivate] fuse/inode: > >> passivating inode(2013354535) lru=1/0 active=1 purge=30 > >> 2008-09-13 08:35:31 D [client-protocol.c:4126:client_protocol_reconnect] > >> gfs-client2: attempting reconnect > >> 2008-09-13 08:35:31 D [socket.c:924:socket_connect] gfs-client2: connect > () > >> called on transport already connected > >> 2008-09-13 08:35:39 D [client-protocol.c:4065:client_setvolume_cbk] > >> gfs-client2: SETVOLUME on remote-host succeeded > >> 2008-09-13 08:35:39 D [afr.c:6309:notify] gfs-afr: GF_EVENT_CHILD_UP > from > >> gfs-client2 > >> 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] > glusterfs-fuse: > >> 458332: WRITE => 13 > >> 1072/131072,1677459456/1677590528 > >> 2008-09-13 08:35:29 D [fuse-bridge.c:1548:fuse_write] glusterfs-fuse: > >> 458333: WRITE (0x516760, > >> size=131072, offset=1677590528) > >> 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk] > glusterfs-fuse: > >> 458333: WRITE => 131072/131072,1677590528/1677721600 > >> 2008-09-13 08:35:29 D [fuse-bridge.c:1572:fuse_flush] glusterfs-fuse: > >> 458334: FLUSH 0x516760 > >> 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: > >> 458334: FLUSH() ERR => 0 > >> 2008-09-13 08:35:29 D [fuse-bridge.c:1599:fuse_release] glusterfs-fuse: > >> 458335: CLOSE 0x516760 > >> 2008-09-13 08:35:29 E [afr.c:3012:afr_close_cbk] gfs-afr: > (path=/testfile > >> child=gfs-client2) op_ret=-1 op_errno=77(File descriptor in bad state) > >> 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse: > >> 458335: CLOSE() ERR => 0 > >> 2008-09-13 08:35:29 D [inode.c:287:__inode_passivate] fuse/inode: > >> passivating inode(2013354535) lru=1/0 active=1 purge=30 > >> 2008-09-13 08:35:31 D [client-protocol.c:4126:client_protocol_reconnect] > >> gfs-client2: attempting reconnect > >> 2008-09-13 08:35:31 D [socket.c:924:socket_connect] gfs-client2: connect > () > >> called on transport already connected > >> 2008-09-13 08:35:39 D [client-protocol.c:4065:client_setvolume_cbk] > >> gfs-client2: SETVOLUME on remote-host succeeded > >> 2008-09-13 08:35:39 D [afr.c:6309:notify] gfs-afr: GF_EVENT_CHILD_UP > from > >> gfs-client2 > >> > >> > >> ==== Server ==== > >> > >> 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine] > gfs-server2: > >> partial header re > >> ad on NB socket. continue later > >> 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine] > gfs-server2: > >> partial header re > >> ad on NB socket. continue later > >> 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine] > gfs-server2: > >> partial header re > >> ad on NB socket. continue later > >> 2008-09-13 08:35:39 W [socket.c:93:__socket_rwv] gfs-server2: EOF from > peer > >> 62.212.91.118:1022 > >> 2008-09-13 08:35:39 D [socket.c:568:socket_proto_state_machine] > gfs-server2: > >> socket read faile > >> d (Transport endpoint is not connected) in state 1 (62.212.91.118:1022) > >> 2008-09-13 08:35:39 D [inode.c:287:__inode_passivate] > gfs-storage2/inode: > >> passivating inode(1879077150) lru=3/1024 active=1 purge=32 > >> 2008-09-13 08:35:39 D [server-protocol.c:6182:server_protocol_cleanup] > >> gfs-server2: cleaned up transport state for client 62.212.91.118:1022 > >> 2008-09-13 08:35:39 D [socket.c:1287:fini] gfs-server2: transport > 0x50d980 > >> destroyed > >> 2008-09-13 08:35:39 D [addr.c:166:gf_auth] gfs-storage2: allowed = "*", > >> received addr = "62.212.91.118" > >> 2008-09-13 08:35:39 D [server-protocol.c:5544:mop_setvolume] > gfs-server2: > >> accepted client from 62.212.91.118:1021 > >> 2008-09-13 08:38:58 D [inode.c:268:__inode_activate] gfs-storage2/inode: > >> activating inode(1879076204), lru=2/1024 active=2 purge=32 > >> 2008-09-13 08:38:58 D [inode.c:287:__inode_passivate] > gfs-storage2/inode: > >> passivating inode(1879076204) lru=3/1024 active=1 purge=32 > >> 2008-09-13 08:38:58 D [inode.c:268:__inode_activate] gfs-storage2/inode: > >> activating inode(1879076204), lru=2/1024 active=2 purge=32 > >> 2008-09-13 08:38:58 D [inode.c:287:__inode_passivate] > gfs-storage2/inode: > >> passivating inode(1879076204) lru=3/1024 active=1 purge=32 > >> > >> > >> Does that mean GlusterFS will always destroy the connection once the > process > >> is complete? > >> > >> > >> KwangErn > >> _______________________________________________ > >> Gluster-devel mailing list > >> Gluster-devel@xxxxxxxxxx > >> http://lists.nongnu.org/mailman/listinfo/gluster-devel > >> > > > > _______________________________________________ > Gluster-devel mailing list > Gluster-devel@xxxxxxxxxx > http://lists.nongnu.org/mailman/listinfo/gluster-devel > > -- Raghavendra G A centipede was happy quite, until a toad in fun, Said, "Prey, which leg comes after which?", This raised his doubts to such a pitch, He fell flat into the ditch, Not knowing how to run. -Anonymous