[Gluster-devel] Re: Bonnie++ bailing out
Raghavendra G
raghavendra.hg at gmail.com
Wed Sep 17 05:43:21 UTC 2008
Hi,
Can you try out with an higher transport-timeout? Currently you are having
transport-timeout configured to 30s. Can you progressively increase the
transport timeout and observe the behaviour? I think 180s can be kept as the
ceiling.
regards,
On Tue, Sep 16, 2008 at 1:01 PM, KwangErn Liew <ke.liew at gmail.com> wrote:
> Hm, a new error occurred when running Bonnie++ at one pass instead of 10.
>
> ====================
> $ bonnie++ -u bonnie -d /home/storage -s 7904m:256k -n 1024:-2:5m:30
> Using uid:100, gid:65534.
> Writing with putc()...done
> Writing intelligently...done
> Rewriting...done
> Reading with getc()...Can't getc(3)
> ====================
>
> The error logs are attached!
>
> Strangely, the server side logs doesn't quite correspond with the
> client...! In terms of the time and its execution.
>
>
> KwangErn
>
> On Sat, Sep 13, 2008 at 1:18 PM, Krishna Srinivas <krishna at zresearch.com>
> wrote:
> > KwangErn,
> > Can you give us the setup details?
> > Thanks
> > Krishna
> >
> > On Sat, Sep 13, 2008 at 2:12 PM, KwangErn Liew <ke.liew at gmail.com>
> wrote:
> >> Hm, it seems to be 'normal'. I have just ran dd across the network and
> this
> >> is what I have...
> >>
> >> $ dd if=/dev/zero of=/home/storage/testfile bs=16384k count=100
> >> 100+0 records in
> >> 100+0 records out
> >> 1677721600 bytes (1.7 GB) copied, 82.5314 seconds, 20.3 MB/s
> >>
> >> $ ls storage -l
> >> total 1638408
> >> drwxr-xr-x 2 bonnie nogroup 21 2008-09-13 08:34 .
> >> drwxr-xr-x 7 root root 4096 2008-09-11 21:28 ..
> >> -rw-r--r-- 1 root root 1677721600 2008-09-13 08:35 testfile
> >>
> >> ==== Client ====
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk]
> glusterfs-fuse:
> >> 458332: WRITE => 13
> >> 1072/131072,1677459456/1677590528
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:1548:fuse_write] glusterfs-fuse:
> >> 458333: WRITE (0x516760,
> >> size=131072, offset=1677590528)
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk]
> glusterfs-fuse:
> >> 458333: WRITE => 131072/131072,1677590528/1677721600
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:1572:fuse_flush] glusterfs-fuse:
> >> 458334: FLUSH 0x516760
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse:
> >> 458334: FLUSH() ERR => 0
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:1599:fuse_release] glusterfs-fuse:
> >> 458335: CLOSE 0x516760
> >> 2008-09-13 08:35:29 E [afr.c:3012:afr_close_cbk] gfs-afr:
> (path=/testfile
> >> child=gfs-client2) op_ret=-1 op_errno=77(File descriptor in bad state)
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse:
> >> 458335: CLOSE() ERR => 0
> >> 2008-09-13 08:35:29 D [inode.c:287:__inode_passivate] fuse/inode:
> >> passivating inode(2013354535) lru=1/0 active=1 purge=30
> >> 2008-09-13 08:35:31 D [client-protocol.c:4126:client_protocol_reconnect]
> >> gfs-client2: attempting reconnect
> >> 2008-09-13 08:35:31 D [socket.c:924:socket_connect] gfs-client2: connect
> ()
> >> called on transport already connected
> >> 2008-09-13 08:35:39 D [client-protocol.c:4065:client_setvolume_cbk]
> >> gfs-client2: SETVOLUME on remote-host succeeded
> >> 2008-09-13 08:35:39 D [afr.c:6309:notify] gfs-afr: GF_EVENT_CHILD_UP
> from
> >> gfs-client2
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk]
> glusterfs-fuse:
> >> 458332: WRITE => 13
> >> 1072/131072,1677459456/1677590528
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:1548:fuse_write] glusterfs-fuse:
> >> 458333: WRITE (0x516760,
> >> size=131072, offset=1677590528)
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:1510:fuse_writev_cbk]
> glusterfs-fuse:
> >> 458333: WRITE => 131072/131072,1677590528/1677721600
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:1572:fuse_flush] glusterfs-fuse:
> >> 458334: FLUSH 0x516760
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse:
> >> 458334: FLUSH() ERR => 0
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:1599:fuse_release] glusterfs-fuse:
> >> 458335: CLOSE 0x516760
> >> 2008-09-13 08:35:29 E [afr.c:3012:afr_close_cbk] gfs-afr:
> (path=/testfile
> >> child=gfs-client2) op_ret=-1 op_errno=77(File descriptor in bad state)
> >> 2008-09-13 08:35:29 D [fuse-bridge.c:857:fuse_err_cbk] glusterfs-fuse:
> >> 458335: CLOSE() ERR => 0
> >> 2008-09-13 08:35:29 D [inode.c:287:__inode_passivate] fuse/inode:
> >> passivating inode(2013354535) lru=1/0 active=1 purge=30
> >> 2008-09-13 08:35:31 D [client-protocol.c:4126:client_protocol_reconnect]
> >> gfs-client2: attempting reconnect
> >> 2008-09-13 08:35:31 D [socket.c:924:socket_connect] gfs-client2: connect
> ()
> >> called on transport already connected
> >> 2008-09-13 08:35:39 D [client-protocol.c:4065:client_setvolume_cbk]
> >> gfs-client2: SETVOLUME on remote-host succeeded
> >> 2008-09-13 08:35:39 D [afr.c:6309:notify] gfs-afr: GF_EVENT_CHILD_UP
> from
> >> gfs-client2
> >>
> >>
> >> ==== Server ====
> >>
> >> 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine]
> gfs-server2:
> >> partial header re
> >> ad on NB socket. continue later
> >> 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine]
> gfs-server2:
> >> partial header re
> >> ad on NB socket. continue later
> >> 2008-09-13 08:34:38 D [socket.c:636:socket_proto_state_machine]
> gfs-server2:
> >> partial header re
> >> ad on NB socket. continue later
> >> 2008-09-13 08:35:39 W [socket.c:93:__socket_rwv] gfs-server2: EOF from
> peer
> >> 62.212.91.118:1022
> >> 2008-09-13 08:35:39 D [socket.c:568:socket_proto_state_machine]
> gfs-server2:
> >> socket read faile
> >> d (Transport endpoint is not connected) in state 1 (62.212.91.118:1022)
> >> 2008-09-13 08:35:39 D [inode.c:287:__inode_passivate]
> gfs-storage2/inode:
> >> passivating inode(1879077150) lru=3/1024 active=1 purge=32
> >> 2008-09-13 08:35:39 D [server-protocol.c:6182:server_protocol_cleanup]
> >> gfs-server2: cleaned up transport state for client 62.212.91.118:1022
> >> 2008-09-13 08:35:39 D [socket.c:1287:fini] gfs-server2: transport
> 0x50d980
> >> destroyed
> >> 2008-09-13 08:35:39 D [addr.c:166:gf_auth] gfs-storage2: allowed = "*",
> >> received addr = "62.212.91.118"
> >> 2008-09-13 08:35:39 D [server-protocol.c:5544:mop_setvolume]
> gfs-server2:
> >> accepted client from 62.212.91.118:1021
> >> 2008-09-13 08:38:58 D [inode.c:268:__inode_activate] gfs-storage2/inode:
> >> activating inode(1879076204), lru=2/1024 active=2 purge=32
> >> 2008-09-13 08:38:58 D [inode.c:287:__inode_passivate]
> gfs-storage2/inode:
> >> passivating inode(1879076204) lru=3/1024 active=1 purge=32
> >> 2008-09-13 08:38:58 D [inode.c:268:__inode_activate] gfs-storage2/inode:
> >> activating inode(1879076204), lru=2/1024 active=2 purge=32
> >> 2008-09-13 08:38:58 D [inode.c:287:__inode_passivate]
> gfs-storage2/inode:
> >> passivating inode(1879076204) lru=3/1024 active=1 purge=32
> >>
> >>
> >> Does that mean GlusterFS will always destroy the connection once the
> process
> >> is complete?
> >>
> >>
> >> KwangErn
> >> _______________________________________________
> >> Gluster-devel mailing list
> >> Gluster-devel at nongnu.org
> >> http://lists.nongnu.org/mailman/listinfo/gluster-devel
> >>
> >
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at nongnu.org
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>
>
--
Raghavendra G
A centipede was happy quite, until a toad in fun,
Said, "Prey, which leg comes after which?",
This raised his doubts to such a pitch,
He fell flat into the ditch,
Not knowing how to run.
-Anonymous
More information about the Gluster-devel
mailing list