[Gluster-devel] Crash with mainline-2.5 patch 240 and dbench

Anand Avati avati at zresearch.com
Mon Jul 2 12:54:29 UTC 2007


Dale,
 this bug is fixed now. please tla update to get the fix.

thanks,
avati

2007/6/28, Anand Avati <avati at zresearch.com>:
>
> Dale,
>   We too got this bug recently. this fix is on the way.
>
> avati
>
> 2007/6/28, Dale Dude <dale at oc3networks.com>:
> >
> > Running dbench produces this backtrace:
> >
> > #0  0xb75b5dff in unify_lookup_cbk (frame=0x86f9bc0, cookie=0x8055e08,
> > this=0x80560b8, op_ret=0, op_errno=146211160, inode=0x8745fe0,
> > buf=0x87ed060) at unify.c:310
> > #1  0xb75c2249 in client_lookup_cbk (frame=0x8ec7598, args=0x8a004c0) at
> >
> > client-protocol.c:3797
> > #2  0xb75c36f5 in notify (this=0x8055e08, event=2, data=0x808dc70) at
> > client-protocol.c:4184
> > #3  0xb7f36ae2 in transport_notify (this=0x8055e08, event=1) at
> > transport.c:152
> > #4  0xb7f3725e in sys_epoll_iteration (ctx=0x8b70158) at epoll.c:54
> > #5  0xb7f36c8d in poll_iteration (ctx=0x8b70158) at transport.c:260
> > #6  0x0804a2e9 in main (argc=3, argv=0xbf95b494) at glusterfs.c:341
> >
> > ==========================
> > glusterfs.log:
> > 2007-06-27 15:51:40 C [ fuse-bridge.c:455:fuse_fd_cbk] glusterfs-fuse:
> > open() got EINTR
> > 2007-06-27 16:04:17 C [common-utils.c:205:gf_print_trace]
> > debug-backtrace: Got signal (11), printing backtrace
> > 2007-06-27 16:04:17 C [common-utils.c :207:gf_print_trace]
> > debug-backtrace: /lib/libglusterfs.so.0(gf_print_trace+0x2d)
> > [0xb7f355f9]
> > 2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
> > debug-backtrace: [0xffffe420]
> > 2007-06-27 16:04:17 C [common-utils.c :207:gf_print_trace]
> > debug-backtrace: /lib/glusterfs/1.3.0-pre5/xlator/protocol/client.so
> > [0xb75c2249]
> > 2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
> > debug-backtrace:
> > /lib/glusterfs/1.3.0-pre5/xlator/protocol/client.so(notify+0x855)
> > [0xb75c36f5]
> > 2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
> > debug-backtrace: /lib/libglusterfs.so.0(transport_notify+0x37)
> > [0xb7f36ae2]
> > 2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
> > debug-backtrace: /lib/libglusterfs.so.0(sys_epoll_iteration+0xd7)
> > [0xb7f3725e]
> > 2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
> > debug-backtrace: /lib/libglusterfs.so.0(poll_iteration+0x1d)
> > [0xb7f36c8d]
> > 2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
> > debug-backtrace: [glusterfs] [0x804a2e9]
> > 2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
> > debug-backtrace: /lib/tls/i686/cmov/libc.so.6(__libc_start_main+0xd2)
> > [0xb7de7ea2]
> > 2007-06-27 16:04:17 C [common-utils.c:207:gf_print_trace]
> > debug-backtrace: [glusterfs] [0x8049c31]
> >
> > =============================
> > glusterfs-client.vol
> > volume server1
> >          type protocol/client
> >          option transport-type tcp/client     # for TCP/IP transport
> >          option remote-host 127.0.0.1     # IP address of the remote
> > brick
> >          option remote-subvolume volumenamespace
> > end-volume
> >
> > volume server1vol1
> >          type protocol/client
> >          option transport-type tcp/client     # for TCP/IP transport
> >          option remote-host 127.0.0.1     # IP address of the remote
> > brick
> >          option remote-subvolume clusterfs1
> > end-volume
> >
> > ###################
> >
> > volume bricks
> >   type cluster/unify
> >   option namespace server1
> >   option readdir-force-success on  # ignore failed mounts
> >   subvolumes server1vol1
> >
> >   option scheduler rr
> >   option rr.limits.min-free-disk 5 #%
> > end-volume
> >
> > volume writebehind   #writebehind improves write performance a lot
> >   type performance/write-behind
> >   option aggregate-size 131072 # in bytes
> >   subvolumes bricks
> > end-volume
> >
> > #volume statprefetch
> >   #type performance/stat-prefetch
> >   #option cache-seconds 20
> >   #subvolumes writebehind
> > #end-volume
> >
> > ================================
> > glusterfs-server.vol:
> > volume clusterfs1
> >   type storage/posix
> >   option directory /volume1
> > end-volume
> >
> > #volume clusterfs1
> >    #type performance/io-threads
> >    #option thread-count 8
> >    #subvolumes volume1
> > #end-volume
> >
> > #######
> >
> > volume volumenamespace
> >   type storage/posix
> >   option directory /volume.namespace
> > end-volume
> >
> > ###
> >
> > volume clusterfs
> >   type protocol/server
> >   option transport-type tcp/server
> >   subvolumes clusterfs1 volumenamespace
> >   option auth.ip.clusterfs1.allow *
> >   option auth.ip.volumenamespace.allow *
> > end-volume
> >
> >
> > _______________________________________________
> > Gluster-devel mailing list
> > Gluster-devel at nongnu.org
> > http://lists.nongnu.org/mailman/listinfo/gluster-devel
> >
>
>
>
> --
> Anand V. Avati




-- 
Anand V. Avati



More information about the Gluster-devel mailing list