[Gluster-devel] Another segfault on client side (only sporadic)

Krishna Srinivas krishna at zresearch.com
Wed Aug 22 15:06:54 UTC 2007


Hi Bernhard,

We are not able to figure out the bug's cause. Is it possible for
you to give us access to your machine for debugging the core?

Thanks
Krishna

On 8/20/07, Bernhard J. M. Grün <bernhard.gruen at googlemail.com> wrote:
> I still have the core dump of the crash I've reported. But I don't
> know if the backtrace is the same every time. The glusterfs client now
> runs perfectly since 2007-08-16. So we have to wait for the next crash
> to analyse that issue further.
> Also the "print child_errno" does not output anything useful. It just
> says that there is no symbol with that name in the current context.
>
> 2007/8/20, Krishna Srinivas <krishna at zresearch.com>:
> > Do you see the same backtrace everytime it crashes?
> > can you do "print child_errno" at the gdb prompt when you have the core?
> >
> > Thanks
> > Krishna
> >
> > On 8/20/07, Bernhard J. M. Grün <bernhard.gruen at googlemail.com> wrote:
> > > Hi Krishna,
> > >
> > > One or also both of our glusterfs clients with that version crash
> > > every 3 to 5 days I think. The problem is that there is much
> > > throughput (about 30MBit/s on each client with about 99.5% file reads,
> > > rest file writes). This makes it hard to debug.
> > > We also have a core file from that crash (If I did not deleted it
> > > because it was quite big) anyway when the next crash occurs I'll save
> > > the core dump for sure.
> > > Do you have some idea how to work around that crash?
> > > .
> > > 2007/8/20, Krishna Srinivas <krishna at zresearch.com>:
> > > > Hi Bernhard,
> > > >
> > > > Sorry for the late response. We are not able to figure out
> > > > the cause for this bug. Do you have the core file?
> > > > Is the bug seen regularly?
> > > >
> > > > Thanks
> > > > Krishna
> > > >
> > > > On 8/16/07, Bernhard J. M. Grün <bernhard.gruen at googlemail.com> wrote:
> > > > > Hello developers,
> > > > >
> > > > > We just discovered another segfault on client side. At the moment we
> > > > > can't give you more information than our version number, a back trace
> > > > > and our client configuration.
> > > > >
> > > > > We use version 1.3.0 with patches up to patch-449.
> > > > >
> > > > > The back trace looks as the follows:
> > > > > Core was generated by `[glusterfs]
> > > > >                               '.
> > > > > Program terminated with signal 11, Segmentation fault.
> > > > > #0  0x00002aaaaacbc2bd in afr_stat (frame=0x2aaabce32cb0,
> > > > >     this=<value optimized out>, loc=0x2aaaac0fe168) at afr.c:2602
> > > > > 2602    afr.c: No such file or directory.
> > > > >         in afr.c
> > > > > (gdb) bt
> > > > > #0  0x00002aaaaacbc2bd in afr_stat (frame=0x2aaabce32cb0,
> > > > >     this=<value optimized out>, loc=0x2aaaac0fe168) at afr.c:2602
> > > > > #1  0x00002aaaaaece1bb in iot_stat (frame=0x2aaabcc00860, this=0x6126d0,
> > > > >     loc=0x2aaaac0fe168) at io-threads.c:651
> > > > > #2  0x00002aaaab0d2252 in wb_stat (frame=0x2aaaad05c5e0, this=0x612fe0,
> > > > >     loc=0x2aaaac0fe168) at write-behind.c:236
> > > > > #3  0x0000000000405fd2 in fuse_getattr (req=<value optimized out>,
> > > > >     ino=<value optimized out>, fi=<value optimized out>) at fuse-bridge.c:496
> > > > > #4  0x0000000000407139 in fuse_transport_notify (xl=<value optimized out>,
> > > > >     event=<value optimized out>, data=<value optimized out>)
> > > > >     at fuse-bridge.c:2067
> > > > > #5  0x00002af562b6a632 in sys_epoll_iteration (ctx=<value optimized out>)
> > > > >     at epoll.c:53
> > > > > #6  0x000000000040356b in main (argc=9, argv=0x7fff48169b78) at glusterfs.c:387
> > > > >
> > > > > And here is our client configuration for that machine:
> > > > > ### Add client feature and attach to remote subvolume
> > > > > volume client1
> > > > >   type protocol/client
> > > > >   option transport-type tcp/client     # for TCP/IP transport
> > > > >   option remote-host 10.1.1.13     # IP address of the remote brick
> > > > >   option remote-port 9999              # default server port is 6996
> > > > >   option remote-subvolume iothreads        # name of the remote volume
> > > > > end-volume
> > > > >
> > > > > ### Add client feature and attach to remote subvolume
> > > > > volume client2
> > > > >   type protocol/client
> > > > >   option transport-type tcp/client     # for TCP/IP transport
> > > > >   option remote-host 10.1.1.14     # IP address of the remote brick
> > > > >   option remote-port 9999              # default server port is 6996
> > > > >   option remote-subvolume iothreads        # name of the remote volume
> > > > > end-volume
> > > > >
> > > > > volume afrbricks
> > > > >   type cluster/afr
> > > > >   subvolumes client1 client2
> > > > >   option replicate *:2
> > > > >   option self-heal off
> > > > > end-volume
> > > > >
> > > > > volume iothreads    #iothreads can give performance a boost
> > > > >    type performance/io-threads
> > > > >    option thread-count 16
> > > > >    subvolumes afrbricks
> > > > > end-volume
> > > > >
> > > > > ### Add writeback feature
> > > > > volume bricks
> > > > >   type performance/write-behind
> > > > >   option aggregate-size 0  # unit in bytes
> > > > >   subvolumes iothreads
> > > > > end-volume
> > > > >
> > > > >
> > > > > We hope you can easily find and fix that error. Thank you in advance
> > > > >
> > > > > Bernhard J. M. Grün
> > > > >
> > > > >
> > > > > _______________________________________________
> > > > > Gluster-devel mailing list
> > > > > Gluster-devel at nongnu.org
> > > > > http://lists.nongnu.org/mailman/listinfo/gluster-devel
> > > > >
> > > >
> > >
> > >
> > > --
> > > Viele Grüße
> > > Bernhard J. M. Grün
> > >
> >
>
>
> --
> Viele Grüße
> Bernhard J. M. Grün
>





More information about the Gluster-devel mailing list