Fwd: [Gluster-devel] gluster client crash

Raghavendra G raghavendra at zresearch.com
Wed Jan 2 06:56:26 UTC 2008


Karl,
 is it possible for us to have the login again to inspect the cores again?
we want to reproduce a similar setup of /data/sharedfast which is causing
the segfault.
Is it possible to mail the specs?



2007/12/17, Karl Bernard <karl at vomba.com>:

>
> The client crashed, if this can be helpful:
>
> 2007-12-15 17:03:59 W [client-protocol.c:289:client_protocol_xfer]
> sxx01: attempting to pipeline request type(0) op(34) with handshake
>
> ---------
> got signal (11), printing backtrace
> ---------
> [0xcc5420]
> /usr/local/lib/glusterfs/1.3.8/xlator/performance/write-behind.so
> [0x196c1d]
> /usr/local/lib/glusterfs/1.3.8/xlator/performance/io-threads.so[0x262dab]
> /usr/local/lib/glusterfs/1.3.8/xlator/cluster/afr.so(afr_close_cbk+0x1d6)[0x118b46]
>
> /usr/local/lib/glusterfs/1.3.8/xlator/protocol/client.so[0x13367d]
>
> /usr/local/lib/glusterfs/1.3.8/xlator/protocol/client.so(notify+0xa84)[0x1374c4]
> /usr/local/lib/libglusterfs.so.0(transport_notify+0x37)[0x6c9717]
> /usr/local/lib/libglusterfs.so.0(sys_epoll_iteration+0xf3)[0x6ca473]
> /usr/local/lib/libglusterfs.so.0(poll_iteration+0x7c)[0x6c984c]
> [glusterfs](main+0x424)[0x804a494]
> /lib/libc.so.6(__libc_start_main+0xdc)[0xa49dec]
> [glusterfs][0x8049fe1]
>
>
> glusterfs 1.3.8
> installed from tla, last patch:
> 2007-12-03 22:29:15 GMT Anand V. Avati <avati at 80x25.org>        patch-594
>
> Config client:
> ----------------------------------------------------------
> volume sxx01
> type protocol/client
> option transport-type tcp/client
> option remote-host sxx01b
> option remote-subvolume brick
> end-volume
>
> volume sxx02
> type protocol/client
> option transport-type tcp/client
> option remote-host sxx02b
> option remote-subvolume brick
> end-volume
>
> volume afr1-2
>   type cluster/afr
>   subvolumes sxx01 sxx02
> end-volume
>
> volume iot
> type performance/io-threads
> subvolumes afr1-2
> option thread-count 8
> end-volume
>
> ## Add writebehind feature
> volume writebehind
>   type performance/write-behind
>   option aggregate-size 128kB
>   subvolumes iot
> end-volume
>
> ## Add readahead feature
> volume readahead
>   type performance/read-ahead
>   option page-size 256kB     #
>   option page-count 16       # cache per file  = (page-count x page-size)
>   subvolumes writebehind
> end-volume
>
> ------------------------------------------------------
>
> Config Server:
> volume brick-posix
>         type storage/posix
>         option directory /data/glusterfs/dataspace
> end-volume
>
> volume brick-ns
>         type storage/posix
>         option directory /data/glusterfs/namespace
> end-volume
>
> volume brick
>   type performance/io-threads
>   option thread-count 2
>   option cache-size 32MB
>   subvolumes brick-posix
> end-volume
>
> volume server
>         type protocol/server
>         option transport-type tcp/server
>         subvolumes brick brick-ns
>         option auth.ip.brick.allow 172.16.93.*
>         option auth.ip.brick-ns.allow 172.16.93.*
> end-volume
>
> ------------------------------------
>
> The client was most likely  checking for the existence of a file or
> writing a new file to the servers.
>



-- 

If I traveled to the end of the rainbow
As Dame Fortune did intend,
Murphy would be there to tell me
The pot's at the other end.



-- 
Raghavendra G

A centipede was happy quite, until a toad in fun,
Said, "Prey, which leg comes after which?",
This raised his doubts to such a pitch,
He fell flat into the ditch,
Not knowing how to run.
-Anonymous



More information about the Gluster-devel mailing list