[Gluster-Maintainers] Build failed in Jenkins: regression-test-burn-in #935
jenkins at build.gluster.org
jenkins at build.gluster.org
Sun May 8 22:06:37 UTC 2016
See <http://build.gluster.org/job/regression-test-burn-in/935/>
------------------------------------------
[...truncated 16809 lines...]
#0 0x00007f6365e9268c in pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1 0x00007f6359f2bc8d in posix_fsyncer_pick (this=0x7f6354007aa0, head=0x7f63383f8e60) at <http://build.gluster.org/job/regression-test-burn-in/ws/xlators/storage/posix/src/posix-helpers.c>:1908
priv = 0x7f6354084600
count = 0
#2 0x00007f6359f2bfb7 in posix_fsyncer (d=0x7f6354007aa0) at <http://build.gluster.org/job/regression-test-burn-in/ws/xlators/storage/posix/src/posix-helpers.c>:2006
this = 0x7f6354007aa0
priv = 0x7f6354084600
stub = 0x0
tmp = 0x0
list = {next = 0x7f63383f8e60, prev = 0x7f63383f8e60}
count = 0
do_fsync = _gf_true
__FUNCTION__ = "posix_fsyncer"
#3 0x00007f6365e8eaa1 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#4 0x00007f63657f793d in clone () from /lib64/libc.so.6
No symbol table info available.
Thread 4 (Thread 0x7f63518fa700 (LWP 16550)):
#0 0x00007f6365e9268c in pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1 0x00007f6358bc9d37 in br_stub_signth (arg=0x7f6354011b70) at <http://build.gluster.org/job/regression-test-burn-in/ws/xlators/features/bit-rot/src/stub/bit-rot-stub.c>:774
this = 0x7f6354011b70
priv = 0x7f6354061a30
sigstub = 0x0
#2 0x00007f6365e8eaa1 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3 0x00007f63657f793d in clone () from /lib64/libc.so.6
No symbol table info available.
Thread 3 (Thread 0x7f63519fb700 (LWP 16549)):
#0 0x00007f6365e92a5e in pthread_cond_timedwait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1 0x00007f6353df95e6 in iot_worker (data=0x7f635404e5e0) at <http://build.gluster.org/job/regression-test-burn-in/ws/xlators/performance/io-threads/src/io-threads.c>:176
conf = 0x7f635404e5e0
this = 0x7f635401a420
stub = 0x7f634c009d2c
sleep_till = {tv_sec = 1462745288, tv_nsec = 0}
ret = 0
pri = -1
timeout = 0 '\000'
bye = 0 '\000'
sleep = {tv_sec = 0, tv_nsec = 0}
__FUNCTION__ = "iot_worker"
#2 0x00007f6365e8eaa1 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3 0x00007f63657f793d in clone () from /lib64/libc.so.6
No symbol table info available.
Thread 2 (Thread 0x7f6358151700 (LWP 16548)):
#0 0x00007f6365e9268c in pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1 0x00007f63537afe31 in index_worker (data=0x7f635401f6c0) at <http://build.gluster.org/job/regression-test-burn-in/ws/xlators/features/index/src/index.c>:209
priv = 0x7f6354044d90
this = 0x7f635401f6c0
stub = 0x0
ret = 0
#2 0x00007f6365e8eaa1 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3 0x00007f63657f793d in clone () from /lib64/libc.so.6
No symbol table info available.
Thread 1 (Thread 0x7f635ab3d700 (LWP 16546)):
#0 0x00007f63658426bf in __strlen_sse42 () from /lib64/libc.so.6
No symbol table info available.
#1 0x00007f63524feae4 in gf_strdup (src=0xcafebabe00007f63 <Address 0xcafebabe00007f63 out of bounds>) at <http://build.gluster.org/job/regression-test-burn-in/ws/libglusterfs/src/mem-pool.h>:183
dup_str = 0x0
len = 0
#2 0x00007f63524ff1a6 in gf_auth (input_params=0x7f6354082a2c, config_params=0x7f635407e34c) at <http://build.gluster.org/job/regression-test-burn-in/ws/xlators/protocol/auth/addr/src/addr.c>:183
result = AUTH_DONT_CARE
ret = 0
name = 0x7f63540838a0 "/d/backends/brick4"
searchstr = 0x7f63540010d0 ""
peer_info = 0x7f6354002a88
peer_info_data = 0x7f635400141c
allow_addr = 0x7f6354025b1c
reject_addr = 0x0
addr_str = 0x0
tmp = 0x0
addr_cpy = 0x0
service = 0x7f635ab3c69f "65424"
peer_port = 65424
is_inet_sdp = 0 '\000'
negate = 0 '\000'
match = 0 '\000'
peer_addr = "23.253.211.104\000\066\065\064\062\064\000\000\000\000&\217xec\177\000\000\020dzZc\177\000\000\360E\002Tc\177\000\000\260\024\000Tc\177\000\000p\024\000Tc\177\000\000\300ٳZ\006\000\000\000\001", '\000' <repeats 34 times>
type = 0x7f6366981ac9 "on"
allow_insecure = _gf_true
__FUNCTION__ = "gf_auth"
#3 0x00007f635314399c in gf_auth_one_method (this=0x7f635403640c, key=0x7f63540366f0 "addr", value=0x7f634c02190c, data=0x7f635ab3c8d0) at <http://build.gluster.org/job/regression-test-burn-in/ws/xlators/protocol/server/src/authenticate.c>:186
args = 0x7f635ab3c8d0
handle = 0x7f634c0214f0
#4 0x00007f6366bacb7b in dict_foreach_match (dict=0x7f635403640c, match=0x7f6366bac9f3 <dict_match_everything>, match_data=0x0, action=0x7f6353143919 <gf_auth_one_method>, action_data=0x7f635ab3c8d0) at <http://build.gluster.org/job/regression-test-burn-in/ws/libglusterfs/src/dict.c>:1236
__FUNCTION__ = "dict_foreach_match"
ret = -1
count = 0
pairs = 0x7f635403684c
next = 0x7f635403646c
#5 0x00007f6366baca53 in dict_foreach (dict=0x7f635403640c, fn=0x7f6353143919 <gf_auth_one_method>, data=0x7f635ab3c8d0) at <http://build.gluster.org/job/regression-test-burn-in/ws/libglusterfs/src/dict.c>:1194
ret = 0
#6 0x00007f6353143a2e in gf_authenticate (input_params=0x7f6354082a2c, config_params=0x7f635407e34c, auth_modules=0x7f635403640c) at <http://build.gluster.org/job/regression-test-burn-in/ws/xlators/protocol/server/src/authenticate.c>:213
name = 0x0
peerinfo_data = 0x0
args = {iparams = 0x7f6354082a2c, cparams = 0x7f635407e34c, result = 2}
__FUNCTION__ = "gf_authenticate"
#7 0x00007f635314286c in server_setvolume (req=0x7f6354098f7c) at <http://build.gluster.org/job/regression-test-burn-in/ws/xlators/protocol/server/src/server-handshake.c>:682
args = {dict = {dict_len = 541, dict_val = 0x7f63540269b0 ""}}
rsp = 0x0
client = 0x7f6354000cd0
serv_ctx = 0x7f6354000fb0
conf = 0x7f6354029fd0
peerinfo = 0x7f6354002a88
reply = 0x7f635408708c
config_params = 0x7f635407e34c
params = 0x7f6354082a2c
name = 0x7f63540838a0 "/d/backends/brick4"
client_uid = 0x7f6354088770 "slave28.cloud.gluster.org-16646-2016/05/08-22:06:08:505764-patchy-client-3-0-0"
clnt_version = 0x7f6354088560 "3.9dev"
xl = 0x7f6354022ae0
msg = 0x0
volfile_key = 0x7f6354088350 "gluster/quotad"
this = 0x7f63540245f0
checksum = 0
ret = 0
op_ret = -1
op_errno = 22
fop_version = 1298437
mgmt_version = 0
lk_version = 1
buf = 0x0
cancelled = _gf_false
__FUNCTION__ = "server_setvolume"
__PRETTY_FUNCTION__ = "server_setvolume"
#8 0x00007f63669721b4 in rpcsvc_handle_rpc_call (svc=0x7f63540376a0, trans=0x7f63540029d0, msg=0x7f635406fe50) at <http://build.gluster.org/job/regression-test-burn-in/ws/rpc/rpc-lib/src/rpcsvc.c>:701
actor = 0x7f6353354640
actor_fn = 0x7f63531419db <server_setvolume>
req = 0x7f6354098f7c
ret = -1
port = 65424
is_unix = _gf_false
unprivileged = _gf_true
reply = 0x0
drc = 0x0
__FUNCTION__ = "rpcsvc_handle_rpc_call"
#9 0x00007f6366972527 in rpcsvc_notify (trans=0x7f63540029d0, mydata=0x7f63540376a0, event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f635406fe50) at <http://build.gluster.org/job/regression-test-burn-in/ws/rpc/rpc-lib/src/rpcsvc.c>:795
ret = -1
msg = 0x7f635406fe50
new_trans = 0x0
svc = 0x7f63540376a0
listener = 0x0
__FUNCTION__ = "rpcsvc_notify"
#10 0x00007f6366977c5a in rpc_transport_notify (this=0x7f63540029d0, event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f635406fe50) at <http://build.gluster.org/job/regression-test-burn-in/ws/rpc/rpc-lib/src/rpc-transport.c>:541
ret = -1
__FUNCTION__ = "rpc_transport_notify"
#11 0x00007f635c16529b in socket_event_poll_in (this=0x7f63540029d0) at <http://build.gluster.org/job/regression-test-burn-in/ws/rpc/rpc-transport/socket/src/socket.c>:2347
ret = 0
pollin = 0x7f635406fe50
priv = 0x7f6354003510
#12 0x00007f635c1657f1 in socket_event_handler (fd=17, idx=6, data=0x7f63540029d0, poll_in=1, poll_out=0, poll_err=0) at <http://build.gluster.org/job/regression-test-burn-in/ws/rpc/rpc-transport/socket/src/socket.c>:2460
this = 0x7f63540029d0
priv = 0x7f6354003510
ret = 0
__FUNCTION__ = "socket_event_handler"
#13 0x00007f6366c21c9c in event_dispatch_epoll_handler (event_pool=0x1560db0, event=0x7f635ab3ce70) at <http://build.gluster.org/job/regression-test-burn-in/ws/libglusterfs/src/event-epoll.c>:571
ev_data = 0x7f635ab3ce74
slot = 0x158eda0
handler = 0x7f635c1655d0 <socket_event_handler>
data = 0x7f63540029d0
idx = 6
gen = 1
ret = -1
fd = 17
__FUNCTION__ = "event_dispatch_epoll_handler"
#14 0x00007f6366c220c8 in event_dispatch_epoll_worker (data=0x15a4920) at <http://build.gluster.org/job/regression-test-burn-in/ws/libglusterfs/src/event-epoll.c>:674
event = {events = 1, data = {ptr = 0x100000006, fd = 6, u32 = 6, u64 = 4294967302}}
ret = 1
ev_data = 0x15a4920
event_pool = 0x1560db0
myindex = 1
timetodie = 0
__FUNCTION__ = "event_dispatch_epoll_worker"
#15 0x00007f6365e8eaa1 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#16 0x00007f63657f793d in clone () from /lib64/libc.so.6
No symbol table info available.
+ echo =========================================================
=========================================================
+ echo ' Finish backtrace'
Finish backtrace
+ echo ' program name : /build/install/sbin/glusterfsd'
program name : /build/install/sbin/glusterfsd
+ echo ' corefile : /glusterfsd-16541.core'
corefile : /glusterfsd-16541.core
+ echo =========================================================
=========================================================
+ echo ''
+ mkdir -p /build/install/cores
+ mv /glusterfsd-16541.core /build/install/cores
+ filename=archived_builds/build-install-20160508:19:47:22.tar
+ rm -f /build/install/cores/liblist.txt
+ rm -f /build/install/cores/liblist.txt.tmp
++ ls /build/install/cores/glusterfsd-16541.core
+ CORELIST=/build/install/cores/glusterfsd-16541.core
+ for corefile in '$CORELIST'
+ getliblistfromcore /build/install/cores/glusterfsd-16541.core
+ rm -f /build/install/cores/gdbout.txt
+ gdb -c /build/install/cores/glusterfsd-16541.core -q -ex 'info sharedlibrary' -ex q
+ set +x
+ rm -f /build/install/cores/gdbout.txt
+ sort /build/install/cores/liblist.txt
+ uniq
+ cat /build/install/cores/liblist.txt.tmp
+ grep -v /build/install
+ tar -cf /archives/archived_builds/build-install-20160508:19:47:22.tar /build/install/sbin /build/install/bin /build/install/lib /build/install/libexec /build/install/cores
tar: Removing leading `/' from member names
+ tar -rhf /archives/archived_builds/build-install-20160508:19:47:22.tar -T /build/install/cores/liblist.txt
tar: Removing leading `/' from member names
+ bzip2 /archives/archived_builds/build-install-20160508:19:47:22.tar
+ rm -f /build/install/cores/liblist.txt
+ rm -f /build/install/cores/liblist.txt.tmp
+ echo Cores and build archived in http://slave28.cloud.gluster.org/archived_builds/build-install-20160508:19:47:22.tar.bz2
Cores and build archived in http://slave28.cloud.gluster.org/archived_builds/build-install-20160508:19:47:22.tar.bz2
+ echo Open core using the following command to get a proper stack...
Open core using the following command to get a proper stack...
+ echo Example: From root of extracted tarball
Example: From root of extracted tarball
+ echo 'gdb -ex '\''set sysroot ./'\'' -ex '\''core-file ./build/install/cores/xxx.core'\'' <target, say ./build/install/sbin/glusterd>'
gdb -ex 'set sysroot ./' -ex 'core-file ./build/install/cores/xxx.core' <target, say ./build/install/sbin/glusterd>
+ RET=1
+ '[' 1 -ne 0 ']'
+ filename=logs/glusterfs-logs-20160508:19:47:22.tgz
+ tar -czf /archives/logs/glusterfs-logs-20160508:19:47:22.tgz /var/log/glusterfs /var/log/messages /var/log/messages-20160417 /var/log/messages-20160424 /var/log/messages-20160501 /var/log/messages-20160508
tar: Removing leading `/' from member names
+ echo Logs archived in http://slave28.cloud.gluster.org/logs/glusterfs-logs-20160508:19:47:22.tgz
Logs archived in http://slave28.cloud.gluster.org/logs/glusterfs-logs-20160508:19:47:22.tgz
+ case $(uname -s) in
++ uname -s
+ /sbin/sysctl -w kernel.core_pattern=/%e-%p.core
kernel.core_pattern = /%e-%p.core
+ exit 1
+ RET=1
+ '[' 1 = 0 ']'
+ V=-1
+ VERDICT=FAILED
+ '[' 0 -eq 1 ']'
+ exit 1
Build step 'Execute shell' marked build as failure
More information about the maintainers
mailing list