[Gluster-Maintainers] Build failed in Jenkins: experimental-periodic #378
jenkins at build.gluster.org
jenkins at build.gluster.org
Sun Jul 8 18:10:34 UTC 2018
See <https://build.gluster.org/job/experimental-periodic/378/display/redirect>
------------------------------------------
[...truncated 2.53 MB...]
#11 0x00007f99d3666bad in clone () from /lib64/libc.so.6
No symbol table info available.
Thread 3 (Thread 0x7f984451c700 (LWP 26544)):
#0 0x00007f99d3fa5995 in pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1 0x00007f99c6284454 in br_stub_worker (data=0x7f99b556cc60) at <https://build.gluster.org/job/experimental-periodic/ws/xlators/features/bit-rot/src/stub/bit-rot-stub-helpers.c>:327
priv = 0x7f99b55c3db0
this = 0x7f99b556cc60
stub = 0x0
#2 0x00007f99d3fa1e25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3 0x00007f99d3666bad in clone () from /lib64/libc.so.6
No symbol table info available.
Thread 2 (Thread 0x7f9803e4c700 (LWP 26688)):
#0 0x00007f99d3fa851d in __lll_lock_wait () from /lib64/libpthread.so.0
No symbol table info available.
#1 0x00007f99d3fa3e1b in _L_lock_812 () from /lib64/libpthread.so.0
No symbol table info available.
#2 0x00007f99d3fa3ce8 in pthread_mutex_lock () from /lib64/libpthread.so.0
No symbol table info available.
#3 0x00007f99d4f8b4a0 in iobuf_get2 (iobuf_pool=0x126c3f0, page_size=188) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/iobuf.c>:645
iobuf = 0x0
iobuf_arena = 0x0
rounded_size = 512
__FUNCTION__ = "iobuf_get2"
#4 0x00007f99c9b2188c in __socket_proto_state_machine (this=0x7f999b751340, pollin=0x7f9803e4bd20) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2290
ret = 0
priv = 0x7f999b7518a0
iobuf = 0x0
iobref = 0x0
vector = {{iov_base = 0x7f990000003c, iov_len = 0}, {iov_base = 0x7f9803e4bcd0, iov_len = 0}}
in = 0x7f999b7518c8
frag = 0x7f999b7518d0
__FUNCTION__ = "__socket_proto_state_machine"
#5 0x00007f99c9b21db9 in socket_proto_state_machine (this=0x7f999b751340, pollin=0x7f9803e4bd20) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2425
priv = 0x7f999b7518a0
ret = 0
__FUNCTION__ = "socket_proto_state_machine"
#6 0x00007f99c9b21e26 in socket_event_poll_in (this=0x7f999b751340, notify_handled=true) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2444
ret = -1
pollin = 0x0
priv = 0x7f999b7518a0
ctx = 0x1267010
#7 0x00007f99c9b22546 in socket_event_handler (fd=450, idx=222, gen=1, data=0x7f999b751340, poll_in=1, poll_out=0, poll_err=0) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2618
this = 0x7f999b751340
priv = 0x7f999b7518a0
ret = 0
ctx = 0x1267010
socket_closed = false
notify_handled = false
__FUNCTION__ = "socket_event_handler"
#8 0x00007f99d4fc695c in event_dispatch_epoll_handler (event_pool=0x129ec30, event=0x7f9803e4bea0) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/event-epoll.c>:587
ev_data = 0x7f9803e4bea4
slot = 0x12d5200
handler = 0x7f99c9b22278 <socket_event_handler>
data = 0x7f999b751340
idx = 222
gen = 1
ret = -1
fd = 450
handled_error_previously = false
__FUNCTION__ = "event_dispatch_epoll_handler"
#9 0x00007f99d4fc6c4f in event_dispatch_epoll_worker (data=0x7f99a8397760) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/event-epoll.c>:663
event = {events = 1, data = {ptr = 0x1000000de, fd = 222, u32 = 222, u64 = 4294967518}}
ret = 1
ev_data = 0x7f99a8397760
event_pool = 0x129ec30
myindex = 57
timetodie = 0
__FUNCTION__ = "event_dispatch_epoll_worker"
#10 0x00007f99d3fa1e25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#11 0x00007f99d3666bad in clone () from /lib64/libc.so.6
No symbol table info available.
Thread 1 (Thread 0x7f9661459700 (LWP 27595)):
#0 0x00007f99c491c7fc in quota_lookup (frame=0x7f99a538f438, this=0x7f99c06df260, loc=0x7f96614588d0, xattr_req=0x0) at <https://build.gluster.org/job/experimental-periodic/ws/xlators/features/quota/src/quota.c>:1663
priv = 0x0
ret = -1
local = 0x0
__FUNCTION__ = "quota_lookup"
#1 0x00007f99c46f823b in io_stats_lookup (frame=0x7f99a538f218, this=0x7f99c06e0ac0, loc=0x7f96614588d0, xdata=0x0) at <https://build.gluster.org/job/experimental-periodic/ws/xlators/debug/io-stats/src/io-stats.c>:2784
_new = 0x7f99a538f438
old_THIS = 0x7f99c06e0ac0
next_xl_fn = 0x7f99c491c7a7 <quota_lookup>
tmp_cbk = 0x7f99c46ec0f2 <io_stats_lookup_cbk>
__FUNCTION__ = "io_stats_lookup"
#2 0x00007f99d5022895 in default_lookup (frame=0x7f99a538f218, this=0x7f99c06e2620, loc=0x7f96614588d0, xdata=0x0) at defaults.c:2714
old_THIS = 0x7f99c06e2620
next_xl = 0x7f99c06e0ac0
next_xl_fn = 0x7f99c46f7e1e <io_stats_lookup>
opn = 27
__FUNCTION__ = "default_lookup"
#3 0x00007f99d4f9ed69 in syncop_lookup (subvol=0x7f99c06e2620, loc=0x7f96614588d0, iatt=0x7f9661458830, parent=0x0, xdata_in=0x0, xdata_out=0x0) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/syncop.c>:1260
_new = 0x7f99a538f218
old_THIS = 0x7f99c002fd40
next_xl_fn = 0x7f99d50226a7 <default_lookup>
tmp_cbk = 0x7f99d4f9e7ab <syncop_lookup_cbk>
task = 0x0
frame = 0x7f99a5198728
args = {op_ret = 0, op_errno = 0, iatt1 = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, iatt2 = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, xattr = 0x0, statvfs_buf = {f_bsize = 0, f_frsize = 0, f_blocks = 0, f_bfree = 0, f_bavail = 0, f_files = 0, f_ffree = 0, f_favail = 0, f_fsid = 0, f_flag = 0, f_namemax = 0, __f_spare = {0, 0, 0, 0, 0, 0}}, vector = 0x0, count = 0, iobref = 0x0, buffer = 0x0, xdata = 0x0, flock = {l_type = 0, l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len = 0, data = '\000' <repeats 1023 times>}}, lease = {cmd = 0, lease_type = NONE, lease_id = '\000' <repeats 15 times>, lease_flags = 0}, dict_out = 0x0, uuid = '\000' <repeats 15 times>, errstr = 0x0, dict = 0x0, lock_dict = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, barrier = {initialized = false, guard = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, cond = {__data = {__lock = 0, __futex = 0, __total_seq = 0, __wakeup_seq = 0, __woken_seq = 0, __mutex = 0x0, __nwaiters = 0, __broadcast_seq = 0}, __size = '\000' <repeats 47 times>, __align = 0}, waitq = {next = 0x0, prev = 0x0}, count = 0, waitfor = 0}, task = 0x0, mutex = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, cond = {__data = {__lock = 0, __futex = 0, __total_seq = 0, __wakeup_seq = 0, __woken_seq = 0, __mutex = 0x0, __nwaiters = 0, __broadcast_seq = 0}, __size = '\000' <repeats 47 times>, __align = 0}, done = 0, entries = {{list = {next = 0x0, prev = 0x0}, {next = 0x0, prev = 0x0}}, d_ino = 0, d_off = 0, d_len = 0, d_type = 0, d_stat = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, dict = 0x0, inode = 0x0, d_name = 0x7f9661458358 ""}, offset = 0, locklist = {list = {next = 0x0, prev = 0x0}, flock = {l_type = 0, l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len = 0, data = '\000' <repeats 1023 times>}}, client_uid = 0x0, lk_flags = 0}}
__FUNCTION__ = "syncop_lookup"
#4 0x00007f99c4274589 in server_first_lookup (this=0x7f99c002fd40, client=0x7f99a539b270, reply=0x7f99a5196788) at <https://build.gluster.org/job/experimental-periodic/ws/xlators/protocol/server/src/server-handshake.c>:382
loc = {path = 0x7f99c429badd "/", name = 0x7f99c429bbe2 "", inode = 0x7f99a5198318, parent = 0x0, gfid = '\000' <repeats 15 times>, "\001", pargfid = '\000' <repeats 15 times>}
iatt = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}
dict = 0x0
ret = 0
xl = 0x7f99c06e2620
msg = 0x0
inode = 0x0
bname = 0x0
str = 0x0
tmp = 0x0
saveptr = 0x0
__FUNCTION__ = "server_first_lookup"
#5 0x00007f99c42760f4 in server_setvolume (req=0x7f99a5398018) at <https://build.gluster.org/job/experimental-periodic/ws/xlators/protocol/server/src/server-handshake.c>:886
args = {dict = {dict_len = 826, dict_val = 0x7f99a5399000 ""}}
rsp = 0x0
client = 0x7f99a539b270
serv_ctx = 0x7f99a539b660
conf = 0x7f99c0037320
peerinfo = 0x7f9690c14870
reply = 0x7f99a5196788
config_params = 0x7f99a5363f58
params = 0x7f99a52d8a18
name = 0x7f99a52d97c0 "/d/backends/vol02/brick5"
client_uid = 0x7f99a5399740 "CTX_ID:a8413b6d-e353-4484-83e6-cb6af8f5cd01-GRAPH_ID:0-PID:20304-HOST:builder107.cloud.gluster.org-PC_NAME:patchy-vol02-client-5-RECON_NO:-1"
clnt_version = 0x7f99a5198d80 "4.2dev"
xl = 0x7f99c06e2620
msg = 0x0
volfile_key = 0x7f99a50e8320 "patchy-vol02"
this = 0x7f99c002fd40
checksum = 0
ret = 0
op_ret = 0
op_errno = 22
buf = 0x0
opversion = 40200
xprt = 0x7f99c0036eb0
fop_version = 1298437
mgmt_version = 0
ctx = 0x1267010
tmp = 0x7f99c088e560
subdir_mount = 0x0
client_name = 0x7f99c429bed3 "unknown"
cleanup_starting = false
__FUNCTION__ = "server_setvolume"
__PRETTY_FUNCTION__ = "server_setvolume"
#6 0x00007f99d4d0a7e2 in rpcsvc_handle_rpc_call (svc=0x7f99c00449a0, trans=0x7f9690c147b0, msg=0x7f99a52d88a0) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-lib/src/rpcsvc.c>:721
actor = 0x7f99c44ae8c0 <gluster_handshake_actors+64>
actor_fn = 0x7f99c427487d <server_setvolume>
req = 0x7f99a5398018
ret = -1
port = 46330
is_unix = false
empty = false
unprivileged = true
reply = 0x0
drc = 0x0
__FUNCTION__ = "rpcsvc_handle_rpc_call"
#7 0x00007f99d4d0ab35 in rpcsvc_notify (trans=0x7f9690c147b0, mydata=0x7f99c00449a0, event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f99a52d88a0) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-lib/src/rpcsvc.c>:815
ret = -1
msg = 0x7f99a52d88a0
new_trans = 0x0
svc = 0x7f99c00449a0
listener = 0x0
__FUNCTION__ = "rpcsvc_notify"
#8 0x00007f99d4d1074b in rpc_transport_notify (this=0x7f9690c147b0, event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f99a52d88a0) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-lib/src/rpc-transport.c>:537
ret = -1
__FUNCTION__ = "rpc_transport_notify"
#9 0x00007f99c9b21ed8 in socket_event_poll_in (this=0x7f9690c147b0, notify_handled=false) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2462
ret = 0
pollin = 0x7f99a52d88a0
priv = 0x7f9690c14d10
ctx = 0x1267010
#10 0x00007f99c9b22546 in socket_event_handler (fd=575, idx=423, gen=1, data=0x7f9690c147b0, poll_in=1, poll_out=0, poll_err=24) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2618
this = 0x7f9690c147b0
priv = 0x7f9690c14d10
ret = 0
ctx = 0x1267010
socket_closed = false
notify_handled = false
__FUNCTION__ = "socket_event_handler"
#11 0x00007f99d4fc695c in event_dispatch_epoll_handler (event_pool=0x129ec30, event=0x7f9661458ea0) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/event-epoll.c>:587
ev_data = 0x7f9661458ea4
slot = 0x12d9d60
handler = 0x7f99c9b22278 <socket_event_handler>
data = 0x7f9690c147b0
idx = 423
gen = 1
ret = -1
fd = 575
handled_error_previously = false
__FUNCTION__ = "event_dispatch_epoll_handler"
#12 0x00007f99d4fc6c4f in event_dispatch_epoll_worker (data=0x7f998e3b96f0) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/event-epoll.c>:663
event = {events = 25, data = {ptr = 0x1000001a7, fd = 423, u32 = 423, u64 = 4294967719}}
ret = 1
ev_data = 0x7f998e3b96f0
event_pool = 0x129ec30
myindex = 116
timetodie = 0
__FUNCTION__ = "event_dispatch_epoll_worker"
#13 0x00007f99d3fa1e25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#14 0x00007f99d3666bad in clone () from /lib64/libc.so.6
No symbol table info available.
=========================================================
Finish backtrace
program name : /build/install/sbin/glusterfsd
corefile : /glusterepoll115-25796.core
=========================================================
+ rm -f /build/install/cores/gdbout.txt
+ sort /build/install/cores/liblist.txt
+ uniq
+ cat /build/install/cores/liblist.txt.tmp
+ grep -v /build/install
+ tar -cf /archives/archived_builds/build-install-experimental-periodic-378.tar /build/install/sbin /build/install/bin /build/install/lib /build/install/libexec /build/install/cores
tar: Removing leading `/' from member names
+ tar -rhf /archives/archived_builds/build-install-experimental-periodic-378.tar -T /build/install/cores/liblist.txt
tar: Removing leading `/' from member names
+ bzip2 /archives/archived_builds/build-install-experimental-periodic-378.tar
+ rm -f /build/install/cores/liblist.txt
+ rm -f /build/install/cores/liblist.txt.tmp
+ find /archives -size +1G -delete -type f
+ echo 'Cores and build archived in http://builder107.cloud.gluster.org/archived_builds/build-install-experimental-periodic-378.tar.bz2'
Cores and build archived in http://builder107.cloud.gluster.org/archived_builds/build-install-experimental-periodic-378.tar.bz2
+ echo 'Open core using the following command to get a proper stack'
Open core using the following command to get a proper stack
+ echo 'Example: From root of extracted tarball'
Example: From root of extracted tarball
+ echo '\t\tgdb -ex '\''set sysroot ./'\'' -ex '\''core-file ./build/install/cores/xxx.core'\'' <target, say ./build/install/sbin/glusterd>'
\t\tgdb -ex 'set sysroot ./' -ex 'core-file ./build/install/cores/xxx.core' <target, say ./build/install/sbin/glusterd>
+ RET=1
+ '[' 1 -ne 0 ']'
+ tar -czf <https://build.gluster.org/job/experimental-periodic/ws/glusterfs-logs.tgz> /var/log/glusterfs /var/log/messages /var/log/messages-20180617 /var/log/messages-20180624 /var/log/messages-20180701 /var/log/messages-20180708
tar: Removing leading `/' from member names
+ scp -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no -i <https://build.gluster.org/job/experimental-periodic/ws/> glusterfs-logs.tgz _logs_collector at http.int.rht.gluster.org:/var/www/glusterfs-logs/experimental-periodic-378.tgz
ssh: connect to host http.int.rht.gluster.org port 22: Connection timed out
lost connection
+ true
+ case $(uname -s) in
++ uname -s
+ /sbin/sysctl -w kernel.core_pattern=/%e-%p.core
kernel.core_pattern = /%e-%p.core
+ exit 1
Build step 'Execute shell' marked build as failure
More information about the maintainers
mailing list