[Gluster-Maintainers] Build failed in Jenkins: experimental-periodic #378

jenkins at build.gluster.org jenkins at build.gluster.org
Sun Jul 8 18:10:34 UTC 2018


See <https://build.gluster.org/job/experimental-periodic/378/display/redirect>

------------------------------------------
[...truncated 2.53 MB...]
#11 0x00007f99d3666bad in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 3 (Thread 0x7f984451c700 (LWP 26544)):
#0  0x00007f99d3fa5995 in pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1  0x00007f99c6284454 in br_stub_worker (data=0x7f99b556cc60) at <https://build.gluster.org/job/experimental-periodic/ws/xlators/features/bit-rot/src/stub/bit-rot-stub-helpers.c>:327
        priv = 0x7f99b55c3db0
        this = 0x7f99b556cc60
        stub = 0x0
#2  0x00007f99d3fa1e25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3  0x00007f99d3666bad in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 2 (Thread 0x7f9803e4c700 (LWP 26688)):
#0  0x00007f99d3fa851d in __lll_lock_wait () from /lib64/libpthread.so.0
No symbol table info available.
#1  0x00007f99d3fa3e1b in _L_lock_812 () from /lib64/libpthread.so.0
No symbol table info available.
#2  0x00007f99d3fa3ce8 in pthread_mutex_lock () from /lib64/libpthread.so.0
No symbol table info available.
#3  0x00007f99d4f8b4a0 in iobuf_get2 (iobuf_pool=0x126c3f0, page_size=188) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/iobuf.c>:645
        iobuf = 0x0
        iobuf_arena = 0x0
        rounded_size = 512
        __FUNCTION__ = "iobuf_get2"
#4  0x00007f99c9b2188c in __socket_proto_state_machine (this=0x7f999b751340, pollin=0x7f9803e4bd20) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2290
        ret = 0
        priv = 0x7f999b7518a0
        iobuf = 0x0
        iobref = 0x0
        vector = {{iov_base = 0x7f990000003c, iov_len = 0}, {iov_base = 0x7f9803e4bcd0, iov_len = 0}}
        in = 0x7f999b7518c8
        frag = 0x7f999b7518d0
        __FUNCTION__ = "__socket_proto_state_machine"
#5  0x00007f99c9b21db9 in socket_proto_state_machine (this=0x7f999b751340, pollin=0x7f9803e4bd20) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2425
        priv = 0x7f999b7518a0
        ret = 0
        __FUNCTION__ = "socket_proto_state_machine"
#6  0x00007f99c9b21e26 in socket_event_poll_in (this=0x7f999b751340, notify_handled=true) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2444
        ret = -1
        pollin = 0x0
        priv = 0x7f999b7518a0
        ctx = 0x1267010
#7  0x00007f99c9b22546 in socket_event_handler (fd=450, idx=222, gen=1, data=0x7f999b751340, poll_in=1, poll_out=0, poll_err=0) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2618
        this = 0x7f999b751340
        priv = 0x7f999b7518a0
        ret = 0
        ctx = 0x1267010
        socket_closed = false
        notify_handled = false
        __FUNCTION__ = "socket_event_handler"
#8  0x00007f99d4fc695c in event_dispatch_epoll_handler (event_pool=0x129ec30, event=0x7f9803e4bea0) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/event-epoll.c>:587
        ev_data = 0x7f9803e4bea4
        slot = 0x12d5200
        handler = 0x7f99c9b22278 <socket_event_handler>
        data = 0x7f999b751340
        idx = 222
        gen = 1
        ret = -1
        fd = 450
        handled_error_previously = false
        __FUNCTION__ = "event_dispatch_epoll_handler"
#9  0x00007f99d4fc6c4f in event_dispatch_epoll_worker (data=0x7f99a8397760) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/event-epoll.c>:663
        event = {events = 1, data = {ptr = 0x1000000de, fd = 222, u32 = 222, u64 = 4294967518}}
        ret = 1
        ev_data = 0x7f99a8397760
        event_pool = 0x129ec30
        myindex = 57
        timetodie = 0
        __FUNCTION__ = "event_dispatch_epoll_worker"
#10 0x00007f99d3fa1e25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#11 0x00007f99d3666bad in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 1 (Thread 0x7f9661459700 (LWP 27595)):
#0  0x00007f99c491c7fc in quota_lookup (frame=0x7f99a538f438, this=0x7f99c06df260, loc=0x7f96614588d0, xattr_req=0x0) at <https://build.gluster.org/job/experimental-periodic/ws/xlators/features/quota/src/quota.c>:1663
        priv = 0x0
        ret = -1
        local = 0x0
        __FUNCTION__ = "quota_lookup"
#1  0x00007f99c46f823b in io_stats_lookup (frame=0x7f99a538f218, this=0x7f99c06e0ac0, loc=0x7f96614588d0, xdata=0x0) at <https://build.gluster.org/job/experimental-periodic/ws/xlators/debug/io-stats/src/io-stats.c>:2784
        _new = 0x7f99a538f438
        old_THIS = 0x7f99c06e0ac0
        next_xl_fn = 0x7f99c491c7a7 <quota_lookup>
        tmp_cbk = 0x7f99c46ec0f2 <io_stats_lookup_cbk>
        __FUNCTION__ = "io_stats_lookup"
#2  0x00007f99d5022895 in default_lookup (frame=0x7f99a538f218, this=0x7f99c06e2620, loc=0x7f96614588d0, xdata=0x0) at defaults.c:2714
        old_THIS = 0x7f99c06e2620
        next_xl = 0x7f99c06e0ac0
        next_xl_fn = 0x7f99c46f7e1e <io_stats_lookup>
        opn = 27
        __FUNCTION__ = "default_lookup"
#3  0x00007f99d4f9ed69 in syncop_lookup (subvol=0x7f99c06e2620, loc=0x7f96614588d0, iatt=0x7f9661458830, parent=0x0, xdata_in=0x0, xdata_out=0x0) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/syncop.c>:1260
        _new = 0x7f99a538f218
        old_THIS = 0x7f99c002fd40
        next_xl_fn = 0x7f99d50226a7 <default_lookup>
        tmp_cbk = 0x7f99d4f9e7ab <syncop_lookup_cbk>
        task = 0x0
        frame = 0x7f99a5198728
        args = {op_ret = 0, op_errno = 0, iatt1 = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, iatt2 = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, xattr = 0x0, statvfs_buf = {f_bsize = 0, f_frsize = 0, f_blocks = 0, f_bfree = 0, f_bavail = 0, f_files = 0, f_ffree = 0, f_favail = 0, f_fsid = 0, f_flag = 0, f_namemax = 0, __f_spare = {0, 0, 0, 0, 0, 0}}, vector = 0x0, count = 0, iobref = 0x0, buffer = 0x0, xdata = 0x0, flock = {l_type = 0, l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len = 0, data = '\000' <repeats 1023 times>}}, lease = {cmd = 0, lease_type = NONE, lease_id = '\000' <repeats 15 times>, lease_flags = 0}, dict_out = 0x0, uuid = '\000' <repeats 15 times>, errstr = 0x0, dict = 0x0, lock_dict = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, barrier = {initialized = false, guard = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, cond = {__data = {__lock = 0, __futex = 0, __total_seq = 0, __wakeup_seq = 0, __woken_seq = 0, __mutex = 0x0, __nwaiters = 0, __broadcast_seq = 0}, __size = '\000' <repeats 47 times>, __align = 0}, waitq = {next = 0x0, prev = 0x0}, count = 0, waitfor = 0}, task = 0x0, mutex = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, cond = {__data = {__lock = 0, __futex = 0, __total_seq = 0, __wakeup_seq = 0, __woken_seq = 0, __mutex = 0x0, __nwaiters = 0, __broadcast_seq = 0}, __size = '\000' <repeats 47 times>, __align = 0}, done = 0, entries = {{list = {next = 0x0, prev = 0x0}, {next = 0x0, prev = 0x0}}, d_ino = 0, d_off = 0, d_len = 0, d_type = 0, d_stat = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, dict = 0x0, inode = 0x0, d_name = 0x7f9661458358 ""}, offset = 0, locklist = {list = {next = 0x0, prev = 0x0}, flock = {l_type = 0, l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len = 0, data = '\000' <repeats 1023 times>}}, client_uid = 0x0, lk_flags = 0}}
        __FUNCTION__ = "syncop_lookup"
#4  0x00007f99c4274589 in server_first_lookup (this=0x7f99c002fd40, client=0x7f99a539b270, reply=0x7f99a5196788) at <https://build.gluster.org/job/experimental-periodic/ws/xlators/protocol/server/src/server-handshake.c>:382
        loc = {path = 0x7f99c429badd "/", name = 0x7f99c429bbe2 "", inode = 0x7f99a5198318, parent = 0x0, gfid = '\000' <repeats 15 times>, "\001", pargfid = '\000' <repeats 15 times>}
        iatt = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}
        dict = 0x0
        ret = 0
        xl = 0x7f99c06e2620
        msg = 0x0
        inode = 0x0
        bname = 0x0
        str = 0x0
        tmp = 0x0
        saveptr = 0x0
        __FUNCTION__ = "server_first_lookup"
#5  0x00007f99c42760f4 in server_setvolume (req=0x7f99a5398018) at <https://build.gluster.org/job/experimental-periodic/ws/xlators/protocol/server/src/server-handshake.c>:886
        args = {dict = {dict_len = 826, dict_val = 0x7f99a5399000 ""}}
        rsp = 0x0
        client = 0x7f99a539b270
        serv_ctx = 0x7f99a539b660
        conf = 0x7f99c0037320
        peerinfo = 0x7f9690c14870
        reply = 0x7f99a5196788
        config_params = 0x7f99a5363f58
        params = 0x7f99a52d8a18
        name = 0x7f99a52d97c0 "/d/backends/vol02/brick5"
        client_uid = 0x7f99a5399740 "CTX_ID:a8413b6d-e353-4484-83e6-cb6af8f5cd01-GRAPH_ID:0-PID:20304-HOST:builder107.cloud.gluster.org-PC_NAME:patchy-vol02-client-5-RECON_NO:-1"
        clnt_version = 0x7f99a5198d80 "4.2dev"
        xl = 0x7f99c06e2620
        msg = 0x0
        volfile_key = 0x7f99a50e8320 "patchy-vol02"
        this = 0x7f99c002fd40
        checksum = 0
        ret = 0
        op_ret = 0
        op_errno = 22
        buf = 0x0
        opversion = 40200
        xprt = 0x7f99c0036eb0
        fop_version = 1298437
        mgmt_version = 0
        ctx = 0x1267010
        tmp = 0x7f99c088e560
        subdir_mount = 0x0
        client_name = 0x7f99c429bed3 "unknown"
        cleanup_starting = false
        __FUNCTION__ = "server_setvolume"
        __PRETTY_FUNCTION__ = "server_setvolume"
#6  0x00007f99d4d0a7e2 in rpcsvc_handle_rpc_call (svc=0x7f99c00449a0, trans=0x7f9690c147b0, msg=0x7f99a52d88a0) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-lib/src/rpcsvc.c>:721
        actor = 0x7f99c44ae8c0 <gluster_handshake_actors+64>
        actor_fn = 0x7f99c427487d <server_setvolume>
        req = 0x7f99a5398018
        ret = -1
        port = 46330
        is_unix = false
        empty = false
        unprivileged = true
        reply = 0x0
        drc = 0x0
        __FUNCTION__ = "rpcsvc_handle_rpc_call"
#7  0x00007f99d4d0ab35 in rpcsvc_notify (trans=0x7f9690c147b0, mydata=0x7f99c00449a0, event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f99a52d88a0) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-lib/src/rpcsvc.c>:815
        ret = -1
        msg = 0x7f99a52d88a0
        new_trans = 0x0
        svc = 0x7f99c00449a0
        listener = 0x0
        __FUNCTION__ = "rpcsvc_notify"
#8  0x00007f99d4d1074b in rpc_transport_notify (this=0x7f9690c147b0, event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7f99a52d88a0) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-lib/src/rpc-transport.c>:537
        ret = -1
        __FUNCTION__ = "rpc_transport_notify"
#9  0x00007f99c9b21ed8 in socket_event_poll_in (this=0x7f9690c147b0, notify_handled=false) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2462
        ret = 0
        pollin = 0x7f99a52d88a0
        priv = 0x7f9690c14d10
        ctx = 0x1267010
#10 0x00007f99c9b22546 in socket_event_handler (fd=575, idx=423, gen=1, data=0x7f9690c147b0, poll_in=1, poll_out=0, poll_err=24) at <https://build.gluster.org/job/experimental-periodic/ws/rpc/rpc-transport/socket/src/socket.c>:2618
        this = 0x7f9690c147b0
        priv = 0x7f9690c14d10
        ret = 0
        ctx = 0x1267010
        socket_closed = false
        notify_handled = false
        __FUNCTION__ = "socket_event_handler"
#11 0x00007f99d4fc695c in event_dispatch_epoll_handler (event_pool=0x129ec30, event=0x7f9661458ea0) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/event-epoll.c>:587
        ev_data = 0x7f9661458ea4
        slot = 0x12d9d60
        handler = 0x7f99c9b22278 <socket_event_handler>
        data = 0x7f9690c147b0
        idx = 423
        gen = 1
        ret = -1
        fd = 575
        handled_error_previously = false
        __FUNCTION__ = "event_dispatch_epoll_handler"
#12 0x00007f99d4fc6c4f in event_dispatch_epoll_worker (data=0x7f998e3b96f0) at <https://build.gluster.org/job/experimental-periodic/ws/libglusterfs/src/event-epoll.c>:663
        event = {events = 25, data = {ptr = 0x1000001a7, fd = 423, u32 = 423, u64 = 4294967719}}
        ret = 1
        ev_data = 0x7f998e3b96f0
        event_pool = 0x129ec30
        myindex = 116
        timetodie = 0
        __FUNCTION__ = "event_dispatch_epoll_worker"
#13 0x00007f99d3fa1e25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#14 0x00007f99d3666bad in clone () from /lib64/libc.so.6
No symbol table info available.
=========================================================
              Finish backtrace
         program name : /build/install/sbin/glusterfsd
         corefile     : /glusterepoll115-25796.core
=========================================================

+ rm -f /build/install/cores/gdbout.txt
+ sort /build/install/cores/liblist.txt
+ uniq
+ cat /build/install/cores/liblist.txt.tmp
+ grep -v /build/install
+ tar -cf /archives/archived_builds/build-install-experimental-periodic-378.tar /build/install/sbin /build/install/bin /build/install/lib /build/install/libexec /build/install/cores
tar: Removing leading `/' from member names
+ tar -rhf /archives/archived_builds/build-install-experimental-periodic-378.tar -T /build/install/cores/liblist.txt
tar: Removing leading `/' from member names
+ bzip2 /archives/archived_builds/build-install-experimental-periodic-378.tar
+ rm -f /build/install/cores/liblist.txt
+ rm -f /build/install/cores/liblist.txt.tmp
+ find /archives -size +1G -delete -type f
+ echo 'Cores and build archived in http://builder107.cloud.gluster.org/archived_builds/build-install-experimental-periodic-378.tar.bz2'
Cores and build archived in http://builder107.cloud.gluster.org/archived_builds/build-install-experimental-periodic-378.tar.bz2
+ echo 'Open core using the following command to get a proper stack'
Open core using the following command to get a proper stack
+ echo 'Example: From root of extracted tarball'
Example: From root of extracted tarball
+ echo '\t\tgdb -ex '\''set sysroot ./'\'' -ex '\''core-file ./build/install/cores/xxx.core'\'' <target, say ./build/install/sbin/glusterd>'
\t\tgdb -ex 'set sysroot ./' -ex 'core-file ./build/install/cores/xxx.core' <target, say ./build/install/sbin/glusterd>
+ RET=1
+ '[' 1 -ne 0 ']'
+ tar -czf <https://build.gluster.org/job/experimental-periodic/ws/glusterfs-logs.tgz> /var/log/glusterfs /var/log/messages /var/log/messages-20180617 /var/log/messages-20180624 /var/log/messages-20180701 /var/log/messages-20180708
tar: Removing leading `/' from member names
+ scp -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no -i <https://build.gluster.org/job/experimental-periodic/ws/> glusterfs-logs.tgz _logs_collector at http.int.rht.gluster.org:/var/www/glusterfs-logs/experimental-periodic-378.tgz
ssh: connect to host http.int.rht.gluster.org port 22: Connection timed out
lost connection
+ true
+ case $(uname -s) in
++ uname -s
+ /sbin/sysctl -w kernel.core_pattern=/%e-%p.core
kernel.core_pattern = /%e-%p.core
+ exit 1
Build step 'Execute shell' marked build as failure


More information about the maintainers mailing list