[Gluster-Maintainers] Build failed in Jenkins: regression-test-burn-in #3899

jenkins at build.gluster.org jenkins at build.gluster.org
Tue Mar 6 12:02:55 UTC 2018


See <https://build.gluster.org/job/regression-test-burn-in/3899/display/redirect?page=changes>

Changes:

[Kaleb S. KEITHLEY] build: address linkage issues

[Raghavendra G] fuse: enable proper "fgetattr"-like semantics

[Poornima] nl-cache: Fix coverity issue RETURN_LOCAL

[Poornima] nl-cache: Fix coverity issue RESOURCE_LEAK

[Pranith Kumar K] cluster/afr: Remove compound-fops usage in afr

[atin] glusterd : memory leak in mgmt_v3 lock functionality

------------------------------------------
[...truncated 2.44 MB...]
#0  0x00007feeade1f7a3 in select () from /lib64/libc.so.6
No symbol table info available.
#1  0x00007feea0c95e3b in changelog_ev_dispatch (data=0x7fed14d91ee8) at <https://build.gluster.org/job/regression-test-burn-in/ws/xlators/features/changelog/src/changelog-ev-handle.c>:351
        ret = 3
        opaque = 0x0
        this = 0x7fed14d32500
        c_clnt = 0x7fed14d91ee8
        tv = {tv_sec = 0, tv_usec = 725462}
        __FUNCTION__ = "changelog_ev_dispatch"
#2  0x00007feeae55be25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3  0x00007feeade2834d in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 6 (Thread 0x7fec68999700 (LWP 4211)):
#0  0x00007feeae55f945 in pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1  0x00007feea0c95829 in changelog_ev_connector (data=0x7fed14d91ee8) at <https://build.gluster.org/job/regression-test-burn-in/ws/xlators/features/changelog/src/changelog-ev-handle.c>:205
        this = 0x7fed14d32500
        c_clnt = 0x7fed14d91ee8
        crpc = 0x0
        __FUNCTION__ = "changelog_ev_connector"
#2  0x00007feeae55be25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3  0x00007feeade2834d in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 5 (Thread 0x7fec68a9a700 (LWP 4210)):
#0  0x00007feeae55f945 in pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1  0x00007feea0a57295 in br_stub_worker (data=0x7fed14d33f10) at <https://build.gluster.org/job/regression-test-burn-in/ws/xlators/features/bit-rot/src/stub/bit-rot-stub-helpers.c>:327
        priv = 0x7fed14d8af00
        this = 0x7fed14d33f10
        stub = 0x0
#2  0x00007feeae55be25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3  0x00007feeade2834d in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 4 (Thread 0x7fec6929b700 (LWP 4209)):
#0  0x00007feeae55f945 in pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1  0x00007feea0a5b454 in br_stub_signth (arg=0x7fed14d33f10) at <https://build.gluster.org/job/regression-test-burn-in/ws/xlators/features/bit-rot/src/stub/bit-rot-stub.c>:868
        this = 0x7fed14d33f10
        priv = 0x7fed14d8af00
        sigstub = 0x0
#2  0x00007feeae55be25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3  0x00007feeade2834d in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 3 (Thread 0x7fec692dc700 (LWP 4208)):
#0  0x00007feeae55fcf2 in pthread_cond_timedwait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1  0x00007fee9b9b3727 in iot_worker (data=0x7fed14d6ed80) at <https://build.gluster.org/job/regression-test-burn-in/ws/xlators/performance/io-threads/src/io-threads.c>:195
        conf = 0x7fed14d6ed80
        this = 0x7fed14d3d920
        stub = 0x0
        sleep_till = {tv_sec = 1520331137, tv_nsec = 234406097}
        ret = 0
        pri = -1
        bye = false
        __FUNCTION__ = "iot_worker"
#2  0x00007feeae55be25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3  0x00007feeade2834d in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 2 (Thread 0x7fec693dd700 (LWP 4207)):
#0  0x00007feeae55f945 in pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1  0x00007fee9b15a329 in index_worker (data=0x7fed14d43af0) at <https://build.gluster.org/job/regression-test-burn-in/ws/xlators/features/index/src/index.c>:218
        priv = 0x7fed14d5e550
        this = 0x7fed14d43af0
        stub = 0x0
        bye = false
#2  0x00007feeae55be25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3  0x00007feeade2834d in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 1 (Thread 0x7fec6acdf700 (LWP 4205)):
#0  0x00007fee9af297d7 in quota_lookup (frame=0x7fee6403c5c8, this=0x7fee6001a950, loc=0x7fec6acde8d0, xattr_req=0x0) at <https://build.gluster.org/job/regression-test-burn-in/ws/xlators/features/quota/src/quota.c>:1641
        priv = 0x0
        ret = -1
        local = 0x0
        __FUNCTION__ = "quota_lookup"
#1  0x00007fee9ad06021 in io_stats_lookup (frame=0x7fee6405bfe8, this=0x7fee6001c130, loc=0x7fec6acde8d0, xdata=0x0) at <https://build.gluster.org/job/regression-test-burn-in/ws/xlators/debug/io-stats/src/io-stats.c>:2784
        _new = 0x7fee6403c5c8
        old_THIS = 0x7fee6001c130
        tmp_cbk = 0x7fee9acfa142 <io_stats_lookup_cbk>
        __FUNCTION__ = "io_stats_lookup"
#2  0x00007feeaf7da0b7 in default_lookup (frame=0x7fee6405bfe8, this=0x7fee6001dc80, loc=0x7fec6acde8d0, xdata=0x0) at defaults.c:2714
        old_THIS = 0x7fee6001dc80
        next_xl = 0x7fee6001c130
        next_xl_fn = 0x7fee9ad05c5c <io_stats_lookup>
        __FUNCTION__ = "default_lookup"
#3  0x00007feeaf75a7a7 in syncop_lookup (subvol=0x7fee6001dc80, loc=0x7fec6acde8d0, iatt=0x7fec6acde830, parent=0x0, xdata_in=0x0, xdata_out=0x0) at <https://build.gluster.org/job/regression-test-burn-in/ws/libglusterfs/src/syncop.c>:1238
        _new = 0x7fee6405bfe8
        old_THIS = 0x7fee9c02fbe0
        tmp_cbk = 0x7feeaf75a235 <syncop_lookup_cbk>
        task = 0x0
        frame = 0x7fee64034578
        args = {op_ret = 0, op_errno = 0, iatt1 = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, iatt2 = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, xattr = 0x0, statvfs_buf = {f_bsize = 0, f_frsize = 0, f_blocks = 0, f_bfree = 0, f_bavail = 0, f_files = 0, f_ffree = 0, f_favail = 0, f_fsid = 0, f_flag = 0, f_namemax = 0, __f_spare = {0, 0, 0, 0, 0, 0}}, vector = 0x0, count = 0, iobref = 0x0, buffer = 0x0, xdata = 0x0, flock = {l_type = 0, l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len = 0, data = '\000' <repeats 1023 times>}}, lease = {cmd = 0, lease_type = NONE, lease_id = '\000' <repeats 15 times>, lease_flags = 0}, uuid = '\000' <repeats 15 times>, errstr = 0x0, dict = 0x0, lock_dict = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, barrier = {guard = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, cond = {__data = {__lock = 0, __futex = 0, __total_seq = 0, __wakeup_seq = 0, __woken_seq = 0, __mutex = 0x0, __nwaiters = 0, __broadcast_seq = 0}, __size = '\000' <repeats 47 times>, __align = 0}, waitq = {next = 0x0, prev = 0x0}, count = 0, waitfor = 0}, task = 0x0, mutex = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, cond = {__data = {__lock = 0, __futex = 0, __total_seq = 0, __wakeup_seq = 0, __woken_seq = 0, __mutex = 0x0, __nwaiters = 0, __broadcast_seq = 0}, __size = '\000' <repeats 47 times>, __align = 0}, done = 0, entries = {{list = {next = 0x0, prev = 0x0}, {next = 0x0, prev = 0x0}}, d_ino = 0, d_off = 0, d_len = 0, d_type = 0, d_stat = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, dict = 0x0, inode = 0x0, d_name = 0x7fec6acde368 ""}, offset = 0, locklist = {list = {next = 0x0, prev = 0x0}, flock = {l_type = 0, l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len = 0, data = '\000' <repeats 1023 times>}}, client_uid = 0x0, lk_flags = 0}}
        __FUNCTION__ = "syncop_lookup"
#4  0x00007fee9a884ee5 in server_first_lookup (this=0x7fee9c02fbe0, client=0x7fecd1794730, reply=0x7fecd1789618) at <https://build.gluster.org/job/regression-test-burn-in/ws/xlators/protocol/server/src/server-handshake.c>:382
        loc = {path = 0x7fee9a8ab29d "/", name = 0x7fee9a8ab3a2 "", inode = 0x7fecd1797e48, parent = 0x0, gfid = '\000' <repeats 15 times>, "\001", pargfid = '\000' <repeats 15 times>}
        iatt = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}
        dict = 0x0
        ret = 0
        xl = 0x7fee6001dc80
        msg = 0x0
        inode = 0x0
        bname = 0x0
        str = 0x0
        tmp = 0x0
        saveptr = 0x0
        __FUNCTION__ = "server_first_lookup"
#5  0x00007fee9a8868a0 in server_setvolume (req=0x7fecd0bee228) at <https://build.gluster.org/job/regression-test-burn-in/ws/xlators/protocol/server/src/server-handshake.c>:862
        args = {dict = {dict_len = 840, dict_val = 0x7fecd1789700 ""}}
        rsp = 0x0
        client = 0x7fecd1794730
        serv_ctx = 0x7fecd1794b20
        conf = 0x7fee9c0371c0
        peerinfo = 0x7fee9000e4d0
        reply = 0x7fecd1789618
        config_params = 0x7fecd1789a78
        params = 0x7fecd13b44e8
        name = 0x7fecd178c5d0 "/d/backends/vol01/brick2"
        client_uid = 0x7fecd178bd40 "CTX_ID:adee0d12-3490-45f8-8a87-688a537f6fae-GRAPH_ID:0-PID:3078-HOST:builder103.cloud.gluster.org-PC_NAME:patchy-vol01-client-2-RECON_NO:-0"
        clnt_version = 0x7fecd178b940 "4.1dev"
        xl = 0x7fee6001dc80
        msg = 0x0
        volfile_key = 0x7fecd178b730 "gluster/glustershd"
        this = 0x7fee9c02fbe0
        checksum = 0
        ret = 0
        op_ret = 0
        op_errno = 22
        buf = 0x0
        opversion = 40100
        xprt = 0x7fee9c036d50
        fop_version = 1298437
        mgmt_version = 0
        ctx = 0x24ca010
        tmp = 0x7fee605c9cd0
        subdir_mount = 0x0
        client_name = 0x7fecd178bb40 "glustershd"
        __FUNCTION__ = "server_setvolume"
        __PRETTY_FUNCTION__ = "server_setvolume"
#6  0x00007feeaf4c88f2 in rpcsvc_handle_rpc_call (svc=0x7fee9c044830, trans=0x7fee9000e410, msg=0x7fecd17894c0) at <https://build.gluster.org/job/regression-test-burn-in/ws/rpc/rpc-lib/src/rpcsvc.c>:721
        actor = 0x7fee9aabd460 <gluster_handshake_actors+64>
        actor_fn = 0x7fee9a8851d9 <server_setvolume>
        req = 0x7fecd0bee228
        ret = -1
        port = 47546
        is_unix = false
        empty = false
        unprivileged = true
        reply = 0x0
        drc = 0x0
        __FUNCTION__ = "rpcsvc_handle_rpc_call"
#7  0x00007feeaf4c8c45 in rpcsvc_notify (trans=0x7fee9000e410, mydata=0x7fee9c044830, event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7fecd17894c0) at <https://build.gluster.org/job/regression-test-burn-in/ws/rpc/rpc-lib/src/rpcsvc.c>:815
        ret = -1
        msg = 0x7fecd17894c0
        new_trans = 0x0
        svc = 0x7fee9c044830
        listener = 0x0
        __FUNCTION__ = "rpcsvc_notify"
#8  0x00007feeaf4ce85b in rpc_transport_notify (this=0x7fee9000e410, event=RPC_TRANSPORT_MSG_RECEIVED, data=0x7fecd17894c0) at <https://build.gluster.org/job/regression-test-burn-in/ws/rpc/rpc-lib/src/rpc-transport.c>:537
        ret = -1
        __FUNCTION__ = "rpc_transport_notify"
#9  0x00007feea42e9f18 in socket_event_poll_in (this=0x7fee9000e410, notify_handled=true) at <https://build.gluster.org/job/regression-test-burn-in/ws/rpc/rpc-transport/socket/src/socket.c>:2462
        ret = 0
        pollin = 0x7fecd17894c0
        priv = 0x7fee9000e970
        ctx = 0x24ca010
#10 0x00007feea42ea586 in socket_event_handler (fd=331, idx=118, gen=1, data=0x7fee9000e410, poll_in=1, poll_out=0, poll_err=0) at <https://build.gluster.org/job/regression-test-burn-in/ws/rpc/rpc-transport/socket/src/socket.c>:2618
        this = 0x7fee9000e410
        priv = 0x7fee9000e970
        ret = 0
        ctx = 0x24ca010
        socket_closed = false
        notify_handled = false
        __FUNCTION__ = "socket_event_handler"
#11 0x00007feeaf781411 in event_dispatch_epoll_handler (event_pool=0x2501c20, event=0x7fec6acdeea0) at <https://build.gluster.org/job/regression-test-burn-in/ws/libglusterfs/src/event-epoll.c>:579
        ev_data = 0x7fec6acdeea4
        slot = 0x2535b00
        handler = 0x7feea42ea2b8 <socket_event_handler>
        data = 0x7fee9000e410
        idx = 118
        gen = 1
        ret = -1
        fd = 331
        handled_error_previously = false
        __FUNCTION__ = "event_dispatch_epoll_handler"
#12 0x00007feeaf781704 in event_dispatch_epoll_worker (data=0x7fecd13b4050) at <https://build.gluster.org/job/regression-test-burn-in/ws/libglusterfs/src/event-epoll.c>:655
        event = {events = 1, data = {ptr = 0x100000076, fd = 118, u32 = 118, u64 = 4294967414}}
        ret = 1
        ev_data = 0x7fecd13b4050
        event_pool = 0x2501c20
        myindex = 78
        timetodie = 0
        __FUNCTION__ = "event_dispatch_epoll_worker"
#13 0x00007feeae55be25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#14 0x00007feeade2834d in clone () from /lib64/libc.so.6
No symbol table info available.
=========================================================
              Finish backtrace
         program name : /build/install/sbin/glusterfsd
         corefile     : /glusterepoll77-3087.core
=========================================================

+ rm -f /build/install/cores/gdbout.txt
+ sort /build/install/cores/liblist.txt
+ uniq
+ cat /build/install/cores/liblist.txt.tmp
+ grep -v /build/install
+ tar -cf /archives/archived_builds/build-install-regression-test-burn-in-3899.tar /build/install/sbin /build/install/bin /build/install/lib /build/install/libexec /build/install/cores
tar: Removing leading `/' from member names
+ tar -rhf /archives/archived_builds/build-install-regression-test-burn-in-3899.tar -T /build/install/cores/liblist.txt
tar: Removing leading `/' from member names
+ bzip2 /archives/archived_builds/build-install-regression-test-burn-in-3899.tar
+ rm -f /build/install/cores/liblist.txt
+ rm -f /build/install/cores/liblist.txt.tmp
+ find /archives -size +1G -delete -type f
+ echo 'Cores and build archived in http://builder103.cloud.gluster.org/archived_builds/build-install-regression-test-burn-in-3899.tar.bz2'
Cores and build archived in http://builder103.cloud.gluster.org/archived_builds/build-install-regression-test-burn-in-3899.tar.bz2
+ echo 'Open core using the following command to get a proper stack'
Open core using the following command to get a proper stack
+ echo 'Example: From root of extracted tarball'
Example: From root of extracted tarball
+ echo '\t\tgdb -ex '\''set sysroot ./'\'' -ex '\''core-file ./build/install/cores/xxx.core'\'' <target, say ./build/install/sbin/glusterd>'
\t\tgdb -ex 'set sysroot ./' -ex 'core-file ./build/install/cores/xxx.core' <target, say ./build/install/sbin/glusterd>
+ RET=1
+ '[' 1 -ne 0 ']'
+ tar -czf <https://build.gluster.org/job/regression-test-burn-in/3899/artifact/glusterfs-logs.tgz> /var/log/glusterfs /var/log/messages /var/log/messages-20180211 /var/log/messages-20180218 /var/log/messages-20180226 /var/log/messages-20180304
tar: Removing leading `/' from member names
+ scp -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no -i <https://build.gluster.org/job/regression-test-burn-in/ws/> glusterfs-logs.tgz _logs_collector at http.int.rht.gluster.org:/var/www/glusterfs-logs/regression-test-burn-in-3899.tgz
ssh: connect to host http.int.rht.gluster.org port 22: Connection timed out
lost connection
+ true
+ case $(uname -s) in
++ uname -s
+ /sbin/sysctl -w kernel.core_pattern=/%e-%p.core
kernel.core_pattern = /%e-%p.core
+ exit 1
Build step 'Execute shell' marked build as failure


More information about the maintainers mailing list