[Gluster-Maintainers] Build failed in Jenkins: regression-test-with-multiplex #652

jenkins at build.gluster.org jenkins at build.gluster.org
Tue Feb 27 17:53:17 UTC 2018


See <https://build.gluster.org/job/regression-test-with-multiplex/652/display/redirect?page=changes>

Changes:

[Amar Tumballi] options: framework for options levels

[Amar Tumballi] libglusterfs: move compat RPC/XDR #defines to eliminate warnings

[Poornima] io-cache: Fix coverity issue

------------------------------------------
[...truncated 1.30 MB...]
        ret = 0
#3  0x00007efbf7771ae1 in ec_shd_index_healer (data=0x7efbf0027ba0) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/xlators/cluster/ec/src/ec-heald.c>:342
        healer = 0x7efbf0027ba0
        this = 0x7efbf0017950
        ec = 0x7efbf0025c90
        __FUNCTION__ = "ec_shd_index_healer"
#4  0x00007efc04187e25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#5  0x00007efc03a5434d in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 5 (Thread 0x7efbf5ab4700 (LWP 30156)):
#0  0x00007efc0418b945 in pthread_cond_wait@@GLIBC_2.3.2 () from /lib64/libpthread.so.0
No symbol table info available.
#1  0x00007efc05389aa4 in syncop_getxattr (subvol=0x7efbf0017950, loc=0x7efbf5ab3aa0, dict=0x0, key=0x7efbf777d4b0 "trusted.ec.heal", xdata_in=0x0, xdata_out=0x0) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/libglusterfs/src/syncop.c>:1690
        task = 0x0
        frame = 0x7efbd4002b78
        args = {op_ret = 0, op_errno = 0, iatt1 = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, iatt2 = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, xattr = 0x0, statvfs_buf = {f_bsize = 0, f_frsize = 0, f_blocks = 0, f_bfree = 0, f_bavail = 0, f_files = 0, f_ffree = 0, f_favail = 0, f_fsid = 0, f_flag = 0, f_namemax = 0, __f_spare = {0, 0, 0, 0, 0, 0}}, vector = 0x0, count = 0, iobref = 0x0, buffer = 0x0, xdata = 0x0, flock = {l_type = 0, l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len = 0, data = '\000' <repeats 1023 times>}}, lease = {cmd = 0, lease_type = NONE, lease_id = '\000' <repeats 15 times>, lease_flags = 0}, uuid = '\000' <repeats 15 times>, errstr = 0x0, dict = 0x0, lock_dict = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, barrier = {guard = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, cond = {__data = {__lock = 0, __futex = 0, __total_seq = 0, __wakeup_seq = 0, __woken_seq = 0, __mutex = 0x0, __nwaiters = 0, __broadcast_seq = 0}, __size = '\000' <repeats 47 times>, __align = 0}, waitq = {next = 0x0, prev = 0x0}, count = 0, waitfor = 0}, task = 0x0, mutex = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 1, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 12 times>, "\001", '\000' <repeats 26 times>, __align = 0}, cond = {__data = {__lock = 0, __futex = 1, __total_seq = 1, __wakeup_seq = 0, __woken_seq = 0, __mutex = 0x7efbf5ab3478, __nwaiters = 2, __broadcast_seq = 0}, __size = "\000\000\000\000\001\000\000\000\001", '\000' <repeats 23 times>, "x4\253\365\373~\000\000\002\000\000\000\000\000\000", __align = 4294967296}, done = 0, entries = {{list = {next = 0x0, prev = 0x0}, {next = 0x0, prev = 0x0}}, d_ino = 0, d_off = 0, d_len = 0, d_type = 0, d_stat = {ia_flags = 0, ia_ino = 0, ia_dev = 0, ia_rdev = 0, ia_size = 0, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_mtime = 0, ia_ctime = 0, ia_btime = 0, ia_atime_nsec = 0, ia_mtime_nsec = 0, ia_ctime_nsec = 0, ia_btime_nsec = 0, ia_attributes = 0, ia_attributes_mask = 0, ia_gfid = '\000' <repeats 15 times>, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, dict = 0x0, inode = 0x0, d_name = 0x7efbf5ab35a8 ""}, offset = 0, locklist = {list = {next = 0x0, prev = 0x0}, flock = {l_type = 0, l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len = 0, data = '\000' <repeats 1023 times>}}, client_uid = 0x0, lk_flags = 0}}
        __FUNCTION__ = "syncop_getxattr"
#2  0x00007efbf777147a in ec_shd_selfheal (healer=0x7efbf0027c10, child=2, loc=0x7efbf5ab3aa0) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/xlators/cluster/ec/src/ec-heald.c>:187
No locals.
#3  0x00007efbf777162e in ec_shd_index_heal (subvol=0x7efbf000d470, entry=0x7efbe804e0b0, parent=0x7efbf5ab3e00, data=0x7efbf0027c10) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/xlators/cluster/ec/src/ec-heald.c>:226
        healer = 0x7efbf0027c10
        ec = 0x7efbf0025c90
        loc = {path = 0x7efbd4004bc0 "/mnt/glusterfs/0/test_dir", name = 0x0, inode = 0x7efbd4002858, parent = 0x0, gfid = "v\316\202p\243\313G\370\273\211ܻia\376\301", pargfid = '\000' <repeats 15 times>}
        ret = 0
        __FUNCTION__ = "ec_shd_index_heal"
#4  0x00007efc053adce6 in syncop_mt_dir_scan (frame=0x0, subvol=0x7efbf000d470, loc=0x7efbf5ab3e00, pid=-6, data=0x7efbf0027c10, fn=0x7efbf777147c <ec_shd_index_heal>, xdata=0x7efbd4001398, max_jobs=1, max_qlen=1024) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/libglusterfs/src/syncop-utils.c>:407
        fd = 0x7efbd4002338
        offset = 12884901884
        last = 0x7efbe80665d0
        ret = 0
        retval = 0
        q = {{list = {next = 0x7efbf5ab3cb0, prev = 0x7efbf5ab3cb0}, {next = 0x7efbf5ab3cb0, prev = 0x7efbf5ab3cb0}}, d_ino = 139620623531296, d_off = 139620884054374, d_len = 4121640240, d_type = 32257, d_stat = {ia_flags = 139620058670536, ia_ino = 139620653716743, ia_dev = 139620058665880, ia_rdev = 139620623531280, ia_size = 17190848, ia_nlink = 3892364040, ia_uid = 2681838932, ia_gid = 0, ia_blksize = 0, ia_blocks = 139620058665984, ia_atime = 3556779464, ia_mtime = 139620623531392, ia_ctime = 139620884054846, ia_btime = 139620623531376, ia_atime_nsec = 3556779464, ia_mtime_nsec = 32507, ia_ctime_nsec = 4121640320, ia_btime_nsec = 32507, ia_attributes = 139620884057965, ia_attributes_mask = 139620528477296, ia_gfid = "\310%\000\324\373~\000\000\a\325w\367\373~\000", ia_type = 3556774808, ia_prot = {suid = 1 '\001', sgid = 1 '\001', sticky = 0 '\000', owner = {read = 0 '\000', write = 1 '\001', exec = 1 '\001'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, dict = 0x7efbd40025c8, inode = 0x53276cb, d_name = 0x7efbf5ab3d80 "\300=\253", <incomplete sequence \365>}
        entry = 0x7efbe804e0b0
        tmp = 0x7efbe804e210
        jobs_running = 0
        qlen = 0
        cond = {__data = {__lock = 0, __futex = 0, __total_seq = 0, __wakeup_seq = 0, __woken_seq = 0, __mutex = 0x0, __nwaiters = 0, __broadcast_seq = 0}, __size = '\000' <repeats 47 times>, __align = 0}
        mut = {__data = {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0, __elision = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}
        cond_init = true
        mut_init = true
        entries = {{list = {next = 0x7efbe804e210, prev = 0x7efbe80665d0}, {next = 0x7efbe804e210, prev = 0x7efbe80665d0}}, d_ino = 1, d_off = 0, d_len = 49, d_type = 91, d_stat = {ia_flags = 139620623530992, ia_ino = 0, ia_dev = 0, ia_rdev = 511101108334, ia_size = 0, ia_nlink = 4121639919, ia_uid = 32507, ia_gid = 2, ia_blksize = 0, ia_blocks = 0, ia_atime = 139620623531088, ia_mtime = 124, ia_ctime = 139620623534528, ia_btime = 139620623533824, ia_atime_nsec = 87476006, ia_mtime_nsec = 32508, ia_ctime_nsec = 45, ia_btime_nsec = 32507, ia_attributes = 139620623531120, ia_attributes_mask = 139620528565568, ia_gfid = " \035\000\324\373~\000\000p<\253\365\373~\000", ia_type = 87474883, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 1 '\001', owner = {read = 0 '\000', write = 1 '\001', exec = 1 '\001'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, dict = 0x66, inode = 0x7efc05408ddd, d_name = 0x7efbf5ab3c40 ""}
#5  0x00007efbf7771860 in ec_shd_index_sweep (healer=0x7efbf0027c10) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/xlators/cluster/ec/src/ec-heald.c>:265
        loc = {path = 0x0, name = 0x0, inode = 0x7efbd4001288, parent = 0x0, gfid = '\000' <repeats 15 times>, pargfid = '\000' <repeats 15 times>}
        ec = 0x7efbf0025c90
        ret = 0
        subvol = 0x7efbf000d470
        xdata = 0x7efbd4001398
        __FUNCTION__ = "ec_shd_index_sweep"
#6  0x00007efbf7771ba5 in ec_shd_index_healer (data=0x7efbf0027c10) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/xlators/cluster/ec/src/ec-heald.c>:351
        healer = 0x7efbf0027c10
        this = 0x7efbf0017950
        ec = 0x7efbf0025c90
        __FUNCTION__ = "ec_shd_index_healer"
#7  0x00007efc04187e25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#8  0x00007efc03a5434d in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 4 (Thread 0x7efbfb930700 (LWP 30115)):
#0  0x00007efc03a1b1ad in nanosleep () from /lib64/libc.so.6
No symbol table info available.
#1  0x00007efc03a1b044 in sleep () from /lib64/libc.so.6
No symbol table info available.
#2  0x00007efc0536cdc6 in pool_sweeper (arg=0x0) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/libglusterfs/src/mem-pool.c>:470
        state = {death_row = {next = 0x0, prev = 0x0}, cold_lists = {0x0 <repeats 1024 times>}, n_cold_lists = 0}
        pool_list = 0x0
        next_pl = 0x0
        pt_pool = 0x0
        i = 0
        begin_time = {tv_sec = 0, tv_usec = 0}
        end_time = {tv_sec = 0, tv_usec = 0}
        elapsed = {tv_sec = 0, tv_usec = 0}
#3  0x00007efc04187e25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#4  0x00007efc03a5434d in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 3 (Thread 0x7efbfc131700 (LWP 30114)):
#0  0x00007efc0418f371 in sigwait () from /lib64/libpthread.so.0
No symbol table info available.
#1  0x000000000040a4ca in ?? ()
No symbol table info available.
#2  0x0000000000000000 in ?? ()
No symbol table info available.

Thread 2 (Thread 0x7efbfc932700 (LWP 30113)):
#0  0x00007efc0418ee4d in nanosleep () from /lib64/libpthread.so.0
No symbol table info available.
#1  0x00007efc0534590c in gf_timer_proc (data=0x10677c0) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/libglusterfs/src/timer.c>:202
        now = 914888061073054
        now_ts = {tv_sec = 914888, tv_nsec = 61073054}
        reg = 0x10677c0
        sleepts = {tv_sec = 1, tv_nsec = 0}
        event = 0x7efbbc218190
        tmp = 0x7efbbc217140
        old_THIS = 0x7efc056382a0 <global_xlator>
#2  0x00007efc04187e25 in start_thread () from /lib64/libpthread.so.0
No symbol table info available.
#3  0x00007efc03a5434d in clone () from /lib64/libc.so.6
No symbol table info available.

Thread 1 (Thread 0x7efbc9ffb700 (LWP 30177)):
#0  0x00007efc039d75bd in malloc_consolidate () from /lib64/libc.so.6
No symbol table info available.
#1  0x00007efc039d84fe in _int_free () from /lib64/libc.so.6
No symbol table info available.
#2  0x00007efc0536cc39 in __gf_free (free_ptr=0x7efbe8096c70) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/libglusterfs/src/mem-pool.c>:360
        ptr = 0x7efbe8096c20
        mem_acct = 0x7efbf0021c00
        header = 0x7efbe8096c20
        __PRETTY_FUNCTION__ = "__gf_free"
#3  0x00007efc0536e238 in gf_dirent_entry_free (entry=0x7efbe8096c70) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/libglusterfs/src/gf-dirent.c>:182
No locals.
#4  0x00007efc0536e295 in gf_dirent_free (entries=0x7efbed9e9950) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/libglusterfs/src/gf-dirent.c>:198
        entry = 0x7efbe8096c70
        tmp = 0x7efbed9e9950
#5  0x00007efc053ae0a2 in syncop_dir_scan (subvol=0x7efbf00126d0, loc=0x7efbed9e9ae0, pid=-6, data=0x7efbed9e9a90, fn=0x7efbf77690b0 <ec_name_heal_handler>) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/libglusterfs/src/syncop-utils.c>:503
        fd = 0x7efbb801c8d8
        offset = 12884901886
        entry = 0x7efbed9e9950
        ret = 0
        entries = {{list = {next = 0x7efbed9e9950, prev = 0x7efbed9e9950}, {next = 0x7efbed9e9950, prev = 0x7efbed9e9950}}, d_ino = 0, d_off = 0, d_len = 0, d_type = 0, d_stat = {ia_flags = 139620875155011, ia_ino = 5186075065184060016, ia_dev = 13978717400155863945, ia_rdev = 0, ia_size = 72057594037927936, ia_nlink = 0, ia_uid = 0, ia_gid = 1311241728, ia_blksize = 3257294259, ia_blocks = 16, ia_atime = 139620488485344, ia_mtime = 0, ia_ctime = 139620884176747, ia_btime = 139620887246224, ia_atime_nsec = 3556780128, ia_mtime_nsec = 32507, ia_ctime_nsec = 4026628432, ia_btime_nsec = 32507, ia_attributes = 139620058671192, ia_attributes_mask = 0, ia_gfid = "`(\000\324\a\000\000\000\060\232\236\355\373~\000", ia_type = 87323569, ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 1 '\001', owner = {read = 0 '\000', write = 1 '\001', exec = 1 '\001'}, group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}}, dict = 0x0, inode = 0x7efbd4002858, d_name = 0x7efbed9e9a20 "\330\310\001\270\373~"}
#6  0x00007efbf776942d in ec_heal_names (frame=0x7efbe000f4b8, ec=0x7efbf0025c90, inode=0x7efbd4002858, participants=0x7efbed9e9b90 "\001\001\001\001\001\001") at <https://build.gluster.org/job/regression-test-with-multiplex/ws/xlators/cluster/ec/src/ec-heal.c>:1469
        i = 4
        j = 6
        loc = {path = 0x0, name = 0x0, inode = 0x7efbd4002858, parent = 0x0, gfid = "v\316\202p\243\313G\370\273\211ܻia\376\301", pargfid = '\000' <repeats 15 times>}
        name_data = {frame = 0x7efbe000f4b8, participants = 0x7efbed9e9b90 "\001\001\001\001\001\001", failed_on = 0x7efbed9e9a50 "", gfidless = 0x0, enoent = 0x0, same = 0x0, name = 0x0, parent = 0x0, replies = 0x0}
        ret = 0
#7  0x00007efbf7769aec in __ec_heal_entry (frame=0x7efbe000f4b8, ec=0x7efbf0025c90, inode=0x7efbd4002858, heal_on=0x7efbed9ef370 "\001\001\001\001\001\001", sources=0x7efbed9f4e30 "", healed_sinks=0x7efbed9f4e10 "\001\001") at <https://build.gluster.org/job/regression-test-with-multiplex/ws/xlators/cluster/ec/src/ec-heal.c>:1536
        locked_on = 0x7efbed9ef210 "\001\001\001\001\001\001"
        output = 0x7efbed9ef1f0 "\001\001\001\001\001\001"
        versions = 0x7efbed9ef1b0
        dirty = 0x7efbed9ef170
        participants = 0x7efbed9e9b90 "\001\001\001\001\001\001"
        replies = 0x7efbed9e9bb0
        ret = 2
        source = 2
        i = 6
        __FUNCTION__ = "__ec_heal_entry"
#8  0x00007efbf776a023 in ec_heal_entry (frame=0x7efbe000f4b8, ec=0x7efbf0025c90, inode=0x7efbd4002858, sources=0x7efbed9f4e30 "", healed_sinks=0x7efbed9f4e10 "\001\001") at <https://build.gluster.org/job/regression-test-with-multiplex/ws/xlators/cluster/ec/src/ec-heal.c>:1585
        locked_on = 0x7efbed9ef370 "\001\001\001\001\001\001"
        up_subvols = 0x7efbed9ef330 "\001\001\001\001\001\001"
        output = 0x7efbed9ef350 ""
        selfheal_domain = "patchy-disperse-0:self-heal", '\000' <repeats 996 times>
        ret = 6
        replies = 0x7efbed9ef390
        __FUNCTION__ = "ec_heal_entry"
#9  0x00007efbf776eadd in ec_heal_do (this=0x7efbf0017950, data=0x7efbd4005ce8, loc=0x7efbd4005fe0, partial=0) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/xlators/cluster/ec/src/ec-heal.c>:2538
        frame = 0x7efbe000f4b8
        participants = 0x7efbed9f4eb0 "\001\001\001\001\001\001"
        msources = 0x7efbed9f4e70 "\001\001\001\001\001\001"
        mhealed_sinks = 0x7efbed9f4e50 ""
        sources = 0x7efbed9f4e30 ""
        healed_sinks = 0x7efbed9f4e10 "\001\001"
        ec = 0x7efbf0025c90
        ret = 0
        op_ret = 0
        op_errno = 0
        mgood = 63
        mbad = 0
        good = 0
        bad = 0
        fop = 0x7efbd4005ce8
        blocking = false
        need_heal = EC_HEAL_MAYBE
        up_subvols = 0x7efbed9f4e90 "\001\001\001\001\001\001"
        up_bricks = '\000' <repeats 31 times>
        __FUNCTION__ = "ec_heal_do"
#10 0x00007efbf776ec7c in ec_synctask_heal_wrap (opaque=0x7efbd4005ce8) at <https://build.gluster.org/job/regression-test-with-multiplex/ws/xlators/cluster/ec/src/ec-heal.c>:2570
        fop = 0x7efbd4005ce8
#11 0x00007efc053837a6 in synctask_wrap () at <https://build.gluster.org/job/regression-test-with-multiplex/ws/libglusterfs/src/syncop.c>:375
        task = 0x7efbd400b650
#12 0x00007efc039a2d40 in ?? () from /lib64/libc.so.6
No symbol table info available.
#13 0x0000000000000000 in ?? ()
No symbol table info available.
=========================================================
              Finish backtrace
         program name : /build/install/sbin/glusterfs
         corefile     : /glustersproc9-30112.core
=========================================================

+ rm -f /build/install/cores/gdbout.txt
+ for corefile in '$CORELIST'
+ getliblistfromcore /build/install/cores/glusteriotwr0-16246.core
+ rm -f /build/install/cores/gdbout.txt
+ gdb -c /build/install/cores/glusteriotwr0-16246.core -q -ex 'set pagination off' -ex 'info sharedlibrary' -ex q
+ set +x
+ rm -f /build/install/cores/gdbout.txt
+ for corefile in '$CORELIST'
+ getliblistfromcore /build/install/cores/glusteriotwr6-30973.core
+ rm -f /build/install/cores/gdbout.txt
+ gdb -c /build/install/cores/glusteriotwr6-30973.core -q -ex 'set pagination off' -ex 'info sharedlibrary' -ex q
+ set +x
+ rm -f /build/install/cores/gdbout.txt
+ for corefile in '$CORELIST'
+ getliblistfromcore /build/install/cores/glustersproc1-27012.core
+ rm -f /build/install/cores/gdbout.txt
+ gdb -c /build/install/cores/glustersproc1-27012.core -q -ex 'set pagination off' -ex 'info sharedlibrary' -ex q
+ set +x
+ rm -f /build/install/cores/gdbout.txt
+ for corefile in '$CORELIST'
+ getliblistfromcore /build/install/cores/glustersproc2-16391.core
+ rm -f /build/install/cores/gdbout.txt
+ gdb -c /build/install/cores/glustersproc2-16391.core -q -ex 'set pagination off' -ex 'info sharedlibrary' -ex q
+ set +x
+ rm -f /build/install/cores/gdbout.txt
+ for corefile in '$CORELIST'
+ getliblistfromcore /build/install/cores/glustersproc9-30112.core
+ rm -f /build/install/cores/gdbout.txt
+ gdb -c /build/install/cores/glustersproc9-30112.core -q -ex 'set pagination off' -ex 'info sharedlibrary' -ex q
+ set +x
+ rm -f /build/install/cores/gdbout.txt
+ sort /build/install/cores/liblist.txt
+ uniq
+ cat /build/install/cores/liblist.txt.tmp
+ grep -v /build/install
+ tar -cf /archives/archived_builds/build-install-regression-test-with-multiplex-652.tar /build/install/sbin /build/install/bin /build/install/lib /build/install/libexec /build/install/cores
tar: Removing leading `/' from member names
+ tar -rhf /archives/archived_builds/build-install-regression-test-with-multiplex-652.tar -T /build/install/cores/liblist.txt
tar: Removing leading `/' from member names
+ bzip2 /archives/archived_builds/build-install-regression-test-with-multiplex-652.tar
+ rm -f /build/install/cores/liblist.txt
+ rm -f /build/install/cores/liblist.txt.tmp
+ find /archives -size +1G -delete -type f
+ echo 'Cores and build archived in http://builder102.cloud.gluster.org/archived_builds/build-install-regression-test-with-multiplex-652.tar.bz2'
Cores and build archived in http://builder102.cloud.gluster.org/archived_builds/build-install-regression-test-with-multiplex-652.tar.bz2
+ echo 'Open core using the following command to get a proper stack'
Open core using the following command to get a proper stack
+ echo 'Example: From root of extracted tarball'
Example: From root of extracted tarball
+ echo '\t\tgdb -ex '\''set sysroot ./'\'' -ex '\''core-file ./build/install/cores/xxx.core'\'' <target, say ./build/install/sbin/glusterd>'
\t\tgdb -ex 'set sysroot ./' -ex 'core-file ./build/install/cores/xxx.core' <target, say ./build/install/sbin/glusterd>
+ RET=1
+ '[' 1 -ne 0 ']'
+ tar -czf <https://build.gluster.org/job/regression-test-with-multiplex/652/artifact/glusterfs-logs.tgz> /var/log/glusterfs /var/log/messages /var/log/messages-20180212 /var/log/messages-20180220 /var/log/messages-20180225
tar: Removing leading `/' from member names
+ scp -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no -i <https://build.gluster.org/job/regression-test-with-multiplex/ws/> glusterfs-logs.tgz _logs_collector at http.int.rht.gluster.org:/var/www/glusterfs-logs/regression-test-with-multiplex-652.tgz
ssh: connect to host http.int.rht.gluster.org port 22: Connection timed out
lost connection
+ true
+ case $(uname -s) in
++ uname -s
+ /sbin/sysctl -w kernel.core_pattern=/%e-%p.core
kernel.core_pattern = /%e-%p.core
+ exit 1
Build step 'Execute shell' marked build as failure


More information about the maintainers mailing list