[Gluster-devel] [Gluster-Maintainers] Release 3.10 spurious(?) regression failures in the past week
Shyam
srangana at redhat.com
Mon Feb 6 19:21:49 UTC 2017
Update from week of: (2017-01-30 to 2017-02-06)
Failure stats and actions:
1) ./tests/basic/tier/tier.t
Core dump needs attention
https://build.gluster.org/job/centos6-regression/3046/consoleText
Looks like the tier rebalance process has crashed (see below for the
stack details)
2) ./tests/basic/ec/ec-background-heals.t
Marked as bad in master, not in release-3.10. May cause unwanted
failures in 3.10 and as a result marked this as bad in 3.10 as well.
Commit: https://review.gluster.org/16549
3) ./tests/bitrot/bug-1373520.t
Marked as bad in master, not in release-3.10. May cause unwanted
failures in 3.10 and as a result marked this as bad in 3.10 as well.
Commit: https://review.gluster.org/16549
Thanks,
Shyam
On 01/30/2017 03:00 PM, Shyam wrote:
> Hi,
>
> The following is a list of spurious(?) regression failures in the 3.10
> branch last week (from fstat.gluster.org).
>
> Request component owner or other devs to take a look at the failures,
> and weed out real issues.
>
> Regression failures 3.10:
>
> Summary:
> 1) https://build.gluster.org/job/centos6-regression/2960/consoleFull
> ./tests/basic/ec/ec-background-heals.t
>
> 2) https://build.gluster.org/job/centos6-regression/2963/consoleFull
> <glusterd Core dumped>
> ./tests/basic/volume-snapshot.t
>
> 3) https://build.gluster.org/job/netbsd7-regression/2694/consoleFull
> ./tests/basic/afr/self-heald.t
>
> 4) https://build.gluster.org/job/centos6-regression/2954/consoleFull
> ./tests/basic/tier/legacy-many.t
>
> 5) https://build.gluster.org/job/centos6-regression/2858/consoleFull
> ./tests/bugs/bitrot/bug-1245981.t
>
> 6) https://build.gluster.org/job/netbsd7-regression/2637/consoleFull
> ./tests/basic/afr/self-heal.t
>
> 7) https://build.gluster.org/job/netbsd7-regression/2624/consoleFull
> ./tests/encryption/crypt.t
>
> Thanks,
> Shyam
Core details from
https://build.gluster.org/job/centos6-regression/3046/consoleText
Core was generated by `/build/install/sbin/glusterfs -s localhost
--volfile-id tierd/patchy -p /var/li'.
Program terminated with signal 11, Segmentation fault.
#0 0x00007ffb62c2c4c4 in __strchr_sse42 () from /lib64/libc.so.6
Thread 1 (Thread 0x7ffb5a169700 (LWP 467)):
#0 0x00007ffb62c2c4c4 in __strchr_sse42 () from /lib64/libc.so.6
No symbol table info available.
#1 0x00007ffb56b7789f in dht_filter_loc_subvol_key
(this=0x7ffb50015930, loc=0x7ffb2c002de4, new_loc=0x7ffb2c413f80,
subvol=0x7ffb2c413fc0) at
/home/jenkins/root/workspace/centos6-regression/xlators/cluster/dht/src/dht-helper.c:307
new_name = 0x0
new_path = 0x0
trav = 0x0
key = '\000' <repeats 1023 times>
ret = 0
#2 0x00007ffb56bb2ce4 in dht_lookup (frame=0x7ffb4c00623c,
this=0x7ffb50015930, loc=0x7ffb2c002de4, xattr_req=0x7ffb4c00949c) at
/home/jenkins/root/workspace/centos6-regression/xlators/cluster/dht/src/dht-common.c:2494
subvol = 0x0
hashed_subvol = 0x0
local = 0x7ffb4c00636c
conf = 0x7ffb5003f380
ret = -1
op_errno = -1
layout = 0x0
i = 0
call_cnt = 0
new_loc = {path = 0x0, name = 0x0, inode = 0x0, parent = 0x0,
gfid = '\000' <repeats 15 times>, pargfid = '\000' <repeats 15 times>}
__FUNCTION__ = "dht_lookup"
#3 0x00007ffb63ff6f5c in syncop_lookup (subvol=0x7ffb50015930,
loc=0x7ffb2c002de4, iatt=0x7ffb2c415af0, parent=0x0,
xdata_in=0x7ffb4c00949c, xdata_out=0x7ffb2c415a50) at
/home/jenkins/root/workspace/centos6-regression/libglusterfs/src/syncop.c:1223
_new = 0x7ffb4c00623c
old_THIS = 0x7ffb50019490
tmp_cbk = 0x7ffb63ff69b3 <syncop_lookup_cbk>
task = 0x7ffb2c009790
frame = 0x7ffb2c001b3c
args = {op_ret = 0, op_errno = 0, iatt1 = {ia_ino = 0, ia_gfid
= '\000' <repeats 15 times>, ia_dev = 0, ia_type = IA_INVAL, ia_prot =
{suid = 0 '\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0
'\000', write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000',
write = 0 '\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0
'\000', exec = 0 '\000'}}, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_rdev
= 0, ia_size = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0,
ia_atime_nsec = 0, ia_mtime = 0, ia_mtime_nsec = 0, ia_ctime = 0,
ia_ctime_nsec = 0}, iatt2 = {ia_ino = 0, ia_gfid = '\000' <repeats 15
times>, ia_dev = 0, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid
= 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0
'\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000',
exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0
'\000'}}, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_rdev = 0, ia_size =
0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_atime_nsec = 0,
ia_mtime = 0, ia_mtime_nsec = 0, ia_ctime = 0, ia_ctime_nsec = 0}, xattr
= 0x0, statvfs_buf = {f_bsize = 0, f_frsize = 0, f_blocks = 0, f_bfree =
0, f_bavail = 0, f_files = 0, f_ffree = 0, f_favail = 0, f_fsid = 0,
f_flag = 0, f_namemax = 0, __f_spare = {0, 0, 0, 0, 0, 0}}, vector =
0x0, count = 0, iobref = 0x0, buffer = 0x0, xdata = 0x0, flock = {l_type
= 0, l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len =
0, data = '\000' <repeats 1023 times>}}, lease = {cmd = 0, lease_type =
NONE, lease_id = '\000' <repeats 15 times>, lease_flags = 0}, uuid =
'\000' <repeats 15 times>, errstr = 0x0, dict = 0x0, lock_dict = {__data
= {__lock = 0, __count = 0, __owner = 0, __nusers = 0, __kind = 0,
__spins = 0, __list = {__prev = 0x0, __next = 0x0}}, __size = '\000'
<repeats 39 times>, __align = 0}, barrier = {guard = {__data = {__lock =
0, __count = 0, __owner = 0, __nusers = 0, __kind = 0, __spins = 0,
__list = {__prev = 0x0, __next = 0x0}}, __size = '\000' <repeats 39
times>, __align = 0}, cond = {__data = {__lock = 0, __futex = 0,
__total_seq = 0, __wakeup_seq = 0, __woken_seq = 0, __mutex = 0x0,
__nwaiters = 0, __broadcast_seq = 0}, __size = '\000' <repeats 47
times>, __align = 0}, waitq = {next = 0x0, prev = 0x0}, count = 0}, task
= 0x7ffb2c009790, mutex = {__data = {__lock = 0, __count = 0, __owner =
0, __nusers = 0, __kind = 0, __spins = 0, __list = {__prev = 0x0, __next
= 0x0}}, __size = '\000' <repeats 39 times>, __align = 0}, cond =
{__data = {__lock = 0, __futex = 0, __total_seq = 0, __wakeup_seq = 0,
__woken_seq = 0, __mutex = 0x0, __nwaiters = 0, __broadcast_seq = 0},
__size = '\000' <repeats 47 times>, __align = 0}, done = 0, entries =
{{list = {next = 0x0, prev = 0x0}, {next = 0x0, prev = 0x0}}, d_ino = 0,
d_off = 0, d_len = 0, d_type = 0, d_stat = {ia_ino = 0, ia_gfid = '\000'
<repeats 15 times>, ia_dev = 0, ia_type = IA_INVAL, ia_prot = {suid = 0
'\000', sgid = 0 '\000', sticky = 0 '\000', owner = {read = 0 '\000',
write = 0 '\000', exec = 0 '\000'}, group = {read = 0 '\000', write = 0
'\000', exec = 0 '\000'}, other = {read = 0 '\000', write = 0 '\000',
exec = 0 '\000'}}, ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_rdev = 0,
ia_size = 0, ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_atime_nsec
= 0, ia_mtime = 0, ia_mtime_nsec = 0, ia_ctime = 0, ia_ctime_nsec = 0},
dict = 0x0, inode = 0x0, d_name = 0x7ffb2c414100 ""}, offset = 0,
locklist = {list = {next = 0x0, prev = 0x0}, flock = {l_type = 0,
l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len = 0,
data = '\000' <repeats 1023 times>}}, client_uid = 0x0, lk_flags = 0}}
__FUNCTION__ = "syncop_lookup"
#4 0x00007ffb568b96c7 in dht_migrate_file (this=0x7ffb50019490,
loc=0x7ffb2c002de4, from=0x7ffb50015930, to=0x7ffb500184a0, flag=1) at
/home/jenkins/root/workspace/centos6-regression/xlators/cluster/dht/src/dht-rebalance.c:1375
ret = 0
new_stbuf = {ia_ino = 0, ia_gfid = '\000' <repeats 15 times>,
ia_dev = 0, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0
'\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000',
exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0
'\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}},
ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_rdev = 0, ia_size = 0,
ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_atime_nsec = 0, ia_mtime
= 0, ia_mtime_nsec = 0, ia_ctime = 0, ia_ctime_nsec = 0}
stbuf = {ia_ino = 0, ia_gfid = '\000' <repeats 15 times>,
ia_dev = 0, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0
'\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000',
exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0
'\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}},
ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_rdev = 0, ia_size = 0,
ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_atime_nsec = 0, ia_mtime
= 0, ia_mtime_nsec = 0, ia_ctime = 0, ia_ctime_nsec = 0}
empty_iatt = {ia_ino = 0, ia_gfid = '\000' <repeats 15 times>,
ia_dev = 0, ia_type = IA_INVAL, ia_prot = {suid = 0 '\000', sgid = 0
'\000', sticky = 0 '\000', owner = {read = 0 '\000', write = 0 '\000',
exec = 0 '\000'}, group = {read = 0 '\000', write = 0 '\000', exec = 0
'\000'}, other = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}},
ia_nlink = 0, ia_uid = 0, ia_gid = 0, ia_rdev = 0, ia_size = 0,
ia_blksize = 0, ia_blocks = 0, ia_atime = 0, ia_atime_nsec = 0, ia_mtime
= 0, ia_mtime_nsec = 0, ia_ctime = 0, ia_ctime_nsec = 0}
src_ia_prot = {suid = 0 '\000', sgid = 0 '\000', sticky = 0
'\000', owner = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'},
group = {read = 0 '\000', write = 0 '\000', exec = 0 '\000'}, other =
{read = 0 '\000', write = 0 '\000', exec = 0 '\000'}}
src_fd = 0x0
dst_fd = 0x0
dict = 0x7ffb4c00949c
xattr = 0x0
xattr_rsp = 0x0
file_has_holes = 0
conf = 0x7ffb5002acd0
rcvd_enoent_from_src = 0
flock = {l_type = 1, l_whence = 0, l_start = 0, l_len = 0,
l_pid = 0, l_owner = {len = 0, data = '\000' <repeats 1023 times>}}
plock = {l_type = 0, l_whence = 0, l_start = 0, l_len = 0,
l_pid = 0, l_owner = {len = 0, data = '\000' <repeats 1023 times>}}
tmp_loc = {path = 0x7ffb4c0083f0 "", name = 0x0, inode =
0x7ffb2c00cf6c, parent = 0x0, gfid =
"\365\267[t\277\205N\370\232\262\206\341o\253:E", pargfid = '\000'
<repeats 15 times>}
locked = _gf_true
p_locked = _gf_false
lk_ret = -1
defrag = 0x7ffb5002b1f0
clean_src = _gf_false
clean_dst = _gf_false
log_level = 9
delete_src_linkto = _gf_true
locklist = {list = {next = 0x0, prev = 0x0}, flock = {l_type =
0, l_whence = 0, l_start = 0, l_len = 0, l_pid = 0, l_owner = {len = 0,
data = '\000' <repeats 1023 times>}}, client_uid = 0x0, lk_flags = 0}
meta_dict = 0x0
meta_locked = _gf_false
__FUNCTION__ = "dht_migrate_file"
#5 0x00007ffb568bb198 in rebalance_task (data=0x7ffb2c00171c) at
/home/jenkins/root/workspace/centos6-regression/xlators/cluster/dht/src/dht-rebalance.c:1915
ret = -1
local = 0x7ffb2c002ddc
frame = 0x7ffb2c00171c
#6 0x00007ffb63ff4fa3 in synctask_wrap (old_task=0x7ffb2c009790) at
/home/jenkins/root/workspace/centos6-regression/libglusterfs/src/syncop.c:375
task = 0x7ffb2c009790
#7 0x00007ffb62b478b0 in ?? () from /lib64/libc.so.6
No symbol table info available.
#8 0x0000000000000000 in ?? ()
No symbol table info available.
More information about the Gluster-devel
mailing list