[Gluster-Maintainers] Build failed in Jenkins: centos8-s390-regression #127
jenkins at build.gluster.org
jenkins at build.gluster.org
Sat May 20 21:12:41 UTC 2023
See <https://build.gluster.org/job/centos8-s390-regression/127/display/redirect>
Changes:
------------------------------------------
[...truncated 3.94 MB...]
205: subvolumes patchy
206: end-volume
207:
+------------------------------------------------------------------------------+
glfs_get_volumeid: returned 16
lstat(/filename2): (-1) No such file or directory
Entries:
.: 64
..: 76
filename2: 3076
setxattr(/filename2): 0 (Success)
setxattr(/filename2): 0 (Success)
getxattr(/filename2): 8 (Success)
listxattr(/filename2): 44 (Success)
symlink(/filename2 /linkfile): Success
readlink(/filename2) : 8 (Success)
lsetxattr(/linkfile) : 0 (Success)
llistxattr(/filename2): 17 (Success)
lgetxattr(/linkfile): 8 (Success)
removexattr(/filename2): 0 (Success)
open(/filename2): (0x3a71ba48) Success
fsetxattr(/filename2): 0 (Success)
fgetxattr(/filename2): 8 (Success)
flistxattr(/filename2): 117 (Success)
fremovexattr(/filename2): 0 (Success)
mkdir(/topdir): Success
mkdir(/dir): Success
[2023-05-20 21:01:52.440848 +0000] W [MSGID: 108001] [afr-common.c:6410:afr_notify] 0-patchy-replicate-0: Client-quorum is not met
[2023-05-20 21:01:52.440977 +0000] W [MSGID: 108001] [afr-common.c:6410:afr_notify] 0-patchy-replicate-1: Client-quorum is not met
[2023-05-20 21:01:52.441005 +0000] E [MSGID: 108006] [afr-common.c:6105:__afr_handle_child_down_event] 0-patchy-replicate-1: All subvolumes are down. Going offline until at least one of them comes back up.
[2023-05-20 21:01:52.441173 +0000] E [MSGID: 108006] [afr-common.c:6105:__afr_handle_child_down_event] 0-patchy-replicate-0: All subvolumes are down. Going offline until at least one of them comes back up.
[2023-05-20 21:01:52.441342 +0000] W [inode.c:1882:inode_table_destroy] (-->/build/install/lib/libgfapi.so.0(glfs_fini+0x54c) [0x3ff8e68e284] -->/build/install/lib/libglusterfs.so.0(inode_table_destroy_all+0xae) [0x3ff8e0d2736] -->/build/install/lib/libglusterfs.so.0(inode_table_destroy+0x23c) [0x3ff8e0d29bc] ) 0-gfapi: Active inode(0x3a71f3c8) with refcount(1) found during cleanup
[2023-05-20 21:01:58.364034 +0000] I [io-stats.c:4200:fini] 0-patchy: io-stats translator unloaded
ok 7 [ 18565/ 2] < 24> 'cleanup_tester ./glfsxmp'
ok 8 [ 12/ 2] < 25> 'rm ./glfsxmp.c'
ok 9 [ 12/ 6089] < 28> 'gluster --mode=script --wignore volume stop patchy'
ok 10 [ 13/ 585] < 30> 'gluster --mode=script --wignore volume delete patchy'
ok
All tests successful.
Files=1, Tests=10, 30 wallclock secs ( 0.02 usr 0.00 sys + 0.60 cusr 0.58 csys = 1.20 CPU)
Result: PASS
Logs preserved in tarball arbiter-coverage-iteration-1.tar.gz
End of test ./tests/line-coverage/arbiter-coverage.t
================================================================================
======================================== (825 / 838) ========================================
[21:02:08] Running tests in file ./tests/line-coverage/cli-negative-case-and-function-coverage.t
./tests/line-coverage/cli-negative-case-and-function-coverage.t ..
1..75
ok 1 [ 145/ 1142] < 9> 'glusterd'
ok 2 [ 12/ 7] < 10> 'pidof glusterd'
ok 3 [ 12/ 53] < 14> '! gluster --mode=script --wignore volume create patchy_1 148.100.84.19-/d/backends/v1 148.100.84.19-/d/backends/v2'
ok 4 [ 11/ 52] < 17> '! gluster --mode=script --wignore volume create patchy_1 :/d/backends/v1 :/d/backends/v2'
ok 5 [ 11/ 52] < 20> '! gluster --mode=script --wignore volume create patchy_1 localhost:/d/backends/v1 localhost:/d/backends/v2'
ok 6 [ 12/ 52] < 24> '! gluster --mode=script --wignore volume inode-quota disable'
ok 7 [ 12/ 52] < 27> '! gluster --mode=script --wignore volume inode-quota patchy_1 disable'
ok 8 [ 11/ 53] < 31> '! gluster --mode=script --wignore volume patchy_1 start'
ok 9 [ 12/ 52] < 32> '! gluster --mode=script --wignore volume patchy_1 limit-usage /random-path 0'
ok 10 [ 11/ 53] < 33> '! gluster --mode=script --wignore volume patchy_1 limit-objects /random-path 0'
ok 11 [ 11/ 52] < 34> '! gluster --mode=script --wignore volume patchy_1 alert-time some-time'
ok 12 [ 11/ 52] < 35> '! gluster --mode=script --wignore volume patchy_1 soft-timeout some-time'
ok 13 [ 11/ 53] < 36> '! gluster --mode=script --wignore volume patchy_1 hard-timeout some-time'
ok 14 [ 11/ 53] < 39> '! gluster --mode=script --wignore volume patchy_1 limit-usage random-path'
ok 15 [ 11/ 53] < 40> '! gluster --mode=script --wignore volume patchy_1 remove random-path'
ok 16 [ 12/ 52] < 41> '! gluster --mode=script --wignore volume patchy_1 remove-objects random-path'
ok 17 [ 12/ 52] < 42> '! gluster --mode=script --wignore volume patchy_1 list random-path'
ok 18 [ 11/ 53] < 45> '! gluster --mode=script --wignore volume patchy_1 remove /random-path'
ok 19 [ 11/ 51] < 46> '! gluster --mode=script --wignore volume patchy_1 remove-objects /random-path'
ok 20 [ 11/ 52] < 47> '! gluster --mode=script --wignore volume patchy_1 alert-time'
ok 21 [ 11/ 52] < 48> '! gluster --mode=script --wignore volume patchy_1 soft-timeout'
ok 22 [ 11/ 53] < 49> '! gluster --mode=script --wignore volume patchy_1 hard-timeout'
ok 23 [ 12/ 53] < 50> '! gluster --mode=script --wignore volume patchy_1 default-soft-limit'
ok 24 [ 12/ 53] < 53> '! gluster --mode=script --wignore nfs-ganesha'
ok 25 [ 11/ 52] < 54> '! gluster --mode=script --wignore nfs-gansha disable'
ok 26 [ 11/ 52] < 55> '! gluster --mode=script --wignore nfs-ganesha stop'
ok 27 [ 11/ 56] < 56> '! gluster --mode=script --wignore nfs-ganesha disable'
ok 28 [ 11/ 53] < 57> '! gluster --mode=script --wignore nfs-ganesha enable'
ok 29 [ 11/ 52] < 60> '! gluster --mode=script --wignore peer probe'
ok 30 [ 12/ 53] < 61> '! gluster --mode=script --wignore peer probe host_name'
ok 31 [ 11/ 52] < 62> '! gluster --mode=script --wignore peer detach'
ok 32 [ 11/ 53] < 63> '! gluster --mode=script --wignore peer detach host-name random-option'
ok 33 [ 11/ 53] < 64> '! gluster --mode=script --wignore peer status host'
ok 34 [ 11/ 52] < 65> '! gluster --mode=script --wignore pool list host'
ok 35 [ 12/ 53] < 68> '! gluster --mode=script --wignore vol sync'
ok 36 [ 12/ 50122] < 69> '! gluster --mode=script --wignore vol sync host-name'
ok 37 [ 14/ 56] < 70> '! gluster --mode=script --wignore vol sync localhost'
ok 38 [ 12/ 53] < 73> '! gluster --mode=script --wignore system:: getspec'
ok 39 [ 12/ 54] < 74> '! gluster --mode=script --wignore system:: portmap brick2port'
ok 40 [ 12/ 53] < 75> '! gluster --mode=script --wignore system:: fsm log random-peer random-value'
ok 41 [ 12/ 54] < 76> '! gluster --mode=script --wignore system:: getwd random-value'
ok 42 [ 12/ 53] < 77> '! gluster --mode=script --wignore system:: mount'
ok 43 [ 12/ 53] < 78> '! gluster --mode=script --wignore system:: umount'
ok 44 [ 12/ 53] < 79> '! gluster --mode=script --wignore system:: uuid get random-value'
ok 45 [ 12/ 54] < 80> '! gluster --mode=script --wignore system:: uuid reset random-value'
ok 46 [ 12/ 56] < 81> '! gluster --mode=script --wignore system:: execute'
ok 47 [ 13/ 53] < 82> '! gluster --mode=script --wignore system:: copy file'
ok 48 [ 12/ 74] < 85> 'gluster --mode=script --wignore volume create patchy_1 replica 3 148.100.84.19:/d/backends/v1 148.100.84.19:/d/backends/v2 148.100.84.19:/d/backends/v3'
ok 49 [ 12/ 1679] < 86> 'gluster --mode=script --wignore volume start patchy_1'
ok 50 [ 13/ 59] < 87> 'Y glustershd_up_status'
ok 51 [ 12/ 53] < 88> 'gluster --mode=script --wignore volume heal patchy_1 statistics'
ok 52 [ 11/ 2399] < 91> 'gluster --mode=script --wignore volume replace-brick patchy_1 148.100.84.19:/d/backends/v1 148.100.84.19:/d/backends/v4 commit force --xml'
ok 53 [ 14/ 74] < 92> 'gluster --mode=script --wignore volume create patchy_2 148.100.84.19:/d/backends/v5 148.100.84.19:/d/backends/v6 --xml'
ok 54 [ 12/ 518] < 93> 'gluster --mode=script --wignore volume delete patchy_2 --xml'
ok 55 [ 14/ 54] < 96> '! gluster --mode=script --wignore volume start'
ok 56 [ 12/ 56] < 97> '! gluster --mode=script --wignore volume start patchy_1 frc'
ok 57 [ 12/ 55] < 98> '! gluster --mode=script --wignore volume info patchy_1 info'
ok 58 [ 12/ 54] < 99> '! gluster --mode=script --wignore volume info patchy_2'
ok 59 [ 12/ 54] < 100> '! gluster --mode=script --wignore volume delete'
ok 60 [ 12/ 53] < 101> '! gluster --mode=script --wignore volume stop'
ok 61 [ 12/ 55] < 102> '! gluster --mode=script --wignore volume stop patchy_1 frc'
ok 62 [ 12/ 53] < 103> '! gluster --mode=script --wignore volume rebalance patchy_1'
ok 63 [ 12/ 53] < 104> '! gluster --mode=script --wignore volume reset'
ok 64 [ 12/ 53] < 105> '! gluster --mode=script --wignore volume profile patchy_1'
ok 65 [ 12/ 53] < 106> '! gluster --mode=script --wignore volume quota all'
ok 66 [ 12/ 53] < 107> '! gluster --mode=script --wignore volume reset-brick patchy_1'
ok 67 [ 12/ 54] < 108> '! gluster --mode=script --wignore volume top patchy_1'
ok 68 [ 12/ 53] < 109> '! gluster --mode=script --wignore volume log rotate'
ok 69 [ 11/ 54] < 110> '! gluster --mode=script --wignore volume status all all'
ok 70 [ 12/ 54] < 111> '! gluster --mode=script --wignore volume heal'
ok 71 [ 12/ 53] < 112> '! gluster --mode=script --wignore volume statedump'
ok 72 [ 12/ 54] < 113> '! gluster --mode=script --wignore volume clear-locks patchy_1 / kid granted entry dir1'
ok 73 [ 12/ 53] < 114> '! gluster --mode=script --wignore volume clear-locks patchy_1 / kind grant entry dir1'
ok 74 [ 12/ 54] < 115> '! gluster --mode=script --wignore volume clear-locks patchy_1 / kind granted ent dir1'
ok 75 [ 12/ 53] < 116> '! gluster --mode=script --wignore volume barrier patchy_1'
ok
All tests successful.
Files=1, Tests=75, 61 wallclock secs ( 0.02 usr 0.00 sys + 3.59 cusr 1.50 csys = 5.11 CPU)
Result: PASS
Logs preserved in tarball cli-negative-case-and-function-coverage-iteration-1.tar.gz
End of test ./tests/line-coverage/cli-negative-case-and-function-coverage.t
================================================================================
======================================== (826 / 838) ========================================
[21:03:09] Running tests in file ./tests/line-coverage/cli-peer-and-volume-operations.t
./tests/line-coverage/cli-peer-and-volume-operations.t ..
1..58
ok 1 [ 152/ 3541] < 13> 'launch_cluster 3'
ok 2 [ 13/ 57] < 15> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log system uuid reset'
ok 3 [ 12/ 76] < 18> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log peer probe 127.1.1.2'
ok 4 [ 12/ 60] < 19> '1 peer_count 1'
ok 5 [ 12/ 58] < 20> '1 peer_count 2'
ok 6 [ 12/ 3] < 23> 'kill_glusterd 3'
ok 7 [ 12/ 58] < 24> '! gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log peer probe 127.1.1.3'
ok 8 [ 12/ 54] < 27> '! gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log peer detach 127.1.1.3'
ok 9 [ 12/ 53] < 28> '! gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log peer detach 127.1.1.3 force'
ok 10 [ 12/ 1151] < 30> 'start_glusterd 3'
ok 11 [ 12/ 237] < 31> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log peer probe 127.1.1.3'
ok 12 [ 13/ 59] < 32> '2 peer_count 1'
ok 13 [ 12/ 57] < 33> '2 peer_count 2'
ok 14 [ 11/ 57] < 34> '2 peer_count 3'
ok 15 [ 12/ 54] < 37> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log peer probe 127.1.1.3'
ok 16 [ 12/ 67] < 40> '! gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log peer probe 1024.1024.1024.1024'
ok 17 [ 12/ 54] < 42> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log pool list'
ok 18 [ 12/ 6] < 44> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log --help'
ok 19 [ 12/ 6] < 45> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log --version'
ok 20 [ 11/ 6] < 46> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log --print-logdir'
ok 21 [ 12/ 6] < 47> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log --print-statedumpdir'
ok 22 [ 11/ 54] < 50> '! gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log volume'
ok 23 [ 12/ 7] < 51> 'pidof glusterd'
ok 24 [ 12/ 53] < 54> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log global help'
ok 25 [ 12/ 53] < 55> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log help'
ok 26 [ 12/ 53] < 57> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log peer help'
ok 27 [ 12/ 54] < 58> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log volume help'
ok 28 [ 12/ 53] < 59> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log volume bitrot help'
ok 29 [ 12/ 53] < 60> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log volume quota help'
ok 30 [ 12/ 53] < 61> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log snapshot help'
ok 31 [ 12/ 107] < 64> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log volume create patchy 127.1.1.1:/d/backends/1/patchy 127.1.1.2:/d/backends/2/patchy 127.1.1.3:/d/backends/3/patchy'
ok 32 [ 12/ 54] < 66> '! gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log volume create patchy 127.1.1.1:/d/backends/1/patchy1 127.1.1.2:/d/backends/2/patchy1'
ok 33 [ 12/ 412] < 67> 'gluster --mode=script --wignore --glusterd-sock=/d/backends/1/glusterd/gd.sock --log-file=/var/log/glusterfs/cli-peer-and-volume-operations.t_cli1.log volume start patchy'
ok 34 [ 12/ 56] < 68> 'Started cluster_volinfo_field 1 patchy Status'
ok 35 [ 11/ 19] < 71> 'glusterfs -s 127.1.1.1 --volfile-id patchy /mnt/glusterfs/1'
ok 36 [ 13/ 127] < 72> 'touch /mnt/glusterfs/1/file1 /mnt/glusterfs/1/file2 /mnt/glusterfs/1/file3 /mnt/glusterfs/1/file4 /mnt/glusterfs/1/file5 /mnt/glusterfs/1/file6 /mnt/glusterfs/1/file7 /mnt/glusterfs/1/file8 /mnt/glusterfs/1/file9 /mnt/glusterfs/1/file10 /mnt/glusterfs/1/file11 /mnt/glusterfs/1/file12 /mnt/glusterfs/1/file13 /mnt/glusterfs/1/file14 /mnt/glusterfs/1/file15 /mnt/glusterfs/1/file16 /mnt/glusterfs/1/file17 /mnt/glusterfs/1/file18 /mnt/glusterfs/1/file19 /mnt/glusterfs/1/file20 /mnt/glusterfs/1/file21 /mnt/glusterfs/1/file22 /mnt/glusterfs/1/file23 /mnt/glusterfs/1/file24 /mnt/glusterfs/1/file25 /mnt/glusterfs/1/file26 /mnt/glusterfs/1/file27 /mnt/glusterfs/1/file28 /mnt/glusterfs/1/file29 /mnt/glusterfs/1/file30 /mnt/glusterfs/1/file31 /mnt/glusterfs/1/file32 /mnt/glusterfs/1/file33 /mnt/glusterfs/1/file34 /mnt/glusterfs/1/file35 /mnt/glusterfs/1/file36 /mnt/glusterfs/1/file37 /mnt/glusterfs/1/file38 /mnt/glusterfs/1/file39 /mnt/glusterfs/1/file40 /mnt/glusterfs/1/file41 /mnt/glusterfs/1/file42 /mnt/glusterfs/1/file43 /mnt/glusterfs/1/file44 /mnt/glusterfs/1/file45 /mnt/glusterfs/1/file46 /mnt/glusterfs/1/file47 /mnt/glusterfs/1/file48 /mnt/glusterfs/1/file49 /mnt/glusterfs/1/file50 /mnt/glusterfs/1/file51 /mnt/glusterfs/1/file52 /mnt/glusterfs/1/file53 /mnt/glusterfs/1/file54 /mnt/glusterfs/1/file55 /mnt/glusterfs/1/file56 /mnt/glusterfs/1/file57 /mnt/glusterfs/1/file58 /mnt/glusterfs/1/file59 /mnt/glusterfs/1/file60 /mnt/glusterfs/1/file61 /mnt/glusterfs/1/file62 /mnt/glusterfs/1/file63 /mnt/glusterfs/1/file64 /mnt/glusterfs/1/file65 /mnt/glusterfs/1/file66 /mnt/glusterfs/1/file67 /mnt/glusterfs/1/file68 /mnt/glusterfs/1/file69 /mnt/glusterfs/1/file70 /mnt/glusterfs/1/file71 /mnt/glusterfs/1/file72 /mnt/glusterfs/1/file73 /mnt/glusterfs/1/file74 /mnt/glusterfs/1/file75 /mnt/glusterfs/1/file76 /mnt/glusterfs/1/file77 /mnt/glusterfs/1/file78 /mnt/glusterfs/1/file79 /mnt/glusterfs/1/file80 /mnt/glusterfs/1/file81 /mnt/glusterfs/1/file82 /mnt/glusterfs/1/file83 /mnt/glusterfs/1/file84 /mnt/glusterfs/1/file85 /mnt/glusterfs/1/file86 /mnt/glusterfs/1/file87 /mnt/glusterfs/1/file88 /mnt/glusterfs/1/file89 /mnt/glusterfs/1/file90 /mnt/glusterfs/1/file91 /mnt/glusterfs/1/file92 /mnt/glusterfs/1/file93 /mnt/glusterfs/1/file94 /mnt/glusterfs/1/file95 /mnt/glusterfs/1/file96 /mnt/glusterfs/1/file97 /mnt/glusterfs/1/file98 /mnt/glusterfs/1/file99 /mnt/glusterfs/1/file100'
ok 37 [ 12/ 55] < 75> '! gluster --mode=script --wiFATAL: command execution failed
java.io.EOFException
at java.base/java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2911)
at java.base/java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:3406)
at java.base/java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:932)
at java.base/java.io.ObjectInputStream.<init>(ObjectInputStream.java:375)
at hudson.remoting.ObjectInputStreamEx.<init>(ObjectInputStreamEx.java:49)
at hudson.remoting.Command.readFrom(Command.java:142)
at hudson.remoting.Command.readFrom(Command.java:128)
at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:35)
at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:61)
Caused: java.io.IOException: Unexpected termination of the channel
at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:75)
Caused: java.io.IOException: Backing channel 'builder-el8-s390x-2.ibm-l1.gluster.org' is disconnected.
at hudson.remoting.RemoteInvocationHandler.channelOrFail(RemoteInvocationHandler.java:215)
at hudson.remoting.RemoteInvocationHandler.invoke(RemoteInvocationHandler.java:285)
at com.sun.proxy.$Proxy150.isAlive(Unknown Source)
at hudson.Launcher$RemoteLauncher$ProcImpl.isAlive(Launcher.java:1215)
at hudson.Launcher$RemoteLauncher$ProcImpl.join(Launcher.java:1207)
at hudson.tasks.CommandInterpreter.join(CommandInterpreter.java:195)
at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:145)
at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:92)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:818)
at hudson.model.Build$BuildExecution.build(Build.java:199)
at hudson.model.Build$BuildExecution.doRun(Build.java:164)
at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:526)
at hudson.model.Run.execute(Run.java:1900)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:44)
at hudson.model.ResourceController.execute(ResourceController.java:101)
at hudson.model.Executor.run(Executor.java:442)
FATAL: Unable to delete script file /tmp/jenkins11785905521952862862.sh
java.io.EOFException
at java.base/java.io.ObjectInputStream$PeekInputStream.readFully(ObjectInputStream.java:2911)
at java.base/java.io.ObjectInputStream$BlockDataInputStream.readShort(ObjectInputStream.java:3406)
at java.base/java.io.ObjectInputStream.readStreamHeader(ObjectInputStream.java:932)
at java.base/java.io.ObjectInputStream.<init>(ObjectInputStream.java:375)
at hudson.remoting.ObjectInputStreamEx.<init>(ObjectInputStreamEx.java:49)
at hudson.remoting.Command.readFrom(Command.java:142)
at hudson.remoting.Command.readFrom(Command.java:128)
at hudson.remoting.AbstractSynchronousByteArrayCommandTransport.read(AbstractSynchronousByteArrayCommandTransport.java:35)
at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:61)
Caused: java.io.IOException: Unexpected termination of the channel
at hudson.remoting.SynchronousCommandTransport$ReaderThread.run(SynchronousCommandTransport.java:75)
Caused: hudson.remoting.ChannelClosedException: Channel "hudson.remoting.Channel at 10f3b8e8:builder-el8-s390x-2.ibm-l1.gluster.org": Remote call on builder-el8-s390x-2.ibm-l1.gluster.org failed. The channel is closing down or has closed down
at hudson.remoting.Channel.call(Channel.java:993)
at hudson.FilePath.act(FilePath.java:1192)
at hudson.FilePath.act(FilePath.java:1181)
at hudson.FilePath.delete(FilePath.java:1728)
at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:163)
at hudson.tasks.CommandInterpreter.perform(CommandInterpreter.java:92)
at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
at hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:818)
at hudson.model.Build$BuildExecution.build(Build.java:199)
at hudson.model.Build$BuildExecution.doRun(Build.java:164)
at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:526)
at hudson.model.Run.execute(Run.java:1900)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:44)
at hudson.model.ResourceController.execute(ResourceController.java:101)
at hudson.model.Executor.run(Executor.java:442)
Build step 'Execute shell' marked build as failure
ERROR: Unable to tear down: null
java.lang.NullPointerException
at hudson.slaves.WorkspaceList.tempDir(WorkspaceList.java:313)
at org.jenkinsci.plugins.credentialsbinding.impl.UnbindableDir.secretsDir(UnbindableDir.java:61)
at org.jenkinsci.plugins.credentialsbinding.impl.UnbindableDir.access$000(UnbindableDir.java:22)
at org.jenkinsci.plugins.credentialsbinding.impl.UnbindableDir$UnbinderImpl.unbind(UnbindableDir.java:83)
at org.jenkinsci.plugins.credentialsbinding.impl.SecretBuildWrapper$1.tearDown(SecretBuildWrapper.java:116)
at hudson.model.AbstractBuild$AbstractBuildExecution.tearDownBuildEnvironments(AbstractBuild.java:566)
at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:530)
at hudson.model.Run.execute(Run.java:1900)
at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:44)
at hudson.model.ResourceController.execute(ResourceController.java:101)
at hudson.model.Executor.run(Executor.java:442)
ERROR: builder-el8-s390x-2.ibm-l1.gluster.org is offline; cannot locate java-1.6.0-openjdk-1.6.0.0.x86_64
More information about the maintainers
mailing list