[Gluster-users] Odd "Transport endpoint is not connected" when trying to gunzip a file

Pat Haley phaley at mit.edu
Wed Jun 15 15:23:10 UTC 2022


Hi,

We have a cluster whose common storage is a gluster volume consisting of 
5 bricks residing on 3 servers.

  * Gluster volume machines
      o mseas-data2:  CentOS release 6.8 (Final)
      o mseas-data3:  CentOS release 6.10 (Final)
      o mseas-data4:  CentOS Linux release 7.9.2009 (Core)
  * Client machines
      o CentOS Linux release 7.9.2009 (Core)

More details on the gluster volume are included below.

We were recently trying to gunzip a file on the gluster volume and got  
a "Transport endpoint is not connected" even though every test we try 
shows that gluster is fully up and running fine.  We traced the file to 
brick 3 in the server mseas-data3.  We have included the relevant 
portions of the various log files on the client (mseas) where we were 
running the gunzip command and the server hosting the file (mseas-data3) 
below the gluster information

What can you suggest we do to further debug and/or solve this issue?

Thanks
Pat

============================================================
Gluster volume information
============================================================

---------------------------------------------------
gluster volume info
-----------------------------------------

Volume Name: data-volume
Type: Distribute
Volume ID: c162161e-2a2d-4dac-b015-f31fd89ceb18
Status: Started
Number of Bricks: 5
Transport-type: tcp
Bricks:
Brick1: mseas-data2:/mnt/brick1
Brick2: mseas-data2:/mnt/brick2
Brick3: mseas-data3:/export/sda/brick3
Brick4: mseas-data3:/export/sdc/brick4
Brick5: mseas-data4:/export/brick5
Options Reconfigured:
diagnostics.client-log-level: ERROR
network.inode-lru-limit: 50000
performance.md-cache-timeout: 60
performance.open-behind: off
disperse.eager-lock: off
auth.allow: *
server.allow-insecure: on
nfs.exports-auth-enable: on
diagnostics.brick-sys-log-level: WARNING
performance.readdir-ahead: on
nfs.disable: on
nfs.export-volumes: off
cluster.min-free-disk: 1%

---------------------------------------------------
gluster volume status
--------------------------------------------

Status of volume: data-volume
Gluster process                             TCP Port  RDMA Port Online  Pid
------------------------------------------------------------------------------
Brick mseas-data2:/mnt/brick1               49154     0 Y       15978
Brick mseas-data2:/mnt/brick2               49155     0 Y       15997
Brick mseas-data3:/export/sda/brick3        49153     0 Y       14221
Brick mseas-data3:/export/sdc/brick4        49154     0 Y       14240
Brick mseas-data4:/export/brick5            49152     0 Y       50569


---------------------------------------------------
gluster peer status
-----------------------------------------

Number of Peers: 2

Hostname: mseas-data3
Uuid: b39d4deb-c291-437e-8013-09050c1fa9e3
State: Peer in Cluster (Connected)

Hostname: mseas-data4
Uuid: 5c4d06eb-df89-4e5c-92e4-441fb401a9ef
State: Peer in Cluster (Connected)

---------------------------------------------------
glusterfs --version
--------------------------------------------

glusterfs 3.7.11 built on Apr 18 2016 13:20:46
Repository revision: git://git.gluster.com/glusterfs.git
Copyright (c) 2006-2013 Red Hat, Inc. <http://www.redhat.com/>
GlusterFS comes with ABSOLUTELY NO WARRANTY.
It is licensed to you under your choice of the GNU Lesser
General Public License, version 3 or any later version (LGPLv3
or later), or the GNU General Public License, version 2 (GPLv2),
in all cases as published by the Free Software Foundation.

============================================================
Relevant sections from log files
============================================================

---------------------------------------------------
mseas: gdata.log
-----------------------------------------

[2022-06-15 14:51:17.263858] C 
[rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 
0-data-volume-client-2: server 172.16.1.113:49153 has not responded in 
the last 42 seconds, disconnecting.
[2022-06-15 14:51:17.264522] E [rpc-clnt.c:362:saved_frames_unwind] (--> 
/usr/local/lib/libglusterfs.so.0(_gf_log_callingfn+0x172)[0x7f84886a0202] 
(--> 
/usr/local/lib/libgfrpc.so.0(saved_frames_unwind+0x1c2)[0x7f848846c3e2] 
(--> 
/usr/local/lib/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f848846c4de] 
(--> 
/usr/local/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7a)[0x7f848846dd2a] 
(--> /usr/local/lib/libgfrpc.so.0(rpc_clnt_notify+0x88)[0x7f848846e538] 
))))) 0-data-volume-client-2: forced unwinding frame type(GlusterFS 3.3) 
op(READ(12)) called at 2022-06-15 14:49:52.113795 (xid=0xb4f49b)
[2022-06-15 14:51:17.264859] E [rpc-clnt.c:362:saved_frames_unwind] (--> 
/usr/local/lib/libglusterfs.so.0(_gf_log_callingfn+0x172)[0x7f84886a0202] 
(--> 
/usr/local/lib/libgfrpc.so.0(saved_frames_unwind+0x1c2)[0x7f848846c3e2] 
(--> 
/usr/local/lib/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f848846c4de] 
(--> 
/usr/local/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7a)[0x7f848846dd2a] 
(--> /usr/local/lib/libgfrpc.so.0(rpc_clnt_notify+0x88)[0x7f848846e538] 
))))) 0-data-volume-client-2: forced unwinding frame type(GF-DUMP) 
op(NULL(2)) called at 2022-06-15 14:49:53.251903 (xid=0xb4f49c)
[2022-06-15 14:51:17.265111] E [rpc-clnt.c:362:saved_frames_unwind] (--> 
/usr/local/lib/libglusterfs.so.0(_gf_log_callingfn+0x172)[0x7f84886a0202] 
(--> 
/usr/local/lib/libgfrpc.so.0(saved_frames_unwind+0x1c2)[0x7f848846c3e2] 
(--> 
/usr/local/lib/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f848846c4de] 
(--> 
/usr/local/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7a)[0x7f848846dd2a] 
(--> /usr/local/lib/libgfrpc.so.0(rpc_clnt_notify+0x88)[0x7f848846e538] 
))))) 0-data-volume-client-2: forced unwinding frame type(GlusterFS 3.3) 
op(FSTAT(25)) called at 2022-06-15 14:50:00.103768 (xid=0xb4f49d)
[root at mseas glusterfs]#

---------------------------------------------------
mseas-data3:  cli.log
-----------------------------------------

[2022-06-15 14:27:12.982510] I [cli.c:721:main] 0-cli: Started running 
gluster with version 3.7.11
[2022-06-15 14:27:13.206046] I [MSGID: 101190] 
[event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread 
with index 1
[2022-06-15 14:27:13.206152] I [socket.c:2356:socket_event_handler] 
0-transport: disconnecting now
[2022-06-15 14:27:13.208711] I [input.c:36:cli_batch] 0-: Exiting with: 0
[2022-06-15 14:27:23.579669] I [cli.c:721:main] 0-cli: Started running 
gluster with version 3.7.11
[2022-06-15 14:27:23.711445] I [MSGID: 101190] 
[event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread 
with index 1
[2022-06-15 14:27:23.711551] I [socket.c:2356:socket_event_handler] 
0-transport: disconnecting now
[2022-06-15 14:27:23.735073] I [input.c:36:cli_batch] 0-: Exiting with: 0

---------------------------------------------------
mseas-data3:  usr-local-etc-glusterfs-glusterd.vol.log
-----------------------------------------

[2022-06-15 14:27:13.208084] I [MSGID: 106487] 
[glusterd-handler.c:1472:__glusterd_handle_cli_list_friends] 0-glusterd: 
Received cli list req
[2022-06-15 14:27:23.721724] I [MSGID: 106499] 
[glusterd-handler.c:4331:__glusterd_handle_status_volume] 0-management: 
Received status volume req for volume data-volume
[2022-06-15 14:27:23.732286] W [MSGID: 106217] 
[glusterd-op-sm.c:4630:glusterd_op_modify_op_ctx] 0-management: Failed 
uuid to hostname conversion
[2022-06-15 14:27:23.732328] W [MSGID: 106387] 
[glusterd-op-sm.c:4734:glusterd_op_modify_op_ctx] 0-management: op_ctx 
modification failed

---------------------------------------------------
mseas-data3:  bricks/export-sda-brick3.log
-----------------------------------------
[2022-06-15 14:50:42.588143] I [MSGID: 115036] 
[server.c:552:server_rpc_notify] 0-data-volume-server: disconnecting 
connection from 
mseas.mit.edu-155483-2022/05/13-03:24:14:618694-data-volume-client-2-0-28
[2022-06-15 14:50:42.588220] I [MSGID: 115013] 
[server-helpers.c:294:do_fd_cleanup] 0-data-volume-server: fd cleanup on 
/projects/posydon/Acoustics_ASA/MSEAS-ParEq-DO/Save/2D/Test_Cases/RI/DO_NAPE_JASA_Paper/Uncertain_Pekeris_Waveguide_DO_MC
[2022-06-15 14:50:42.588259] I [MSGID: 115013] 
[server-helpers.c:294:do_fd_cleanup] 0-data-volume-server: fd cleanup on 
/projects/dri_calypso/PE/2019/Apr09/Ens3R200deg001/pe_out.nc.gz
[2022-06-15 14:50:42.588288] I [MSGID: 101055] 
[client_t.c:420:gf_client_unref] 0-data-volume-server: Shutting down 
connection 
mseas.mit.edu-155483-2022/05/13-03:24:14:618694-data-volume-client-2-0-28
[2022-06-15 14:50:53.605215] I [MSGID: 115029] 
[server-handshake.c:690:server_setvolume] 0-data-volume-server: accepted 
client from 
mseas.mit.edu-155483-2022/05/13-03:24:14:618694-data-volume-client-2-0-29 
(version: 3.7.11)
[2022-06-15 14:50:42.588247] I [MSGID: 115013] 
[server-helpers.c:294:do_fd_cleanup] 0-data-volume-server: fd cleanup on 
/projects/posydon/Acoustics_ASA/MSEAS-ParEq-DO/Save/2D/Test_Cases/RI/DO_NAPE_JASA_Paper/Uncertain_Pekeris_Waveguide_DO_MC

-- 

-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
Pat Haley                          Email:phaley at mit.edu
Center for Ocean Engineering       Phone:  (617) 253-6824
Dept. of Mechanical Engineering    Fax:    (617) 253-8125
MIT, Room 5-213http://web.mit.edu/phaley/www/
77 Massachusetts Avenue
Cambridge, MA  02139-4301
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20220615/4ea75de7/attachment.html>


More information about the Gluster-users mailing list