[Gluster-devel] gluster write million of lines: WRITE => -1 (Transport endpoint is not connected)

Ben England bengland at redhat.com
Mon Oct 27 16:33:36 UTC 2014


Sergio, 

I agree, excessive logging is a performance issue and can potentially fill a system disk partition or LVM volume over a long enough period of time, resulting in other errors.  See bz 1156624 for another example that I encountered.  Does it happen in glusterfs-3.6?  Is there a logging option/interface that would rate-limit a particular logging call per unit of time to N messages/sec, and when that limit is exceeded a message is logged saying that M more events of that type were seen in the last second?

-Ben England

----- Original Message -----
> From: "Sergio Traldi" <sergio.traldi at pd.infn.it>
> To: gluster-users at gluster.org, gluster-devel at gluster.org
> Sent: Monday, October 27, 2014 9:51:37 AM
> Subject: [Gluster-devel] gluster write million of lines: WRITE => -1 (Transport endpoint is not connected)
> 
> Hi all,
> One server Redhat 6 with this rpms set:
> 
> [ ~]# rpm -qa | grep gluster | sort
> glusterfs-3.5.2-1.el6.x86_64
> glusterfs-api-3.5.2-1.el6.x86_64
> glusterfs-cli-3.5.2-1.el6.x86_64
> glusterfs-fuse-3.5.2-1.el6.x86_64
> glusterfs-geo-replication-3.5.2-1.el6.x86_64
> glusterfs-libs-3.5.2-1.el6.x86_64
> glusterfs-server-3.5.2-1.el6.x86_64
> 
> I have a gluster volume with 1 server and 1 brick:
> 
> [ ~]# gluster volume info volume-nova-pp
> Volume Name: volume-nova-pp
> Type: Distribute
> Volume ID: b5ec289b-9a54-4df1-9c21-52ca556aeead
> Status: Started
> Number of Bricks: 1
> Transport-type: tcp
> Bricks:
> Brick1: 192.168.61.100:/brick-nova-pp/mpathc
> Options Reconfigured:
> storage.owner-gid: 162
> storage.owner-uid: 162
> 
> There are four clients attached to this volume with same O.S. and same
> fuse gluster rpms set:
> [ ~]# rpm -qa | grep gluster | sort
> glusterfs-3.5.0-2.el6.x86_64
> glusterfs-api-3.5.0-2.el6.x86_64
> glusterfs-fuse-3.5.0-2.el6.x86_64
> glusterfs-libs-3.5.0-2.el6.x86_6
> 
> Last week, but it happens also two weeks ago, I found the disk almost
> full and I found the gluster logs
> /var/log/glusterfs/var-lib-nova-instances.log of 68GB:
> In the log there was the starting problem:
> 
> [2014-10-10 07:29:43.730792] W [socket.c:522:__socket_rwv] 0-glusterfs:
> readv on 192.168.61.100:24007 failed (No data available)
> [2014-10-10 07:29:54.022608] E [socket.c:2161:socket_connect_finish]
> 0-glusterfs: connection to 192.168.61.100:24007 failed (Connection refused)
> [2014-10-10 07:30:05.271825] W
> [client-rpc-fops.c:866:client3_3_writev_cbk] 0-volume-nova-pp-client-0:
> remote operation failed: Input/output error
> [2014-10-10 07:30:08.783145] W [fuse-bridge.c:2201:fuse_writev_cbk]
> 0-glusterfs-fuse: 3661260: WRITE => -1 (Input/output error)
> [2014-10-10 07:30:08.783368] W [fuse-bridge.c:2201:fuse_writev_cbk]
> 0-glusterfs-fuse: 3661262: WRITE => -1 (Input/output error)
> [2014-10-10 07:30:08.806553] W [fuse-bridge.c:2201:fuse_writev_cbk]
> 0-glusterfs-fuse: 3661649: WRITE => -1 (Input/output error)
> [2014-10-10 07:30:08.844415] W [fuse-bridge.c:2201:fuse_writev_cbk]
> 0-glusterfs-fuse: 3662235: WRITE => -1 (Input/output error)
> 
> and a lot of these lines:
> 
> [2014-10-15 14:41:15.895105] W [fuse-bridge.c:2201:fuse_writev_cbk]
> 0-glusterfs-fuse: 951700230: WRITE => -1 (Transport endpoint is not
> connected)
> [2014-10-15 14:41:15.896205] W [fuse-bridge.c:2201:fuse_writev_cbk]
> 0-glusterfs-fuse: 951700232: WRITE => -1 (Transport endpoint is not
> connected)
> 
> This second line log with different "sector" number has been written
> every millisecond so in about 1 minute we have 1GB write in O.S. disk.
> 
> I search for a solution but I didn't find nobody having the same problem.
> 
> I think there was a network problem  but why does gluster write in logs
> million of:
> [2014-10-15 14:41:15.895105] W [fuse-bridge.c:2201:fuse_writev_cbk]
> 0-glusterfs-fuse: 951700230: WRITE => -1 (Transport endpoint is not
> connected) ?
> 
> Thanks in advance.
> Cheers
> Sergio
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-devel
> 


More information about the Gluster-devel mailing list