[Gluster-users] Hanging writes after upgrading "clients" to debian squeeze

Stefan Becker sbecker at rapidsoft.de
Sun Feb 5 20:49:47 UTC 2012


- no ip tables involved
- the server is running 3.2.0 as well, of course I could upgrade but this probably means some downtime which I can not afford right now
- did not find something in the logs, but there are a lot of files so I might miss something, logs on the client or server side?
- some debug flags or verbose logging possible?

the bricks log on the server side just says "client connected" so there is not a lot of value in that. On the client side I have the following:

[2012-02-05 19:38:53.324172] I [fuse-bridge.c:3214:fuse_thread_proc] 0-fuse: unmounting /home/XXXstorage
[2012-02-05 19:38:53.324221] I [glusterfsd.c:712:cleanup_and_exit] 0-glusterfsd: shutting down
[2012-02-05 19:38:58.709783] W [write-behind.c:3023:init] 0-XXXstorage-write-behind: disabling write-behind for first 0 bytes
[2012-02-05 19:38:58.711289] I [client.c:1935:notify] 0-XXXstorage-client-0: parent translators are ready, attempting connect on transport
[2012-02-05 19:38:58.711489] I [client.c:1935:notify] 0-XXXstorage-client-1: parent translators are ready, attempting connect on transport
Given volfile:
+------------------------------------------------------------------------------+
  1: volume XXXstorage-client-0
  2:     type protocol/client
  3:     option remote-host 10.10.100.40
  4:     option remote-subvolume /brick1
  5:     option transport-type tcp
  6: end-volume
  7:
  8: volume XXXstorage-client-1
  9:     type protocol/client
 10:     option remote-host 10.10.100.41
 11:     option remote-subvolume /brick1
 12:     option transport-type tcp
 13: end-volume
 14:
 15: volume XXXstorage-replicate-0
 16:     type cluster/replicate
 17:     subvolumes XXXstorage-client-0 XXXstorage-client-1
 18: end-volume
 19:
 20: volume XXXstorage-write-behind
 21:     type performance/write-behind
 22:     subvolumes XXXstorage-replicate-0
 23: end-volume
 24:
 25: volume XXXstorage-read-ahead
 26:     type performance/read-ahead
 27:     subvolumes XXXstorage-write-behind
 28: end-volume
 29:
 30: volume XXXstorage-io-cache
 31:     type performance/io-cache
 32:     subvolumes XXXstorage-read-ahead
 33: end-volume
 34:
 35: volume XXXstorage-stat-prefetch
 36:     type performance/stat-prefetch
 37:     subvolumes XXXstorage-io-cache
 38: end-volume
 39:
 40: volume XXXstorage
 41:     type debug/io-stats
 42:     option latency-measurement off
 43:     option count-fop-hits off
 44:     subvolumes XXXstorage-stat-prefetch
 45: end-volume

+------------------------------------------------------------------------------+
[2012-02-05 19:38:58.712460] I [rpc-clnt.c:1531:rpc_clnt_reconfig] 0-XXXstorage-client-1: changing port to 24015 (from 0)
[2012-02-05 19:38:58.712527] I [rpc-clnt.c:1531:rpc_clnt_reconfig] 0-XXXstorage-client-0: changing port to 24012 (from 0)
[2012-02-05 19:39:02.709882] I [client-handshake.c:1080:select_server_supported_programs] 0-XXXstorage-client-1: Using Program GlusterFS-3.1.0, Num (1298437), Version (310)
[2012-02-05 19:39:02.710112] I [client-handshake.c:1080:select_server_supported_programs] 0-XXXstorage-client-0: Using Program GlusterFS-3.1.0, Num (1298437), Version (310)
[2012-02-05 19:39:02.710355] I [client-handshake.c:913:client_setvolume_cbk] 0-XXXstorage-client-1: Connected to 10.10.100.41:24015, attached to remote volume '/brick1'.
[2012-02-05 19:39:02.710395] I [afr-common.c:2514:afr_notify] 0-XXXstorage-replicate-0: Subvolume 'XXXstorage-client-1' came back up; going online.
[2012-02-05 19:39:02.712314] I [fuse-bridge.c:3316:fuse_graph_setup] 0-fuse: switched to graph 0
[2012-02-05 19:39:02.712387] I [client-handshake.c:913:client_setvolume_cbk] 0-XXXstorage-client-0: Connected to 10.10.100.40:24012, attached to remote volume '/brick1'.
[2012-02-05 19:39:02.712436] I [fuse-bridge.c:2897:fuse_init] 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.13 kernel 7.13
[2012-02-05 19:39:02.713253] I [afr-common.c:836:afr_fresh_lookup_cbk] 0-XXXstorage-replicate-0: added root inode

I cannot see any problems. I was tailing a few logs while I issued a write which hangs. Nothing gets logged.

-----Ursprüngliche Nachricht-----
Von: Whit Blauvelt [mailto:whit.gluster at transpect.com] 
Gesendet: Sonntag, 5. Februar 2012 21:25
An: Brian Candler
Cc: Stefan Becker; gluster-users at gluster.org
Betreff: Re: [Gluster-users] Hanging writes after upgrading "clients" to debian squeeze

On Sun, Feb 05, 2012 at 07:36:55PM +0000, Brian Candler wrote:
> On Sun, Feb 05, 2012 at 08:02:08PM +0100, Stefan Becker wrote:
> >    After the debian upgrade I can
> >    still mount my volumes. Reading is fine as well but it hangs on writes.
> 
> Could it be that on the post-upgrade machines one brick is reachable but not
> the other?  Compare iptables rules between the pre-upgrade and post-upgrade
> machines?  Compare tcpdump or ntop between them?

If you can, try dropping iptables out of the picture entirely. If you are
running it, and have it logging what it drops, the docs say "Ensure that TCP
ports 111, 24007,24008, 24009-(24009 + number of bricks across all volumes)
are open on all Gluster servers. If you will be using NFS, open additional
ports 38465 to 38467." So I'd check your logs to see if iptables is dropping
any traffic to/from the IPs in question on those ports.

Or us "netstat -tc" while doing some file operations, and you should see the
traffic on the IPs/ports. Another utility to see the same thing is "iptraf."  

Whit	



More information about the Gluster-users mailing list