[Gluster-devel] Whole failure when one glusterfsd brought down.

Dale Dude dale at oc3networks.com
Wed Mar 21 20:14:39 UTC 2007


Using code pulled March 21st @ 1am EST. Kernel 2.6.20-10 on Ubuntu 
Feisty 32bit.

I have 3 machines serving with glusterfsd and mounted the cluster from 
the first server. If I kill one of the glusterfsd's on any of the 
machines, the mount point becomes broken with the 'Transport' error 
below. Also, glusterfs will produce this error even if I unmount and 
remount with the 1 glusterfsd server still down. Is this expected 
results or shouldn't the mount continue to work even though one of the 
servers has "died"?

ls: reading directory local/: Transport endpoint is not connected

glusterfs.log produces this:
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] 
protocol/client: client_protocol_xfer: :transport_submit failed
[Mar 21 06:25:17] [ERROR/tcp-client.c:284/tcp_connect()] 
tcp/client:non-blocking connect() returned: 111 (Connection refused)
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] 
protocol/client: client_protocol_xfer: :transport_submit failed
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] 
protocol/client: client_protocol_xfer: :transport_submit failed
[Mar 21 06:25:17] [ERROR/tcp-client.c:284/tcp_connect()] 
tcp/client:non-blocking connect() returned: 111 (Connection refused)
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] 
protocol/client: client_protocol_xfer: :transport_submit failed
[Mar 21 06:25:17] [ERROR/client-protocol.c:183/client_protocol_xfer()] 
protocol/client: client_protocol_xfer: :transport_submit failed


======================================
glusterfs-server.vol used by all the servers:
(ignore my bad volume naming, was just testing)

volume brick
  type storage/posix                   # POSIX FS translator
  option directory /home/export        # Export this directory
end-volume

volume iothreads
   type performance/io-threads
   option thread-count 8
   subvolumes brick
end-volume

volume server
  type protocol/server
  option transport-type tcp/server     # For TCP/IP transport
  subvolumes iothreads
  option auth.ip.brick.allow * # Allow access to "brick" volume
end-volume


======================================
glusterfs-client.vol used on server1:
(ignore my bad volume naming, was just testing)

volume client1
  type protocol/client
  option transport-type tcp/client     # for TCP/IP transport
  option remote-host 1.1.1.1     # IP address of the remote brick
  option remote-subvolume brick        # name of the remote volume
end-volume

volume client2
  type protocol/client
  option transport-type tcp/client     # for TCP/IP transport
  option remote-host 2.2.2.2     # IP address of the remote brick
  option remote-subvolume brick        # name of the remote volume
end-volume

volume client3
  type protocol/client
  option transport-type tcp/client     # for TCP/IP transport
  option remote-host 3.3.3.3     # IP address of the remote brick
  option remote-subvolume brick        # name of the remote volume
end-volume

volume bricks
  type cluster/unify
  subvolumes client1 client2 client3
  option scheduler alu
  option alu.limits.min-free-disk  60GB              # Stop creating 
files when free-space lt 60GB
  option alu.limits.max-open-files 10000
  option alu.order 
disk-usage:read-usage:write-usage:open-files-usage:disk-speed-usage
  option alu.disk-usage.entry-threshold 2GB          # Units in KB, MB 
and GB are allowed
  option alu.disk-usage.exit-threshold  60MB         # Units in KB, MB 
and GB are allowed
  option alu.open-files-usage.entry-threshold 1024
  option alu.open-files-usage.exit-threshold 32
  option alu.stat-refresh.interval 10sec
end-volume

volume writebehind   #writebehind improves write performance a lot
  type performance/write-behind
  option aggregate-size 131072 # in bytes
  subvolumes bricks
end-volume





More information about the Gluster-devel mailing list