[Gluster-devel] glusterfs-1.3.8pre1

Dan Podeanu pdan at camcontacts.net
Fri Feb 22 17:52:37 UTC 2008


Hello,

*Q1: *

I've installed glusterfs-1.3.8pre1 on one node of a cluster running 
1.3.7, but glusterfsd 1.3.8 drops incoming connections from 1.3.7 
clients. Is this by design ? Do I need to upgrade everything to 
1.3.8pre1 at once ?

Here's a part of /var/log/glusterfs/glusterfsd.log

2008-02-22 17:17:33 C [tcp.c:87:tcp_disconnect] server: connection 
disconnected
2008-02-22 17:17:33 E [server-protocol.c:183:generic_reply] server: 
transport_writev failed
2008-02-22 17:17:33 C [tcp.c:87:tcp_disconnect] server: connection 
disconnected
2008-02-22 17:17:33 E [server-protocol.c:183:generic_reply] server: 
transport_writev failed
2008-02-22 17:17:33 C [tcp.c:87:tcp_disconnect] server: connection 
disconnected
2008-02-22 17:17:33 C [tcp.c:87:tcp_disconnect] server: connection 
disconnected


*Q2: *

The reasons for trying to upgrade to 1.3.8 are the following:

Current configuration:

- 16 clients / 16 servers (one client/server on each machine)
- servers are dual opteron, some of them quad core, 8 or 12 gb ram
- kernel 2.6.24-2, linux gentoo (can provide gluster ebuilds)
- fuse 2.7.2glfs8, glusterfs 1.3.7 - see config files- basicly a simple 
unify with no ra/wt cache

Configs are here: http://gluster.pastebin.com/m7f61927f
All servers are stable and the problems below are in normal running 
conditions.

Inside the gluster filesystem we store ~3 million pictures, in a 
directory tree that guarantees up
to 1k pictures or subdirectories per directory, with ~30 writes per 
second, and ~300 reads per
second. Files are relatively small, 4-5k/picture.


1. glusterfs (client) appears to memory leak in our configuration - 300 
mb RAM eaten over 2 days.

2. frequent files with size 0, ctime 0 (1970) even if all servers are up 
and running.

3. occasional files with correct size/ctime that cannot be read, and 
sometimes they can be
read from other servers.

4. back when I was using AFR for mirrored namespace (which I gave up, 
trying to alleviate the
other errors), crash in AFR in glusterfs (client) when one of the 
servers was shutting down.

These errors appear in glusterfs.log when a file cannot be read:

2008-02-22 17:37:36 E [unify.c:837:unify_open] 
nowb-nora-client-stable-www: /tmpfs/small/1/70/92/7092182.jpg: 
entry_count is 4
2008-02-22 17:34:51 E [unify.c:790:unify_open_cbk] 
nowb-nora-client-stable-www: Open success on namespace, failed on child node


*Q3:*

Nice-to-haves:

1. Redundant namespace => no single point of failure.

2. A way to see a diagram of the cluster, it's connected nodes, etc. for 
anyone running more than 2-3 servers - pulling live data from one of the 
servers/clients.


As a side question, our organization could commit a part time developer 
dedicated to helping out with glusterfs; are you interested ?

Best regards
Dan



More information about the Gluster-devel mailing list