[Gluster-users] Transport endpoint not connected
Joe Warren-Meeks
joe at encoretickets.co.uk
Mon Apr 26 11:30:54 UTC 2010
Here is the relevant crash section:
patchset: v3.0.4
signal received: 11
time of crash: 2010-04-23 21:40:40
configuration details:
argp 1
backtrace 1
dlfcn 1
fdatasync 1
libpthread 1
llistxattr 1
setfsid 1
spinlock 1
epoll.h 1
xattr.h 1
st_atim.tv_nsec 1
package-string: glusterfs 3.0.4
/lib/libc.so.6[0x7ffd0d809100]
/usr/local/lib/glusterfs/3.0.4/xlator/performance/read-ahead.so(ra_fstat
+0x82)[0
x7ffd0c968d22]
/usr/local/lib/libglusterfs.so.0(default_fstat+0xcb)[0x7ffd0df7411b]
/usr/local/lib/glusterfs/3.0.4/xlator/performance/quick-read.so(qr_fstat
+0x113)[
0x7ffd0c5570a3]
/usr/local/lib/glusterfs/3.0.4/xlator/performance/write-behind.so(wb_fst
at_helpe
r+0xcb)[0x7ffd0c346adb]
/usr/local/lib/libglusterfs.so.0(call_resume+0x390)[0x7ffd0df7cf60]
/usr/local/lib/glusterfs/3.0.4/xlator/performance/write-behind.so(wb_res
ume_othe
r_requests+0x58)[0x7ffd0c349938]
/usr/local/lib/glusterfs/3.0.4/xlator/performance/write-behind.so(wb_pro
cess_que
ue+0xe1)[0x7ffd0c348251]
/usr/local/lib/glusterfs/3.0.4/xlator/performance/write-behind.so(wb_fst
at+0x20a
)[0x7ffd0c34a87a]
/usr/local/lib/libglusterfs.so.0(default_fstat+0xcb)[0x7ffd0df7411b]
/usr/local/lib/glusterfs/3.0.4/xlator/mount/fuse.so[0x7ffd0bf23a36]
/usr/local/lib/glusterfs/3.0.4/xlator/mount/fuse.so[0x7ffd0bf246b6]
/lib/libpthread.so.0[0x7ffd0db3f3f7]
/lib/libc.so.6(clone+0x6d)[0x7ffd0d8aeb4d]
And Startup section:
---------
========================================================================
========
Version : glusterfs 3.0.4 built on Apr 19 2010 16:37:50
git: v3.0.4
Starting Time: 2010-04-26 10:00:59
Command line : /usr/local/sbin/glusterfs --log-level=NORMAL
--volfile=/etc/glust
erfs/repstore1-tcp.vol /data/import
PID : 5910
System name : Linux
Nodename : w2
Kernel Release : 2.6.24-27-server
Hardware Identifier: x86_64
Given volfile:
+-----------------------------------------------------------------------
-------+
1: ## file auto generated by /usr/local/bin/glusterfs-volgen
(mount.vol)
2: # Cmd line:
3: # $ /usr/local/bin/glusterfs-volgen --name repstore1 --raid 1
10.10.130.11:/data/export 10.10.130.12:/data/export
4:
5: # RAID 1
6: # TRANSPORT-TYPE tcp
7: volume 10.10.130.12-1
8: type protocol/client
9: option transport-type tcp
10: option remote-host 10.10.130.12
11: option transport.socket.nodelay on
12: option transport.remote-port 6996
13: option remote-subvolume brick1
14: end-volume
15:
16: volume 10.10.130.11-1
17: type protocol/client
18: option transport-type tcp
19: option remote-host 10.10.130.11
20: option transport.socket.nodelay on
21: option transport.remote-port 6996
22: option remote-subvolume brick1
23: end-volume
24:
25: volume mirror-0
26: type cluster/replicate
27: subvolumes 10.10.130.11-1 10.10.130.12-1
28: end-volume
29:
30: volume readahead
31: type performance/read-ahead
32: option page-count 4
33: subvolumes mirror-0
34: end-volume
35:
36: volume iocache
37: type performance/io-cache
38: option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo |
sed 's/[^0-9]//g') / 5120 ))`MB
39: option cache-timeout 1
40: subvolumes readahead
41: end-volume
42:
43: volume quickread
44: type performance/quick-read
45: option cache-timeout 1
46: option max-file-size 64kB
47: subvolumes iocache
48: end-volume
49:
50: volume writebehind
51: type performance/write-behind
52: option cache-size 4MB
53: subvolumes quickread
54: end-volume
55:
56: volume statprefetch
57: type performance/stat-prefetch
58: subvolumes writebehind
59: end-volume
60:
> -----Original Message-----
> From: Vijay Bellur [mailto:vijay at gluster.com]
> Sent: 22 April 2010 18:40
> To: Joe Warren-Meeks
> Cc: gluster-users at gluster.org
> Subject: Re: [Gluster-users] Transport endpoint not connected
>
> Hi Joe,
>
> Can you please share the complete client log file?
>
> Thanks,
> Vijay
>
>
> Joe Warren-Meeks wrote:
> > Hey guys,
> >
> >
> >
> > I've recently implemented gluster to share webcontent read-write
> between
> > two servers.
> >
> >
> >
> > Version : glusterfs 3.0.4 built on Apr 19 2010 16:37:50
> >
> > Fuse : 2.7.2-1ubuntu2.1
> >
> > Platform : ubuntu 8.04LTS
> >
> >
> >
> > I used the following command to generate my configs:
> >
> > /usr/local/bin/glusterfs-volgen --name repstore1 --raid 1
> > 10.10.130.11:/data/export 10.10.130.12:/data/export
> >
> >
> >
> > And mount them on each of the servers as so:
> >
> > /etc/fstab:
> >
> > /etc/glusterfs/repstore1-tcp.vol /data/import glusterfs defaults
> 0
> > 0
> >
> >
> >
> >
> >
> > Every 12 hours or so, one or other of the servers will lose the
mount
> > and error with:
> >
> > df: `/data/import': Transport endpoint is not connected
> >
> >
> >
> > And I get the following in my logfile:
> >
> > patchset: v3.0.4
> >
> > signal received: 11
> >
> > time of crash: 2010-04-22 11:41:10
> >
> > configuration details:
> >
> > argp 1
> >
> > backtrace 1
> >
> > dlfcn 1
> >
> > fdatasync 1
> >
> > libpthread 1
> >
> > llistxattr 1
> >
> > setfsid 1
> >
> > spinlock 1
> >
> > epoll.h 1
> >
> > xattr.h 1
> >
> > st_atim.tv_nsec 1
> >
> > package-string: glusterfs 3.0.4
> >
> > /lib/libc.so.6[0x7f2eca39a100]
> >
> > /usr/local/lib/glusterfs/3.0.4/xlator/performance/read-
> ahead.so(ra_fstat
> > +0x82
> >
> > )[0x7f2ec94f9d22]
> >
> > /usr/local/lib/libglusterfs.so.0(default_fstat+0xcb)[0x7f2ecab0511b]
> >
> > /usr/local/lib/glusterfs/3.0.4/xlator/performance/quick-
> read.so(qr_fstat
> > +0x11
> >
> > 3)[0x7f2ec90e80a3]
> >
> > /usr/local/lib/glusterfs/3.0.4/xlator/performance/write-
> behind.so(wb_fst
> > at_he
> >
> > lper+0xcb)[0x7f2ec8ed7adb]
> >
> > /usr/local/lib/libglusterfs.so.0(call_resume+0x390)[0x7f2ecab0df60]
> >
> > /usr/local/lib/glusterfs/3.0.4/xlator/performance/write-
> behind.so(wb_res
> > ume_o
> >
> > ther_requests+0x58)[0x7f2ec8eda938]
> >
> > /usr/local/lib/glusterfs/3.0.4/xlator/performance/write-
> behind.so(wb_pro
> > cess_queue+0xe1)[0x7f2ec8ed9251]
> >
> > /usr/local/lib/glusterfs/3.0.4/xlator/performance/write-
> behind.so(wb_fst
> > at+0x20a)[0x7f2ec8edb87a]
> >
> > /usr/local/lib/libglusterfs.so.0(default_fstat+0xcb)[0x7f2ecab0511b]
> >
> > /usr/local/lib/glusterfs/3.0.4/xlator/mount/fuse.so[0x7f2ec8ab4a36]
> >
> > /usr/local/lib/glusterfs/3.0.4/xlator/mount/fuse.so[0x7f2ec8ab56b6]
> >
> > /lib/libpthread.so.0[0x7f2eca6d03f7]
> >
> > /lib/libc.so.6(clone+0x6d)[0x7f2eca43fb4d]
> >
> >
> >
> >
> >
> > If I umount and remount, things work again, but it isn't ideal..
> >
> >
> >
> > Any clues, pointers, hints?
> >
> >
> >
> > Kind regards
> >
> >
> >
> > -- joe.
> >
> >
> >
> > Joe Warren-Meeks
> >
> > Director Of Systems Development
> >
> > ENCORE TICKETS LTD
> >
> > Encore House, 50-51 Bedford Row, London WC1R 4LR
> >
> > Direct line: +44 (0)20 7492 1506
> >
> > Reservations: +44 (0)20 7492 1500
> >
> > Fax: +44 (0)20 7831 4410
> >
> > Email: joe at encoretickets.co.uk
> > <mailto:joe at encoretickets.co.uk>
> >
> > web: www.encoretickets.co.uk
> > <http://www.encoretickets.co.uk/>
> >
> >
> >
> >
> >
> > Copyright in this message and any attachments remains with us. It is
> > confidential and may be legally privileged. If this message is not
> > intended for you it must not be read, copied or used by you or
> disclosed
> > to anyone else. Please advise the sender immediately if you have
> > received this message in error. Although this message and any
> > attachments are believed to be free of any virus or other defect
that
> > might affect any computer system into which it is received and
opened
> it
> > is the responsibility of the recipient to ensure that it is virus
> free
> > and no responsibility is accepted by Encore Tickets Limited for any
> loss
> > or damage in any way arising from its use.
> >
> >
> >
> >
> >
> >
---------------------------------------------------------------------
> ---
> >
> > _______________________________________________
> > Gluster-users mailing list
> > Gluster-users at gluster.org
> > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
> >
>
More information about the Gluster-users
mailing list