[Gluster-users] NFS problem
anthony garnier
sokar6012 at hotmail.com
Thu Jun 9 15:09:37 UTC 2011
Hi,
I got the same problem as Juergen,
My volume is a simple replicated volume with 2 host and GlusterFS 3.2.0
Volume Name: poolsave
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: ylal2950:/soft/gluster-data
Brick2: ylal2960:/soft/gluster-data
Options Reconfigured:
diagnostics.brick-log-level: DEBUG
network.ping-timeout: 20
performance.cache-size: 512MB
nfs.port: 2049
I'm running this command :
I get those error :
tar: ./uvs00: owner not changed
tar: could not stat ./uvs00/log/0906uvsGESEC.log
tar: ./uvs00: group not changed
tar: could not stat ./uvs00/log/0306uvsGESEC.log
tar: ./uvs00/log: Input/output error
cannot change back?: Unknown error 526
tar: ./uvs00/log: owner not changed
tar: ./uvs00/log: group not changed
tar: tape blocksize error
And then I tried to "ls" in gluster mount :
/bin/ls: .: Input/output error
only way is to restart the volume
Here is the logfile in Debug mod :
Given volfile:
+------------------------------------------------------------------------------+
1: volume poolsave-client-0
2: type protocol/client
3: option remote-host ylal2950
4: option remote-subvolume /soft/gluster-data
5: option transport-type tcp
6: option ping-timeout 20
7: end-volume
8:
9: volume poolsave-client-1
10: type protocol/client
11: option remote-host ylal2960
12: option remote-subvolume /soft/gluster-data
13: option transport-type tcp
14: option ping-timeout 20
15: end-volume
16:
17: volume poolsave-replicate-0
18: type cluster/replicate
19: subvolumes poolsave-client-0 poolsave-client-1
20: end-volume
21:
22: volume poolsave-write-behind
23: type performance/write-behind
24: subvolumes poolsave-replicate-0
25: end-volume
26:
27: volume poolsave-read-ahead
28: type performance/read-ahead
29: subvolumes poolsave-write-behind
30: end-volume
31:
32: volume poolsave-io-cache
33: type performance/io-cache
34: option cache-size 512MB
35: subvolumes poolsave-read-ahead
36: end-volume
37:
38: volume poolsave-quick-read
39: type performance/quick-read
40: option cache-size 512MB
41: subvolumes poolsave-io-cache
42: end-volume
43:
44: volume poolsave-stat-prefetch
45: type performance/stat-prefetch
46: subvolumes poolsave-quick-read
47: end-volume
48:
49: volume poolsave
50: type debug/io-stats
51: option latency-measurement off
52: option count-fop-hits off
53: subvolumes poolsave-stat-prefetch
54: end-volume
55:
56: volume nfs-server
57: type nfs/server
58: option nfs.dynamic-volumes on
59: option rpc-auth.addr.poolsave.allow *
60: option nfs3.poolsave.volume-id 71e0dabf-4620-4b6d-b138-3266096b93b6
61: option nfs.port 2049
62: subvolumes poolsave
63: end-volume
+------------------------------------------------------------------------------+
[2011-06-09 16:52:23.709018] I [rpc-clnt.c:1531:rpc_clnt_reconfig] 0-poolsave-client-0: changing port to 24014 (from 0)
[2011-06-09 16:52:23.709211] I [rpc-clnt.c:1531:rpc_clnt_reconfig] 0-poolsave-client-1: changing port to 24011 (from 0)
[2011-06-09 16:52:27.716417] I [client-handshake.c:1080:select_server_supported_programs] 0-poolsave-client-0: Using Program GlusterFS-3.1.0, Num (1298437), Version (310)
[2011-06-09 16:52:27.716650] I [client-handshake.c:913:client_setvolume_cbk] 0-poolsave-client-0: Connected to 10.68.217.85:24014, attached to remote volume '/soft/gluster-data'.
[2011-06-09 16:52:27.716679] I [afr-common.c:2514:afr_notify] 0-poolsave-replicate-0: Subvolume 'poolsave-client-0' came back up; going online.
[2011-06-09 16:52:27.717020] I [afr-common.c:836:afr_fresh_lookup_cbk] 0-poolsave-replicate-0: added root inode
[2011-06-09 16:52:27.729719] I [client-handshake.c:1080:select_server_supported_programs] 0-poolsave-client-1: Using Program GlusterFS-3.1.0, Num (1298437), Version (310)
[2011-06-09 16:52:27.730014] I [client-handshake.c:913:client_setvolume_cbk] 0-poolsave-client-1: Connected to 10.68.217.86:24011, attached to remote volume '/soft/gluster-data'.
[2011-06-09 17:01:35.537084] W [stat-prefetch.c:178:sp_check_and_create_inode_ctx] (-->/usr/local/lib/glusterfs/3.2.0/xlator/nfs/server.so(nfs_fop_mkdir+0x1cc) [0x2aaaab3b88fc] (-->/usr/local/lib/glusterfs/3.2.0/xlator/debug/io-stats.so(io_stats_mkdir+0x151) [0x2aaaab2948e1] (-->/usr/local/lib/glusterfs/3.2.0/xlator/performance/stat-prefetch.so(sp_mkdir+0xd2) [0x2aaaab1856c2]))) 0-poolsave-stat-prefetch: stat-prefetch context is present in inode (ino:0 gfid:00000000-0000-0000-0000-000000000000) when it is supposed to be not present
[2011-06-09 17:01:35.546601] W [stat-prefetch.c:178:sp_check_and_create_inode_ctx] (-->/usr/local/lib/glusterfs/3.2.0/xlator/nfs/server.so(nfs_fop_create+0x1db) [0x2aaaab3b95bb] (-->/usr/local/lib/glusterfs/3.2.0/xlator/debug/io-stats.so(io_stats_create+0x165) [0x2aaaab294ad5] (-->/usr/local/lib/glusterfs/3.2.0/xlator/performance/stat-prefetch.so(sp_create+0xbc) [0x2aaaab185c9c]))) 0-poolsave-stat-prefetch: stat-prefetch context is present in inode (ino:0 gfid:00000000-0000-0000-0000-000000000000) when it is supposed to be not present
[2011-06-09 17:01:35.569755] I [client3_1-fops.c:547:client3_1_rmdir_cbk] 0-poolsave-client-0: remote operation failed: Directory not empty
[2011-06-09 17:01:35.569881] I [client3_1-fops.c:547:client3_1_rmdir_cbk] 0-poolsave-client-1: remote operation failed: Directory not empty
[2011-06-09 17:01:35.579674] W [stat-prefetch.c:178:sp_check_and_create_inode_ctx] (-->/usr/local/lib/glusterfs/3.2.0/xlator/nfs/server.so(nfs_fop_mkdir+0x1cc) [0x2aaaab3b88fc] (-->/usr/local/lib/glusterfs/3.2.0/xlator/debug/io-stats.so(io_stats_mkdir+0x151) [0x2aaaab2948e1] (-->/usr/local/lib/glusterfs/3.2.0/xlator/performance/stat-prefetch.so(sp_mkdir+0xd2) [0x2aaaab1856c2]))) 0-poolsave-stat-prefetch: stat-prefetch context is present in inode (ino:0 gfid:00000000-0000-0000-0000-000000000000) when it is supposed to be not present
[2011-06-09 17:01:35.587907] W [stat-prefetch.c:178:sp_check_and_create_inode_ctx] (-->/usr/local/lib/glusterfs/3.2.0/xlator/nfs/server.so(nfs_fop_create+0x1db) [0x2aaaab3b95bb] (-->/usr/local/lib/glusterfs/3.2.0/xlator/debug/io-stats.so(io_stats_create+0x165) [0x2aaaab294ad5] (-->/usr/local/lib/glusterfs/3.2.0/xlator/performance/stat-prefetch.so(sp_create+0xbc) [0x2aaaab185c9c]))) 0-poolsave-stat-prefetch: stat-prefetch context is present in inode (ino:0 gfid:00000000-0000-0000-0000-000000000000) when it is supposed to be not present
[2011-06-09 17:01:35.612918] W [stat-prefetch.c:178:sp_check_and_create_inode_ctx] (-->/usr/local/lib/glusterfs/3.2.0/xlator/nfs/server.so(nfs_fop_create+0x1db) [0x2aaaab3b95bb] (-->/usr/local/lib/glusterfs/3.2.0/xlator/debug/io-stats.so(io_stats_create+0x165) [0x2aaaab294ad5] (-->/usr/local/lib/glusterfs/3.2.0/xlator/performance/stat-prefetch.so(sp_create+0xbc) [0x2aaaab185c9c]))) 0-poolsave-stat-prefetch: stat-prefetch context is present in inode (ino:0 gfid:00000000-0000-0000-0000-000000000000) when it is supposed to be not present
[2011-06-09 17:01:35.645357] W [stat-prefetch.c:178:sp_check_and_create_inode_ctx] (-->/usr/local/lib/glusterfs/3.2.0/xlator/nfs/server.so(nfs_fop_create+0x1db) [0x2aaaab3b95bb] (-->/usr/local/lib/glusterfs/3.2.0/xlator/debug/io-stats.so(io_stats_create+0x165) [0x2aaaab294ad5] (-->/usr/local/lib/glusterfs/3.2.0/xlator/performance/stat-prefetch.so(sp_create+0xbc) [0x2aaaab185c9c]))) 0-poolsave-stat-prefetch: stat-prefetch context is present in inode (ino:0 gfid:00000000-0000-0000-0000-000000000000) when it is supposed to be not present
[2011-06-09 17:01:35.660873] I [client3_1-fops.c:547:client3_1_rmdir_cbk] 0-poolsave-client-0: remote operation failed: Directory not empty
[2011-06-09 17:01:35.660955] I [client3_1-fops.c:547:client3_1_rmdir_cbk] 0-poolsave-client-1: remote operation failed: Directory not empty
[2011-06-09 17:01:35.665933] I [client3_1-fops.c:547:client3_1_rmdir_cbk] 0-poolsave-client-0: remote operation failed: Directory not empty
[2011-06-09 17:01:35.666057] I [client3_1-fops.c:547:client3_1_rmdir_cbk] 0-poolsave-client-1: remote operation failed: Directory not empty
[2011-06-09 17:01:35.671199] I [client3_1-fops.c:547:client3_1_rmdir_cbk] 0-poolsave-client-0: remote operation failed: Directory not empty
[2011-06-09 17:01:35.671241] I [client3_1-fops.c:547:client3_1_rmdir_cbk] 0-poolsave-client-1: remote operation failed: Directory not empty
[2011-06-09 17:01:35.680959] W [stat-prefetch.c:178:sp_check_and_create_inode_ctx] (-->/usr/local/lib/glusterfs/3.2.0/xlator/nfs/server.so(nfs_fop_create+0x1db) [0x2aaaab3b95bb] (-->/usr/local/lib/glusterfs/3.2.0/xlator/debug/io-stats.so(io_stats_create+0x165) [0x2aaaab294ad5] (-->/usr/local/lib/glusterfs/3.2.0/xlator/performance/stat-prefetch.so(sp_create+0xbc) [0x2aaaab185c9c]))) 0-poolsave-stat-prefetch: stat-prefetch context is present in inode (ino:0 gfid:00000000-0000-0000-0000-000000000000) when it is supposed to be not present
[2011-06-09 17:01:35.715633] W [stat-prefetch.c:178:sp_check_and_create_inode_ctx] (-->/usr/local/lib/glusterfs/3.2.0/xlator/nfs/server.so(nfs_fop_create+0x1db) [0x2aaaab3b95bb] (-->/usr/local/lib/glusterfs/3.2.0/xlator/debug/io-stats.so(io_stats_create+0x165) [0x2aaaab294ad5] (-->/usr/local/lib/glusterfs/3.2.0/xlator/performance/stat-prefetch.so(sp_create+0xbc) [0x2aaaab185c9c]))) 0-poolsave-stat-prefetch: stat-prefetch context is present in inode (ino:0 gfid:00000000-0000-0000-0000-000000000000) when it is supposed to be not present
[2011-06-09 17:01:35.732798] I [client3_1-fops.c:547:client3_1_rmdir_cbk] 0-poolsave-client-0: remote operation failed: Permission denied
[2011-06-09 17:01:35.733044] I [client3_1-fops.c:547:client3_1_rmdir_cbk] 0-poolsave-client-1: remote operation failed: Permission denied
[2011-06-09 17:01:35.750009] W [stat-prefetch.c:178:sp_check_and_create_inode_ctx] (-->/usr/local/lib/glusterfs/3.2.0/xlator/nfs/server.so(nfs_fop_create+0x1db) [0x2aaaab3b95bb] (-->/usr/local/lib/glusterfs/3.2.0/xlator/debug/io-stats.so(io_stats_create+0x165) [0x2aaaab294ad5] (-->/usr/local/lib/glusterfs/3.2.0/xlator/performance/stat-prefetch.so(sp_create+0xbc) [0x2aaaab185c9c]))) 0-poolsave-stat-prefetch: stat-prefetch context is present in inode (ino:0 gfid:00000000-0000-0000-0000-000000000000) when it is supposed to be not present
[2011-06-09 17:01:35.784610] W [socket.c:1494:__socket_proto_state_machine] 0-poolsave-client-0: reading from socket failed. Error (Transport endpoint is not connected), peer (10.68.217.85:24014)
[2011-06-09 17:01:35.784745] E [rpc-clnt.c:338:saved_frames_unwind] (-->/usr/local/lib/libgfrpc.so.0(rpc_clnt_notify+0xb9) [0x2ab58145f7f9] (-->/usr/local/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e) [0x2ab58145ef8e] (-->/usr/local/lib/libgfrpc.so.0(saved_frames_destroy+0xe) [0x2ab58145eefe]))) 0-poolsave-client-0: forced unwinding frame type(GlusterFS 3.1) op(SETATTR(38)) called at 2011-06-09 17:01:35.752080
[2011-06-09 17:01:35.784770] I [client3_1-fops.c:1640:client3_1_setattr_cbk] 0-poolsave-client-0: remote operation failed: Transport endpoint is not connected
[2011-06-09 17:01:35.784811] E [rpc-clnt.c:338:saved_frames_unwind] (-->/usr/local/lib/libgfrpc.so.0(rpc_clnt_notify+0xb9) [0x2ab58145f7f9] (-->/usr/local/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e) [0x2ab58145ef8e] (-->/usr/local/lib/libgfrpc.so.0(saved_frames_destroy+0xe) [0x2ab58145eefe]))) 0-poolsave-client-0: forced unwinding frame type(GlusterFS 3.1) op(STAT(1)) called at 2011-06-09 17:01:35.752414
[2011-06-09 17:01:35.784828] I [client3_1-fops.c:411:client3_1_stat_cbk] 0-poolsave-client-0: remote operation failed: Transport endpoint is not connected
[2011-06-09 17:01:35.784875] I [client.c:1883:client_rpc_notify] 0-poolsave-client-0: disconnected
[2011-06-09 17:01:35.785400] W [socket.c:204:__socket_rwv] 0-poolsave-client-1: readv failed (Connection reset by peer)
[2011-06-09 17:01:35.785435] W [socket.c:1494:__socket_proto_state_machine] 0-poolsave-client-1: reading from socket failed. Error (Connection reset by peer), peer (10.68.217.86:24011)
[2011-06-09 17:01:35.785496] E [rpc-clnt.c:338:saved_frames_unwind] (-->/usr/local/lib/libgfrpc.so.0(rpc_clnt_notify+0xb9) [0x2ab58145f7f9] (-->/usr/local/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e) [0x2ab58145ef8e] (-->/usr/local/lib/libgfrpc.so.0(saved_frames_destroy+0xe) [0x2ab58145eefe]))) 0-poolsave-client-1: forced unwinding frame type(GlusterFS 3.1) op(SETATTR(38)) called at 2011-06-09 17:01:35.752089
[2011-06-09 17:01:35.785516] I [client3_1-fops.c:1640:client3_1_setattr_cbk] 0-poolsave-client-1: remote operation failed: Transport endpoint is not connected
[2011-06-09 17:01:35.785542] W [client3_1-fops.c:4379:client3_1_xattrop] 0-poolsave-client-0: failed to send the fop: Transport endpoint is not connected
[2011-06-09 17:01:35.817662] I [socket.c:2272:socket_submit_request] 0-poolsave-client-1: not connected (priv->connected = 0)
[2011-06-09 17:01:35.817698] W [rpc-clnt.c:1411:rpc_clnt_submit] 0-poolsave-client-1: failed to submit rpc-request (XID: 0x576x Program: GlusterFS 3.1, ProgVers: 310, Proc: 33) to rpc-transport (poolsave-client-1)
[2011-06-09 17:01:35.817721] W [client3_1-fops.c:4735:client3_1_inodelk] 0-poolsave-client-0: failed to send the fop: Transport endpoint is not connected
[2011-06-09 17:01:35.817744] W [rpc-clnt.c:1411:rpc_clnt_submit] 0-poolsave-client-1: failed to submit rpc-request (XID: 0x577x Program: GlusterFS 3.1, ProgVers: 310, Proc: 29) to rpc-transport (poolsave-client-1)
[2011-06-09 17:01:35.817780] I [client3_1-fops.c:1226:client3_1_inodelk_cbk] 0-poolsave-client-1: remote operation failed: Transport endpoint is not connected
[2011-06-09 17:01:35.817897] E [rpc-clnt.c:338:saved_frames_unwind] (-->/usr/local/lib/libgfrpc.so.0(rpc_clnt_notify+0xb9) [0x2ab58145f7f9] (-->/usr/local/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e) [0x2ab58145ef8e] (-->/usr/local/lib/libgfrpc.so.0(saved_frames_destroy+0xe) [0x2ab58145eefe]))) 0-poolsave-client-1: forced unwinding frame type(GlusterFS 3.1) op(STAT(1)) called at 2011-06-09 17:01:35.784870
[2011-06-09 17:01:35.817918] I [client3_1-fops.c:411:client3_1_stat_cbk] 0-poolsave-client-1: remote operation failed: Transport endpoint is not connected
[2011-06-09 17:01:35.817969] I [client.c:1883:client_rpc_notify] 0-poolsave-client-1: disconnected
[2011-06-09 17:01:35.817988] E [afr-common.c:2546:afr_notify] 0-poolsave-replicate-0: All subvolumes are down. Going offline until atleast one of them comes back up.
[2011-06-09 17:01:35.818007] E [socket.c:1685:socket_connect_finish] 0-poolsave-client-1: connection to 10.68.217.86:24011 failed (Connection refused)
[2011-06-09 17:01:35.818606] I [afr.h:838:AFR_LOCAL_INIT] 0-poolsave-replicate-0: no subvolumes up
[2011-06-09 17:01:35.819129] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /uvs00/log: no child is up
[2011-06-09 17:01:35.819354] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /uvs00/log: no child is up
[2011-06-09 17:01:35.820090] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /uvs00: no child is up
[2011-06-09 17:01:35.820760] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
[2011-06-09 17:01:35.821212] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
[2011-06-09 17:01:35.821600] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
[2011-06-09 17:01:35.822123] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
[2011-06-09 17:01:35.822511] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
[2011-06-09 17:01:35.822975] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
[2011-06-09 17:01:35.823286] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
[2011-06-09 17:01:35.823583] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
[2011-06-09 17:01:35.823857] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
[2011-06-09 17:01:47.518006] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
[2011-06-09 17:01:49.39204] E [socket.c:1685:socket_connect_finish] 0-poolsave-client-0: connection to 10.68.217.85:24014 failed (Connection refused)
[2011-06-09 17:01:49.136932] I [afr-inode-read.c:270:afr_stat] 0-poolsave-replicate-0: /: no child is up
> Message: 7
> Date: Thu, 9 Jun 2011 12:56:39 +0530
> From: Shehjar Tikoo <shehjart at gluster.com>
> Subject: Re: [Gluster-users] Glusterfs 3.2.0 NFS Problem
> To: J?rgen Winkler <juergen.winkler at xidras.com>
> Cc: gluster-users at gluster.org
> Message-ID: <4DF075AF.3040509 at gluster.com>
> Content-Type: text/plain; charset="us-ascii"; format=flowed
>
> This can happen if all your servers were unreachable for a few seconds. The
> situation must have rectified during the restart. We could confirm if you
> change the log level on nfs to DEBUG and send us the log.
>
> Thanks
> -Shehjar
>
> Ju"rgen Winkler wrote:
> > Hi,
> >
> > i noticed a strange behavior with NFS and Glusterfs 3.2.0 , 3 of our
> > Servers are loosing the Mount but when you restart the Volume on the
> > Server it works again without a remount.
> >
> > On the server i noticed this entries in the Glusterfs/Nfs log-file when
> > the mount on the Client becomes unavailable :
> >
> > [2011-06-08 14:37:02.568693] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:02.569212] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:02.611910] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:02.624477] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.288272] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.296150] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.309247] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.320939] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.321786] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.333609] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.334089] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.344662] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.352666] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.354195] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.360446] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.369331] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.471556] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:04.480013] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:05.639700] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:05.652535] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:07.578469] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:07.588949] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:07.590395] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:07.591414] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:07.591932] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:07.592596] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:07.639317] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:07.652919] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:09.332435] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:09.340622] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:09.349360] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:09.349550] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:09.360445] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:09.369497] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:09.369752] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:09.382097] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> > [2011-06-08 14:37:09.382387] I [afr-inode-read.c:270:afr_stat]
> > 0-ksc-replicate-0: /: no child is up
> >
> >
> > Thx for the help
> >
> > _______________________________________________
> > Gluster-users mailing list
> > Gluster-users at gluster.org
> > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
>
>
> ------------------------------
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>
>
> End of Gluster-users Digest, Vol 38, Issue 14
> *********************************************
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20110609/4ff7aa4c/attachment.html>
More information about the Gluster-users
mailing list