[Gluster-users] smth wrongs with glusterfs

Alexey Shalin cash at hoster.kg
Fri Dec 5 07:41:18 UTC 2014


Hello, again
Smth wrong with my install gluster

OS:Debian
cat /etc/debian_version 7.6
Package: glusterfs-server
Versions:
3.2.7-3+deb7u1

Description:   I   have   3   servers   with   bricks  (192.168.1.1  -
node1,192.168.1.2 - node2, 192.168.1.3 - node3)
volume create by:
gluster volume create opennebula transport tcp node1:/data node2:/data node3:/data

192.168.1.4 - client

# volume info
gluster volume info

Volume Name: opennebula
Type: Replicate
Status: Started
Number of Bricks: 3
Transport-type: tcp
Bricks:
Brick1: node1:/data
Brick2: node2:/data
Brick3: node3:/data
Options Reconfigured:
server.allow-insecure: on


#peer info
gluster peer show
unrecognized word: show (position 1)
root at node1:/data# gluster peer status
Number of Peers: 2

Hostname: node3
Uuid: 355f676d-044c-453d-8e82-13b810c089bb
State: Peer in Cluster (Connected)

Hostname: node2
Uuid: bfed0b59-6b2f-474e-a3d7-18b0eb0b1c77
State: Peer in Cluster (Connected)


# on client i mounted volume by:
mount.glusterfs node1:/opennebula /var/lib/one/

ls -al /var/lib/one - show  files, but after 1 mins
ls -al /var/lib/one - hangs up


log

----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
[2014-12-05 13:28:53.290981] I [fuse-bridge.c:3461:fuse_graph_setup] 0-fuse: switched to graph 0
[2014-12-05 13:28:53.291223] I [fuse-bridge.c:3049:fuse_init] 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.13 kernel 7.17
[2014-12-05 13:28:53.291800] I [afr-common.c:1522:afr_set_root_inode_on_first_lookup] 0-opennebula-replicate-0: added root inode
[2014-12-05 13:29:16.355469] C [client-handshake.c:121:rpc_client_ping_timer_expired] 0-opennebula-client-0: server 192.168.1.1:24009 has not responded in the last 42 seconds, disconnecting.
[2014-12-05 13:29:16.355684] E [rpc-clnt.c:341:saved_frames_unwind] (-->/usr/lib/libgfrpc.so.0(rpc_clnt_notify+0xb0) [0x7f7020ccec60] (-->/usr/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e) [0x7f7020cce8fe] (-->/usr/lib/libgfrpc.so.0(saved_frames_destroy+0xe) [0x7f7020cce85e]))) 0-opennebula-client-0: forced unwinding frame type(GlusterFS 3.1) op(READDIRP(40)) called at 2014-12-05 13:27:10.345569
[2014-12-05 13:29:16.355754] E [client3_1-fops.c:1937:client3_1_readdirp_cbk] 0-opennebula-client-0: remote operation failed: Transport endpoint is not connected
[2014-12-05 13:29:16.355772] I [afr-self-heal-entry.c:1846:afr_sh_entry_impunge_readdir_cbk] 0-opennebula-replicate-0: readdir of / on subvolume opennebula-client-0 failed (Transport endpoint is not connected)
[2014-12-05 13:29:16.356073] I [socket.c:2275:socket_submit_request] 0-opennebula-client-0: not connected (priv->connected = 0)
[2014-12-05 13:29:16.356091] W [rpc-clnt.c:1417:rpc_clnt_submit] 0-opennebula-client-0: failed to submit rpc-request (XID: 0x112x Program: GlusterFS 3.1, ProgVers: 310, Proc: 33) to rpc-transport (opennebula-client-0)
[2014-12-05 13:29:16.356107] I [afr-self-heal-entry.c:129:afr_sh_entry_erase_pending_cbk] 0-opennebula-replicate-0: /: failed to erase pending xattrs on opennebula-client-0 (Transport endpoint is not connected)
[2014-12-05 13:29:16.356209] E [rpc-clnt.c:341:saved_frames_unwind] (-->/usr/lib/libgfrpc.so.0(rpc_clnt_notify+0xb0) [0x7f7020ccec60] (-->/usr/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e) [0x7f7020cce8fe] (-->/usr/lib/libgfrpc.so.0(saved_frames_destroy+0xe) [0x7f7020cce85e]))) 0-opennebula-client-0: forced unwinding frame type(GlusterFS Handshake) op(PING(3)) called at 2014-12-05 13:27:52.348889
[2014-12-05 13:29:16.356227] W [client-handshake.c:264:client_ping_cbk] 0-opennebula-client-0: timer must have expired
[2014-12-05 13:29:16.356257] E [rpc-clnt.c:341:saved_frames_unwind] (-->/usr/lib/libgfrpc.so.0(rpc_clnt_notify+0xb0) [0x7f7020ccec60] (-->/usr/lib/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x7e) [0x7f7020cce8fe] (-->/usr/lib/libgfrpc.so.0(saved_frames_destroy+0xe) [0x7f7020cce85e]))) 0-opennebula-client-0: forced unwinding frame type(GlusterFS 3.1) op(STATFS(14)) called at 2014-12-05 13:28:20.214777
[2014-12-05 13:29:16.356274] I [client3_1-fops.c:637:client3_1_statfs_cbk] 0-opennebula-client-0: remote operation failed: Transport endpoint is not connected
[2014-12-05 13:29:16.356304] I [client.c:1883:client_rpc_notify] 0-opennebula-client-0: disconnected
[2014-12-05 13:29:16.356663] I [client-handshake.c:1090:select_server_supported_programs] 0-opennebula-client-0: Using Program GlusterFS 3.2.7, Num (1298437), Version (310)
[2014-12-05 13:29:16.356966] W [rpc-common.c:64:xdr_to_generic] (-->/usr/lib/libgfrpc.so.0(rpc_clnt_notify+0x85) [0x7f7020ccec35] (-->/usr/lib/libgfrpc.so.0(rpc_clnt_handle_reply+0xa5) [0x7f7020cce295] (-->/usr/lib/glusterfs/3.2.7/xlator/protocol/client.so(client3_1_entrylk_cbk+0x52) [0x7f701da44122]))) 0-xdr: XDR decoding failed
[2014-12-05 13:29:16.356993] E [client3_1-fops.c:1292:client3_1_entrylk_cbk] 0-opennebula-client-0: error
[2014-12-05 13:29:16.357015] E [client3_1-fops.c:1303:client3_1_entrylk_cbk] 0-opennebula-client-0: remote operation failed: Invalid argument
[2014-12-05 13:29:16.357036] I [afr-self-heal-common.c:2193:afr_self_heal_completion_cbk] 0-opennebula-replicate-0: background  entry self-heal completed on /
[2014-12-05 13:29:16.357229] I [client-handshake.c:913:client_setvolume_cbk] 0-opennebula-client-0: Connected to 192.168.1.1:24009, attached to remote volume '/data'.
[2014-12-05 13:29:16.357246] I [client-handshake.c:779:client_post_handshake] 0-opennebula-client-0: 2 fds open - Delaying child_up until they are re-opened
[2014-12-05 13:29:16.357617] I [client-handshake.c:536:client3_1_reopendir_cbk] 0-opennebula-client-0: reopendir on / succeeded (fd = 0)
[2014-12-05 13:29:16.357651] I [client-handshake.c:536:client3_1_reopendir_cbk] 0-opennebula-client-0: reopendir on / succeeded (fd = 1)
[2014-12-05 13:29:16.357666] I [client-lk.c:617:decrement_reopen_fd_count] 0-opennebula-client-0: last fd open'd/lock-self-heal'd - notifying CHILD-UP
[2014-12-05 13:29:16.357681] I [client3_1-fops.c:2355:client_fdctx_destroy] 0-opennebula-client-0: sending releasedir on fd
[2014-12-05 13:29:16.377961] I [afr-common.c:1039:afr_launch_self_heal] 0-opennebula-replicate-0: background  entry self-heal triggered. path: /
[2014-12-05 13:29:16.378081] I [afr-common.c:1039:afr_launch_self_heal] 0-opennebula-replicate-0: background  entry self-heal triggered. path: /
[2014-12-05 13:29:16.378333] E [afr-self-heal-entry.c:2189:afr_sh_post_nonblocking_entry_cbk] 0-opennebula-replicate-0: Non Blocking entrylks failed for /.
[2014-12-05 13:29:16.378359] E [afr-self-heal-common.c:2190:afr_self_heal_completion_cbk] 0-opennebula-replicate-0: background  entry self-heal failed on /
----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------

and now .. df -h - also hangs up :(

node1
root at node1:/data# ps aux | grep gluster
root      2391  0.0  0.3  67676 18428 ?        Ssl  13:21   0:00 /usr/sbin/glusterd -p /var/run/glusterd.pid
root      2959  0.0  0.2 218924 14900 ?        Ssl  13:21   0:00 /usr/sbin/glusterfsd --xlator-option opennebula-server.listen-port=24009 -s localhost --volfile-id opennebula.node1.data -p /etc/glusterd/vols/opennebula/run/node1-data.pid -S /tmp/41ee3506b47079b17ab7acbda6b5b459.socket --brick-name /data --brick-port 24009 -l /var/log/glusterfs/bricks/data.log
root      2963  0.0  0.6 168580 41336 ?        Ssl  13:21   0:00 /usr/sbin/glusterfs -f /etc/glusterd/nfs/nfs-server.vol -p /etc/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log
root      3129  0.0  0.0   6304   596 pts/0    S+   13:38   0:00 grep gluster
root at node1:/data#


node2
root at node2:~#  ps aux | grep gluster
root      2335  0.0  0.2  67676 18424 ?        Ssl  13:20   0:00 /usr/sbin/glusterd -p /var/run/glusterd.pid
root      2961  0.0  0.1 149112 14656 ?        Ssl  13:20   0:00 /usr/sbin/glusterfsd --xlator-option opennebula-server.listen-port=24009 -s localhost --volfile-id opennebula.node2.data -p /etc/glusterd/vols/opennebula/run/node2-data.pid -S /tmp/191be92428c92005cd8acf75ec50fdb9.socket --brick-name /data --brick-port 24009 -l /var/log/glusterfs/bricks/data.log
root      2966  0.0  0.5 103044 41324 ?        Ssl  13:20   0:00 /usr/sbin/glusterfs -f /etc/glusterd/nfs/nfs-server.vol -p /etc/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log
root      3190  0.0  0.0   7832   880 pts/0    S+   13:38   0:00 grep gluster
root at node2:~#

root at node3:~#  ps aux | grep gluster
root      2394  0.0  0.2  67676 18428 ?        Ssl  11:50   0:00 /usr/sbin/glusterd -p /var/run/glusterd.pid
root      2964  0.0  0.1 149196 14648 ?        Ssl  11:51   0:00 /usr/sbin/glusterfsd --xlator-option opennebula-server.listen-port=24009 -s localhost --volfile-id opennebula.node3.data -p /etc/glusterd/vols/opennebula/run/node3-data.pid -S /tmp/14356b3499622409f8bfb31f38493f06.socket --brick-name /data --brick-port 24009 -l /var/log/glusterfs/bricks/data.log
root      2970  0.0  0.5 103044 41572 ?        Ssl  11:51   0:00 /usr/sbin/glusterfs -f /etc/glusterd/nfs/nfs-server.vol -p /etc/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log
root      3422  0.0  0.0   7832   880 pts/0    S+   13:38   0:00 grep gluster
root at node3:~#


Can you please help me to resolve it ?



-------------------------------------------------------
Старший Системный Администратор
Алексей Шалин
ОсОО "Хостер kg" - http://www.hoster.kg
ул. Ахунбаева 123 (здание БГТС)
help at hoster.kg



More information about the Gluster-users mailing list