[Gluster-users] State: Peer Rejected (Connected)
YANG ChengFu
youngseph at gmail.com
Fri Jan 11 17:39:26 UTC 2013
furthermore, when I stop gluster, and restartd glusterfs, in the log, I
have
==> etc-glusterfs-glusterd.vol.log <==
[2013-01-11 16:39:55.438506] I [glusterfsd.c:1666:main]
0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 3.3.1
[2013-01-11 16:39:55.440098] I [glusterd.c:807:init] 0-management: Using
/var/lib/glusterd as working directory
[2013-01-11 16:39:55.440797] C [rdma.c:4102:gf_rdma_init]
0-rpc-transport/rdma: Failed to get IB devices
[2013-01-11 16:39:55.440859] E [rdma.c:4993:init] 0-rdma.management: Failed
to initialize IB Device
[2013-01-11 16:39:55.440881] E [rpc-transport.c:316:rpc_transport_load]
0-rpc-transport: 'rdma' initialization failed
[2013-01-11 16:39:55.440901] W [rpcsvc.c:1356:rpcsvc_transport_create]
0-rpc-service: cannot create listener, initing the transport failed
[2013-01-11 16:39:55.440992] I [glusterd.c:95:glusterd_uuid_init]
0-glusterd: retrieved UUID: eece061b-1cd0-4f30-ad17-61809297aba9
[2013-01-11 16:39:56.050996] E
[glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key:
brick-0
[2013-01-11 16:39:56.051041] E
[glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key:
brick-1
[2013-01-11 16:39:56.235444] E
[glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key:
brick-0
[2013-01-11 16:39:56.235482] E
[glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key:
brick-1
[2013-01-11 16:39:56.235810] E
[glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key:
brick-0
[2013-01-11 16:39:56.235831] E
[glusterd-store.c:2080:glusterd_store_retrieve_volume] 0-: Unknown key:
brick-1
[2013-01-11 16:39:56.236277] I [rpc-clnt.c:968:rpc_clnt_connection_init]
0-management: setting frame-timeout to 600
[2013-01-11 16:39:56.236843] I
[glusterd-handler.c:2227:glusterd_friend_add] 0-management: connect
returned 0
[2013-01-11 16:39:56.241266] E
[glusterd-store.c:2586:glusterd_resolve_all_bricks] 0-glusterd: resolve
brick failed in restore
[2013-01-11 16:39:56.243958] E [glusterd-utils.c:3418:glusterd_brick_start]
0-glusterd: cannot resolve brick: irene.mdc:/opt/gluster-data/puppet/ssl
[2013-01-11 16:39:56.247827] E [glusterd-utils.c:3418:glusterd_brick_start]
0-glusterd: cannot resolve brick: irene.mdc:/opt/gluster-data/puppet/dist
[2013-01-11 16:39:56.251832] E [glusterd-utils.c:3418:glusterd_brick_start]
0-glusterd: cannot resolve brick: irene.mdc:/opt/gluster-data/puppet/bucket
[2013-01-11 16:39:56.258909] I [rpc-clnt.c:968:rpc_clnt_connection_init]
0-management: setting frame-timeout to 600
==> nfs.log.1 <==
[2013-01-11 16:39:56.259055] W [socket.c:410:__socket_keepalive] 0-socket:
failed to set keep idle on socket 7
[2013-01-11 16:39:56.259108] W [socket.c:1876:socket_server_event_handler]
0-socket.glusterfsd: Failed to set keep-alive: Operation not supported
[2013-01-11 16:39:56.259154] W [socket.c:410:__socket_keepalive] 0-socket:
failed to set keep idle on socket 8
[2013-01-11 16:39:56.259172] W [socket.c:1876:socket_server_event_handler]
0-socket.glusterfsd: Failed to set keep-alive: Operation not supported
==> etc-glusterfs-glusterd.vol.log <==
[2013-01-11 16:39:56.266390] I [rpc-clnt.c:968:rpc_clnt_connection_init]
0-management: setting frame-timeout to 600
==> glustershd.log.1 <==
[2013-01-11 16:39:56.266520] W [socket.c:410:__socket_keepalive] 0-socket:
failed to set keep idle on socket 7
[2013-01-11 16:39:56.266562] W [socket.c:1876:socket_server_event_handler]
0-socket.glusterfsd: Failed to set keep-alive: Operation not supported
==> etc-glusterfs-glusterd.vol.log <==
Given volfile:
+------------------------------------------------------------------------------+
1: volume management
2: type mgmt/glusterd
3: option working-directory /var/lib/glusterd
4: option transport-type socket,rdma
5: option transport.socket.keepalive-time 10
6: option transport.socket.keepalive-interval 2
7: option transport.socket.read-fail-log off
8: end-volume
+------------------------------------------------------------------------------+
==> glustershd.log.1 <==
[2013-01-11 16:39:56.266610] W [socket.c:410:__socket_keepalive] 0-socket:
failed to set keep idle on socket 8
[2013-01-11 16:39:56.266624] W [socket.c:1876:socket_server_event_handler]
0-socket.glusterfsd: Failed to set keep-alive: Operation not supported
==> etc-glusterfs-glusterd.vol.log <==
[2013-01-11 16:39:56.267030] I
[glusterd-handshake.c:397:glusterd_set_clnt_mgmt_program] 0-: Using Program
glusterd mgmt, Num (1238433), Version (2)
[2013-01-11 16:39:56.267053] I
[glusterd-handshake.c:403:glusterd_set_clnt_mgmt_program] 0-: Using Program
Peer mgmt, Num (1238437), Version (2)
==> nfs.log.1 <==
[2013-01-11 16:39:58.148908] W [nfs.c:735:nfs_init_state] 1-nfs:
/sbin/rpc.statd not found. Disabling NLM
==> etc-glusterfs-glusterd.vol.log <==
[2013-01-11 16:39:58.149702] I
[glusterd-handler.c:1486:glusterd_handle_incoming_friend_req] 0-glusterd:
Received probe from uuid: 184a81f4-ff0f-48d6-adb8-798b98957b1a
[2013-01-11 16:39:58.149818] E
[glusterd-utils.c:1926:glusterd_compare_friend_volume] 0-: Cksums of volume
puppet-bucket differ. local cksum = 1273524870, remote cksum = 1932840611
[2013-01-11 16:39:58.149858] I
[glusterd-handler.c:2395:glusterd_xfer_friend_add_resp] 0-glusterd:
Responded to bastille.mdc (0), ret: 0
==> nfs.log.1 <==
[2013-01-11 16:39:58.179450] E [socket.c:333:__socket_server_bind]
1-socket.nfs-server: binding to failed: Address already in use
[2013-01-11 16:39:58.179512] E [socket.c:336:__socket_server_bind]
1-socket.nfs-server: Port is already in use
[2013-01-11 16:39:58.179535] W [rpcsvc.c:1363:rpcsvc_transport_create]
1-rpc-service: listening on transport failed
[2013-01-11 16:39:58.179663] E
[rpcsvc.c:1135:rpcsvc_program_register_portmap] 1-rpc-service: Could not
register with portmap
[2013-01-11 16:39:58.179710] E [socket.c:333:__socket_server_bind]
1-socket.nfs-server: binding to failed: Address already in use
[2013-01-11 16:39:58.179727] E [socket.c:336:__socket_server_bind]
1-socket.nfs-server: Port is already in use
[2013-01-11 16:39:58.179743] W [rpcsvc.c:1363:rpcsvc_transport_create]
1-rpc-service: listening on transport failed
[2013-01-11 16:39:58.179815] E
[rpcsvc.c:1135:rpcsvc_program_register_portmap] 1-rpc-service: Could not
register with portmap
[2013-01-11 16:39:58.180193] E [socket.c:333:__socket_server_bind]
1-socket.nfs-server: binding to failed: Address already in use
[2013-01-11 16:39:58.180214] E [socket.c:336:__socket_server_bind]
1-socket.nfs-server: Port is already in use
[2013-01-11 16:39:58.180230] W [rpcsvc.c:1363:rpcsvc_transport_create]
1-rpc-service: listening on transport failed
[2013-01-11 16:39:58.180300] E
[rpcsvc.c:1135:rpcsvc_program_register_portmap] 1-rpc-service: Could not
register with portmap
[2013-01-11 16:39:58.180319] I [nfs.c:821:init] 1-nfs: NFS service started
[2013-01-11 16:39:58.186245] W [graph.c:316:_log_if_unknown_option]
1-nfs-server: option 'rpc-auth.auth-glusterfs' is not recognized
[2013-01-11 16:39:58.186346] W [graph.c:316:_log_if_unknown_option]
1-nfs-server: option 'rpc-auth-allow-insecure' is not recognized
[2013-01-11 16:39:58.186366] W [graph.c:316:_log_if_unknown_option]
1-nfs-server: option 'transport-type' is not recognized
[2013-01-11 16:39:58.186400] I [client.c:2142:notify]
1-puppet-ssl-client-0: parent translators are ready, attempting connect on
transport
[2013-01-11 16:39:58.187286] I [client.c:2142:notify]
1-puppet-ssl-client-1: parent translators are ready, attempting connect on
transport
[2013-01-11 16:39:58.188173] I [client.c:2142:notify]
1-puppet-dist-client-0: parent translators are ready, attempting connect on
transport
[2013-01-11 16:39:58.189031] I [client.c:2142:notify]
1-puppet-dist-client-1: parent translators are ready, attempting connect on
transport
[2013-01-11 16:39:58.189703] I [client.c:2142:notify]
1-puppet-bucket-client-0: parent translators are ready, attempting connect
on transport
[2013-01-11 16:39:58.190559] I [client.c:2142:notify]
1-puppet-bucket-client-1: parent translators are ready, attempting connect
on transport
Given volfile:
+------------------------------------------------------------------------------+
1: volume puppet-bucket-client-0
2: type protocol/client
3: option remote-host sandy.mdc
4: option remote-subvolume /opt/gluster-data/snake-puppet/bucket
5: option transport-type tcp
6: end-volume
7:
8: volume puppet-bucket-client-1
9: type protocol/client
10: option remote-host irene.mdc
11: option remote-subvolume /opt/gluster-data/puppet/bucket
12: option transport-type tcp
13: end-volume
14:
15: volume puppet-bucket-replicate-0
16: type cluster/replicate
17: subvolumes puppet-bucket-client-0 puppet-bucket-client-1
18: end-volume
19:
20: volume puppet-bucket
21: type debug/io-stats
22: option latency-measurement off
23: option count-fop-hits off
24: subvolumes puppet-bucket-replicate-0
25: end-volume
26:
27: volume puppet-dist-client-0
28: type protocol/client
29: option remote-host sandy.mdc
30: option remote-subvolume /opt/gluster-data/snake-puppet/dist
31: option transport-type tcp
32: end-volume
33:
34: volume puppet-dist-client-1
35: type protocol/client
36: option remote-host irene.mdc
37: option remote-subvolume /opt/gluster-data/puppet/dist
38: option transport-type tcp
39: end-volume
40:
41: volume puppet-dist-replicate-0
42: type cluster/replicate
43: option data-self-heal-algorithm full
44: subvolumes puppet-dist-client-0 puppet-dist-client-1
45: end-volume
46:
47: volume puppet-dist
48: type debug/io-stats
49: option latency-measurement off
50: option count-fop-hits off
51: subvolumes puppet-dist-replicate-0
52: end-volume
53:
54: volume puppet-ssl-client-0
55: type protocol/client
56: option remote-host sandy.mdc
57: option remote-subvolume /opt/gluster-data/snake-puppet/ssl
58: option transport-type tcp
59: end-volume
60:
61: volume puppet-ssl-client-1
62: type protocol/client
63: option remote-host irene.mdc
64: option remote-subvolume /opt/gluster-data/puppet/ssl
65: option transport-type tcp
66: end-volume
67:
68: volume puppet-ssl-replicate-0
69: type cluster/replicate
70: option metadata-change-log on
71: option data-self-heal-algorithm full
72: subvolumes puppet-ssl-client-0 puppet-ssl-client-1
73: end-volume
74:
75: volume puppet-ssl
76: type debug/io-stats
77: option latency-measurement off
78: option count-fop-hits off
79: subvolumes puppet-ssl-replicate-0
80: end-volume
81:
82: volume nfs-server
83: type nfs/server
84: option nfs.dynamic-volumes on
85: option nfs.nlm on
86: option rpc-auth.addr.puppet-ssl.allow *
87: option nfs3.puppet-ssl.volume-id
bb2ffdd5-f00c-4016-ab07-301a6ede3042
88: option rpc-auth.addr.puppet-dist.allow *
89: option nfs3.puppet-dist.volume-id
376220d6-dcdd-4f3f-9809-397046a78f5a
90: option rpc-auth.addr.puppet-bucket.allow *
91: option nfs3.puppet-bucket.volume-id
3a7e146c-7c37-41ea-baa5-5262c79b1232
92: subvolumes puppet-ssl puppet-dist puppet-bucket
93: end-volume
+------------------------------------------------------------------------------+
[2013-01-11 16:39:58.191727] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-ssl-client-1: changing port to 24010 (from 0)
[2013-01-11 16:39:58.191806] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-dist-client-1: changing port to 24012 (from 0)
[2013-01-11 16:39:58.191844] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-bucket-client-1: changing port to 24014 (from 0)
[2013-01-11 16:39:58.191881] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-ssl-client-0: changing port to 24012 (from 0)
[2013-01-11 16:39:58.191974] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-dist-client-0: changing port to 24010 (from 0)
[2013-01-11 16:39:58.192024] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-bucket-client-0: changing port to 24014 (from 0)
==> glustershd.log.1 <==
[2013-01-11 16:39:58.381647] I [graph.c:241:gf_add_cmdline_options]
0-puppet-ssl-replicate-0: adding option 'node-uuid' for volume
'puppet-ssl-replicate-0' with value 'eece061b-1cd0-4f30-ad17-61809297aba9'
[2013-01-11 16:39:58.381673] I [graph.c:241:gf_add_cmdline_options]
0-puppet-dist-replicate-0: adding option 'node-uuid' for volume
'puppet-dist-replicate-0' with value 'eece061b-1cd0-4f30-ad17-61809297aba9'
[2013-01-11 16:39:58.381686] I [graph.c:241:gf_add_cmdline_options]
0-puppet-bucket-replicate-0: adding option 'node-uuid' for volume
'puppet-bucket-replicate-0' with value
'eece061b-1cd0-4f30-ad17-61809297aba9'
[2013-01-11 16:39:58.390396] I [client.c:2142:notify]
1-puppet-ssl-client-0: parent translators are ready, attempting connect on
transport
[2013-01-11 16:39:58.391487] I [client.c:2142:notify]
1-puppet-ssl-client-1: parent translators are ready, attempting connect on
transport
[2013-01-11 16:39:58.392209] I [client.c:2142:notify]
1-puppet-dist-client-0: parent translators are ready, attempting connect on
transport
[2013-01-11 16:39:58.392995] I [client.c:2142:notify]
1-puppet-dist-client-1: parent translators are ready, attempting connect on
transport
[2013-01-11 16:39:58.393804] I [client.c:2142:notify]
1-puppet-bucket-client-0: parent translators are ready, attempting connect
on transport
[2013-01-11 16:39:58.394598] I [client.c:2142:notify]
1-puppet-bucket-client-1: parent translators are ready, attempting connect
on transport
Given volfile:
+------------------------------------------------------------------------------+
1: volume puppet-bucket-client-0
2: type protocol/client
3: option remote-host sandy.mdc
4: option remote-subvolume /opt/gluster-data/snake-puppet/bucket
5: option transport-type tcp
6: end-volume
7:
8: volume puppet-bucket-client-1
9: type protocol/client
10: option remote-host irene.mdc
11: option remote-subvolume /opt/gluster-data/puppet/bucket
12: option transport-type tcp
13: end-volume
14:
15: volume puppet-bucket-replicate-0
16: type cluster/replicate
17: option background-self-heal-count 0
18: option metadata-self-heal on
19: option data-self-heal on
20: option entry-self-heal on
21: option self-heal-daemon on
22: option iam-self-heal-daemon yes
23: subvolumes puppet-bucket-client-0 puppet-bucket-client-1
24: end-volume
25:
26: volume puppet-dist-client-0
27: type protocol/client
28: option remote-host sandy.mdc
29: option remote-subvolume /opt/gluster-data/snake-puppet/dist
30: option transport-type tcp
31: end-volume
32:
33: volume puppet-dist-client-1
34: type protocol/client
35: option remote-host irene.mdc
36: option remote-subvolume /opt/gluster-data/puppet/dist
37: option transport-type tcp
38: end-volume
39:
40: volume puppet-dist-replicate-0
41: type cluster/replicate
42: option background-self-heal-count 0
43: option metadata-self-heal on
44: option data-self-heal on
45: option entry-self-heal on
46: option self-heal-daemon on
47: option data-self-heal-algorithm full
48: option iam-self-heal-daemon yes
49: subvolumes puppet-dist-client-0 puppet-dist-client-1
50: end-volume
51:
52: volume puppet-ssl-client-0
53: type protocol/client
54: option remote-host sandy.mdc
55: option remote-subvolume /opt/gluster-data/snake-puppet/ssl
56: option transport-type tcp
57: end-volume
58:
59: volume puppet-ssl-client-1
60: type protocol/client
61: option remote-host irene.mdc
62: option remote-subvolume /opt/gluster-data/puppet/ssl
63: option transport-type tcp
64: end-volume
65:
66: volume puppet-ssl-replicate-0
67: type cluster/replicate
68: option background-self-heal-count 0
69: option metadata-self-heal on
70: option data-self-heal on
71: option entry-self-heal on
72: option self-heal-daemon on
73: option metadata-change-log on
74: option data-self-heal-algorithm full
75: option iam-self-heal-daemon yes
76: subvolumes puppet-ssl-client-0 puppet-ssl-client-1
77: end-volume
78:
79: volume glustershd
80: type debug/io-stats
81: subvolumes puppet-ssl-replicate-0 puppet-dist-replicate-0
puppet-bucket-replicate-0
82: end-volume
+------------------------------------------------------------------------------+
[2013-01-11 16:39:58.395877] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-ssl-client-1: changing port to 24010 (from 0)
[2013-01-11 16:39:58.395978] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-bucket-client-0: changing port to 24014 (from 0)
[2013-01-11 16:39:58.396048] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-dist-client-1: changing port to 24012 (from 0)
[2013-01-11 16:39:58.396106] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-bucket-client-1: changing port to 24014 (from 0)
[2013-01-11 16:39:58.396161] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-ssl-client-0: changing port to 24012 (from 0)
[2013-01-11 16:39:58.396223] I [rpc-clnt.c:1657:rpc_clnt_reconfig]
1-puppet-dist-client-0: changing port to 24010 (from 0)
==> nfs.log.1 <==
[2013-01-11 16:40:02.148931] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-ssl-client-1: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.149212] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-ssl-client-1:
Connected to 10.136.200.16:24010, attached to remote volume
'/opt/gluster-data/puppet/ssl'.
[2013-01-11 16:40:02.149238] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-ssl-client-1:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.149289] I [afr-common.c:3628:afr_notify]
1-puppet-ssl-replicate-0: Subvolume 'puppet-ssl-client-1' came back up;
going online.
[2013-01-11 16:40:02.149382] I
[client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-ssl-client-1:
Server lk version = 1
[2013-01-11 16:40:02.149711] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-dist-client-1: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.149931] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-dist-client-1:
Connected to 10.136.200.16:24012, attached to remote volume
'/opt/gluster-data/puppet/dist'.
[2013-01-11 16:40:02.149951] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-dist-client-1:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.149995] I [afr-common.c:3628:afr_notify]
1-puppet-dist-replicate-0: Subvolume 'puppet-dist-client-1' came back up;
going online.
[2013-01-11 16:40:02.150086] I
[client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-dist-client-1:
Server lk version = 1
[2013-01-11 16:40:02.150727] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-bucket-client-1: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.151013] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-bucket-client-1:
Connected to 10.136.200.16:24014, attached to remote volume
'/opt/gluster-data/puppet/bucket'.
[2013-01-11 16:40:02.151042] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-bucket-client-1:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.151091] I [afr-common.c:3628:afr_notify]
1-puppet-bucket-replicate-0: Subvolume 'puppet-bucket-client-1' came back
up; going online.
[2013-01-11 16:40:02.151187] I
[client-handshake.c:453:client_set_lk_version_cbk]
1-puppet-bucket-client-1: Server lk version = 1
[2013-01-11 16:40:02.151623] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-ssl-client-0: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.151924] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-ssl-client-0:
Connected to 10.136.200.27:24012, attached to remote volume
'/opt/gluster-data/snake-puppet/ssl'.
[2013-01-11 16:40:02.151950] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-ssl-client-0:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.152166] I
[client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-ssl-client-0:
Server lk version = 1
[2013-01-11 16:40:02.152472] I
[afr-common.c:1965:afr_set_root_inode_on_first_lookup]
1-puppet-ssl-replicate-0: added root inode
[2013-01-11 16:40:02.152566] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-dist-client-0: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.152807] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-dist-client-0:
Connected to 10.136.200.27:24010, attached to remote volume
'/opt/gluster-data/snake-puppet/dist'.
[2013-01-11 16:40:02.152827] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-dist-client-0:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.152991] I
[client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-dist-client-0:
Server lk version = 1
[2013-01-11 16:40:02.153187] I
[afr-common.c:1965:afr_set_root_inode_on_first_lookup]
1-puppet-dist-replicate-0: added root inode
[2013-01-11 16:40:02.153403] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-bucket-client-0: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.153644] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-bucket-client-0:
Connected to 10.136.200.27:24014, attached to remote volume
'/opt/gluster-data/snake-puppet/bucket'.
[2013-01-11 16:40:02.153665] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-bucket-client-0:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.153797] I
[client-handshake.c:453:client_set_lk_version_cbk]
1-puppet-bucket-client-0: Server lk version = 1
[2013-01-11 16:40:02.154054] I
[afr-common.c:1965:afr_set_root_inode_on_first_lookup]
1-puppet-bucket-replicate-0: added root inode
==> glustershd.log.1 <==
[2013-01-11 16:40:02.381825] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-ssl-client-1: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.382098] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-ssl-client-1:
Connected to 10.136.200.16:24010, attached to remote volume
'/opt/gluster-data/puppet/ssl'.
[2013-01-11 16:40:02.382119] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-ssl-client-1:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.382203] I [afr-common.c:3628:afr_notify]
1-puppet-ssl-replicate-0: Subvolume 'puppet-ssl-client-1' came back up;
going online.
[2013-01-11 16:40:02.382321] I
[client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-ssl-client-1:
Server lk version = 1
[2013-01-11 16:40:02.382889] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-bucket-client-0: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.383190] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-bucket-client-0:
Connected to 10.136.200.27:24014, attached to remote volume
'/opt/gluster-data/snake-puppet/bucket'.
[2013-01-11 16:40:02.383213] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-bucket-client-0:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.383284] I [afr-common.c:3628:afr_notify]
1-puppet-bucket-replicate-0: Subvolume 'puppet-bucket-client-0' came back
up; going online.
[2013-01-11 16:40:02.384825] I
[client-handshake.c:453:client_set_lk_version_cbk]
1-puppet-bucket-client-0: Server lk version = 1
[2013-01-11 16:40:02.384999] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-dist-client-1: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.385614] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-dist-client-1:
Connected to 10.136.200.16:24012, attached to remote volume
'/opt/gluster-data/puppet/dist'.
[2013-01-11 16:40:02.385646] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-dist-client-1:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.385725] I [afr-common.c:3628:afr_notify]
1-puppet-dist-replicate-0: Subvolume 'puppet-dist-client-1' came back up;
going online.
[2013-01-11 16:40:02.386268] I
[client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-dist-client-1:
Server lk version = 1
[2013-01-11 16:40:02.386381] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-bucket-client-1: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.386710] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-ssl-client-0: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.386817] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-bucket-client-1:
Connected to 10.136.200.16:24014, attached to remote volume
'/opt/gluster-data/puppet/bucket'.
[2013-01-11 16:40:02.386842] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-bucket-client-1:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.387051] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-ssl-client-0:
Connected to 10.136.200.27:24012, attached to remote volume
'/opt/gluster-data/snake-puppet/ssl'.
[2013-01-11 16:40:02.387087] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-ssl-client-0:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.387222] I
[client-handshake.c:453:client_set_lk_version_cbk]
1-puppet-bucket-client-1: Server lk version = 1
[2013-01-11 16:40:02.387345] I
[client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-ssl-client-0:
Server lk version = 1
[2013-01-11 16:40:02.387427] I
[client-handshake.c:1636:select_server_supported_programs]
1-puppet-dist-client-0: Using Program GlusterFS 3.3.1, Num (1298437),
Version (330)
[2013-01-11 16:40:02.388029] I
[client-handshake.c:1433:client_setvolume_cbk] 1-puppet-dist-client-0:
Connected to 10.136.200.27:24010, attached to remote volume
'/opt/gluster-data/snake-puppet/dist'.
[2013-01-11 16:40:02.388058] I
[client-handshake.c:1445:client_setvolume_cbk] 1-puppet-dist-client-0:
Server and Client lk-version numbers are not same, reopening the fds
[2013-01-11 16:40:02.389682] I
[client-handshake.c:453:client_set_lk_version_cbk] 1-puppet-dist-client-0:
Server lk version = 1
^C
--
Yang
Orange Key: 35745318S1
On Fri, Jan 11, 2013 at 11:00 AM, YANG ChengFu <youngseph at gmail.com> wrote:
> Hello Fu Yong Tao,
>
> thanks for your suggest, after I did your steps, I got the following:
>
> gluster> volume sync new-host
> please delete all the volumes before full sync
> gluster> peer status
> Number of Peers: 1
>
> Hostname: 10.136.200.27
> Uuid: 184a81f4-ff0f-48d6-adb8-798b98957b1a
> State: Accepted peer request (Connected)
>
> I still can not put the server in the truested pool !
>
> --
> Yang
> Orange Key: 35745318S1
>
>
> On Fri, Jan 11, 2013 at 5:22 AM, 符永涛 <yongtaofu at gmail.com> wrote:
>
>> Reinstall gluster server or upgrade is a dangerous task before it it's
>> better to backup /etc/glusterfs /var/lib/glusterd.
>>
>> /var/lib/glusterd/glusterd.info contains the uuid of current server
>> and /var/lib/glusterd/peers contain it's peers
>> make sure above two files are all correct
>>
>> If other servers status are fine then with only above configuration
>> files you can start current host and gluster volumes files will
>> automatically sync to current host.
>>
>> Always remember backup
>>
>> 2013/1/11, YANG ChengFu <youngseph at gmail.com>:
>> > Hello,
>> >
>> > I did an upgrade glusterfs from 3.0.5 to 3.3.1, before I did it, I have
>> > other two 3.3.1 hosts(new-host) ready and made a cluster.
>> >
>> > After I upgraded old hosts, I tried to add them to the cluster, I
>> > got State: Peer Rejected (Connected), for sure it could be about same
>> > volumes on the old hosts, but I have tried to stop glusterd, remove
>> > everything from the old host, such /etc/glusterd, /etc/glusterfs
>> > and /var/lib/glusterd/, and start glusterd, then I readded it to
>> cluster,
>> > the problem is still there.
>> >
>> > I also did 'volume sync', but I failed, because of the following error
>> > message
>> >
>> > gluster> volume sync new-hosts
>> > please delete all the volumes before full sync
>> >
>> > I can not do it, or I will lose all my data!
>> >
>> > The most funny thing I found, even if the peer status is rejected, but I
>> > can mount the volume from the old host.
>> >
>> > Any ideas !
>> >
>> > --
>> > Yang
>> > Orange Key: 35745318S1
>> >
>>
>>
>> --
>> 符永涛
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130111/d3e959ea/attachment.html>
More information about the Gluster-users
mailing list