[Gluster-users] gluster peer probe error (v3.6.2)

Andreas andreas.hollaus at ericsson.com
Mon Mar 23 13:14:30 UTC 2015


Hi,

# gluster peer detach 10.32.1.144
(No output here. Similar to the problem with 'gluster peer probe'.)
# gluster peer detach 10.32.1.144 force
peer detach: failed: Peer is already being detached from cluster.
Check peer status by running gluster peer status
# gluster peer status
Number of Peers: 1

Hostname: 10.32.1.144
Uuid: 82cdb873-28cc-4ed0-8cfe-2b6275770429
State: Probe Sent to Peer (Connected)

# ping 10.32.1.144
PING 10.32.1.144 (10.32.1.144): 56 data bytes
64 bytes from 10.32.1.144: seq=0 ttl=64 time=1.811 ms
64 bytes from 10.32.1.144: seq=1 ttl=64 time=1.834 ms
^C
--- 10.32.1.144 ping statistics ---
2 packets transmitted, 2 packets received, 0% packet loss
round-trip min/avg/max = 1.811/1.822/1.834 ms


As previously stated, this problem seems to be similar to what I experienced with
'gluster peer probe'. I can reboot the server, but the situation will be the same
(I've tried this many times).
Any ideas of which ports to investigate and how to do it to get the most reliable result?
Anything else that could cause this?



Regards
Andreas


On 03/23/15 11:10, Atin Mukherjee wrote:
>
> On 03/23/2015 03:28 PM, Andreas Hollaus wrote:
>> 2Hi,
>>
>> This network problem is persistent. However, I can ping the server so guess it
>> depends on the port no, right?
>> I tried to telnet to port 24007, but I was not sure how to interpret the result as I
>> got no respons and no timeout (it just seemed to be waiting for something).
>> That's why I decided to install nmap, but according to that tool the port was
>> accessible. Are there any other ports that are vital to gluster peer probe?
>>
>> When you say 'deprobe', I guess you mean 'gluster peer detach'? That command shows
>> similar behaviour to gluster peer probe.
> Yes I meant peer detach. How about gluster peer detach force?


>>
>> Regards
>> Andreas
>>
>> On 03/23/15 05:34, Atin Mukherjee wrote:
>>> On 03/22/2015 07:11 PM, Andreas Hollaus wrote:
>>>> Hi,
>>>>
>>>> I hope that these are the logs that you requested.
>>>>
>>>> Logs from 10.32.0.48:
>>>> ------------------------------
>>>> # more /var/log/glusterfs/.cmd_log_history
>>>> [2015-03-19 13:52:03.277438]  : peer probe 10.32.1.144 : FAILED : Probe returned
>>>>  with unknown errno -1
>>>>
>>>> # more /var/log/glusterfs/etc-glusterfs-glusterd.vol.log
>>>> [2015-03-19 13:41:31.241768] I [MSGID: 100030] [glusterfsd.c:2018:main] 0-/usr/s
>>>> bin/glusterd: Started running /usr/sbin/glusterd version 3.6.2 (args: /usr/sbin/
>>>> glusterd -p /var/run/glusterd.pid)
>>>> [2015-03-19 13:41:31.245352] I [glusterd.c:1214:init] 0-management: Maximum allo
>>>> wed open file descriptors set to 65536
>>>> [2015-03-19 13:41:31.245432] I [glusterd.c:1259:init] 0-management: Using /var/l
>>>> ib/glusterd as working directory
>>>> [2015-03-19 13:41:31.247826] I [glusterd-store.c:2063:glusterd_restore_op_versio
>>>> n] 0-management: Detected new install. Setting op-version to maximum : 30600
>>>> [2015-03-19 13:41:31.247902] I [glusterd-store.c:3497:glusterd_store_retrieve_mi
>>>> ssed_snaps_list] 0-management: No missed snaps list.
>>>> Final graph:
>>>> +------------------------------------------------------------------------------+
>>>>   1: volume management
>>>>   2:     type mgmt/glusterd
>>>>   3:     option rpc-auth.auth-glusterfs on
>>>>   4:     option rpc-auth.auth-unix on
>>>>   5:     option rpc-auth.auth-null on
>>>>   6:     option transport.socket.listen-backlog 128
>>>>   7:     option ping-timeout 30
>>>>   8:     option transport.socket.read-fail-log off
>>>>   9:     option transport.socket.keepalive-interval 2
>>>>  10:     option transport.socket.keepalive-time 10
>>>>  11:     option transport-type socket
>>>>  12:     option working-directory /var/lib/glusterd
>>>>  13: end-volume
>>>>  14: 
>>>> +------------------------------------------------------------------------------+
>>>> [2015-03-19 13:42:02.258403] I [glusterd-handler.c:1015:__glusterd_handle_cli_pr
>>>> obe] 0-glusterd: Received CLI probe req 10.32.1.144 24007
>>>> [2015-03-19 13:42:02.259456] I [glusterd-handler.c:3165:glusterd_probe_begin] 0-
>>>> glusterd: Unable to find peerinfo for host: 10.32.1.144 (24007)
>>>> [2015-03-19 13:42:02.259664] I [rpc-clnt.c:969:rpc_clnt_connection_init] 0-manag
>>>> ement: setting frame-timeout to 600
>>>> [2015-03-19 13:42:02.260488] I [glusterd-handler.c:3098:glusterd_friend_add] 0-m
>>>> anagement: connect returned 0
>>>> [2015-03-19 13:42:02.270316] I [glusterd.c:176:glusterd_uuid_generate_save] 0-ma
>>>> nagement: generated UUID: 4441e237-89d6-4cdf-a212-f17ecb953b58
>>>> [2015-03-19 13:42:02.273427] I [glusterd-rpc-ops.c:244:__glusterd_probe_cbk] 0-m
>>>> anagement: Received probe resp from uuid: 82cdb873-28cc-4ed0-8cfe-2b6275770429,
>>>> host: 10.32.1.144
>>>> [2015-03-19 13:42:02.273681] I [glusterd-rpc-ops.c:386:__glusterd_probe_cbk] 0-g
>>>> lusterd: Received resp to probe req
>>>> [2015-03-19 13:42:02.278863] I [glusterd-handshake.c:1119:__glusterd_mgmt_hndsk_
>>>> versions_ack] 0-management: using the op-version 30600
>>>> [2015-03-19 13:52:03.277422] E [rpc-clnt.c:201:call_bail] 0-management: bailing
>>>> out frame type(Peer mgmt) op(--(2)) xid = 0x6 sent = 2015-03-19 13:42:02.273482.
>>>>  timeout = 600 for 10.32.1.144:24007
>>> Here is the issue, there was some problem in the network at the time
>>> when peer probe was issued. This is why the call bail is seen. Could you
>>> try to deprobe and then probe it back again?
>>>> [2015-03-19 13:52:03.277453] I [socket.c:3366:socket_submit_reply] 0-socket.mana
>>>> gement: not connected (priv->connected = 255)
>>>> [2015-03-19 13:52:03.277468] E [rpcsvc.c:1247:rpcsvc_submit_generic] 0-rpc-servi
>>>> ce: failed to submit message (XID: 0x1, Program: GlusterD svc cli, ProgVers: 2,
>>>> Proc: 1) to rpc-transport (socket.management)
>>>> [2015-03-19 13:52:03.277483] E [glusterd-utils.c:387:glusterd_submit_reply] 0-:
>>>> Reply submission failed
>>>>
>>>>
>>>>
>>>> Logs from 10.32.1.144:
>>>> ---------------------------------
>>>> # more ./.cmd_log_history
>>>>
>>>> # more ./etc-glusterfs-glusterd.vol.log
>>>> [1970-01-01 00:00:53.225739] I [MSGID: 100030] [glusterfsd.c:2018:main] 0-/usr/s
>>>> bin/glusterd: Started running /usr/sbin/glusterd version 3.6.2 (args: /usr/sbin/
>>>> glusterd -p /var/run/glusterd.pid)
>>>> [1970-01-01 00:00:53.229222] I [glusterd.c:1214:init] 0-management: Maximum allo
>>>> wed open file descriptors set to 65536
>>>> [1970-01-01 00:00:53.229301] I [glusterd.c:1259:init] 0-management: Using /var/l
>>>> ib/glusterd as working directory
>>>> [1970-01-01 00:00:53.231653] I [glusterd-store.c:2063:glusterd_restore_op_versio
>>>> n] 0-management: Detected new install. Setting op-version to maximum : 30600
>>>> [1970-01-01 00:00:53.231730] I [glusterd-store.c:3497:glusterd_store_retrieve_mi
>>>> ssed_snaps_list] 0-management: No missed snaps list.
>>>> Final graph:
>>>> +------------------------------------------------------------------------------+
>>>>   1: volume management
>>>>   2:     type mgmt/glusterd
>>>>   3:     option rpc-auth.auth-glusterfs on
>>>>   4:     option rpc-auth.auth-unix on
>>>>   5:     option rpc-auth.auth-null on
>>>>   6:     option transport.socket.listen-backlog 128
>>>>   7:     option ping-timeout 30
>>>>   8:     option transport.socket.read-fail-log off
>>>>   9:     option transport.socket.keepalive-interval 2
>>>>  10:     option transport.socket.keepalive-time 10
>>>>  11:     option transport-type socket
>>>>  12:     option working-directory /var/lib/glusterd
>>>>  13: end-volume
>>>>  14: 
>>>> +------------------------------------------------------------------------------+
>>>> [1970-01-01 00:01:24.417689] I [glusterd-handshake.c:1119:__glusterd_mgmt_hndsk_
>>>> versions_ack] 0-management: using the op-version 30600
>>>> [1970-01-01 00:01:24.417736] I [glusterd.c:176:glusterd_uuid_generate_save] 0-ma
>>>> nagement: generated UUID: 82cdb873-28cc-4ed0-8cfe-2b6275770429
>>>> [1970-01-01 00:01:24.420067] I [glusterd-handler.c:2523:__glusterd_handle_probe_
>>>> query] 0-glusterd: Received probe from uuid: 4441e237-89d6-4cdf-a212-f17ecb953b5
>>>> 8
>>>> [1970-01-01 00:01:24.420158] I [glusterd-handler.c:2551:__glusterd_handle_probe_
>>>> query] 0-glusterd: Unable to find peerinfo for host: 10.32.0.48 (24007)
>>>> [1970-01-01 00:01:24.420379] I [rpc-clnt.c:969:rpc_clnt_connection_init] 0-manag
>>>> ement: setting frame-timeout to 600
>>>> [1970-01-01 00:01:24.421140] I [glusterd-handler.c:3098:glusterd_friend_add] 0-m
>>>> anagement: connect returned 0
>>>> [1970-01-01 00:01:24.421167] I [glusterd-handler.c:2575:__glusterd_handle_probe_
>>>> query] 0-glusterd: Responded to 10.32.0.48, op_ret: 0, op_errno: 0, ret: 0
>>>> [1970-01-01 00:01:24.422991] I [glusterd-handler.c:2216:__glusterd_handle_incomi
>>>> ng_friend_req] 0-glusterd: Received probe from uuid: 4441e237-89d6-4cdf-a212-f17
>>>> ecb953b58
>>>> [1970-01-01 00:01:24.423024] E [glusterd-utils.c:5760:glusterd_compare_friend_da
>>>> ta] 0-management: Importing global options failed
>>>> [1970-01-01 00:01:24.423036] E [glusterd-sm.c:1078:glusterd_friend_sm] 0-gluster
>>>> d: handler returned: -2
>>>>  
>>>>
>>>> Regards
>>>> Andreas
>>>>
>>>>
>>>> On 03/22/15 07:33, Atin Mukherjee wrote:
>>>>> On 03/22/2015 12:09 AM, Andreas Hollaus wrote:
>>>>>> Hi,
>>>>>>
>>>>>> I get a strange result when I execute 'gluster peer probe'. The command hangs and
>>>>>> seems to timeout without any message (I can ping the address):
>>>>>> # gluster peer probe 10.32.1.144
>>>>>> # echo $?
>>>>>> 146
>>>>> Could you provide the glusterd log and .cmd_log_history for all the
>>>>> nodes in the cluster?
>>>>>> The status looks promising, but there's a differences between this output and what
>>>>>> you normally get from a successful call:
>>>>>> # gluster peer status
>>>>>> Number of Peers: 1
>>>>>>
>>>>>> Hostname: 10.32.1.144
>>>>>> Uuid: 0b008d3e-c51b-4243-ad19-c79c869ba9f2
>>>>>> State: Probe Sent to Peer (Connected)
>>>>>>
>>>>>> (instead of 'State: Peer in Cluster (Connected)')
>>>>>>
>>>>>> Running the command again will tell you that it is connected:
>>>>>>
>>>>>> # gluster peer probe 10.32.1.144
>>>>>> peer probe: success. Host 10.32.1.144 port 24007 already in peer list
>>>>> This means that this peer was added locally but peer handshake was not
>>>>> completed for previous peer probe transaction. I would be interested to
>>>>> see the logs and then can comment on what went wrong.
>>>>>> But when you try to add a brick from that server it fails:
>>>>>>
>>>>>> # gluster volume add-brick c_test replica 2 10.32.1.144:/opt/lvmdir/c2 force
>>>>>> volume add-brick: failed: Host 10.32.1.144 is not in 'Peer in Cluster' state
>>>>>>
>>>>>> The volume was previously created using the following commands:
>>>>>> # gluster volume create c_test 10.32.0.48:/opt/lvmdir/c2 force
>>>>>> volume create: c_test: success: please start the volume to access data
>>>>>> # gluster volume start c_test
>>>>>> volume start: c_test: success
>>>>>>
>>>>>> What could be the reason for this problem?
>>>>>>
>>>>>>
>>>>>> Regards
>>>>>> Andreas
>>>>>>
>>>>>> _______________________________________________
>>>>>> Gluster-users mailing list
>>>>>> Gluster-users at gluster.org
>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>>>
>>>>
>>



More information about the Gluster-users mailing list