[Gluster-devel] Another peer probe problem after quota enable

박성식 mulgo79 at gmail.com
Thu Dec 3 07:01:39 UTC 2015


Thanks to Vijay answer.

Volume 'info' and 'cksum' file copy already know.

I wanted to know if there is another solution for the 'quota-version'.

It's now planning to create test automation scripts for basic full-features.

I'll post another when issue is detected.

Thank you once again for answers.

Sungsik, Park.


------------------------------------------------------------------------------------------

Sung Sik, Park [박성식, 朴宬植]

Email: mulgo79 at gmail.com <sypark at wisetodd.com>

KR: +82 10-4518-2647

------------------------------------------------------------------------------------------

On Thu, Dec 3, 2015 at 2:56 PM, Vijaikumar Mallikarjuna <vmallika at redhat.com
> wrote:

> Hi Sung Sik Park,
>
> Please find my comments inline.
>
>
>
> On Thu, Dec 3, 2015 at 7:19 AM, 박성식 <mulgo79 at gmail.com> wrote:
>
>> Hi all
>>
>> Gluster-3.7.6 in 'Peer Rejected' problem occurs in the following test
>> case.(it doesn't occur if don't enable the volume quota)
>>
>> The problem is caused from the commit point
>> '3d3176958b7da48dbacb1a5a0fedf26322a38297'.
>>
>> Can you help with the following problems?
>>
>> # Test enviroment
>>
>> 1. Ubuntu 14.04 Server
>>    (1) c00002-dn00001: 192.16.63.21
>>    (2) c00002-dn00002: 192.16.63.22
>>    (3) c00002-dn00003: 192.16.63.23
>> 2. Glusterfs-3.7.6
>>
>> # Test procedure
>>
>> 1. Gluster installation
>>
>> $ git clone git://git.gluster.com/glusterfs.git git.gluster.com
>> $ cd git.gluster.com/
>> $ git checkout v3.7.6
>> $ ./autogen.sh ; ./configure ; make ; make install ; ldconfig
>>
>> 2. Glusterd startup
>>
>> $ /etc/init.d/glusterd start
>>
>> 3. Gluster peer: 2 Server peer
>>
>> [c00002-dn00001] $ gluster peer probe 192.16.63.22
>> [c00002-dn00001] $ gluster volume create tvol0001
>> c00002-dn00001:/brick/tvol0001 c00002-dn00002:/brick/tvol0001 force
>> [c00002-dn00001] $ gluster volume start tvol0001
>> [c00002-dn00001] $ gluster volume quota tvol0001 enable
>>
>> 4. Another server peer
>>
>> [c00002-dn00001] $ gluster peer probe 192.16.63.23
>> peer probe: success.
>>
>> [c00002-dn00001] $ gluster peer status
>> Uuid: dd749b9b-d4e1-4b94-8f36-c03064bc5a57
>> State: Peer Rejected (Connected)
>>
>> # c00002-dn00001 log
>> [2015-12-03 00:37:52.464600] I [MSGID: 106487]
>> [glusterd-handler.c:1422:__glusterd_handle_cli_list_friends] 0-glusterd:
>> Received cli list req
>> [2015-12-03 00:38:51.035018] I [MSGID: 106487]
>> [glusterd-handler.c:1301:__glusterd_handle_cli_deprobe] 0-glusterd:
>> Received CLI deprobe req
>> [2015-12-03 00:38:51.036338] I [MSGID: 106493]
>> [glusterd-rpc-ops.c:599:__glusterd_friend_remove_cbk] 0-glusterd: Received
>> ACC from uuid: dd749b9b-d4e1-4b94-8f36-c03064bc5a57, host: c00002-dn00001,
>> port: 0
>> [2015-12-03 00:38:51.041859] I [MSGID: 106493]
>> [glusterd-rpc-ops.c:695:__glusterd_friend_update_cbk] 0-management:
>> Received ACC from uuid: cb22d272-d563-4bda-802b-c93798130108
>> [2015-12-03 00:38:56.441189] I [MSGID: 106487]
>> [glusterd-handler.c:1189:__glusterd_handle_cli_probe] 0-glusterd: Received
>> CLI probe req 192.16.63.23 24007
>> [2015-12-03 00:38:56.441828] I [MSGID: 106129]
>> [glusterd-handler.c:3611:glusterd_probe_begin] 0-glusterd: Unable to find
>> peerinfo for host: 192.16.63.23 (24007)
>> [2015-12-03 00:38:56.445471] I [rpc-clnt.c:984:rpc_clnt_connection_init]
>> 0-management: setting frame-timeout to 600
>> [2015-12-03 00:38:56.445954] W [socket.c:869:__socket_keepalive]
>> 0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 23, Invalid
>> argument
>> [2015-12-03 00:38:56.445977] E [socket.c:2965:socket_connect]
>> 0-management: Failed to set keep-alive: Invalid argument
>> [2015-12-03 00:38:56.446232] I [MSGID: 106498]
>> [glusterd-handler.c:3539:glusterd_friend_add] 0-management: connect
>> returned 0
>> [2015-12-03 00:38:56.457933] I [MSGID: 106511]
>> [glusterd-rpc-ops.c:257:__glusterd_probe_cbk] 0-management: Received probe
>> resp from uuid: dd749b9b-d4e1-4b94-8f36-c03064bc5a57, host: 192.16.63.23
>> [2015-12-03 00:38:56.457992] I [MSGID: 106511]
>> [glusterd-rpc-ops.c:417:__glusterd_probe_cbk] 0-glusterd: Received resp to
>> probe req
>> [2015-12-03 00:38:56.519048] I [MSGID: 106493]
>> [glusterd-rpc-ops.c:481:__glusterd_friend_add_cbk] 0-glusterd: Received ACC
>> from uuid: dd749b9b-d4e1-4b94-8f36-c03064bc5a57, host: 192.16.63.23, port: 0
>> [2015-12-03 00:38:56.524556] I [MSGID: 106163]
>> [glusterd-handshake.c:1193:__glusterd_mgmt_hndsk_versions_ack]
>> 0-management: using the op-version 30707
>> [2015-12-03 00:38:56.527701] I [MSGID: 106490]
>> [glusterd-handler.c:2895:__glusterd_handle_probe_query] 0-glusterd:
>> Received probe from uuid: dd749b9b-d4e1-4b94-8f36-c03064bc5a57
>> [2015-12-03 00:38:56.528598] I [MSGID: 106493]
>> [glusterd-handler.c:2958:__glusterd_handle_probe_query] 0-glusterd:
>> Responded to c00002-dn00003, op_ret: 0, op_errno: 0, ret: 0
>> [2015-12-03 00:38:56.529154] I [MSGID: 106490]
>> [glusterd-handler.c:2550:__glusterd_handle_incoming_friend_req] 0-glusterd:
>> Received probe from uuid: dd749b9b-d4e1-4b94-8f36-c03064bc5a57
>> [2015-12-03 00:38:56.529743] E [MSGID: 106010]
>> [glusterd-utils.c:2727:glusterd_compare_friend_volume] 0-management:
>> Version of Cksums tvol0001 differ. local cksum = 870323923, remote cksum =
>> 870586066 on peer 192.16.63.23
>> [2015-12-03 00:38:56.529905] I [MSGID: 106493]
>> [glusterd-handler.c:3791:glusterd_xfer_friend_add_resp] 0-glusterd:
>> Responded to 192.16.63.23 (0), ret: 0
>> [2015-12-03 00:38:59.243087] I [MSGID: 106487]
>> [glusterd-handler.c:1422:__glusterd_handle_cli_list_friends] 0-glusterd:
>> Received cli list req
>>
>> # c00002-dn00003 log
>> [2015-12-03 00:38:45.556047] I [MSGID: 106491]
>> [glusterd-handler.c:2600:__glusterd_handle_incoming_unfriend_req]
>> 0-glusterd: Received unfriend from uuid:
>> 090e2d18-7a6a-4d11-89cd-7c583180cba3
>> [2015-12-03 00:38:45.556835] I [MSGID: 106493]
>> [glusterd-handler.c:3758:glusterd_xfer_friend_remove_resp] 0-glusterd:
>> Responded to c00002-dn00001 (0), ret: 0
>> [2015-12-03 00:38:45.556863] I [MSGID: 106132]
>> [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management:  already stopped
>> [2015-12-03 00:38:45.556873] I [MSGID: 106510]
>> [glusterd-sm.c:637:glusterd_peer_detach_cleanup] 0-management: Deleting
>> stale volume tvol0001
>> [2015-12-03 00:38:45.571032] W [socket.c:588:__socket_rwv] 0-nfs: readv
>> on /var/run/gluster/ce041cf16b3a2a739b2381df8b2eb141.socket failed (No data
>> available)
>> [2015-12-03 00:38:46.557936] I [MSGID: 106132]
>> [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: glustershd
>> already stopped
>> [2015-12-03 00:38:46.563409] W [socket.c:588:__socket_rwv] 0-quotad:
>> readv on /var/run/gluster/a3c5c6ab5ef0542c80d4dcee7b6f6b00.socket failed
>> (No data available)
>> [2015-12-03 00:38:47.559114] I [MSGID: 106132]
>> [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: bitd already
>> stopped
>> [2015-12-03 00:38:47.559217] I [MSGID: 106132]
>> [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: scrub already
>> stopped
>> [2015-12-03 00:38:47.565412] I [MSGID: 101053]
>> [mem-pool.c:616:mem_pool_destroy] 0-management: size=588 max=3 total=6
>> [2015-12-03 00:38:47.565589] I [MSGID: 101053]
>> [mem-pool.c:616:mem_pool_destroy] 0-management: size=124 max=3 total=6
>> [2015-12-03 00:38:50.969101] I [MSGID: 106163]
>> [glusterd-handshake.c:1193:__glusterd_mgmt_hndsk_versions_ack]
>> 0-management: using the op-version 30707
>> [2015-12-03 00:38:50.972274] I [MSGID: 106490]
>> [glusterd-handler.c:2895:__glusterd_handle_probe_query] 0-glusterd:
>> Received probe from uuid: 090e2d18-7a6a-4d11-89cd-7c583180cba3
>> [2015-12-03 00:38:50.973833] I [MSGID: 106129]
>> [glusterd-handler.c:2930:__glusterd_handle_probe_query] 0-glusterd: Unable
>> to find peerinfo for host: c00002-dn00001 (24007)
>> [2015-12-03 00:38:50.976661] I [rpc-clnt.c:984:rpc_clnt_connection_init]
>> 0-management: setting frame-timeout to 600
>> [2015-12-03 00:38:50.977655] W [socket.c:869:__socket_keepalive]
>> 0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 12, Invalid
>> argument
>> [2015-12-03 00:38:50.977679] E [socket.c:2965:socket_connect]
>> 0-management: Failed to set keep-alive: Invalid argument
>> [2015-12-03 00:38:50.978170] I [MSGID: 106498]
>> [glusterd-handler.c:3539:glusterd_friend_add] 0-management: connect
>> returned 0
>> [2015-12-03 00:38:50.978682] I [MSGID: 106493]
>> [glusterd-handler.c:2958:__glusterd_handle_probe_query] 0-glusterd:
>> Responded to c00002-dn00001, op_ret: 0, op_errno: 0, ret: 0
>> [2015-12-03 00:38:50.979553] I [MSGID: 106490]
>> [glusterd-handler.c:2550:__glusterd_handle_incoming_friend_req] 0-glusterd:
>> Received probe from uuid: 090e2d18-7a6a-4d11-89cd-7c583180cba3
>> [2015-12-03 00:38:50.987658] E [MSGID: 106187]
>> [glusterd-utils.c:4541:glusterd_brick_start] 0-management: Could not find
>> peer on which brick c00002-dn00002:/brick/tvol0001 resides
>> [2015-12-03 00:38:50.987678] E [MSGID: 106005]
>> [glusterd-op-sm.c:2184:glusterd_start_bricks] 0-management: Failed to start
>> c00002-dn00002:/brick/tvol0001 for tvol0001
>> [2015-12-03 00:38:51.017178] I [MSGID: 106132]
>> [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: nfs already
>> stopped
>> [2015-12-03 00:38:51.026769] W [socket.c:3009:socket_connect] 0-nfs:
>> Ignore failed connection attempt on
>> /var/run/gluster/ce041cf16b3a2a739b2381df8b2eb141.socket, (No such file or
>> directory)
>> [2015-12-03 00:38:51.026992] I [MSGID: 106132]
>> [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: glustershd
>> already stopped
>> [2015-12-03 00:38:51.032213] I [MSGID: 106132]
>> [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: quotad already
>> stopped
>> [2015-12-03 00:38:51.036689] W [socket.c:3009:socket_connect] 0-quotad:
>> Ignore failed connection attempt on
>> /var/run/gluster/a3c5c6ab5ef0542c80d4dcee7b6f6b00.socket, (No such file or
>> directory)
>> [2015-12-03 00:38:51.036778] I [MSGID: 106132]
>> [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: bitd already
>> stopped
>> [2015-12-03 00:38:51.036828] I [MSGID: 106132]
>> [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: scrub already
>> stopped
>> [2015-12-03 00:38:51.039228] I [MSGID: 106493]
>> [glusterd-handler.c:3791:glusterd_xfer_friend_add_resp] 0-glusterd:
>> Responded to c00002-dn00001 (0), ret: 0
>> [2015-12-03 00:38:51.044296] W [socket.c:588:__socket_rwv] 0-quotad:
>> readv on /var/run/gluster/a3c5c6ab5ef0542c80d4dcee7b6f6b00.socket failed
>> (Invalid argument)
>> [2015-12-03 00:38:51.044652] I [MSGID: 106006]
>> [glusterd-svc-mgmt.c:323:glusterd_svc_common_rpc_notify] 0-management:
>> quotad has disconnected from glusterd.
>> [2015-12-03 00:38:51.044868] W [socket.c:588:__socket_rwv] 0-nfs: readv
>> on /var/run/gluster/ce041cf16b3a2a739b2381df8b2eb141.socket failed (Invalid
>> argument)
>> [2015-12-03 00:38:51.044902] I [MSGID: 106006]
>> [glusterd-svc-mgmt.c:323:glusterd_svc_common_rpc_notify] 0-management: nfs
>> has disconnected from glusterd.
>> [2015-12-03 00:38:51.049332] I [MSGID: 106511]
>> [glusterd-rpc-ops.c:257:__glusterd_probe_cbk] 0-management: Received probe
>> resp from uuid: 090e2d18-7a6a-4d11-89cd-7c583180cba3, host: c00002-dn00001
>> [2015-12-03 00:38:51.049365] I [MSGID: 106511]
>> [glusterd-rpc-ops.c:417:__glusterd_probe_cbk] 0-glusterd: Received resp to
>> probe req
>> [2015-12-03 00:38:51.050941] I [MSGID: 106493]
>> [glusterd-rpc-ops.c:481:__glusterd_friend_add_cbk] 0-glusterd: Received RJT
>> from uuid: 090e2d18-7a6a-4d11-89cd-7c583180cba3, host: c00002-dn00001,
>> port: 0
>>
>> # c00002-dn00001 /var/lib/glusterd/vols/tvol0001/info
>> type=0
>> count=2
>> status=1
>> sub_count=0
>> stripe_count=1
>> replica_count=1
>> disperse_count=0
>> redundancy_count=0
>> version=3
>> transport-type=0
>> volume-id=cb9864b7-0bb4-439d-b60b-595d2e1e0cc3
>> username=eb29fba6-a51a-4e65-abbd-8ef03d7bb22d
>> password=c987bfd7-8da3-4379-abe1-7dcbcda7d49d
>> op-version=3
>> client-op-version=3
>> quota-version=1
>> parent_volname=N/A
>> restored_from_snap=00000000-0000-0000-0000-000000000000
>> snap-max-hard-limit=256
>> features.quota-deem-statfs=on
>> features.inode-quota=on
>> features.quota=on
>> performance.readdir-ahead=on
>> brick-0=c00002-dn00001:-brick-tvol0001
>> brick-1=c00002-dn00002:-brick-tvol0001
>>
>> # c00002-dn00003 /var/lib/glusterd/vols/tvol0001/info
>> type=0
>> count=2
>> status=1
>> sub_count=0
>> stripe_count=1
>> replica_count=1
>> disperse_count=0
>> redundancy_count=0
>> version=3
>> transport-type=0
>> volume-id=cb9864b7-0bb4-439d-b60b-595d2e1e0cc3
>> username=eb29fba6-a51a-4e65-abbd-8ef03d7bb22d
>> password=c987bfd7-8da3-4379-abe1-7dcbcda7d49d
>> op-version=3
>> client-op-version=3
>> quota-version=0
>>
>
> Somehow, parameter 'quota-version' is not updated here. I will try to
> re-create this problem in-house and fix this issue ASAP.
> For a workaround, you can try below steps
> 1) stop glusterd on 'c00002-dn00003'
> 2) copy volume 'info' and 'cksum' file from 'c00002-dn00001' to
> ''c00002-dn00003'
> 3) start glusterd
> 4) try peer probe
>
> Thanks,
> Vijay
>
>
> parent_volname=N/A
>> restored_from_snap=00000000-0000-0000-0000-000000000000
>> snap-max-hard-limit=256
>> features.quota-deem-statfs=on
>> features.inode-quota=on
>> features.quota=on
>> performance.readdir-ahead=on
>> brick-0=c00002-dn00001:-brick-tvol0001
>> brick-1=c00002-dn00002:-brick-tvol0001
>>
>>
>> ------------------------------------------------------------------------------------------
>>
>> Sung Sik, Park [박성식, 朴宬植]
>>
>> Email: mulgo79 at gmail.com <sypark at wisetodd.com>
>>
>>
>> ------------------------------------------------------------------------------------------
>>
>> _______________________________________________
>> Gluster-devel mailing list
>> Gluster-devel at gluster.org
>> http://www.gluster.org/mailman/listinfo/gluster-devel
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-devel/attachments/20151203/dc4f9550/attachment-0001.html>


More information about the Gluster-devel mailing list