[Gluster-users] GlusterFS peer probe hangs on local area network
    Rahul51 S 
    rahul51.s at tcs.com
       
    Fri Jun 21 15:13:54 UTC 2013
    
    
  
Hi All,
I am trying to "peer probe" on the node in the LAN, but it hangs for a 
while and  when the command is completed after sometime, it displays the 
Uuid as all zero's 
Below is the output of peer status on both the nodes
root at typhoon-base-unit0:/var/lib/glusterd> gluster peer status
Number of Peers: 1
Hostname: 172.24.132.1
Port: 24007
Uuid: 00000000-0000-0000-0000-000000000000
State: Establishing Connection (Connected)
root at typhoon-base-unit1:/var/lib/glusterd/peers> gluster peer status
Number of Peers: 1
Hostname: 172.24.132.0
Uuid: 00000000-0000-0000-0000-000000000000
State: Connected to Peer (Connected)
After this, when I try to create a replicated volume it fails with the 
error 
root at typhoon-base-unit0:/root> gluster volume create testvol replica 2 
172.24.132.0:/.krfs/_home 172.24.132.1:/.krfs/_home
volume create: testvol: failed: Failed to find host 172.24.132.1
Please note that this node is a ATCA blade which has multiple ethernet 
interfaces. The above failure is occurring when I try to peer probe on a 
ethernet interfaces which are connected in a Local Area network.(bond 0 on 
both the nodes )
There is one other ethernet interface(front 0) which is connected to the 
router for both the nodes. If I peer probe on the other node using this 
interface, then peer probe is successful.
Below is the output of the ifconfig command on node 0 
root at typhoon-base-unit0:/root> ifconfig
bond0     Link encap:Ethernet  HWaddr EC:9E:CD:07:DC:0A
          inet addr:172.24.132.0  Bcast:172.24.255.255  Mask:255.255.0.0
          inet6 addr: fe80::ee9e:cdff:fe07:dc0a/64 Scope:Link
          UP BROADCAST RUNNING MASTER MULTICAST  MTU:9000  Metric:1
          RX packets:290442 errors:0 dropped:0 overruns:0 frame:0
          TX packets:309604 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:37604741 (35.8 MiB)  TX bytes:33336286 (31.7 MiB)
front0    Link encap:Ethernet  HWaddr EC:9E:CD:07:DC:0E
          inet addr:172.17.23.117  Bcast:172.17.23.255  Mask:255.255.255.0
          inet6 addr: fe80::ee9e:cdff:fe07:dc0e/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:800920 errors:0 dropped:0 overruns:0 frame:0
          TX packets:333044 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:1000
          RX bytes:963382278 (918.7 MiB)  TX bytes:45300199 (43.2 MiB)
lo        Link encap:Local Loopback
          inet addr:127.0.0.1  Mask:255.0.0.0
          inet6 addr: ::1/128 Scope:Host
          UP LOOPBACK RUNNING  MTU:16436  Metric:1
          RX packets:2724222 errors:0 dropped:0 overruns:0 frame:0
          TX packets:2724222 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:2918838588 (2.7 GiB)  TX bytes:2918838588 (2.7 GiB)
 Below is the output of the ifconfig command on node 1 
root at typhoon-base-unit1:/root> ifconfig
bond0     Link encap:Ethernet  HWaddr EC:9E:CD:08:43:82
          inet addr:172.24.132.1  Bcast:0.0.0.0  Mask:255.255.0.0
          inet6 addr: fe80::ee9e:cdff:fe08:4382/64 Scope:Link
          UP BROADCAST RUNNING MASTER MULTICAST  MTU:9000  Metric:1
          RX packets:3236373 errors:0 dropped:0 overruns:0 frame:0
          TX packets:2955309 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:336930560 (321.3 MiB)  TX bytes:379249481 (361.6 MiB)
front0    Link encap:Ethernet  HWaddr EC:9E:CD:08:43:86
          inet addr:172.17.23.119  Bcast:172.17.23.255  Mask:255.255.255.0
          inet6 addr: fe80::ee9e:cdff:fe08:4386/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:2092076 errors:0 dropped:0 overruns:0 frame:0
          TX packets:426900 errors:780 dropped:0 overruns:0 carrier:780
          collisions:134320 txqueuelen:1000
          RX bytes:1263074540 (1.1 GiB)  TX bytes:49395506 (47.1 MiB)
lo        Link encap:Local Loopback
          inet addr:127.0.0.1  Mask:255.0.0.0
          inet6 addr: ::1/128 Scope:Host
          UP LOOPBACK RUNNING  MTU:16436  Metric:1
          RX packets:17223778 errors:0 dropped:0 overruns:0 frame:0
          TX packets:17223778 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:4539999675 (4.2 GiB)  TX bytes:4539999675 (4.2 GiB)
I compared the success logs with the failure logs and found that node 1 
never makes a transition from "Connected to Peer" to "Peer is connected 
and accepted" when I use bond0 interfaces
Could you please shed some light on this.
I am attaching both the sucess logs and failure logs for both the nodes
Regards
Rahul Shrivastava
=====-----=====-----=====
Notice: The information contained in this e-mail
message and/or attachments to it may contain 
confidential or privileged information. If you are 
not the intended recipient, any dissemination, use, 
review, distribution, printing or copying of the 
information contained in this e-mail message 
and/or attachments to it are strictly prohibited. If 
you have received this communication in error, 
please notify us by reply e-mail or telephone and 
immediately and permanently delete the message 
and any attachments. Thank you
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130621/3d44dc34/attachment.html>
-------------- next part --------------
An embedded and charset-unspecified text was scrubbed...
Name: node0_failure_logs.txt
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130621/3d44dc34/attachment.txt>
-------------- next part --------------
An embedded and charset-unspecified text was scrubbed...
Name: node0_success_logs.txt
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130621/3d44dc34/attachment-0001.txt>
-------------- next part --------------
An embedded and charset-unspecified text was scrubbed...
Name: node1_failure_logs.txt
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130621/3d44dc34/attachment-0002.txt>
-------------- next part --------------
An embedded and charset-unspecified text was scrubbed...
Name: node1_success_logs.txt
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130621/3d44dc34/attachment-0003.txt>
    
    
More information about the Gluster-users
mailing list