[Gluster-users] Cannot start Gluster -- resolve brick failed in restore

Atin Mukherjee amukherj at redhat.com
Mon Jun 8 09:25:56 UTC 2015



On 06/08/2015 01:38 PM, shacky wrote:
> Hi.
> I have a GlusterFS cluster running on a Debian Wheezy with GlusterFS
> 3.6.2, with one volume on all three bricks (web1, web2, web3).
> All was working good until I changed the IP addresses of bricks,
> because after then only the GlusterFS daemon on web1 is starting well,
> and the deamons on web2 and web3 are exiting with these errors:
> 
> [2015-06-08 07:59:15.929330] I [MSGID: 100030]
> [glusterfsd.c:2018:main] 0-/usr/sbin/glusterd: Started running
> /usr/sbin/glusterd version 3.6.2 (args: /usr/sbin/glusterd -p
> /var/run/glusterd.pid)
> [2015-06-08 07:59:15.932417] I [glusterd.c:1214:init] 0-management:
> Maximum allowed open file descriptors set to 65536
> [2015-06-08 07:59:15.932482] I [glusterd.c:1259:init] 0-management:
> Using /var/lib/glusterd as working directory
> [2015-06-08 07:59:15.933772] W [rdma.c:4221:__gf_rdma_ctx_create]
> 0-rpc-transport/rdma: rdma_cm event channel creation failed (No such
> device)
> [2015-06-08 07:59:15.933815] E [rdma.c:4519:init] 0-rdma.management:
> Failed to initialize IB Device
> [2015-06-08 07:59:15.933838] E
> [rpc-transport.c:333:rpc_transport_load] 0-rpc-transport: 'rdma'
> initialization failed
> [2015-06-08 07:59:15.933887] W [rpcsvc.c:1524:rpcsvc_transport_create]
> 0-rpc-service: cannot create listener, initing the transport failed
> [2015-06-08 07:59:17.354500] I
> [glusterd-store.c:2043:glusterd_restore_op_version] 0-glusterd:
> retrieved op-version: 30600
> [2015-06-08 07:59:17.527377] I
> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
> 0-management: connect returned 0
> [2015-06-08 07:59:17.527446] I
> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
> 0-management: connect returned 0
> [2015-06-08 07:59:17.527499] I
> [rpc-clnt.c:969:rpc_clnt_connection_init] 0-management: setting
> frame-timeout to 600
> [2015-06-08 07:59:17.528139] I
> [rpc-clnt.c:969:rpc_clnt_connection_init] 0-management: setting
> frame-timeout to 600
> [2015-06-08 07:59:17.528861] E
> [glusterd-store.c:4244:glusterd_resolve_all_bricks] 0-glusterd:
> resolve brick failed in restore
> [2015-06-08 07:59:17.528891] E [xlator.c:425:xlator_init]
> 0-management: Initialization of volume 'management' failed, review
> your volfile again
> [2015-06-08 07:59:17.528906] E [graph.c:322:glusterfs_graph_init]
> 0-management: initializing translator failed
> [2015-06-08 07:59:17.528917] E [graph.c:525:glusterfs_graph_activate]
> 0-graph: init failed
> [2015-06-08 07:59:17.529257] W [glusterfsd.c:1194:cleanup_and_exit]
> (--> 0-: received signum (0), shutting down
> 
> Please note that bricks name are setted in /etc/hosts and all of them
> are resolving well with the new IP addresses, so I cannot find out
> where the problem is.
> 
> Could you help me please?

Here is what you can do on the nodes where glusterD fails to start:

1. cd /var/lib/glusterd
2. grep -irns "old ip"
The output will be similar like this :

vols/test-vol/info:20:brick-0=172.17.0.2:-tmp-b1
vols/test-vol/info:21:brick-1=172.17.0.2:-tmp-b2
vols/test-vol/test-vol.tcp-fuse.vol:6:    option remote-host 172.17.0.2
vols/test-vol/test-vol.tcp-fuse.vol:15:    option remote-host 172.17.0.2
vols/test-vol/trusted-test-vol.tcp-fuse.vol:8:    option remote-host
172.17.0.2
vols/test-vol/trusted-test-vol.tcp-fuse.vol:19:    option remote-host
172.17.0.2
vols/test-vol/test-vol-rebalance.vol:6:    option remote-host 172.17.0.2
vols/test-vol/test-vol-rebalance.vol:15:    option remote-host 172.17.0.2
vols/test-vol/bricks/172.17.0.1:-tmp-b1:1:hostname=172.17.0.2
vols/test-vol/bricks/172.17.0.1:-tmp-b2:1:hostname=172.17.0.2
nfs/nfs-server.vol:8:    option remote-host 172.17.0.2
nfs/nfs-server.vol:19:    option remote-host 172.17.0.2

3. find . * -exec sed -i "s/<old ip>/<new ip>/g" {} \;

4. You would need to manually rename few files (for eg : mv
vols/test-vol/bricks/172.17.0.1:-tmp-b1
vols/test-vol/bricks/172.17.0.2:-tmp-b1)

Do this exercise on all the failed nodes and recheck and let me know if
it works.

~Atin
> 
> Thank you very much!
> Bye
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
> 
> 

-- 
~Atin


More information about the Gluster-users mailing list