[Gluster-users] Peer Rejected(Connected) and Self heal daemon is not running causing split brain

Atin Mukherjee amukherj at redhat.com
Thu Feb 26 09:40:12 UTC 2015


Could you check the N/W firewall setting? Flush iptable setting using
iptables -F and retry.

~Atin

On 02/26/2015 02:55 PM, Kaamesh Kamalaaharan wrote:
> Hi guys,
> 
> I managed to get gluster running but im having a couple of issues with my
> setup 1) my peer status is rejected but connected 2) my self heal daemon is
> not running on one server and im getting split brain files.
> My setup is two gluster volumes  (gfs1 and gfs2) on replicate each with a
> brick
> 
> 1) My peer status doesnt go into Peer in Cluster. running a peer status
> command gives me State:Peer Rejected (Connected) . At this point, the brick
> on gfs2 does not go online and i get this output
> 
> 
> #gluster volume status
> 
> Status of volume: gfsvolume
> 
> Gluster process Port Online Pid
> 
> ------------------------------------------------------------------------------
> 
> Brick gfs1:/export/sda/brick 49153 Y 15025
> 
> NFS Server on localhost 2049 Y 15039
> 
> Self-heal Daemon on localhost N/A Y 15044
> 
> 
> 
> Task Status of Volume gfsvolume
> 
> ------------------------------------------------------------------------------
> 
> There are no active volume tasks
> 
> 
> 
> I have followed the methods used in one of the threads and performed the
> following
> 
>    a) stop glusterd
>    b) rm all files in /var/lib/glusterd/  except for glusterd.info
>    c) start glusterd and probe gfs1 from gfs2 and peer status which gives me
> 
> 
> # gluster peer status
> 
> Number of Peers: 1
> 
> 
> Hostname: gfs1
> 
> Uuid: 49acc9c2-4809-4da5-a6f0-6a3d48314070
> 
> State: Sent and Received peer request (Connected)
> 
> 
> the same thread mentioned that changing the status of the peer in
> /var/lib/glusterd/peer/{UUID} from status=5 to status=3 fixes this and on
> restart of gfs1 the peer status goes to
> 
> #gluster peer status
> 
> Number of Peers: 1
> 
> 
> Hostname: gfs1
> 
> Uuid: 49acc9c2-4809-4da5-a6f0-6a3d48314070
> 
> State: Peer in Cluster (Connected)
> 
> This fixes the connection between the peers and the volume status shows
> 
> 
> Status of volume: gfsvolume
> 
> Gluster process Port Online Pid
> 
> ------------------------------------------------------------------------------
> 
> Brick gfs1:/export/sda/brick 49153 Y 10852
> 
> Brick gfs2:/export/sda/brick 49152 Y 17024
> 
> NFS Server on localhost N/A N N/A
> 
> Self-heal Daemon on localhost N/A N N/A
> 
> NFS Server on gfs2 N/A N N/A
> 
> Self-heal Daemon on gfs2 N/A N N/A
> 
> 
> 
> Task Status of Volume gfsvolume
> 
> ------------------------------------------------------------------------------
> 
> There are no active volume tasks
> 
> 
> Which brings us to problem 2
> 
> 2) My self-heal demon is not alive
> 
> I fixed the self heal on gfs1 by running
> 
>  #find <gluster-mount> -noleaf -print0 | xargs --null stat >/dev/null
> 2>/var/log/gluster/<gluster-mount>-selfheal.log
> 
> and running a volume status command gives me
> 
> # gluster volume status
> 
> Status of volume: gfsvolume
> 
> Gluster process Port Online Pid
> 
> ------------------------------------------------------------------------------
> 
> Brick gfs1:/export/sda/brick 49152 Y 16660
> 
> Brick gfs2:/export/sda/brick 49152 Y 21582
> 
> NFS Server on localhost 2049 Y 16674
> 
> Self-heal Daemon on localhost N/A Y 16679
> 
> NFS Server on gfs2 N/A N 21596
> 
> Self-heal Daemon on gfs2 N/A N 21600
> 
> 
> 
> Task Status of Volume gfsvolume
> 
> ------------------------------------------------------------------------------
> 
> There are no active volume tasks
> 
> 
> 
> However, running this on gfs2 doesnt fix the daemon.
> 
> Restarting the gfs2 server brings me back to problem 1 and the cycle
> continues..
> 
> Can anyone assist me with this issue(s).. thank you.
> 
> Thank You Kindly,
> Kaamesh
> 
> 
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
> 

-- 
~Atin


More information about the Gluster-users mailing list