[Gluster-users] Peer Rejected(Connected) and Self heal daemon is not running causing split brain
Kaamesh Kamalaaharan
kaamesh at novocraft.com
Thu Feb 26 09:25:46 UTC 2015
Hi guys,
I managed to get gluster running but im having a couple of issues with my
setup 1) my peer status is rejected but connected 2) my self heal daemon is
not running on one server and im getting split brain files.
My setup is two gluster volumes (gfs1 and gfs2) on replicate each with a
brick
1) My peer status doesnt go into Peer in Cluster. running a peer status
command gives me State:Peer Rejected (Connected) . At this point, the brick
on gfs2 does not go online and i get this output
#gluster volume status
Status of volume: gfsvolume
Gluster process Port Online Pid
------------------------------------------------------------------------------
Brick gfs1:/export/sda/brick 49153 Y 15025
NFS Server on localhost 2049 Y 15039
Self-heal Daemon on localhost N/A Y 15044
Task Status of Volume gfsvolume
------------------------------------------------------------------------------
There are no active volume tasks
I have followed the methods used in one of the threads and performed the
following
a) stop glusterd
b) rm all files in /var/lib/glusterd/ except for glusterd.info
c) start glusterd and probe gfs1 from gfs2 and peer status which gives me
# gluster peer status
Number of Peers: 1
Hostname: gfs1
Uuid: 49acc9c2-4809-4da5-a6f0-6a3d48314070
State: Sent and Received peer request (Connected)
the same thread mentioned that changing the status of the peer in
/var/lib/glusterd/peer/{UUID} from status=5 to status=3 fixes this and on
restart of gfs1 the peer status goes to
#gluster peer status
Number of Peers: 1
Hostname: gfs1
Uuid: 49acc9c2-4809-4da5-a6f0-6a3d48314070
State: Peer in Cluster (Connected)
This fixes the connection between the peers and the volume status shows
Status of volume: gfsvolume
Gluster process Port Online Pid
------------------------------------------------------------------------------
Brick gfs1:/export/sda/brick 49153 Y 10852
Brick gfs2:/export/sda/brick 49152 Y 17024
NFS Server on localhost N/A N N/A
Self-heal Daemon on localhost N/A N N/A
NFS Server on gfs2 N/A N N/A
Self-heal Daemon on gfs2 N/A N N/A
Task Status of Volume gfsvolume
------------------------------------------------------------------------------
There are no active volume tasks
Which brings us to problem 2
2) My self-heal demon is not alive
I fixed the self heal on gfs1 by running
#find <gluster-mount> -noleaf -print0 | xargs --null stat >/dev/null
2>/var/log/gluster/<gluster-mount>-selfheal.log
and running a volume status command gives me
# gluster volume status
Status of volume: gfsvolume
Gluster process Port Online Pid
------------------------------------------------------------------------------
Brick gfs1:/export/sda/brick 49152 Y 16660
Brick gfs2:/export/sda/brick 49152 Y 21582
NFS Server on localhost 2049 Y 16674
Self-heal Daemon on localhost N/A Y 16679
NFS Server on gfs2 N/A N 21596
Self-heal Daemon on gfs2 N/A N 21600
Task Status of Volume gfsvolume
------------------------------------------------------------------------------
There are no active volume tasks
However, running this on gfs2 doesnt fix the daemon.
Restarting the gfs2 server brings me back to problem 1 and the cycle
continues..
Can anyone assist me with this issue(s).. thank you.
Thank You Kindly,
Kaamesh
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150226/0b054c00/attachment.html>
More information about the Gluster-users
mailing list