[Bugs] [Bug 1537362] glustershd/ glusterd is not using right port when connecting to glusterfsd process
bugzilla at redhat.com
bugzilla at redhat.com
Mon Apr 16 04:47:20 UTC 2018
https://bugzilla.redhat.com/show_bug.cgi?id=1537362
--- Comment #3 from zhou lin <zz.sh.cynthia at gmail.com> ---
(In reply to Worker Ant from comment #2)
> COMMIT: https://review.gluster.org/19263 committed in master by \"Atin
> Mukherjee\" <amukherj at redhat.com> with a commit message- glusterd: process
> pmap sign in only when port is marked as free
>
> Because of some crazy race in volume start code path because of friend
> handshaking with volumes with quorum enabled we might end up into a situation
> where glusterd would start a brick and get a disconnect and then immediately
> try
> to start the same brick instance based on another friend update request. And
> then if for the very first brick even if the process doesn't come up at the
> end
> sign in event gets sent and we end up having two duplicate portmap entries
> for
> the same brick. Since in brick start we mark the previous port as free, its
> better to consider a sign in request as no op if the corresponding port type
> is
> marked as free.
>
> Change-Id: I995c348c7b6988956d24b06bf3f09ab64280fc32
> BUG: 1537362
> Signed-off-by: Atin Mukherjee <amukherj at redhat.com>
this issue appeared again, in my local test env
gluster volume heal log info
Brick sn-0.local:/mnt/bricks/log/brick
Status: Transport endpoint is not connected
Number of entries: -
Brick sn-1.local:/mnt/bricks/log/brick
/master/fsaudit/internal-ssh.log
/
/master/fsaudit/auth.log
/blackboxes
/master/fsaudit/alarms
/rcploglib/sn-0/DHAAgent.level
/rcploglib/sn-0/DVMAgent.level
/master/syslog
Status: Connected
Number of entries: 8
[root at sn-0:/root]
#
gluster volume heal mstate info
Brick sn-0.local:/mnt/bricks/mstate/brick
Status: Transport endpoint is not connected
Number of entries: -
Brick sn-1.local:/mnt/bricks/mstate/brick
/
Status: Connected
Number of entries: 1
[root at sn-0:/root]
#ps aux | grep glustershd | grep -v grep | wc -l glustershd process exist.
1
[root at sn-0:/root]
#
_gluster volume status log
Status of volume: log
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick sn-0.local:/mnt/bricks/log/brick 49157 0 Y 1402
Brick sn-1.local:/mnt/bricks/log/brick 49155 0 Y 3817
Self-heal Daemon on localhost N/A N/A Y 1428
Self-heal Daemon on sn-1.local N/A N/A Y 20316
Self-heal Daemon on sn-2.local N/A N/A Y 15565
Task Status of Volume log
------------------------------------------------------------------------------
There are no active volume tasks
[root at sn-0:/root]
#_gluster volume status mstate
Status of volume: mstate
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick sn-0.local:/mnt/bricks/mstate/brick 49158 0 Y 1403
Brick sn-1.local:/mnt/bricks/mstate/brick 49153 0 Y 2967
Self-heal Daemon on localhost N/A N/A Y 1428
Self-heal Daemon on sn-1.local N/A N/A Y 20316
Self-heal Daemon on sn-2.local N/A N/A Y 15565
Task Status of Volume mstate
------------------------------------------------------------------------------
There are no active volume tasks
[root at sn-0:/root]
--
You are receiving this mail because:
You are on the CC list for the bug.
Unsubscribe from this bug https://bugzilla.redhat.com/token.cgi?t=UvZVBLWmoC&a=cc_unsubscribe
More information about the Bugs
mailing list