[Gluster-users] [Gluster-devel] Messup with peer status!!

Atin Mukherjee amukherj at redhat.com
Mon Mar 14 11:57:49 UTC 2016



On 03/14/2016 05:09 PM, ABHISHEK PALIWAL wrote:
> I am not getting you which glusterd directory you are asking. if you are
> asking about the /var/lib/glusterd directory then which I shared earlier
> is the same.
1. Go to /var/log/glusterfs directory
2. Look for glusterd log file
3. attach the log
Do it for both the boards.
> 
> I have two directories related to gluster
> 
> 1. /var/log/glusterfs
> 2./var/lib/glusterd
> 
> On Mon, Mar 14, 2016 at 4:12 PM, Atin Mukherjee <amukherj at redhat.com
> <mailto:amukherj at redhat.com>> wrote:
> 
> 
> 
>     On 03/14/2016 03:59 PM, ABHISHEK PALIWAL wrote:
>     > I have only these glusterd files available on the nodes
>     Look for etc-*-glusterd*.log in /var/log/glusterfs, that represents the
>     glusterd log file.
>     >
>     > Regards,
>     > Abhishek
>     >
>     > On Mon, Mar 14, 2016 at 3:43 PM, Atin Mukherjee <amukherj at redhat.com <mailto:amukherj at redhat.com>
>     > <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>>> wrote:
>     >
>     >
>     >
>     >     On 03/14/2016 02:18 PM, ABHISHEK PALIWAL wrote:
>     >     >
>     >     >
>     >     > On Mon, Mar 14, 2016 at 12:12 PM, Atin Mukherjee <amukherj at redhat.com <mailto:amukherj at redhat.com>
>     <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>>
>     >     > <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>
>     <mailto:amukherj at redhat.com <mailto:amukherj at redhat.com>>>> wrote:
>     >     >
>     >     >
>     >     >
>     >     >     On 03/14/2016 10:52 AM, ABHISHEK PALIWAL wrote:
>     >     >     > Hi Team,
>     >     >     >
>     >     >     > I am facing some issue with peer status and because of
>     that
>     >     remove-brick
>     >     >     > on replica volume is getting failed.
>     >     >     >
>     >     >     > Here. is the scenario what I am doing with gluster:
>     >     >     >
>     >     >     > 1. I have two boards A & B and gluster is running on
>     both of
>     >     the boards.
>     >     >     > 2. On  board I have created a replicated volume with one
>     >     brick on each
>     >     >     > board.
>     >     >     > 3. Created one glusterfs mount point where both of
>     brick are
>     >     mounted.
>     >     >     > 4. start the volume with nfs.disable=true.
>     >     >     > 5. Till now everything is in sync between both of bricks.
>     >     >     >
>     >     >     > Now when I manually plug-out the board B from the slot and
>     >     plug-in it again.
>     >     >     >
>     >     >     > 1. After bootup the board B I have started the glusted on
>     >     the board B.
>     >     >     >
>     >     >     > Following are the some gluster command output on Board B
>     >     after the step 1.
>     >     >     >
>     >     >     > # gluster peer status
>     >     >     > Number of Peers: 2
>     >     >     >
>     >     >     > Hostname: 10.32.0.48
>     >     >     > Uuid: f4ebe3c5-b6a4-4795-98e0-732337f76faf
>     >     >     > State: Accepted peer request (Connected)
>     >     >     >
>     >     >     > Hostname: 10.32.0.48
>     >     >     > Uuid: 4bf982c0-b21b-415c-b870-e72f36c7f2e7
>     >     >     > State: Peer is connected and Accepted (Connected)
>     >     >     >
>     >     >     > Why this peer status is showing two peer with
>     different UUID?
>     >     >     GlusterD doesn't generate a new UUID on init if it has
>     already
>     >     generated
>     >     >     an UUID earlier. This clearly indicates that on reboot
>     of board B
>     >     >     content of /var/lib/glusterd were wiped off. I've asked this
>     >     question to
>     >     >     you multiple times that is it the case?
>     >     >
>     >     >
>     >     > Yes I am following the same which is mentioned in the link:
>     >     >
>     >     >
>     >   
>      http://www.gluster.org/community/documentation/index.php/Resolving_Peer_Rejected
>     >     >
>     >     > but why it is showing two peer enteries?
>     >     >
>     >     >     >
>     >     >     > # gluster volume info
>     >     >     >
>     >     >     > Volume Name: c_glusterfs
>     >     >     > Type: Replicate
>     >     >     > Volume ID: c11f1f13-64a0-4aca-98b5-91d609a4a18d
>     >     >     > Status: Started
>     >     >     > Number of Bricks: 1 x 2 = 2
>     >     >     > Transport-type: tcp
>     >     >     > Bricks:
>     >     >     > Brick1: 10.32.0.48:/opt/lvmdir/c2/brick
>     >     >     > Brick2: 10.32.1.144:/opt/lvmdir/c2/brick
>     >     >     > Options Reconfigured:
>     >     >     > performance.readdir-ahead: on
>     >     >     > network.ping-timeout: 4
>     >     >     > nfs.disable: on
>     >     >     > # gluster volume heal c_glusterfs info
>     >     >     > c_glusterfs: Not able to fetch volfile from glusterd
>     >     >     > Volume heal failed.
>     >     >     > # gluster volume status c_glusterfs
>     >     >     > Status of volume: c_glusterfs
>     >     >     > Gluster process                             TCP Port 
>     RDMA Port
>     >     >     Online
>     >     >     > Pid
>     >     >     >
>     >     >
>     >     
>     ------------------------------------------------------------------------------
>     >     >     >
>     >     >     > Brick 10.32.1.144:/opt/lvmdir/c2/brick      N/A       N/A
>     >         N
>     >     >     > N/A
>     >     >     > Self-heal Daemon on localhost               N/A       N/A
>     >         Y
>     >     >     > 3922
>     >     >     >
>     >     >     > Task Status of Volume c_glusterfs
>     >     >     >
>     >     >
>     >     
>     ------------------------------------------------------------------------------
>     >     >     >
>     >     >     > There are no active volume tasks
>     >     >     > --
>     >     >     >
>     >     >     > At the same time Board A have the following gluster
>     commands
>     >     outcome:
>     >     >     >
>     >     >     > # gluster peer status
>     >     >     > Number of Peers: 1
>     >     >     >
>     >     >     > Hostname: 10.32.1.144
>     >     >     > Uuid: c6b64e36-76da-4e98-a616-48e0e52c7006
>     >     >     > State: Peer in Cluster (Connected)
>     >     >     >
>     >     >     > Why it is showing the older UUID of host 10.32.1.144
>     when this
>     >     >     UUID has
>     >     >     > been changed and new UUID is
>     >     267a92c3-fd28-4811-903c-c1d54854bda9
>     >     >     >
>     >     >     >
>     >     >     > # gluster volume heal c_glusterfs info
>     >     >     > c_glusterfs: Not able to fetch volfile from glusterd
>     >     >     > Volume heal failed.
>     >     >     > # gluster volume status c_glusterfs
>     >     >     > Status of volume: c_glusterfs
>     >     >     > Gluster process                             TCP Port 
>     RDMA Port
>     >     >     Online
>     >     >     > Pid
>     >     >     >
>     >     >
>     >     
>     ------------------------------------------------------------------------------
>     >     >     >
>     >     >     > Brick 10.32.0.48:/opt/lvmdir/c2/brick       49169     0
>     >         Y
>     >     >     > 2427
>     >     >     > Brick 10.32.1.144:/opt/lvmdir/c2/brick      N/A       N/A
>     >         N
>     >     >     > N/A
>     >     >     > Self-heal Daemon on localhost               N/A       N/A
>     >         Y
>     >     >     > 3388
>     >     >     > Self-heal Daemon on 10.32.1.144             N/A       N/A
>     >         Y
>     >     >     > 3922
>     >     >     >
>     >     >     > Task Status of Volume c_glusterfs
>     >     >     >
>     >     >
>     >     
>     ------------------------------------------------------------------------------
>     >     >     >
>     >     >     > There are no active volume tasks
>     >     >     >
>     >     >     > As you see in the "gluster volume status" showing that
>     Brick
>     >     >     > "10.32.1.144:/opt/lvmdir/c2/brick " is offline so We have
>     >     tried to
>     >     >     > remove it but getting "volume remove-brick c_glusterfs
>     replica 1
>     >     >     > 10.32.1.144:/opt/lvmdir/c2/brick force : FAILED :
>     Incorrect
>     >     brick
>     >     >     > 10.32.1.144:/opt/lvmdir/c2/brick for volume c_glusterfs"
>     >     error on the
>     >     >     > Board A.
>     >     >     >
>     >     >     > Please reply on this post because I am always getting
>     this error
>     >     >     in this
>     >     >     > scenario.
>     >     >     >
>     >     >     > For more detail I am also adding the logs of both of the
>     >     board which
>     >     >     > having some manual created file in which you can find the
>     >     output of
>     >     >     > glulster command from both of the boards
>     >     >     >
>     >     >     > in logs
>     >     >     > 00030 is board A
>     >     >     > 00250 is board B.
>     >     >     This attachment doesn't help much. Could you attach full
>     >     glusterd log
>     >     >     files from both the nodes?
>     >     >     >
>     >     >
>     >     > inside this attachment you will found full glusterd log file
>     >     > 00300/glusterd/ and 002500/glusterd/
>     >     No, that contains the configuration files.
>     >     >
>     >     >     > Thanks in advance waiting for the reply.
>     >     >     >
>     >     >     > Regards,
>     >     >     > Abhishek
>     >     >     >
>     >     >     >
>     >     >     > Regards
>     >     >     > Abhishek Paliwal
>     >     >     >
>     >     >     >
>     >     >     > _______________________________________________
>     >     >     > Gluster-devel mailing list
>     >     >     > Gluster-devel at gluster.org
>     <mailto:Gluster-devel at gluster.org> <mailto:Gluster-devel at gluster.org
>     <mailto:Gluster-devel at gluster.org>>
>     >     <mailto:Gluster-devel at gluster.org
>     <mailto:Gluster-devel at gluster.org> <mailto:Gluster-devel at gluster.org
>     <mailto:Gluster-devel at gluster.org>>>
>     >     >     > http://www.gluster.org/mailman/listinfo/gluster-devel
>     >     >     >
>     >     >
>     >     >
>     >     >
>     >     >
>     >     > --
>     >     >
>     >     >
>     >     >
>     >     >
>     >     > Regards
>     >     > Abhishek Paliwal
>     >
>     >
>     >
>     >
>     > --
>     >
>     >
>     >
>     >
>     > Regards
>     > Abhishek Paliwal
> 
> 
> 
> 
> -- 
> 
> 
> 
> 
> Regards
> Abhishek Paliwal


More information about the Gluster-users mailing list