[Gluster-devel] query about a split-brain problem found in glusterfs3.12.3

Ravishankar N ravishankar at redhat.com
Thu Feb 8 11:52:31 UTC 2018



On 02/08/2018 01:08 PM, Zhou, Cynthia (NSB - CN/Hangzhou) wrote:
>
> Hi,
>
> I check the link you provided. It does not mention the the “dirty” 
> attribute, if I try to fix this split-brain by manually setfattr 
> command, should I only set the “trusted.afr.export-client-0” command?
>
Manually resetting xattrs is not recommended. Use the gluster CLI to 
resolve it.
>
> By the way, I feel it is quite strange that the output of “gluster 
> volume heal export info” command there is two entries with the same 
> name, how does this happen?
>
Maybe the same entry is listed in different subfolders of 
.glusterfs/indices?
>
> gluster volume heal export info
> Brick sn-0.local:/mnt/bricks/export/brick
> Status: Connected
> Number of entries: 0
>
> Brick sn-1.local:/mnt/bricks/export/brick
> /testdir - Is in split-brain
>
> /testdir - Possibly undergoing heal
>
> Status: Connected
> Number of entries: 2
>
> I also do some other test, when sn-0 side file/dir does not has 
> “dirty” and “trusted.afr.export-client-*” attribute and sn-1 side 
> file/dir has both “dirty” and “trusted.afr.export-client-*” non-zero. 
> The gluster could self heal such scenario. But in this case the it 
> could never self heal.
>
> *From:*Ravishankar N [mailto:ravishankar at redhat.com]
> *Sent:* Thursday, February 08, 2018 11:56 AM
> *To:* Zhou, Cynthia (NSB - CN/Hangzhou) 
> <cynthia.zhou at nokia-sbell.com>; Gluster-devel at gluster.org
> *Subject:* Re: query about a split-brain problem found in glusterfs3.12.3
>
> On 02/08/2018 07:16 AM, Zhou, Cynthia (NSB - CN/Hangzhou) wrote:
>
>     Hi,
>
>     Thanks for responding?
>
>     If split-brain happen in such kind of test is reasonable, how to
>     fix this split-brain situation?
>
> If you are using replica 2, then there is no prevention. Once they 
> occur, you can resolve them using 
> http://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/
>
> If you want to prevent split-brain, you would need to use replica 3 or 
> arbiter volume.
>
> Regards,
> Ravi
>
>     *From:*Ravishankar N [mailto:ravishankar at redhat.com]
>     *Sent:* Thursday, February 08, 2018 12:12 AM
>     *To:* Zhou, Cynthia (NSB - CN/Hangzhou)
>     <cynthia.zhou at nokia-sbell.com>
>     <mailto:cynthia.zhou at nokia-sbell.com>; Gluster-devel at gluster.org
>     <mailto:Gluster-devel at gluster.org>
>     *Subject:* Re: query about a split-brain problem found in
>     glusterfs3.12.3
>
>     On 02/07/2018 10:39 AM, Zhou, Cynthia (NSB - CN/Hangzhou) wrote:
>
>         Hi glusterfs expert:
>
>         Good day.
>
>         Lately, we meet a glusterfs split brain problem in our env in
>         /mnt/export/testdir. We start 3 ior process (IOR tool) from
>         non-sn nodes, which is creating/removing files repeatedly in
>         testdir. then we reboot sn nodes(sn0 and sn1) by sequence.
>         Then we meet following problem.
>
>         Do you have some comments on how this could happen? And how to
>         fix it in this situation? Thanks!
>
>
>     Is the problem that split-brain is happening? Is this a replica 2
>     volume? If yes, then it looks like it is expected behavior?
>     Regards
>     Ravi
>
>
>         gluster volume heal export info
>         Brick sn-0.local:/mnt/bricks/export/brick
>         Status: Connected
>         Number of entries: 0
>
>         Brick sn-1.local:/mnt/bricks/export/brick
>         /testdir - Is in split-brain
>
>         /testdir - Possibly undergoing heal
>
>         Status: Connected
>         Number of entries: 2
>
>         wait for a while …..
>
>         gluster volume heal export info
>         Brick sn-0.local:/mnt/bricks/export/brick
>         Status: Connected
>         Number of entries: 0
>
>         Brick sn-1.local:/mnt/bricks/export/brick
>         /testdir - Possibly undergoing heal
>
>         /testdir - Possibly undergoing heal
>
>         and finally:
>
>         [root at sn-0:/root <http://sn-0/root>]
>         # gluster v heal export info
>         Brick sn-0.local:/mnt/bricks/export/brick
>         <http://local/mnt/bricks/export/brick>
>         Status: Connected
>         Number of entries: 0
>
>         Brick sn-1.local:/mnt/bricks/export/brick
>         <http://local/mnt/bricks/export/brick>
>         /testdir - Is in split-brain
>
>         Status: Connected
>         Number of entries: 1
>
>         [root at sn-0:/root <http://sn-0/root>]
>
>         # getfattr -m .* -d -e  hex /mnt/bricks/export/brick/testdir
>
>         getfattr: Removing leading '/' from absolute path names
>
>         # file: mnt/bricks/export/brick/testdir
>
>         trusted.gfid=0x5622cff893b3484dbdb6a20a0edb0e77
>
>         trusted.glusterfs.dht=0x000000010000000000000000ffffffff
>
>         [root at sn-1:/root <http://sn-1/root>]
>
>         # getfattr -m .* -d -e  hex /mnt/bricks/export/brick/testdir
>
>         getfattr: Removing leading '/' from absolute path names
>
>         # file: mnt/bricks/export/brick/testdir
>
>         trusted.afr.dirty=0x000000000000000000000001
>
>         trusted.afr.export-client-0=0x000000000000000000000038
>
>         trusted.gfid=0x5622cff893b3484dbdb6a20a0edb0e77
>
>         trusted.glusterfs.dht=0x000000010000000000000000ffffffff
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-devel/attachments/20180208/89106226/attachment-0001.html>


More information about the Gluster-devel mailing list