[Gluster-users] Determine og force healing occurring

Jesper Led Lauridsen TS Infra server jly at dr.dk
Sat Jul 16 16:23:50 UTC 2016


On 07/16/2016 04:10 AM, Pranith Kumar Karampuri wrote:
>
>
> On Fri, Jul 15, 2016 at 5:20 PM, Jesper Led Lauridsen TS Infra server
> <JLY at dr.dk <mailto:JLY at dr.dk>> wrote:
>
>     Hi,
>
>     How do I determine in which log etc. that a healing is in progress
>     or startet and how do I if not startet force it.
>
>     Additional info, is that I have some problem with a volume if I
>     execute 'gluster volume heal <volume> info' the command just hangs
>     but if I execute  'gluster volume heal <volume> info split-brain'
>     it return that no file are in split-brain. Yet there is and I have
>     successfully recovered another one.
>
>
> If the command hangs there is a chance that operations on the file may
> have lead to stale locks. Could you give the output of statedump?
> You can follow
> https://gluster.readthedocs.io/en/latest/Troubleshooting/statedump/ to
> generat the files.

Thanks for you response. You are right there was a stale lock. But I am
sorry I booted all my cluster nodes, so I guess (without knowing) that
there is no reason to give you the output of a statedump?

What I can confirm and give of information is:
  * All the servers failed to reboot so I had to push the button. They
all failed with the message
     "Unmounting pipe file system: Cannot create link /etc/mtab~
     Perhaps there is a stale lock file?"
  * After 2 nodes had rebooted the command executed without any problem
and reported a couple off split-brain (Both Directory and Files)
  * strace the command showed that it was just looping, so basically the
command didn't hanging. It just couldn't finish.
  * I am using "glusterfs-3.6.2-1.el6.x86_64". But hoping to upgrade to
3.6.9 this weekend.
  * The file I refereed to here. Now has the same output on both
replicas when getting getfattr information. The
grusted.afr.glu_rhevtst_dr2_data_01-client-[0,1] and trusted.afr.dirty
are now all zero
 
>
>
>     I just have problem with this one. I can determine if there is a
>     healing process running or not
>
>     I have change 'trusted.afr.glu_rhevtst_dr2_data_01-client-1' to
>     0x000000000000000000000000 on the file located on glustertst03 and
>     executed a 'ls -lrt' on the file on the gluster-mount.
>
>     [root at glustertst04 ]# getfattr -d -m . -e hex
>     /bricks/brick1/glu_rhevtst_dr2_data_01/6bdc67d1-4ae5-47e3-86c3-ef0916996862/images/7669ca25-028e-40a5-9dc8-06c716101709/a1ae3612-bb89-45d8-8041-134c34592eab
>     getfattr: Removing leading '/' from absolute path names
>     # file:
>     bricks/brick1/glu_rhevtst_dr2_data_01/6bdc67d1-4ae5-47e3-86c3-ef0916996862/images/7669ca25-028e-40a5-9dc8-06c716101709/a1ae3612-bb89-45d8-8041-134c34592eab
>     security.selinux=0x73797374656d5f753a6f626a6563745f723a66696c655f743a733000
>     trusted.afr.dirty=0x000000000000000000000000
>     trusted.afr.glu_rhevtst_dr2_data_01-client-0=0x00004c700000000000000000
>     trusted.afr.glu_rhevtst_dr2_data_01-client-1=0x000000000000000000000000
>     trusted.gfid=0x7575f870875b4c899fd81ef16be3b1a1
>     trusted.glusterfs.quota.70145d52-bb80-42ce-b437-64be6ee4a7d4.contri=0x00000001606dc000
>     trusted.pgfid.70145d52-bb80-42ce-b437-64be6ee4a7d4=0x00000001
>
>      [root at glustertst03 ]# getfattr -d -m . -e hex
>     /bricks/brick1/glu_rhevtst_dr2_data_01/6bdc67d1-4ae5-47e3-86c3-ef0916996862/images/7669ca25-028e-40a5-9dc8-06c716101709/a1ae3612-bb89-45d8-8041-134c34592eab
>     getfattr: Removing leading '/' from absolute path names
>     # file:
>     bricks/brick1/glu_rhevtst_dr2_data_01/6bdc67d1-4ae5-47e3-86c3-ef0916996862/images/7669ca25-028e-40a5-9dc8-06c716101709/a1ae3612-bb89-45d8-8041-134c34592eab
>     security.selinux=0x73797374656d5f753a6f626a6563745f723a66696c655f743a733000
>     trusted.afr.dirty=0x000000270000000000000000
>     trusted.afr.glu_rhevtst_dr2_data_01-client-0=0x000000000000000000000000
>     trusted.afr.glu_rhevtst_dr2_data_01-client-1=0x000000000000000000000000
>     trusted.gfid=0x7575f870875b4c899fd81ef16be3b1a1
>     trusted.glusterfs.quota.70145d52-bb80-42ce-b437-64be6ee4a7d4.contri=0x0000000160662000
>     trusted.pgfid.70145d52-bb80-42ce-b437-64be6ee4a7d4=0x00000001
>
>     [root at glustertst04 ]# stat
>     /var/run/gluster/glu_rhevtst_dr2_data_01/6bdc67d1-4ae5-47e3-86c3-ef0916996862/images/7669ca25-028e-40a5-9dc8-06c716101709/a1ae3612-bb89-45d8-8041-134c34592eab
>       File:
>     `/var/run/gluster/glu_rhevtst_dr2_data_01/6bdc67d1-4ae5-47e3-86c3-ef0916996862/images/7669ca25-028e-40a5-9dc8-06c716101709/a1ae3612-bb89-45d8-8041-134c34592eab'
>       Size: 21474836480     Blocks: 11548384   IO Block: 131072
>     regular file
>     Device: 31h/49d Inode: 11517990069246079393  Links: 1
>     Access: (0660/-rw-rw----)  Uid: (   36/    vdsm)   Gid: (   36/   
>      kvm)
>     Access: 2016-07-15 13:33:47.860224289 +0200
>     Modify: 2016-07-15 13:34:44.396125458 +0200
>     Change: 2016-07-15 13:34:44.397125492 +0200
>
>     [root at glustertst03 ]# stat
>     /bricks/brick1/glu_rhevtst_dr2_data_01/6bdc67d1-4ae5-47e3-86c3-ef0916996862/images/7669ca25-028e-40a5-9dc8-06c716101709/a1ae3612-bb89-45d8-8041-134c34592eab
>       File:
>     `/bricks/brick1/glu_rhevtst_dr2_data_01/6bdc67d1-4ae5-47e3-86c3-ef0916996862/images/7669ca25-028e-40a5-9dc8-06c716101709/a1ae3612-bb89-45d8-8041-134c34592eab'
>       Size: 21474836480     Blocks: 11547408   IO Block: 4096 
>      regular file
>     Device: fd02h/64770d    Inode: 159515      Links: 2
>     Access: (0660/-rw-rw----)  Uid: (   36/    vdsm)   Gid: (   36/   
>      kvm)
>     Access: 2016-07-13 08:33:00.000561984 +0200
>     Modify: 2016-07-13 08:32:59.969561154 +0200
>     Change: 2016-07-15 12:52:28.414192052 +0200
>
>     Thanks
>     Jesper
>     _______________________________________________
>     Gluster-users mailing list
>     Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>     http://www.gluster.org/mailman/listinfo/gluster-users
>
>
>
>
> -- 
> Pranith

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160716/8206d8d2/attachment.html>


More information about the Gluster-users mailing list