[Gluster-users] [Gluster-devel] failed heal

Pranith Kumar Karampuri pkarampu at redhat.com
Fri Feb 6 02:03:46 UTC 2015



----- Original Message -----
> From: "Niels de Vos" <ndevos at redhat.com>
> To: "Pranith Kumar Karampuri" <pkarampu at redhat.com>
> Cc: gluster-users at gluster.org, "Gluster Devel" <gluster-devel at gluster.org>
> Sent: Friday, February 6, 2015 2:32:36 AM
> Subject: Re: [Gluster-devel] failed heal
> 
> On Thu, Feb 05, 2015 at 11:21:58AM +0530, Pranith Kumar Karampuri wrote:
> > 
> > On 02/04/2015 11:52 PM, David F. Robinson wrote:
> > >I don't recall if that was before or after my upgrade.
> > >I'll forward you an email thread for the current heal issues which are
> > >after the 3.6.2 upgrade...
> > This is executed after the upgrade on just one machine. 3.6.2 entry locks
> > are not compatible with versions <= 3.5.3 and 3.6.1 that is the reason.
> > From
> > 3.5.4 and releases >=3.6.2 it should work fine.
> 
> Oh, I was not aware of this requirement. Does it mean we should not mix
> deployments with these versions (what about 3.4?) any longer? 3.5.4 has
> not been released yet, so anyone with a mixed 3.5/3.6.2 environment will
> hit these issues? Is this only for the self-heal daemon, or are the
> triggered/stat self-heal procedures affected too?
> 
> It should be noted *very* clearly in the release notes, and I think an
> announcement (email+blog) as a warning/reminder would be good. Could you
> get some details and advice written down, please?
Will do today.

Pranith
> 
> Thanks,
> Niels
> 
> 
> > 
> > Pranith
> > >David
> > >------ Original Message ------
> > >From: "Pranith Kumar Karampuri" <pkarampu at redhat.com
> > ><mailto:pkarampu at redhat.com>>
> > >To: "David F. Robinson" <david.robinson at corvidtec.com
> > ><mailto:david.robinson at corvidtec.com>>; "gluster-users at gluster.org"
> > ><gluster-users at gluster.org <mailto:gluster-users at gluster.org>>; "Gluster
> > >Devel" <gluster-devel at gluster.org <mailto:gluster-devel at gluster.org>>
> > >Sent: 2/4/2015 2:33:20 AM
> > >Subject: Re: [Gluster-devel] failed heal
> > >>
> > >>On 02/02/2015 03:34 AM, David F. Robinson wrote:
> > >>>I have several files that gluster says it cannot heal. I deleted the
> > >>>files from all of the bricks
> > >>>(/data/brick0*/hpc_shared/motorsports/gmics/Raven/p3/*) and ran a full
> > >>>heal using 'gluster volume heal homegfs full'.  Even after the full
> > >>>heal, the entries below still show up.
> > >>>How do I clear these?
> > >>3.6.1 Had an issue where files undergoing I/O will also be shown in the
> > >>output of 'gluster volume heal <volname> info', we addressed that in
> > >>3.6.2. Is this output from 3.6.1 by any chance?
> > >>
> > >>Pranith
> > >>>[root at gfs01a ~]# gluster volume heal homegfs info
> > >>>Gathering list of entries to be healed on volume homegfs has been
> > >>>successful
> > >>>Brick gfsib01a.corvidtec.com:/data/brick01a/homegfs
> > >>>Number of entries: 10
> > >>>/hpc_shared/motorsports/gmics/Raven/p3/70_rke/Movies
> > >>><gfid:a6fc9011-74ad-4128-a232-4ccd41215ac8>
> > >>><gfid:bc17fa79-c1fd-483d-82b1-2c0d3564ddc5>
> > >>><gfid:ec804b5c-8bfc-4e7b-91e3-aded7952e609>
> > >>><gfid:ba62e340-4fad-477c-b450-704133577cbb>
> > >>><gfid:4843aa40-8361-4a97-88d5-d37fc28e04c0>
> > >>><gfid:c90a8f1c-c49e-4476-8a50-2bfb0a89323c>
> > >>><gfid:090042df-855a-4f5d-8929-c58feec10e33>
> > >>>/hpc_shared/motorsports/gmics/Raven/p3/70_rke/.Convrg.swp
> > >>>/hpc_shared/motorsports/gmics/Raven/p3/70_rke
> > >>>Brick gfsib01b.corvidtec.com:/data/brick01b/homegfs
> > >>>Number of entries: 2
> > >>><gfid:f96b4ddf-8a75-4abb-a640-15dbe41fdafa>
> > >>>/hpc_shared/motorsports/gmics/Raven/p3/70_rke
> > >>>Brick gfsib01a.corvidtec.com:/data/brick02a/homegfs
> > >>>Number of entries: 7
> > >>><gfid:5d08fe1d-17b3-4a76-ab43-c708e346162f>
> > >>>/hpc_shared/motorsports/gmics/Raven/p3/70_rke/PICTURES/.tmpcheck
> > >>>/hpc_shared/motorsports/gmics/Raven/p3/70_rke/PICTURES
> > >>>/hpc_shared/motorsports/gmics/Raven/p3/70_rke/Movies
> > >>><gfid:427d3738-3a41-4e51-ba2b-f0ba7254d013>
> > >>><gfid:8ad88a4d-8d5e-408f-a1de-36116cf6d5c1>
> > >>><gfid:0e034160-cd50-4108-956d-e45858f27feb>
> > >>>Brick gfsib01b.corvidtec.com:/data/brick02b/homegfs
> > >>>Number of entries: 0
> > >>>Brick gfsib02a.corvidtec.com:/data/brick01a/homegfs
> > >>>Number of entries: 0
> > >>>Brick gfsib02b.corvidtec.com:/data/brick01b/homegfs
> > >>>Number of entries: 0
> > >>>Brick gfsib02a.corvidtec.com:/data/brick02a/homegfs
> > >>>Number of entries: 0
> > >>>Brick gfsib02b.corvidtec.com:/data/brick02b/homegfs
> > >>>Number of entries: 0
> > >>>===============================
> > >>>David F. Robinson, Ph.D.
> > >>>President - Corvid Technologies
> > >>>704.799.6944 x101 [office]
> > >>>704.252.1310 [cell]
> > >>>704.799.7974 [fax]
> > >>>David.Robinson at corvidtec.com <mailto:David.Robinson at corvidtec.com>
> > >>>http://www.corvidtechnologies.com <http://www.corvidtechnologies.com/>
> > >>>
> > >>>
> > >>>_______________________________________________
> > >>>Gluster-devel mailing list
> > >>>Gluster-devel at gluster.org
> > >>>http://www.gluster.org/mailman/listinfo/gluster-devel
> > >>
> > 
> 
> > _______________________________________________
> > Gluster-devel mailing list
> > Gluster-devel at gluster.org
> > http://www.gluster.org/mailman/listinfo/gluster-devel
> 
> 


More information about the Gluster-users mailing list