[Gluster-users] VM disks corruption on 3.7.11
ajneil.tech at gmail.com
Thu May 19 20:04:49 UTC 2016
I am slightly confused you say you have image file corruption but then you
say the qemu-img check says there is no corruption. If what you mean is
that you see I/O errors during a heal this is likely to be due to io
starvation, something that is a well know issue.
There is work happening to improve this in version 3.8:
On 19 May 2016 at 09:58, Kevin Lemonnier <lemonnierk at ulrar.net> wrote:
> That's a different problem then, I have corruption without removing or
> adding bricks,
> as mentionned. Might be two separate issue
> On Thu, May 19, 2016 at 11:25:34PM +1000, Lindsay Mathieson wrote:
> > On 19/05/2016 12:17 AM, Lindsay Mathieson wrote:
> > One thought - since the VM's are active while the brick is
> > removed/re-added, could it be the shards that are written while the
> > brick is added that are the reverse healing shards?
> > I tested by:
> > - removing brick 3
> > - erasing brick 3
> > - closing down all VM's
> > - adding new brick 3
> > - waiting until heal number reached its max and started decreasing
> > There were no reverse heals
> > - Started the VM's backup. No real issues there though one showed IO
> > errors, presumably due to shards being locked as they were healed.
> > - VM's started ok, no reverse heals were noted and eventually Brick 3
> > fully healed. The VM's do not appear to be corrupted.
> > So it would appear the problem is adding a brick while the volume is
> > written to.
> > Cheers,
> > --
> > Lindsay Mathieson
> > _______________________________________________
> > Gluster-users mailing list
> > Gluster-users at gluster.org
> > http://www.gluster.org/mailman/listinfo/gluster-users
> Kevin Lemonnier
> PGP Fingerprint : 89A5 2283 04A0 E6E9 0111
> Gluster-users mailing list
> Gluster-users at gluster.org
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the Gluster-users