[Gluster-users] [URGENT] Add-bricks to a volume corrupted the files
Kevin Lemonnier
lemonnierk at ulrar.net
Tue Sep 6 10:38:54 UTC 2016
Hi,
Here is the info :
Volume Name: VMs
Type: Replicate
Volume ID: c5272382-d0c8-4aa4-aced-dd25a064e45c
Status: Started
Number of Bricks: 1 x 3 = 3
Transport-type: tcp
Bricks:
Brick1: ips4adm.name:/mnt/storage/VMs
Brick2: ips5adm.name:/mnt/storage/VMs
Brick3: ips6adm.name:/mnt/storage/VMs
Options Reconfigured:
performance.readdir-ahead: on
cluster.quorum-type: auto
cluster.server-quorum-type: server
network.remote-dio: enable
cluster.eager-lock: enable
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
features.shard: on
features.shard-block-size: 64MB
cluster.data-self-heal-algorithm: full
network.ping-timeout: 15
For the logs I'm sending that over to you in private.
On Tue, Sep 06, 2016 at 09:48:07AM +0530, Krutika Dhananjay wrote:
> Could you please attach the glusterfs client and brick logs?
> Also provide output of `gluster volume info`.
> -Krutika
> On Tue, Sep 6, 2016 at 4:29 AM, Kevin Lemonnier <lemonnierk at ulrar.net>
> wrote:
>
> >A A - What was the original (and current) geometry? (status and info)
>
> It was a 1x3 that I was trying to bump to 2x3.
> >A A - what parameters did you use when adding the bricks?
> >
>
> Just a simple add-brick node1:/path node2:/path node3:/path
> Then a fix-layout when everything started going wrong.
>
> I was able to salvage some VMs by stopping them then starting them
> again,
> but most won't start for various reasons (disk corrupted, grub not found
> ...).
> For those we are deleting the disks then importing them from backups,
> that's
> a huge loss but everything has been down for so long, no choice ..
> >A A On 6/09/2016 8:00 AM, Kevin Lemonnier wrote:
> >
> >A I tried a fix-layout, and since that didn't work I removed the brick
> (start then commit when it showed
> >A completed). Not better, the volume is now running on the 3 original
> bricks (replica 3) but the VMs
> >A are still corrupted. I have 880 Mb of shards on the bricks I removed
> for some reason, thos shards do exist
> >A (and are bigger) on the "live" volume. I don't understand why now
> that I have removed the new bricks
> >A everything isn't working like before ..
> >
> >A On Mon, Sep 05, 2016 at 11:06:16PM +0200, Kevin Lemonnier wrote:
> >
> >A Hi,
> >
> >A I just added 3 bricks to a volume and all the VMs are doing I/O
> errors now.
> >A I rebooted a VM to see and it can't start again, am I missing
> something ? Is the reblance required
> >A to make everything run ?
> >
> >A That's urgent, thanks.
> >
> >A --
> >A Kevin Lemonnier
> >A PGP Fingerprint : 89A5 2283 04A0 E6E9 0111
> >
> >
> >
> >
> >A _______________________________________________
> >A Gluster-users mailing list
> >A Gluster-users at gluster.org
> >A http://www.gluster.org/mailman/listinfo/gluster-users
> >
> >
> >
> >A _______________________________________________
> >A Gluster-users mailing list
> >A Gluster-users at gluster.org
> >A http://www.gluster.org/mailman/listinfo/gluster-users
> >
> >A --
> >A Lindsay Mathieson
>
> > _______________________________________________
> > Gluster-users mailing list
> > Gluster-users at gluster.org
> > http://www.gluster.org/mailman/listinfo/gluster-users
>
> --
> Kevin Lemonnier
> PGP Fingerprint : 89A5 2283 04A0 E6E9 0111
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
--
Kevin Lemonnier
PGP Fingerprint : 89A5 2283 04A0 E6E9 0111
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 819 bytes
Desc: Digital signature
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160906/3b88037d/attachment.sig>
More information about the Gluster-users
mailing list