[Gluster-users] More Peculiar heal behaviour after removing brick
Krutika Dhananjay
kdhananj at redhat.com
Mon Jan 25 07:40:53 UTC 2016
Could you share the logs?
I'd like to look at the glustershd logs and etc-glusterfs-glusterd.vol.log files.
-Krutika
----- Original Message -----
> From: "Lindsay Mathieson" <lindsay.mathieson at gmail.com>
> To: "gluster-users" <Gluster-users at gluster.org>
> Sent: Saturday, January 23, 2016 9:20:52 AM
> Subject: [Gluster-users] More Peculiar heal behaviour after removing brick
> Maybe I'm doing some wrong here but I'm not sure what, or maybe this is
> normal behaviour?
> All of the following is performed from my vna node which has the highest
> numbered uuid. Indenting applied by me for readability.
> Spoiler, because it happens at the end: removing the vng brick followed by a
> full heal gives this error:
> " Commit failed on vng.proxmox.softlog. Please check log file for details."
> Sets to recreate:
> 1. Create a test volume:
> > vna$ gluster volume create test3 rep 3 transport tcp
> > vnb.proxmox.softlog:/vmdata/test3 vng.proxmox.softlog:/vmdata/test3
> > vna.proxmox.softlog:/vmdata/test3
>
> > vna$ gluster volume set test3 group softlog
>
> > vna$ gluster volume info test3
>
> > > Volume Name: test3
> >
>
> > > Type: Replicate
> >
>
> > > Volume ID: 0be89d63-775c-4eb5-9d98-0a4a87f30fbf
> >
>
> > > Status: Created
> >
>
> > > Number of Bricks: 1 x 3 = 3
> >
>
> > > Transport-type: tcp
> >
>
> > > Bricks:
> >
>
> > > Brick1: vnb.proxmox.softlog:/vmdata/test3
> >
>
> > > Brick2: vng.proxmox.softlog:/vmdata/test3
> >
>
> > > Brick3: vna.proxmox.softlog:/vmdata/test3
> >
>
> > > Options Reconfigured:
> >
>
> > > cluster.data-self-heal-algorithm: full
> >
>
> > > network.remote-dio: enable
> >
>
> > > cluster.eager-lock: enable
> >
>
> > > performance.io-cache: off
> >
>
> > > performance.read-ahead: off
> >
>
> > > performance.quick-read: off
> >
>
> > > performance.stat-prefetch: off
> >
>
> > > performance.strict-write-ordering: on
> >
>
> > > performance.write-behind: off
> >
>
> > > nfs.enable-ino32: off
> >
>
> > > nfs.addr-namelookup: off
> >
>
> > > nfs.disable: on
> >
>
> > > performance.cache-refresh-timeout: 4
> >
>
> > > performance.io-thread-count: 32
> >
>
> > > performance.low-prio-threads: 32
> >
>
> > > cluster.server-quorum-type: server
> >
>
> > > cluster.quorum-type: auto
> >
>
> > > client.event-threads: 4
> >
>
> > > server.event-threads: 4
> >
>
> > > cluster.self-heal-window-size: 256
> >
>
> > > features.shard-block-size: 512MB
> >
>
> > > features.shard: on
> >
>
> > > performance.readdir-ahead: off
> >
>
> > vna$ gluster volume start test3
>
> 2. Immediately remove the vng brick:
> > vna$ gluster volume remove-brick test3 replica 2
> > vng.proxmox.softlog:/vmdata/test3 force
>
> > vna$ gluster volume info test3
>
> > > Volume Name: test3
> >
>
> > > Type: Replicate
> >
>
> > > Volume ID: 36421a23-68c4-455d-8d4c-e21d9428e1da
> >
>
> > > Status: Started
> >
>
> > > Number of Bricks: 1 x 2 = 2
> >
>
> > > Transport-type: tcp
> >
>
> > > Bricks:
> >
>
> > > Brick1: vnb.proxmox.softlog:/vmdata/test3
> >
>
> > > Brick2: vna.proxmox.softlog:/vmdata/test3
> >
>
> > > Options Reconfigured:
> >
>
> > > cluster.data-self-heal-algorithm: full
> >
>
> > > network.remote-dio: enable
> >
>
> > > cluster.eager-lock: enable
> >
>
> > > performance.io-cache: off
> >
>
> > > performance.read-ahead: off
> >
>
> > > performance.quick-read: off
> >
>
> > > performance.stat-prefetch: off
> >
>
> > > performance.strict-write-ordering: on
> >
>
> > > performance.write-behind: off
> >
>
> > > nfs.enable-ino32: off
> >
>
> > > nfs.addr-namelookup: off
> >
>
> > > nfs.disable: on
> >
>
> > > performance.cache-refresh-timeout: 4
> >
>
> > > performance.io-thread-count: 32
> >
>
> > > performance.low-prio-threads: 32
> >
>
> > > cluster.server-quorum-type: server
> >
>
> > > cluster.quorum-type: auto
> >
>
> > > client.event-threads: 4
> >
>
> > > server.event-threads: 4
> >
>
> > > cluster.self-heal-window-size: 256
> >
>
> > > features.shard-block-size: 512MB
> >
>
> > > features.shard: on
> >
>
> > > performance.readdir-ahead: off
> >
>
> 3. Then run a full heal:
> > vna$ gluster volume heal test3 full
>
> > Commit failed on vng.proxmox.softlog. Please check log file for details.
>
> Weird, because of cause the vng brick has been removed. This happens every
> time.
> I have preserved the glustershd logs from vna & vng if needed. There were no
> heal logs.
> --
> Lindsay Mathieson
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160125/64424380/attachment.html>
More information about the Gluster-users
mailing list