[Gluster-users] More Peculiar heal behaviour after removing brick

Anuradha Talur atalur at redhat.com
Mon Jan 25 07:41:20 UTC 2016



----- Original Message -----
> From: "Lindsay Mathieson" <lindsay.mathieson at gmail.com>
> To: "gluster-users" <Gluster-users at gluster.org>
> Sent: Saturday, January 23, 2016 9:20:52 AM
> Subject: [Gluster-users] More Peculiar heal behaviour after removing brick
> 
> Maybe I'm doing some wrong here but I'm not sure what, or maybe this is
> normal behaviour?
> 
> 
> All of the following is performed from my vna node which has the highest
> numbered uuid. Indenting applied by me for readability.
> 
> Spoiler, because it happens at the end: removing the vng brick followed by a
> full heal gives this error:
> " Commit failed on vng.proxmox.softlog. Please check log file for details."
> 
> 
> Sets to recreate:
> 1. Create a test volume:
> 
> 
> vna$ gluster volume create test3 rep 3 transport tcp
> vnb.proxmox.softlog:/vmdata/test3 vng.proxmox.softlog:/vmdata/test3
> vna.proxmox.softlog:/vmdata/test3
> vna$ gluster volume set test3 group softlog
> vna$ gluster volume info test3
> 
> 
> Volume Name: test3
> Type: Replicate
> Volume ID: 0be89d63-775c-4eb5-9d98-0a4a87f30fbf
> Status: Created
> Number of Bricks: 1 x 3 = 3
> Transport-type: tcp
> Bricks:
> Brick1: vnb.proxmox.softlog:/vmdata/test3
> Brick2: vng.proxmox.softlog:/vmdata/test3
> Brick3: vna.proxmox.softlog:/vmdata/test3
> Options Reconfigured:
> cluster.data-self-heal-algorithm: full
> network.remote-dio: enable
> cluster.eager-lock: enable
> performance.io-cache: off
> performance.read-ahead: off
> performance.quick-read: off
> performance.stat-prefetch: off
> performance.strict-write-ordering: on
> performance.write-behind: off
> nfs.enable-ino32: off
> nfs.addr-namelookup: off
> nfs.disable: on
> performance.cache-refresh-timeout: 4
> performance.io-thread-count: 32
> performance.low-prio-threads: 32
> cluster.server-quorum-type: server
> cluster.quorum-type: auto
> client.event-threads: 4
> server.event-threads: 4
> cluster.self-heal-window-size: 256
> features.shard-block-size: 512MB
> features.shard: on
> performance.readdir-ahead: off
> vna$ gluster volume start test3
> 
> 2. Immediately remove the vng brick:
> 
> 
> vna$ gluster volume remove-brick test3 replica 2
> vng.proxmox.softlog:/vmdata/test3 force
> vna$ gluster volume info test3
> 
> 
> Volume Name: test3
> Type: Replicate
> Volume ID: 36421a23-68c4-455d-8d4c-e21d9428e1da
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: vnb.proxmox.softlog:/vmdata/test3
> Brick2: vna.proxmox.softlog:/vmdata/test3
> Options Reconfigured:
> cluster.data-self-heal-algorithm: full
> network.remote-dio: enable
> cluster.eager-lock: enable
> performance.io-cache: off
> performance.read-ahead: off
> performance.quick-read: off
> performance.stat-prefetch: off
> performance.strict-write-ordering: on
> performance.write-behind: off
> nfs.enable-ino32: off
> nfs.addr-namelookup: off
> nfs.disable: on
> performance.cache-refresh-timeout: 4
> performance.io-thread-count: 32
> performance.low-prio-threads: 32
> cluster.server-quorum-type: server
> cluster.quorum-type: auto
> client.event-threads: 4
> server.event-threads: 4
> cluster.self-heal-window-size: 256
> features.shard-block-size: 512MB
> features.shard: on
> performance.readdir-ahead: off
> 
> 3. Then run a full heal:
> 
> 
> 
> vna$ gluster volume heal test3 full
> Commit failed on vng.proxmox.softlog. Please check log file for details.
> 
> 
Hi,

Could you provide glusterd logs from vna? It will be in the same directory as glustershd logs.

> Weird, because of cause the vng brick has been removed. This happens every
> time.
> 
> I have preserved the glustershd logs from vna & vng if needed. There were no
> heal logs.
> 
> 
> 
> 
> 
> 
> 
> 
> --
> Lindsay Mathieson
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users

-- 
Thanks,
Anuradha.


More information about the Gluster-users mailing list