[Gluster-users] Extra work in gluster volume rebalance and odd reporting
Joel Young
jdy at cryregarder.com
Fri Oct 25 21:29:10 UTC 2013
A couple more things:
1. For the work volume, the failures are caused by hard links that
can't be rebalanced. It is odd thought that the hardlinks show up in
the rebalanced files count even though they failed.
2. Output of gluster volume info:
Volume Name: home
Type: Distributed-Replicate
Volume ID: 83fa39a6-6e68-4e1c-8fae-3c3e30b1bd66
Status: Started
Number of Bricks: 2 x 2 = 4
Transport-type: tcp
Bricks:
Brick1: ir0:/lhome/gluster_home
Brick2: ir1:/lhome/gluster_home
Brick3: ir2:/lhome/gluster_home
Brick4: ir3:/raid/gluster_home
Options Reconfigured:
cluster.lookup-unhashed: no
performance.client-io-threads: on
performance.cache-size: 512MB
server.statedump-path: /tmp
Volume Name: work
Type: Distribute
Volume ID: 823816bb-2e60-4b37-a142-ba464a77bfdc
Status: Started
Number of Bricks: 3
Transport-type: tcp
Bricks:
Brick1: ir0:/raid/gluster_work
Brick2: ir1:/raid/gluster_work
Brick3: ir2:/raid/gluster_work
Options Reconfigured:
performance.client-io-threads: on
performance.cache-size: 1GB
performance.write-behind-window-size: 3MB
performance.flush-behind: on
server.statedump-path: /tmp
Thanks,
Joel
On Fri, Oct 25, 2013 at 11:49 AM, Joel Young <jdy at cryregarder.com> wrote:
> Folks,
>
> With gluster 1.4.0 on fedora 19:
>
> I have a four node gluster peer group (ir0, ir1, ir2, ir3). I've got
> two distributed filesystems on the cluster.
>
> One (work) distributed with bricks on ir0, ir1, and ir2. The other
> (home) replicated and distributed with replication across the
> distribution pairs (ir0, ir3) and (ir1, ir2).
>
> When doing a gluster volume rebalance home start and gluster volume
> rebalance work start, it does rebalance operations on every node in
> the peer group. For work, it ran a rebalance on ir3 even though there
> is no brick on ir3. For home, it ran a rebalance on ir1 and ir3 and
> did no work on those.
>
> [root at ir0]# gluster volume rebalance home status; gluster volume
> rebalance work status
> Node Rebalanced-files
> size scanned failures status run time in secs
> --------- -----------
> ----------- ----------- ----------- ------------
> --------------
> localhost 33441
> 2.3GB 120090 0 in progress 67154.00
> ir2 12878
> 32.7GB 234395 0 completed 29569.00
> ir3 0
> 0Bytes 234367 0 completed 1581.00
> ir1 0
> 0Bytes 234367 0 completed 1569.00
> volume rebalance: home: success:
> Node Rebalanced-files
> size scanned failures status run time in secs
> --------- -----------
> ----------- ----------- ----------- ------------
> --------------
> localhost 0
> 0Bytes 1862936 0 completed 4444.00
> ir2 417
> 10.4GB 1862936 417 completed 4466.00
> ir3 0
> 0Bytes 1862936 0 completed 4454.00
> ir1 4
> 282.8MB 1862936 4 completed 4438.00
>
>
> Sometimes I would get:
>
> volume rebalance: work: success:
> [root at ir0 ghenders]# gluster volume rebalance home status; gluster
> volume rebalance work status
> Node Rebalanced-files
> size scanned failures status run time in secs
> --------- -----------
> ----------- ----------- ----------- ------------
> --------------
> localhost 31466
> 2.3GB 114290 0 in progress 63194.00
> localhost 31466
> 2.3GB 114290 0 in progress 63194.00
> localhost 31466
> 2.3GB 114290 0 in progress 63194.00
> localhost 31466
> 2.3GB 114290 0 in progress 63194.00
> ir3 0
> 0Bytes 234367 0 completed 1581.00
> volume rebalance: home: success:
> Node Rebalanced-files
> size scanned failures status run time in secs
> --------- -----------
> ----------- ----------- ----------- ------------
> --------------
> localhost 0
> 0Bytes 1862936 0 completed 4444.00
> localhost 0
> 0Bytes 1862936 0 completed 4444.00
> localhost 0
> 0Bytes 1862936 0 completed 4444.00
> localhost 0
> 0Bytes 1862936 0 completed 4444.00
> ir1 4
> 282.8MB 1862936 4 completed 4438.00
>
>
> Where it only reports progress on one node.
>
> Should I file bugs on these?
>
> Joel
More information about the Gluster-users
mailing list