[Gluster-users] remove-brick question
james.bellinger at icecube.wisc.edu
james.bellinger at icecube.wisc.edu
Mon Sep 16 21:56:07 UTC 2013
I inherited a system with a wide mix of array sizes (no replication) in
3.2.2, and wanted to drain data from a failing array.
I upgraded to 3.3.2, and began a
gluster volume remove-brick scratch "gfs-node01:/sda" start
After some time I got this:
gluster volume remove-brick scratch "gfs-node01:/sda" status
Node Rebalanced-files size scanned failures
status
--------- ----------- ----------- ----------- -----------
------------
localhost 0 0Bytes 0 0
not started
gfs-node06 0 0Bytes 0 0
not started
gfs-node03 0 0Bytes 0 0
not started
gfs-node05 0 0Bytes 0 0
not started
gfs-node01 2257394624 2.8TB 5161640 208878
completed
Two things jump instantly to mind:
1) The number of failures is rather large
2) A _different_ disk seems to have been _partially_ drained.
/dev/sda 2.8T 2.7T 12G 100% /sda
/dev/sdb 2.8T 769G 2.0T 28% /sdb
/dev/sdc 2.8T 2.1T 698G 75% /sdc
/dev/sdd 2.8T 2.2T 589G 79% /sdd
When I mount the system it is read-only (another problem I want to fix
ASAP) so I'm pretty sure the failures aren't due to users changing the
system underneath me.
Thanks for any pointers.
James Bellinger
More information about the Gluster-users
mailing list