[Gluster-users] Kosher admin practices
joe at julianfamily.org
Tue Jul 23 18:43:32 UTC 2013
I kill (SIGTERM) the glusterfsd process for the specific brick.
To restart that brick, "gluster volume start $vol force".
Michael Peek <peek at nimbios.org> wrote:
>I have a cluster with replication (four machines, two drives in each)
>for testing that I've been beating on. I've just simulated one type of
>hardware failure by remounting a drive read-only.
>The manual covers many useful things: Adding/removing peers;
>Starting/stopping, creating, expanding, shrinking, and deleting
>etc. But it doesn't cover what you should do to replace a failed brick
>to minimize frustration and chances of data loss.
>I can't unmount the brick because glusterfs still has open files on it.
>If I stop the glusterfs-server then that takes the other brick in the
>machine out of commission too.
>I have the same problem if I reboot the machine -- I take the other
>brick out of service.
>What's the correct way to deal with this? Is there a way to tell
>gluster to take a brick out of commission for replacement without
>interrupting access to other bricks in the same machine?
>Thanks for your help,
>Gluster-users mailing list
>Gluster-users at gluster.org
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the Gluster-users