[Gluster-users] File Corruption when adding bricks to live replica volumes
Lindsay Mathieson
lindsay.mathieson at gmail.com
Thu Jan 21 00:54:58 UTC 2016
On 19/01/16 22:06, Krutika Dhananjay wrote:
> As far as the reverse heal is concerned, there is one issue with
> add-brick where replica count is increased, which is still under review.
> Could you instead try the following steps at the time of add-brick and
> tell me if it works fine:
>
> 1. Run 'gluster volume add-brick datastore1 replica 3
> vng.proxmox.softlog:/vmdata/datastore1' as usual.
>
> 2. Kill the glusterfsd process corresponding to newly added brick (the
> brick in vng in your case). You should be able to get its pid in the
> output of 'gluster volume status datastore1'.
> 3. Create a dummy file on the root of the volume from the mount point.
> This can be any random name.
> 4. Delete the dummy file created in step 3.
> 5. Bring the killed brick back up. For this, you can run 'gluster
> volume start datastore1 force'.
> 6. Then execute 'gluster volume heal datastore1 full' on the node with
> the highest uuid (this we know how to do from the previous thread on
> the same topic).
>
> Then monitor heal-info output to track heal progress.
I'm afraid it didn't work Krutika, I still got the reverse heal problem.
nb. I am starting from a replica 3 store, removing a brick, cleaning it,
then re-adding it. Possibly that affects the process?
--
Lindsay Mathieson
More information about the Gluster-users
mailing list