[Gluster-users] self-heal failed
Liang Ma
ma.satops at gmail.com
Thu Jan 10 17:50:48 UTC 2013
Looks like replace-brick in a corrupted replicate brick with 3.3 doesn't
work any more. To rebuild the cluster is exactly what we want to avoid when
we choose to use glusterfs at the first place.
I assume to replace a failed replicate disk or node should be a standard
procedure, isn't it? I could find anything related to this in the 3.3
manual.
On Thu, Jan 10, 2013 at 12:18 PM, Daniel Taylor <dtaylor at vocalabs.com>wrote:
> I've run replace-brick on missing bricks before, it should still work.
>
> On the other hand, data corruption is the worst case failure mode.
> The one time I hit data corruption on a node my final answer ended up
> being to rebuild the cluster from scratch and restore the best copy of the
> data I had (mix of backups and live data).
>
>
> On 01/10/2013 11:12 AM, Liang Ma wrote:
>
>>
>> Thank you Daniel for you more comments.
>>
>> Now I can remove the damaged zfs brick after rebooting the system. But
>> then what can I do to rejoin a new brick? I can't run gluster volume
>> replace-brick because the old brick is gone. I can't even remove the old
>> brick because the gluster's replicate count is 2. So what is the right
>> procedure to replace a failed brick for replicate gluster volume?
>>
>> Liang
>>
>>
>> On Thu, Jan 10, 2013 at 11:57 AM, Daniel Taylor <dtaylor at vocalabs.com<mailto:
>> dtaylor at vocalabs.com>> wrote:
>>
>> I'm not familiar with zfs in particular, but it should have given
>> you a message saying why it won't unmount.
>>
>> In the worst case you can indeed remove the mount point from
>> /etc/fstab and reboot. A hard reboot may be necessary in a case
>> like this.
>>
>>
>> On 01/10/2013 10:43 AM, Liang Ma wrote:
>>
>>
>> Yes, I stopped the glusterfs service on the damaged system but
>> zfs still won't allow me to umount the filesystem. Maybe I
>> should try to shutdown the entire system.
>>
>>
>> On Wed, Jan 9, 2013 at 10:28 AM, Daniel Taylor
>> <dtaylor at vocalabs.com <mailto:dtaylor at vocalabs.com>
>> <mailto:dtaylor at vocalabs.com <mailto:dtaylor at vocalabs.com>>**>
>>
>> wrote:
>>
>>
>> On 01/09/2013 08:31 AM, Liang Ma wrote:
>>
>>
>> Hi Daniel,
>>
>> Ok, if gluster can't self-heal from this situation, I
>> hope at
>> least I can manually restore the volume by using the good
>> brick available. So would you please tell me how can I
>> "simply
>> rebuild the filesystem and let gluster attempt to
>> restore it
>> from a *clean* filesystem"?
>>
>>
>> Trimmed for space.
>>
>> You could do as Tom Pfaff suggests, but given the odds of data
>> corruption carrying forward I'd do the following:
>> Shut down gluster on the damaged system.
>> Unmount the damaged filesystem.
>> Reformat the damaged filesystem as new (throwing away any
>> potential corruption that might not get caught on rebuild)
>> Mount the new filesystem at the original mount point
>> Restart gluster
>>
>> In the event of corruption due to hardware failure you'd
>> be doing
>> this on replacement hardware.
>> The key is you have to have a functional filesystem for
>> gluster to
>> work with.
>>
>>
>> -- Daniel Taylor VP Operations Vocal
>> Laboratories, Inc
>> dtaylor at vocalabs.com <mailto:dtaylor at vocalabs.com>
>> <mailto:dtaylor at vocalabs.com <mailto:dtaylor at vocalabs.com>>
>> 612-235-5711 <tel:612-235-5711>
>> <tel:612-235-5711 <tel:612-235-5711>>
>>
>>
>> ______________________________**_________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.**org<Gluster-users at gluster.org>
>> >
>> <mailto:Gluster-users at gluster.**org <Gluster-users at gluster.org>
>>
>> <mailto:Gluster-users at gluster.**org <Gluster-users at gluster.org>>>
>> http://supercolony.gluster.**org/mailman/listinfo/gluster-**users<http://supercolony.gluster.org/mailman/listinfo/gluster-users>
>>
>>
>>
>> -- Daniel Taylor VP Operations Vocal
>> Laboratories, Inc
>> dtaylor at vocalabs.com <mailto:dtaylor at vocalabs.com> 612-235-5711
>> <tel:612-235-5711>
>>
>> ______________________________**_________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org <mailto:Gluster-users at gluster.**org<Gluster-users at gluster.org>
>> >
>> http://supercolony.gluster.**org/mailman/listinfo/gluster-**users<http://supercolony.gluster.org/mailman/listinfo/gluster-users>
>>
>>
>>
> --
> Daniel Taylor VP Operations Vocal Laboratories, Inc
> dtaylor at vocalabs.com 612-235-5711
>
> ______________________________**_________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.**org/mailman/listinfo/gluster-**users<http://supercolony.gluster.org/mailman/listinfo/gluster-users>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20130110/55793cbc/attachment.html>
More information about the Gluster-users
mailing list