[Gluster-users] Gluster Self Heal
toby.corkindale at strategicdata.com.au
Wed Jul 10 08:01:16 UTC 2013
On 09/07/13 18:17, 符永涛 wrote:
> Hi Toby,
> What's the bug #? I want to have a look and backport it to our
> production server if it helps. Thank you.
I think it was this one:
The bug being that the daemons were crashing out if you had a lot of
volumes defined, I think?
> 2013/7/9 Toby Corkindale <toby.corkindale at strategicdata.com.au
> <mailto:toby.corkindale at strategicdata.com.au>>
> On 09/07/13 15:38, Bobby Jacob wrote:
> I have a 2-node gluster with 3 TB storage.
> 1)I believe the “glusterfsd” is responsible for the self healing
> the 2 nodes.
> 2)Due to some network error, the replication stopped for some
> reason but
> the application was accessing the data from node1. When I
> manually try
> to start “glusterfsd” service, its not starting.
> Please advice on how I can maintain the integrity of the data so
> that we
> have all the data in both the locations. ??
> There were some bugs in the self-heal daemon present in 3.3.0 and
> 3.3.1. Our systems see the SHD crash out with segfaults quite often,
> and it does not recover.
> I reported this bug a long time ago, and it was fixed in trunk
> relatively quickly -- however version 3.3.2 has still not been
> released, despite the fix being found six months ago.
> I find this quite disappointing.
> Gluster-users mailing list
> Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
More information about the Gluster-users