[Gluster-users] possible gluster error causing kvm to shutdown
Daniele Antolini
lantuin at gmail.com
Tue Feb 21 15:59:46 UTC 2017
Is it possible that self-heal process on the kvm VM runs intensively and
shutdown automatically the vm?
Daniele
2017-02-21 16:53 GMT+01:00 Alessandro Briosi <ab1 at metalit.com>:
> Il 21/02/2017 09:53, Alessandro Briosi ha scritto:
> > Hi all,
> > I have had a couple of times now a KVM VM which suddenly was shutdown
> > (whithout any apparent reason)
> >
> > At the time this happened the only thing I can find in logs are related
> > to gluster:
> >
> > Stops have happened at 16.19 on the 13th and at 03.34 on the 19th. (time
> > is local time which is GMT+1)
> > I though tink that gluster logs are in GMT.
> >
> > This is from 1st node (which also runs the kvm and basically should be
> > client of itself)
> >
> > [2017-02-07 22:29:15.030197] I [MSGID: 114035]
> > [client-handshake.c:202:client_set_lk_version_cbk]
> > 0-datastore1-client-1: Server lk version = 1
> > [2017-02-19 05:22:07.747187] I [MSGID: 108026]
> > [afr-self-heal-common.c:1173:afr_log_selfheal] 0-datastore1-replicate-0:
> > Completed data selfheal on 9e66f0d2-501
> > b-4cf9-80db-f423e2e2ef0f. sources=[1] sinks=0
> > r
> > This is from 2nd node:
> >
> > [2017-02-07 22:29:15.044422] I [MSGID: 114035]
> > [client-handshake.c:202:client_set_lk_version_cbk]
> > 0-datastore1-client-0: Server lk version = 1
> > [2017-02-08 00:13:58.612483] I [MSGID: 108026]
> > [afr-self-heal-common.c:1173:afr_log_selfheal] 0-datastore1-replicate-0:
> > Completed data selfheal on b32ccae9-01e
> > d-406c-988f-64394e4cb37c. sources=[0] sinks=1
> > [2017-02-13 16:44:10.570176] I [MSGID: 108026]
> > [afr-self-heal-common.c:1173:afr_log_selfheal] 0-datastore1-replicate-0:
> > Completed data selfheal on bc8f6a7e-31e
> > 5-4b48-946c-f779a4b2e64f. sources=[1] sinks=0
> > [2017-02-19 04:30:46.049524] I [MSGID: 108026]
> > [afr-self-heal-common.c:1173:afr_log_selfheal] 0-datastore1-replicate-0:
> > Completed data selfheal on bc8f6a7e-31e
> > 5-4b48-946c-f779a4b2e64f. sources=[1] sinks=0
> >
> > Could this be the cause?
> >
> > This is current volume configuration, I'll be adding an additional node
> > in the near future, but need to have this stable before.
> >
> > Volume Name: datastore1
> > Type: Replicate
> > Volume ID: e4dbbf6e-11e6-4b36-bab0-c37647ef6ad6
> > Status: Started
> > Snapshot Count: 0
> > Number of Bricks: 1 x 2 = 2
> > Transport-type: tcp
> > Bricks:
> > Brick1: srvpve1g:/data/brick1/brick
> > Brick2: srvpve2g:/data/brick1/brick
> > Options Reconfigured:
> > nfs.disable: on
> > performance.readdir-ahead: on
> > transport.address-family: inet
> >
> >
>
> nobody has any clue on this?
> Should I provide more information/logs?
>
> For what I understand there was a healing triggered, but I have no idea
> on why this happened, and why the kvm was shutdown.
> Gluster client is supposed to be client of both servers for failover.
> Also there are other vm running and they did not get shutdown.
>
> Alessandro
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170221/2871bc4d/attachment.html>
More information about the Gluster-users
mailing list