[Gluster-users] Reboot Issue with 6.5 on Ubuntu 18.04
Alexander Iliev
ailiev+gluster at mamul.org
Sun Sep 8 18:53:36 UTC 2019
Hi all,
I am running a GlusterFS server 6.3 on three Ubuntu 18.04 nodes
installed from the https://launchpad.net/~gluster PPA.
I tried upgrading to 6.5 today and ran into an issue with the first (and
only) node that has been upgraded so far. When I rebooted the node the
underlying brick filesystems failed to mount because of a `pvscan`
process timing out on boot.
I did some experimenting and the issue seems to be that on reboot the
glusterfsd processes (that expose the bricks as far as I understand) are
not being shut down which leads to the underlying filesystems show up as
busy and not getting properly unmounted.
Then I found out that `systemctl stop glusterd.service` doesn't stop the
brick processes by design and it also seems that for Fedora/RHEL this
has been worked around by having a separate `glusterfsd.service` unit
that only acts on shutdown.
This however does not seem to be the case on Ubuntu and I can't figure
out what is the expected flow there.
So I guess my question is - is this normal/expected behaviour on Ubuntu?
How is one supposed to set things up so that bricks get properly
unmounted on reboot and properly mounted at startup?
I am also considering migrating from Ubuntu to CentOS now as the
upstream support seems much better there. If I decide to switch can I
re-use the existing bricks or do I need to spin up a clean node, join
the cluster and get the data synced to it?
Thanks!
Best regards,
--
alexander iliev
More information about the Gluster-users
mailing list