[Gluster-users] gluster processes won't start when a single node is booted

Atin Mukherjee atin.mukherjee83 at gmail.com
Sun Sep 20 11:50:31 UTC 2015


This behaviour is expected and in a 2 node cluster setup brick processes
are started only when the other nodes come up. The same would be true in
3.5.x series as well. Adding a dummy node into the cluster would help you
to solve your problem.

-Atin
Sent from one plus one
On Sep 20, 2015 4:56 PM, "Mauro M." <gluster at ezplanet.net> wrote:

> Hi all,
>
> I hope you might help.
>
> I just upgraded from 3.5.6 to 3.7.4
>
> My configuration is 1 volume with 2 x bricks replicated.
>
> Normally I have brick1 running and brick2 turned off so that when I want
> to do maintenance on brick1 I turn on brick2, wait for synchronization to
> complete and turn off brick1.
>
> Often I just reboot brick1 with brick2 still turned off.
>
> With glusterfs version 3.5 I could do all of the above.
>
> After the upgrade to 3.7.4 if I boot brick1 (or brick2) without the other
> node, glusterd starts, but the gluster network processes won't start.
>
> Here is the output of gluster volume info:
> Volume Name: gv_home
> Type: Replicate
> Volume ID: ef806153-2a02-4db9-a54e-c2f89f79b52e
> Status: Started
> Number of Bricks: 1 x 2 = 2
> Transport-type: tcp
> Bricks:
> Brick1: brick1:/brick0/gv_home
> Brick2: brick2:/brick0/gv_home
> Options Reconfigured:
> nfs.disable: on
> config.transport: tcp
>
> ... and gluster volume status:
>
> Status of volume: gv_home
> Gluster process                         TCP Port  RDMA Port  Online  Pid
> ---------------------------------------------------------------------------
> Brick brick1:/brick0/gv_home            N/A       N/A        N       N/A
> NFS Server on localhost                 N/A       N/A        N       N/A
>
> Task Status of Volume gv_home
> ---------------------------------------------------------------------------
> There are no active volume tasks
>
> Under this condition gv_home cannot be mounted.
>
> Only if I start brick2, once glusterd starts on brick2 the gluster
> processes also start on brick1 and gv_home can be mounted:
>
> Status of volume: gv_home
> Gluster process                         TCP Port  RDMA Port  Online  Pid
> --------------------------------------------------------------------------
> Brick brick1:/brick0/gv_home            49158     0          Y       30049
> Brick brick2:/brick0/gv_home            49158     0          Y       14797
> Self-heal Daemon on localhost           N/A       N/A        Y       30044
> Self-heal Daemon on brick2              N/A       N/A        Y       14792
>
> Task Status of Volume gv_home
> --------------------------------------------------------------------------
> There are no active volume tasks
>
> Once I turn off brick2 then the volume remains available and mounted
> without issues (for as long that the relative gluster processes remain
> active, if I kill them I am back without volume).
>
> The issue is that I would like to safely boot one the bricks without
> having to boot both to get the volume back and mountable which is what I
> was able to do with glusterfs version 3.5.
>
> Please could you help?
> Is there any parameter to set that would enable the same behaviour as in
> 3.5?
>
> Thank you in advance,
> Mauro
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150920/5f334b24/attachment.html>


More information about the Gluster-users mailing list