[Gluster-users] How-to start gluster when only one node is up ?

Diego Remolina dijuremo at gmail.com
Fri Oct 30 13:11:00 UTC 2015


Install and configure gluster, make sure the firewall openings are
proper in all nodes and then run:

gluster peer probe (new node IP)

Then you should see it on a gluster volume status [volname]

# gluster v status export
Status of volume: export
Gluster process                                         Port    Online  Pid
------------------------------------------------------------------------------
Brick 10.0.1.7:/bricks/hdds/brick                       49152   Y       21068
Brick 10.0.1.6:/bricks/hdds/brick                       49154   Y       21197
Self-heal Daemon on localhost                           N/A     Y       21217
Self-heal Daemon on 10.0.1.7                            N/A     Y       21087
Self-heal Daemon on 10.0.1.5                            N/A     Y       12956

Task Status of Volume export
------------------------------------------------------------------------------
There are no active volume tasks

As you can see above, I have two nodes with bricks, 10.0.1.6 and
10.0.1.7 and the third dummy node with no bricks is 10.0.1.5

HTH,

Diego


On Fri, Oct 30, 2015 at 8:54 AM, Peter Michael Calum <pemca at tdc.dk> wrote:
> Hi,
>
>
>
> How do i  add a ’dummy’ node ?
>
>
>
> thanks,
>
> Peter
>
>
>
> Fra: gluster-users-bounces at gluster.org
> [mailto:gluster-users-bounces at gluster.org] På vegne af Atin Mukherjee
> Sendt: 30. oktober 2015 13:14
> Til: Mauro Mozzarelli
> Cc: gluster-users at gluster.org
> Emne: Re: [Gluster-users] How-to start gluster when only one node is up ?
>
>
>
> -Atin
> Sent from one plus one
> On Oct 30, 2015 5:28 PM, "Mauro Mozzarelli" <mauro at ezplanet.net> wrote:
>>
>> Hi,
>>
>> Atin keeps giving the same answer: "it is by design"
>>
>> I keep saying "the design is wrong and it should be changed to cater for
>> standby servers"
> Every design has got its own set of limitations and i would say this is a
> limitation instead of mentioning the overall design itself wrong. I would
> again stand with my points as correctness is always a priority in a
> distributed system. This behavioural change was introduced in 3.5 and if
> this was not included as part of release note I apologize on behalf of the
> release management.
> As communicated earlier, we will definitely resolve this issue in GlusterD2.
>>
>> In the meantime this is the workaround I am using:
>> When the single node starts I stop and start the volume, and then it
>> becomes mountable. On CentOS 6 and CentOS 7 it works with release up to
>> 3.7.4. Release 3.7.5 is broken so I reverted back to 3.7.4.
> This is where I am not convinced. An explicit volume start should start the
> bricks, can you raise a BZ with all the relevant details?
>>
>> In my experience glusterfs releases are a bit of a hit and miss. Often
>> something stops working with newer releases, then after a few more
>> releases it works again or there is a workaround ... Not quite the
>> stability one would want for commercial use, and thus at the moment I can
>> risk using it only for my home servers, hence the cluster with a node
>> always ON and the second as STANDBY.
>>
>> MOUNT=/home
>> LABEL="GlusterFS:"
>> if grep -qs $MOUNT /proc/mounts; then
>>     echo "$LABEL $MOUNT is mounted";
>>     gluster volume start gv_home 2>/dev/null
>> else
>>     echo "$LABEL $MOUNT is NOT mounted";
>>     echo "$LABEL Restarting gluster volume ..."
>>     yes|gluster volume stop gv_home > /dev/null
>>     gluster volume start gv_home
>>     mount -t glusterfs sirius-ib:/gv_home $MOUNT;
>>     if grep -qs $MOUNT /proc/mounts; then
>>         echo "$LABEL $MOUNT is mounted";
>>         gluster volume start gv_home 2>/dev/null
>>     else
>>         echo "$LABEL failure to mount $MOUNT";
>>     fi
>> fi
>>
>> I hope this helps.
>> Mauro
>>
>> On Fri, October 30, 2015 11:48, Atin Mukherjee wrote:
>> > -Atin
>> > Sent from one plus one
>> > On Oct 30, 2015 4:35 PM, "Remi Serrano" <rserrano at pros.com> wrote:
>> >>
>> >> Hello,
>> >>
>> >>
>> >>
>> >> I setup a gluster file cluster with 2 nodes. It works fine.
>> >>
>> >> But, when I shut down the 2 nodes, and startup only one node, I cannot
>> > mount the share :
>> >>
>> >>
>> >>
>> >> [root at xxx ~]#  mount -t glusterfs 10.32.0.11:/gv0 /glusterLocalShare
>> >>
>> >> Mount failed. Please check the log file for more details.
>> >>
>> >>
>> >>
>> >> Log says :
>> >>
>> >> [2015-10-30 10:33:26.147003] I [MSGID: 100030] [glusterfsd.c:2318:main]
>> > 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.7.5
>> > (args: /usr/sbin/glusterfs -127.0.0.1 --volfile-id=/gv0
>> > /glusterLocalShare)
>> >>
>> >> [2015-10-30 10:33:26.171964] I [MSGID: 101190]
>> > [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread
>> > with index 1
>> >>
>> >> [2015-10-30 10:33:26.185685] I [MSGID: 101190]
>> > [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread
>> > with index 2
>> >>
>> >> [2015-10-30 10:33:26.186972] I [MSGID: 114020] [client.c:2118:notify]
>> > 0-gv0-client-0: parent translators are ready, attempting connect on
>> > transport
>> >>
>> >> [2015-10-30 10:33:26.191823] I [MSGID: 114020] [client.c:2118:notify]
>> > 0-gv0-client-1: parent translators are ready, attempting connect on
>> > transport
>> >>
>> >> [2015-10-30 10:33:26.192209] E [MSGID: 114058]
>> > [client-handshake.c:1524:client_query_portmap_cbk] 0-gv0-client-0:
>> > failed
>> > to get the port number for remote subvolume. Please ume status' on
>> > server
>> > to see if brick process is running.
>> >>
>> >> [2015-10-30 10:33:26.192339] I [MSGID: 114018]
>> > [client.c:2042:client_rpc_notify] 0-gv0-client-0: disconnected from
>> > gv0-client-0. Client process will keep trying to connect t brick's port
>> > is
>> > available
>> >>
>> >>
>> >>
>> >> And when I check the volumes I get:
>> >>
>> >> [root at xxx ~]# gluster volume status
>> >>
>> >> Status of volume: gv0
>> >>
>> >> Gluster process                             TCP Port  RDMA Port  Online
>> > Pid
>> >>
>> >>
>> >
>> > ------------------------------------------------------------------------------
>> >>
>> >> Brick 10.32.0.11:/glusterBrick1/gv0         N/A       N/A        N
>> > N/A
>> >>
>> >> NFS Server on localhost                     N/A       N/A        N
>> > N/A
>> >>
>> >> NFS Server on localhost                     N/A       N/A        N
>> > N/A
>> >>
>> >>
>> >>
>> >> Task Status of Volume gv0
>> >>
>> >>
>> >
>> > ------------------------------------------------------------------------------
>> >>
>> >> There are no active volume tasks
>> >>
>> >>
>> >>
>> >> If I start th second node, all is OK.
>> >>
>> >>
>> >>
>> >> Is this normal ?
>> > This behaviour is by design. In a multi node cluster when GlusterD comes
>> > up
>> > it doesn't start the bricks until it receives the configuration from its
>> > one of the friends to ensure that stale information is not been
>> > referred.
>> > In your case since the other node is down bricks are not started and
>> > hence
>> > mount fails.
>> > As a workaround, we recommend to add a dummy node to the cluster to
>> > avoid
>> > this issue.
>> >>
>> >>
>> >>
>> >> Regards,
>> >>
>> >>
>> >>
>> >> Rémi
>> >>
>> >>
>> >>
>> >>
>> >> _______________________________________________
>> >> Gluster-users mailing list
>> >> Gluster-users at gluster.org
>> >> http://www.gluster.org/mailman/listinfo/gluster-users
>> > _______________________________________________
>> > Gluster-users mailing list
>> > Gluster-users at gluster.org
>> > http://www.gluster.org/mailman/listinfo/gluster-users
>>
>>
>> --
>> Mauro Mozzarelli
>> Phone: +44 7941 727378
>> eMail: mauro at ezplanet.net
>>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users


More information about the Gluster-users mailing list