[Gluster-users] How-to start gluster when only one node is up ?

Atin Mukherjee atin.mukherjee83 at gmail.com
Fri Oct 30 13:06:24 UTC 2015


gluster peer probe <node ip>

-Atin
Sent from one plus one
On Oct 30, 2015 6:34 PM, "Peter Michael Calum" <pemca at tdc.dk> wrote:

> Hi,
>
>
>
> How do i  add a ’dummy’ node ?
>
>
>
> thanks,
>
> Peter
>
>
>
> *Fra:* gluster-users-bounces at gluster.org [mailto:
> gluster-users-bounces at gluster.org] *På vegne af *Atin Mukherjee
> *Sendt:* 30. oktober 2015 13:14
> *Til:* Mauro Mozzarelli
> *Cc:* gluster-users at gluster.org
> *Emne:* Re: [Gluster-users] How-to start gluster when only one node is up
> ?
>
>
>
> -Atin
> Sent from one plus one
> On Oct 30, 2015 5:28 PM, "Mauro Mozzarelli" <mauro at ezplanet.net> wrote:
> >
> > Hi,
> >
> > Atin keeps giving the same answer: "it is by design"
> >
> > I keep saying "the design is wrong and it should be changed to cater for
> > standby servers"
> Every design has got its own set of limitations and i would say this is a
> limitation instead of mentioning the overall design itself wrong. I would
> again stand with my points as correctness is always a priority in a
> distributed system. This behavioural change was introduced in 3.5 and if
> this was not included as part of release note I apologize on behalf of the
> release management.
> As communicated earlier, we will definitely resolve this issue in
> GlusterD2.
> >
> > In the meantime this is the workaround I am using:
> > When the single node starts I stop and start the volume, and then it
> > becomes mountable. On CentOS 6 and CentOS 7 it works with release up to
> > 3.7.4. Release 3.7.5 is broken so I reverted back to 3.7.4.
> This is where I am not convinced. An explicit volume start should start
> the bricks, can you raise a BZ with all the relevant details?
> >
> > In my experience glusterfs releases are a bit of a hit and miss. Often
> > something stops working with newer releases, then after a few more
> > releases it works again or there is a workaround ... Not quite the
> > stability one would want for commercial use, and thus at the moment I can
> > risk using it only for my home servers, hence the cluster with a node
> > always ON and the second as STANDBY.
> >
> > MOUNT=/home
> > LABEL="GlusterFS:"
> > if grep -qs $MOUNT /proc/mounts; then
> >     echo "$LABEL $MOUNT is mounted";
> >     gluster volume start gv_home 2>/dev/null
> > else
> >     echo "$LABEL $MOUNT is NOT mounted";
> >     echo "$LABEL Restarting gluster volume ..."
> >     yes|gluster volume stop gv_home > /dev/null
> >     gluster volume start gv_home
> >     mount -t glusterfs sirius-ib:/gv_home $MOUNT;
> >     if grep -qs $MOUNT /proc/mounts; then
> >         echo "$LABEL $MOUNT is mounted";
> >         gluster volume start gv_home 2>/dev/null
> >     else
> >         echo "$LABEL failure to mount $MOUNT";
> >     fi
> > fi
> >
> > I hope this helps.
> > Mauro
> >
> > On Fri, October 30, 2015 11:48, Atin Mukherjee wrote:
> > > -Atin
> > > Sent from one plus one
> > > On Oct 30, 2015 4:35 PM, "Remi Serrano" <rserrano at pros.com> wrote:
> > >>
> > >> Hello,
> > >>
> > >>
> > >>
> > >> I setup a gluster file cluster with 2 nodes. It works fine.
> > >>
> > >> But, when I shut down the 2 nodes, and startup only one node, I cannot
> > > mount the share :
> > >>
> > >>
> > >>
> > >> [root at xxx ~]#  mount -t glusterfs 10.32.0.11:/gv0 /glusterLocalShare
> > >>
> > >> Mount failed. Please check the log file for more details.
> > >>
> > >>
> > >>
> > >> Log says :
> > >>
> > >> [2015-10-30 10:33:26.147003] I [MSGID: 100030]
> [glusterfsd.c:2318:main]
> > > 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version
> 3.7.5
> > > (args: /usr/sbin/glusterfs -127.0.0.1 --volfile-id=/gv0
> > > /glusterLocalShare)
> > >>
> > >> [2015-10-30 10:33:26.171964] I [MSGID: 101190]
> > > [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread
> > > with index 1
> > >>
> > >> [2015-10-30 10:33:26.185685] I [MSGID: 101190]
> > > [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread
> > > with index 2
> > >>
> > >> [2015-10-30 10:33:26.186972] I [MSGID: 114020] [client.c:2118:notify]
> > > 0-gv0-client-0: parent translators are ready, attempting connect on
> > > transport
> > >>
> > >> [2015-10-30 10:33:26.191823] I [MSGID: 114020] [client.c:2118:notify]
> > > 0-gv0-client-1: parent translators are ready, attempting connect on
> > > transport
> > >>
> > >> [2015-10-30 10:33:26.192209] E [MSGID: 114058]
> > > [client-handshake.c:1524:client_query_portmap_cbk] 0-gv0-client-0:
> failed
> > > to get the port number for remote subvolume. Please ume status' on
> server
> > > to see if brick process is running.
> > >>
> > >> [2015-10-30 10:33:26.192339] I [MSGID: 114018]
> > > [client.c:2042:client_rpc_notify] 0-gv0-client-0: disconnected from
> > > gv0-client-0. Client process will keep trying to connect t brick's
> port is
> > > available
> > >>
> > >>
> > >>
> > >> And when I check the volumes I get:
> > >>
> > >> [root at xxx ~]# gluster volume status
> > >>
> > >> Status of volume: gv0
> > >>
> > >> Gluster process                             TCP Port  RDMA Port
> Online
> > > Pid
> > >>
> > >>
> > >
> ------------------------------------------------------------------------------
> > >>
> > >> Brick 10.32.0.11:/glusterBrick1/gv0         N/A       N/A        N
> > > N/A
> > >>
> > >> NFS Server on localhost                     N/A       N/A        N
> > > N/A
> > >>
> > >> NFS Server on localhost                     N/A       N/A        N
> > > N/A
> > >>
> > >>
> > >>
> > >> Task Status of Volume gv0
> > >>
> > >>
> > >
> ------------------------------------------------------------------------------
> > >>
> > >> There are no active volume tasks
> > >>
> > >>
> > >>
> > >> If I start th second node, all is OK.
> > >>
> > >>
> > >>
> > >> Is this normal ?
> > > This behaviour is by design. In a multi node cluster when GlusterD
> comes
> > > up
> > > it doesn't start the bricks until it receives the configuration from
> its
> > > one of the friends to ensure that stale information is not been
> referred.
> > > In your case since the other node is down bricks are not started and
> hence
> > > mount fails.
> > > As a workaround, we recommend to add a dummy node to the cluster to
> avoid
> > > this issue.
> > >>
> > >>
> > >>
> > >> Regards,
> > >>
> > >>
> > >>
> > >> Rémi
> > >>
> > >>
> > >>
> > >>
> > >> _______________________________________________
> > >> Gluster-users mailing list
> > >> Gluster-users at gluster.org
> > >> http://www.gluster.org/mailman/listinfo/gluster-users
> > > _______________________________________________
> > > Gluster-users mailing list
> > > Gluster-users at gluster.org
> > > http://www.gluster.org/mailman/listinfo/gluster-users
> >
> >
> > --
> > Mauro Mozzarelli
> > Phone: +44 7941 727378
> > eMail: mauro at ezplanet.net
> >
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://www.gluster.org/mailman/listinfo/gluster-users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20151030/d2867519/attachment.html>


More information about the Gluster-users mailing list