[Gluster-users] Gluster volume not automounted when peer is down

A F alex at cartell.ie
Wed Nov 26 00:59:27 UTC 2014


Hi,

This is interesting, I think this is the same problem as mine.

Quoting myself from 1st message:
If I shutdown one of the servers (192.168.2.10), and I reboot the
remaining one (192.168.0.10), it won't come up as fast as it should. It
lags a few minutes waiting for gluster. After it eventually starts,
mount point is not mounted and volume is stopped:
[...]
This didn't happen before, so fine, I first have to stop the volume and
then start it again. It now shows as online:
Brick 192.168.0.10:/var/aloha                           49155 Y       3473
Self-heal Daemon on localhost                           N/A Y       3507

# time mount -a
real    2m7.307s

# time mount -t glusterfs 192.168.0.10:/aloha /var/www/hawaii
real    2m7.365s
------------------------------------------------------------------------
Quoting myself from 2nd message:
Steps to reproduce:
1. probe server2, create and start volume
2. do not mount volume
3. reboot/poweroff server2; or add server1 to its iptables (with -j
DROP, not -j REJECT)
4. on server1 (while server2 is rebooting or dropping traffic from
server1): time mount -t glusterfs server1:/volume /some/path

PS: with -j REJECT it mounts instantly. with -j DROP it always waits
2mins 7secs
------------------------------------------------------------------------

Links at
http://supercolony.gluster.org/pipermail/gluster-users/2014-November/019415.html
http://supercolony.gluster.org/pipermail/gluster-users/2014-November/019488.html

Can you also take a look and confirm it is the same? Also, while keeping one node down, please try that "time mount" command.



On Nov 24, 2014 8:07 PM, Eric Ewanco wrote:

Hi all,

We're trying to use gluster as a replicated volume.  It works OK when both peers are up but when one peer is down and the other reboots, the "surviving" peer does not automount glusterfs.  Furthermore, after the boot sequence is complete, it can be mounted without issue.  It automounts fine when the peer is up during startup.  I tried to google this and while I found some similar issues, I haven't found any solutions to my problem.  Any insight would be appreciated.  Thanks.

gluster volume info output (after startup):
Volume Name: rel-vol
Type: Replicate
Volume ID: 90cbe313-e9f9-42d9-a947-802315ab72b0
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 10.250.1.1:/export/brick1
Brick2: 10.250.1.2:/export/brick1

gluster peer status output (after startup):
Number of Peers: 1

Hostname: 10.250.1.2
Uuid: 8d49b929-4660-4b1e-821b-bfcd6291f516
State: Peer in Cluster (Disconnected)

Original volume create command:
gluster volume create rel-vol rep 2 transport tcp 10.250.1.1:/export/brick1 10.250.1.2:/export/brick1

I am running Gluster 3.4.5 on OpenSuSE 12.2.
gluster --version:
glusterfs 3.4.5 built on Jul 25 2014 08:31:19
Repository revision:git://git.gluster.com/glusterfs.git
Copyright (c) 2006-2011 Gluster Inc. <http://www.gluster.com>
GlusterFS comes with ABSOLUTELY NO WARRANTY.
You may redistribute copies of GlusterFS under the terms of the GNU General Public License.

The fstab line is:

localhost:/rel-vol              /home        glusterfs  defaults,_netdev      0 0

lsof -i :24007-24100:
COMMAND    PID USER   FD   TYPE DEVICE SIZE/OFF NODE NAME
glusterd  4073 root    6u  IPv4  82170      0t0  TCP s1:24007->s1:1023 (ESTABLISHED)
glusterd  4073 root    9u  IPv4  13816      0t0  TCP *:24007 (LISTEN)
glusterd  4073 root   10u  IPv4  88106      0t0  TCP s1:exp2->s2:24007 (SYN_SENT)
glusterfs 4097 root    8u  IPv4  16751      0t0  TCP s1:1023->s1:24007 (ESTABLISHED)

This is shorter than it is when it works, but maybe that's because the mount spawns some more processes.


Some ports are down:



root at q50-s1  <http://supercolony.gluster.org/mailman/listinfo/gluster-users>:/root> telnet localhost 24007

Trying ::1...

telnet: connect to address ::1: Connection refused

Trying 127.0.0.1...

Connected to localhost.

Escape character is '^]'.



telnet> close

Connection closed.

root at q50-s1  <http://supercolony.gluster.org/mailman/listinfo/gluster-users>:/root> telnet localhost 24009

Trying ::1...

telnet: connect to address ::1: Connection refused

Trying 127.0.0.1...

telnet: connect to address 127.0.0.1: Connection refused

ps axww | fgrep glu:

4073 ?        Ssl    0:10 /usr/sbin/glusterd -p /run/glusterd.pid

4097 ?        Ssl    0:00 /usr/sbin/glusterfsd -s 10.250.1.1 --volfile-id rel-vol.10.250.1.1.export-brick1 -p /var/lib/glusterd/vols/rel-vol/run/10.250.1.1-export-brick1.pid -S /var/run/89ba432ed09e07e107723b4b266e18f9.socket --brick-name /export/brick1 -l /var/log/glusterfs/bricks/export-brick1.log --xlator-option *-posix.glusterd-uuid=3b02a581-8fb9-4c6a-8323-9463262f23bc --brick-port 49152 --xlator-option rel-vol-server.listen-port=49152

5949 ttyS0    S+     0:00 fgrep glu

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20141126/cb1935f9/attachment.html>


More information about the Gluster-users mailing list