[Gluster-users] One node goes offline, the other node can't see the replicated volume anymore
Ben Turner
bturner at redhat.com
Mon Jul 15 21:32:00 UTC 2013
Hi Greg. I don't know if this is the thread I replied to before but it still sound to me like your NICs aren't fully up when the gluster mount is getting mounted. The _netdev(at least the version in RHEL 6, I haven't looked at others) doesn't check if the NIC is fully up, it only looks to see if the NW manager lock file exists. When I saw this happen in my tests the lockfile existed but the NIC was still initializing and unable to send/receive traffic to mount the FS. I was able to put a sleep in the initscript to work around this:
# diff -pruN /etc/rc.d/init.d/netfs /tmp/initrd/netfs
--- /etc/rc.d/init.d/netfs 2013-04-26 14:32:28.759283055 -0400
+++ /tmp/initrd/netfs 2013-04-26 14:31:38.320059175 -0400
@@ -32,8 +32,6 @@ NETDEVMTAB=$(LC_ALL=C awk '$4 ~ /_netdev
# See how we were called.
case "$1" in
start)
- echo "Sleeping 30 seconds for NW init workaround -benT"
- sleep 30
[ ! -f /var/lock/subsys/network ] && ! nm-online -x >/dev/null 2>&1 && exit 0
[ "$EUID" != "0" ] && exit 4
[ -n "$NFSFSTAB" ] &&
I just used the sleep for testing, the preferred way of dealing with this is probably using the LINKDELAY option in your /etc/sysconfig/network-scripts/ifcfg-* script. This variable will cause the network scripts to delay $LINKDELAY number of seconds. Can you try testing with either of those options to see if you are able to mount at boot?
-b
----- Original Message -----
> From: "Greg Scott" <GregScott at infrasupport.com>
> To: "Joe Julian" <joe at julianfamily.org>
> Cc: gluster-users at gluster.org
> Sent: Monday, July 15, 2013 4:44:10 PM
> Subject: Re: [Gluster-users] One node goes offline, the other node can't see the replicated volume anymore
>
> This time after rebooting both nodes, neither one shows /firewall-scripts
> mounted after a login. But mount -av by hand is successful on both nodes.
> Fw1 and fw2 both behave identically. Here is what fw1 looks like. Fw2 is
> identical. This aspect of the problem is screaming timing glitch.
>
> login as: root
> root at 10.10.10.71's password:
> Last login: Mon Jul 15 15:19:41 2013 from tinahp100b.infrasupport.local
> [root at chicago-fw1 ~]# df -h
> Filesystem Size Used Avail Use% Mounted on
> /dev/mapper/fedora-root 14G 3.9G 8.7G 31% /
> devtmpfs 990M 0 990M 0% /dev
> tmpfs 996M 0 996M 0% /dev/shm
> tmpfs 996M 888K 996M 1% /run
> tmpfs 996M 0 996M 0% /sys/fs/cgroup
> tmpfs 996M 0 996M 0% /tmp
> /dev/sda2 477M 87M 365M 20% /boot
> /dev/sda1 200M 9.4M 191M 5% /boot/efi
> /dev/mapper/fedora-gluster--fw1 7.9G 33M 7.8G 1% /gluster-fw1
> [root at chicago-fw1 ~]# mount -av
> / : ignored
> /boot : already mounted
> /boot/efi : already mounted
> /gluster-fw1 : already mounted
> swap : ignored
> extra arguments at end (ignored)
> /firewall-scripts : successfully mounted
> [root at chicago-fw1 ~]# df -h
> Filesystem Size Used Avail Use% Mounted on
> /dev/mapper/fedora-root 14G 3.9G 8.7G 31% /
> devtmpfs 990M 0 990M 0% /dev
> tmpfs 996M 0 996M 0% /dev/shm
> tmpfs 996M 888K 996M 1% /run
> tmpfs 996M 0 996M 0% /sys/fs/cgroup
> tmpfs 996M 0 996M 0% /tmp
> /dev/sda2 477M 87M 365M 20% /boot
> /dev/sda1 200M 9.4M 191M 5% /boot/efi
> /dev/mapper/fedora-gluster--fw1 7.9G 33M 7.8G 1% /gluster-fw1
> 192.168.253.1:/firewall-scripts 7.6G 19M 7.2G 1% /firewall-scripts
> [root at chicago-fw1 ~]#
>
> - Greg
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
More information about the Gluster-users
mailing list