[Gluster-users] One node goes offline, the other node can't see the replicated volume anymore

Joe Julian joe at julianfamily.org
Mon Jul 15 21:36:06 UTC 2013


Ben may be on to something. If you're using NetworkManager, there may be 
a chicken and egg problem due to the lack of a hardware link being 
established (no switch). What if you mount from localhost?

On 07/15/2013 02:32 PM, Ben Turner wrote:
> Hi Greg.  I don't know if this is the thread I replied to before but it still sound to me like your NICs aren't fully up when the gluster mount is getting mounted.  The _netdev(at least the version in RHEL 6, I haven't looked at others) doesn't check if the NIC is fully up, it only looks to see if the NW manager lock file exists.  When I saw this happen in my tests the lockfile existed but the NIC was still initializing and unable to send/receive traffic to mount the FS.  I was able to put a sleep in the initscript to work around this:
>
> # diff -pruN /etc/rc.d/init.d/netfs /tmp/initrd/netfs
> --- /etc/rc.d/init.d/netfs        2013-04-26 14:32:28.759283055 -0400
> +++ /tmp/initrd/netfs        2013-04-26 14:31:38.320059175 -0400
> @@ -32,8 +32,6 @@ NETDEVMTAB=$(LC_ALL=C awk '$4 ~ /_netdev
>   # See how we were called.
>   case "$1" in
>     start)
> -        echo "Sleeping 30 seconds for NW init workaround -benT"
> -        sleep 30
>           [ ! -f /var/lock/subsys/network ] && ! nm-online -x >/dev/null 2>&1 && exit 0
>           [ "$EUID" != "0" ] && exit 4
>           [ -n "$NFSFSTAB" ] &&
>
> I just used the sleep for testing, the preferred way of dealing with this is probably using the LINKDELAY option in your /etc/sysconfig/network-scripts/ifcfg-* script.  This variable will cause the network scripts to delay $LINKDELAY number of seconds.  Can you try testing with either of those options to see if you are able to mount at boot?
>
> -b
>
> ----- Original Message -----
>> From: "Greg Scott" <GregScott at infrasupport.com>
>> To: "Joe Julian" <joe at julianfamily.org>
>> Cc: gluster-users at gluster.org
>> Sent: Monday, July 15, 2013 4:44:10 PM
>> Subject: Re: [Gluster-users] One node goes offline, the other node can't see the replicated volume anymore
>>
>> This time after rebooting both nodes, neither one shows /firewall-scripts
>> mounted after a login.   But mount -av by hand is successful on both nodes.
>> Fw1 and fw2 both behave identically.  Here is what fw1 looks like.  Fw2 is
>> identical.   This aspect of the problem is screaming timing glitch.
>>
>> login as: root
>> root at 10.10.10.71's password:
>> Last login: Mon Jul 15 15:19:41 2013 from tinahp100b.infrasupport.local
>> [root at chicago-fw1 ~]# df -h
>> Filesystem                       Size  Used Avail Use% Mounted on
>> /dev/mapper/fedora-root           14G  3.9G  8.7G  31% /
>> devtmpfs                         990M     0  990M   0% /dev
>> tmpfs                            996M     0  996M   0% /dev/shm
>> tmpfs                            996M  888K  996M   1% /run
>> tmpfs                            996M     0  996M   0% /sys/fs/cgroup
>> tmpfs                            996M     0  996M   0% /tmp
>> /dev/sda2                        477M   87M  365M  20% /boot
>> /dev/sda1                        200M  9.4M  191M   5% /boot/efi
>> /dev/mapper/fedora-gluster--fw1  7.9G   33M  7.8G   1% /gluster-fw1
>> [root at chicago-fw1 ~]# mount -av
>> /                        : ignored
>> /boot                    : already mounted
>> /boot/efi                : already mounted
>> /gluster-fw1             : already mounted
>> swap                     : ignored
>> extra arguments at end (ignored)
>> /firewall-scripts        : successfully mounted
>> [root at chicago-fw1 ~]# df -h
>> Filesystem                       Size  Used Avail Use% Mounted on
>> /dev/mapper/fedora-root           14G  3.9G  8.7G  31% /
>> devtmpfs                         990M     0  990M   0% /dev
>> tmpfs                            996M     0  996M   0% /dev/shm
>> tmpfs                            996M  888K  996M   1% /run
>> tmpfs                            996M     0  996M   0% /sys/fs/cgroup
>> tmpfs                            996M     0  996M   0% /tmp
>> /dev/sda2                        477M   87M  365M  20% /boot
>> /dev/sda1                        200M  9.4M  191M   5% /boot/efi
>> /dev/mapper/fedora-gluster--fw1  7.9G   33M  7.8G   1% /gluster-fw1
>> 192.168.253.1:/firewall-scripts  7.6G   19M  7.2G   1% /firewall-scripts
>> [root at chicago-fw1 ~]#
>>
>> - Greg
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users




More information about the Gluster-users mailing list