[Gluster-users] Mount problems when secondary node down
Pranith Kumar Karampuri
pkarampu at redhat.com
Tue Nov 11 01:19:50 UTC 2014
On 11/10/2014 11:47 PM, A F wrote:
> Hello,
>
> I have two servers, 192.168.0.10 and 192.168.2.10. I'm using gluster
> 3.6.1 (installed from gluster repo) on AWS Linux. Both servers are
> completely reachable in LAN.
> # rpm -qa|grep gluster
> glusterfs-3.6.1-1.el6.x86_64
> glusterfs-server-3.6.1-1.el6.x86_64
> glusterfs-libs-3.6.1-1.el6.x86_64
> glusterfs-api-3.6.1-1.el6.x86_64
> glusterfs-cli-3.6.1-1.el6.x86_64
> glusterfs-fuse-3.6.1-1.el6.x86_64
>
> These are the commands I ran:
> # gluster peer probe 192.168.2.10
> # gluster volume create aloha replica 2 transport tcp
> 192.168.0.10:/var/aloha 192.168.2.10:/var/aloha force
> # gluster volume start aloha
> # gluster volume set aloha network.ping-timeout 5
> # gluster volume set aloha nfs.disable on
>
> Problem number 1:
> tail -f /var/log/glusterfs/etc-glusterfs-glusterd.vol.log shows log
> cluttering with:
> [2014-11-10 17:41:26.328796] W [socket.c:611:__socket_rwv]
> 0-management: readv on
> /var/run/38c520c774793c9cdae8ace327512027.socket failed (Invalid
> argument)
> this happens every 3 seconds on both servers. It is related to NFS and
> probably rpcbind, but I absolutely want them disabled. As you see,
> I've set gluster to disable nfs - why doesn't it keep quiet about it
> then?
>
> Problem number 2:
> in fstab on server 192.168.0.10: 192.168.0.10:/aloha /var/www/hawaii
> glusterfs defaults,_netdev 0 0
> in fstab on server 192.168.2.10: 192.168.2.10:/aloha
> /var/www/hawaii glusterfs defaults,_netdev 0 0
>
> If I shutdown one of the servers (192.168.2.10), and I reboot the
> remaining one (192.168.0.10), it won't come up as fast as it should.
> It lags a few minutes waiting for gluster. After it eventually starts,
> mount point is not mounted and volume is stopped:
> # gluster volume status
> Status of volume: aloha
> Gluster process Port Online Pid
> ------------------------------------------------------------------------------
>
> Brick 192.168.0.10:/var/aloha N/A N N/A
> Self-heal Daemon on localhost N/A N N/A
>
> Task Status of Volume aloha
> ------------------------------------------------------------------------------
>
> There are no active volume tasks
>
> This didn't happen before, so fine, I first have to stop the volume
> and then start it again. It now shows as online:
> Brick 192.168.0.10:/var/aloha 49155 Y
> 3473
> Self-heal Daemon on localhost N/A Y 3507
>
> # time mount -a
> real 2m7.307s
>
> # time mount -t glusterfs 192.168.0.10:/aloha /var/www/hawaii
> real 2m7.365s
>
> # strace mount -t glusterfs 192.168.0.10:/aloha /var/www/hawaii
> (attached)
>
> # tail /var/log/glusterfs/* -f|grep -v readv
> (attached)
>
> I've done this setup before, so I'm amazed it doesn't work. I even
> have it in production at the moment, with the same options and setup,
> and for example I'm not getting readv errors. I'm unable to test the
> mount part though, but I feel I have covered it way back when I was
> testing the environment.
> Any help is kindly appreciated.
CC glusterd folks
Pranith
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20141111/b43ae35f/attachment.html>
More information about the Gluster-users
mailing list