[Gluster-users] New to GlusterFS
haiwei.xie-soulinfo
haiwei.xie at soulinfo.com
Tue Oct 22 13:32:52 UTC 2013
hi,
I use 5 second to enable gluster-nfs ip shift when nodes poweroff,
but in heavy IO load, client often timeout. maybe need longer timeout.
> The reason for the long (42 second) ping-timeout is because
> re-establishing fd's and locks can be a very expensive operation.
> Allowing a longer time to reestablish connections is logical, unless you
> have servers that frequently die.
>
> If you shut down a server through the normal kill process, the TCP
> connections will be closed properly. The client will be aware that the
> server is going away and there will be no timeout. This allows server
> maintenance without encountering that issue.
>
> One issue with a 42 second timeout is that ext4 may detect an error and
> remount itself read only should that happen while the VM is running. You
> can override this behavior by specifying the mount option,
> "errors=continue" in fstab ("errors=remount-ro" is the default). The
> default can be changed, as well, by changing the superblock option with
> tune2fs.
>
> On 10/22/2013 03:12 AM, John Mark Walker wrote:
> >
> > Hi JC,
> >
> > Yes, the default is a 42-second timeout for failover. You can
> > configure that to be a smaller window.
> >
> > -JM
> >
> > On Oct 22, 2013 10:57 AM, "JC Putter" <jcputter at gmail.com
> > <mailto:jcputter at gmail.com>> wrote:
> >
> > Hi,
> >
> > I am new to GlusterFS, i am trying to accomplish something which i am
> > not 100% sure is the correct use case but hear me out.
> >
> > I want to use GlusterFS to host KVM VM's, from what I've read this was
> > not recommended due to poor write performance however since
> > libgfapi/qemu 1.3 this is now viable ?
> >
> >
> > Currently i'am testing out GlusterFS with two nodes, both running as
> > server and client
> >
> > i have the following Volume:
> >
> > Volume Name: DATA
> > Type: Replicate
> > Volume ID: eaa7746b-a1c1-4959-ad7d-743ac519f86a
> > Status: Started
> > Number of Bricks: 1 x 2 = 2
> > Transport-type: tcp
> > Bricks:
> > Brick1: glusterfs1.example.com:/data
> > Brick2: glusterfs2.example.com:/data
> >
> >
> > and mounting the brick locally on each server as /mnt/gluster,
> > replication works and everything but as soon as i kill one node, the
> > directory /mnt/gluster/ becomes unavailable for 30/40 seconds
> >
> > log shows
> >
> > [2013-10-22 11:55:48.055571] W [socket.c:514:__socket_rwv]
> > 0-DATA-client-0: readv failed (No data available)
> >
> >
> > Thanks in advance!
> > _______________________________________________
> > Gluster-users mailing list
> > Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
> > http://supercolony.gluster.org/mailman/listinfo/gluster-users
> >
> >
> >
> > _______________________________________________
> > Gluster-users mailing list
> > Gluster-users at gluster.org
> > http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
More information about the Gluster-users
mailing list