[Gluster-users] RE : Frequent connect and disconnect messages flooded in logs

Micha Ober micha2k at gmail.com
Tue Mar 14 11:31:59 UTC 2017


Hi,

no, the workload does not include deletion of any files or directories at
all.
The workload consists mainly of writing (creating) small files or appending
to files.​

If it is of any use, I could profile the workload.

Regards,
Micha

2017-03-09 11:46 GMT+01:00 Mohammed Rafi K C <rkavunga at redhat.com>:

> I'm sorry that you had to downgrade. We will work on it and hopefully
> will see you soon in 3.8 ;) .
>
>
> Just one question, does your workload include lot of delete either files
> or directories. We just want to see if the delayed deletes (Janitor
> thread) causing any issue .
>
>
> Regards
>
> Rafi KC
>
>
> On 03/09/2017 01:53 PM, Amar Tumballi wrote:
> >
> > ----- Original Message -----
> >> From: "Micha Ober" <micha2k at gmail.com>
> >>
> >> ​Just to let you know: I have reverted back to glusterfs 3.4.2 and
> everything
> >> is working again. No more disconnects, no more errors in the kernel
> log. So
> >> there *has* to be some kind of regression in the newer versions​.
> Sadly, I
> >> guess, it will be hard to find.
> >>
> > Thanks for the update Micha. This helps to corner the issue a little at
> least.
> >
> > Regards,
> > Amar
> >
> >
> >> 2016-12-20 13:31 GMT+01:00 Micha Ober < micha2k at gmail.com > :
> >>
> >>
> >>
> >> Hi Rafi,
> >>
> >> here are the log files:
> >>
> >> NFS: http://paste.ubuntu.com/23658653/
> >> Brick: http://paste.ubuntu.com/23658656/
> >>
> >> The brick log is of the brick which has caused the last disconnect at
> >> 2016-12-20 06:46:36 (0-gv0-client-7).
> >>
> >> For completeness, here is also dmesg output:
> >> http://paste.ubuntu.com/23658691/
> >>
> >> Regards,
> >> Micha
> >>
> >> 2016-12-19 7:28 GMT+01:00 Mohammed Rafi K C < rkavunga at redhat.com > :
> >>
> >>
> >>
> >>
> >>
> >> Hi Micha,
> >>
> >> Sorry for the late reply. I was busy with some other things.
> >>
> >> If you have still the setup available Can you enable TRACE log level
> [1],[2]
> >> and see if you could find any log entries when the network start
> >> disconnecting. Basically I'm trying to find out any disconnection had
> >> occurred other than ping timer expire issue.
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >> [1] : gluster volume <volname> diagnostics.brick-log-level TRACE
> >>
> >> [2] : gluster volume <volname> diagnostics.client-log-level TRACE
> >>
> >>
> >>
> >>
> >>
> >> Regards
> >>
> >> Rafi KC
> >>
> >> On 12/08/2016 07:59 PM, Atin Mukherjee wrote:
> >>
> >>
> >>
> >>
> >>
> >> On Thu, Dec 8, 2016 at 4:37 PM, Micha Ober < micha2k at gmail.com > wrote:
> >>
> >>
> >>
> >> Hi Rafi,
> >>
> >> thank you for your support. It is greatly appreciated.
> >>
> >> Just some more thoughts from my side:
> >>
> >> There have been no reports from other users in *this* thread until now,
> but I
> >> have found at least one user with a very simiar problem in an older
> thread:
> >>
> >> https://www.gluster.org/pipermail/gluster-users/2014-
> November/019637.html
> >>
> >> He is also reporting disconnects with no apparent reasons, althogh his
> setup
> >> is a bit more complicated, also involving a firewall. In our setup, all
> >> servers/clients are connected via 1 GbE with no firewall or anything
> that
> >> might block/throttle traffic. Also, we are using exactly the same
> software
> >> versions on all nodes.
> >>
> >>
> >> I can also find some reports in the bugtracker when searching for
> >> "rpc_client_ping_timer_expired" and "rpc_clnt_ping_timer_expired"
> (looks
> >> like spelling changed during versions).
> >>
> >> https://bugzilla.redhat.com/show_bug.cgi?id=1096729
> >>
> >> Just FYI, this is a different issue, here GlusterD fails to handle the
> volume
> >> of incoming requests on time since MT-epoll is not enabled here.
> >>
> >>
> >>
> >>
> >>
> >> https://bugzilla.redhat.com/show_bug.cgi?id=1370683
> >>
> >> But both reports involve large traffic/load on the bricks/disks, which
> is not
> >> the case for out setup.
> >> To give a ballpark figure: Over three days, 30 GiB were written. And
> the data
> >> was not written at once, but continuously over the whole time.
> >>
> >>
> >> Just to be sure, I have checked the logfiles of one of the other
> clusters
> >> right now, which are sitting in the same building, in the same rack,
> even on
> >> the same switch, running the same jobs, but with glusterfs 3.4.2 and I
> can
> >> see no disconnects in the logfiles. So I can definitely rule out our
> >> infrastructure as problem.
> >>
> >> Regards,
> >> Micha
> >>
> >>
> >>
> >> Am 07.12.2016 um 18:08 schrieb Mohammed Rafi K C:
> >>
> >>
> >>
> >>
> >> Hi Micha,
> >>
> >> This is great. I will provide you one debug build which has two fixes
> which I
> >> possible suspect for a frequent disconnect issue, though I don't have
> much
> >> data to validate my theory. So I will take one more day to dig in to
> that.
> >>
> >> Thanks for your support, and opensource++
> >>
> >> Regards
> >>
> >> Rafi KC
> >> On 12/07/2016 05:02 AM, Micha Ober wrote:
> >>
> >>
> >>
> >> Hi,
> >>
> >> thank you for your answer and even more for the question!
> >> Until now, I was using FUSE. Today I changed all mounts to NFS using
> the same
> >> 3.7.17 version.
> >>
> >> But: The problem is still the same. Now, the NFS logfile contains lines
> like
> >> these:
> >>
> >> [2016-12-06 15:12:29.006325] C
> >> [rpc-clnt-ping.c:165:rpc_clnt_ping_timer_expired] 0-gv0-client-7:
> server
> >> X.X.18.62:49153 has not responded in the last 42 seconds, disconnecting.
> >>
> >> Interestingly enough, the IP address X.X.18.62 is the same machine! As I
> >> wrote earlier, each node serves both as a server and a client, as each
> node
> >> contributes bricks to the volume. Every server is connecting to itself
> via
> >> its hostname. For example, the fstab on the node "giant2" looks like:
> >>
> >> #giant2:/gv0 /shared_data glusterfs defaults,noauto 0 0
> >> #giant2:/gv2 /shared_slurm glusterfs defaults,noauto 0 0
> >>
> >> giant2:/gv0 /shared_data nfs defaults,_netdev,vers=3 0 0
> >> giant2:/gv2 /shared_slurm nfs defaults,_netdev,vers=3 0 0
> >>
> >> So I understand the disconnects even less.
> >>
> >> I don't know if it's possible to create a dummy cluster which exposes
> the
> >> same behaviour, because the disconnects only happen when there are
> compute
> >> jobs running on those nodes - and they are GPU compute jobs, so that's
> >> something which cannot be easily emulated in a VM.
> >>
> >> As we have more clusters (which are running fine with an ancient 3.4
> version
> >> :-)) and we are currently not dependent on this particular cluster
> (which
> >> may stay like this for this month, I think) I should be able to deploy
> the
> >> debug build on the "real" cluster, if you can provide a debug build.
> >>
> >> Regards and thanks,
> >> Micha
> >>
> >>
> >>
> >> Am 06.12.2016 um 08:15 schrieb Mohammed Rafi K C:
> >>
> >>
> >>
> >>
> >>
> >>
> >> On 12/03/2016 12:56 AM, Micha Ober wrote:
> >>
> >>
> >>
> >> ** Update: ** I have downgraded from 3.8.6 to 3.7.17 now, but the
> problem
> >> still exists.
> >>
> >>
> >>
> >>
> >> Client log: http://paste.ubuntu.com/ 23569065/
> >> Brick log: http://paste.ubuntu.com/ 23569067/
> >>
> >> Please note that each server has two bricks.
> >> Whereas, according to the logs, one brick loses the connection to all
> other
> >> hosts:
> >> [2016-12-02 18:38:53.703301] W [socket.c:596:__socket_rwv]
> 0-tcp.gv0-server:
> >> writev on X.X.X.219:49121 failed (Broken pipe)
> >> [2016-12-02 18:38:53.703381] W [socket.c:596:__socket_rwv]
> 0-tcp.gv0-server:
> >> writev on X.X.X.62:49118 failed (Broken pipe)
> >> [2016-12-02 18:38:53.703380] W [socket.c:596:__socket_rwv]
> 0-tcp.gv0-server:
> >> writev on X.X.X.107:49121 failed (Broken pipe)
> >> [2016-12-02 18:38:53.703424] W [socket.c:596:__socket_rwv]
> 0-tcp.gv0-server:
> >> writev on X.X.X.206:49120 failed (Broken pipe)
> >> [2016-12-02 18:38:53.703359] W [socket.c:596:__socket_rwv]
> 0-tcp.gv0-server:
> >> writev on X.X.X.58:49121 failed (Broken pipe)
> >>
> >> The SECOND brick on the SAME host is NOT affected, i.e. no disconnects!
> >> As I said, the network connection is fine and the disks are idle.
> >> The CPU always has 2 free cores.
> >>
> >> It looks like I have to downgrade to 3.4 now in order for the
> disconnects to
> >> stop.
> >>
> >> Hi Micha,
> >>
> >> Thanks for the update and sorry for what happened with gluster higher
> >> versions. I can understand the need for downgrade as it is a production
> >> setup.
> >>
> >> Can you tell me the clients used here ? whether it is a
> fuse,nfs,nfs-ganesha,
> >> smb or libgfapi ?
> >>
> >> Since I'm not able to reproduce the issue (I have been trying from last
> >> 3days) and the logs are not much helpful here (we don't have much logs
> in
> >> socket layer), Could you please create a dummy cluster and try to
> reproduce
> >> the issue? If then we can play with that volume and I could provide some
> >> debug build which we can use for further debugging?
> >>
> >> If you don't have bandwidth for this, please leave it ;).
> >>
> >> Regards
> >> Rafi KC
> >>
> >>
> >>
> >>
> >> - Micha
> >>
> >> Am 30.11.2016 um 06:57 schrieb Mohammed Rafi K C:
> >>
> >>
> >>
> >>
> >> Hi Micha,
> >>
> >> I have changed the thread and subject so that your original thread
> remain
> >> same for your query. Let's try to fix the problem what you observed with
> >> 3.8.4, So I have started a new thread to discuss the frequent disconnect
> >> problem.
> >>
> >> If any one else has experienced the same problem, please respond to the
> mail.
> >>
> >>
> >> It would be very helpful if you could give us some more logs from
> clients and
> >> bricks. Also any reproducible steps will surely help to chase the
> problem
> >> further.
> >>
> >> Regards
> >>
> >> Rafi KC
> >> On 11/30/2016 04:44 AM, Micha Ober wrote:
> >>
> >>
> >>
> >> I had opened another thread on this mailing list (Subject: "After
> upgrade
> >> from 3.4.2 to 3.8.5 - High CPU usage resulting in disconnects and
> >> split-brain").
> >>
> >> The title may be a bit misleading now, as I am no longer observing high
> CPU
> >> usage after upgrading to 3.8.6, but the disconnects are still happening
> and
> >> the number of files in split-brain is growing.
> >>
> >> Setup: 6 compute nodes, each serving as a glusterfs server and client,
> Ubuntu
> >> 14.04, two bricks per node, distribute-replicate
> >>
> >> I have two gluster volumes set up (one for scratch data, one for the
> slurm
> >> scheduler). Only the scratch data volume shows critical errors "[...]
> has
> >> not responded in the last 42 seconds, disconnecting.". So I can rule out
> >> network problems, the gigabit link between the nodes is not saturated at
> >> all. The disks are almost idle (<10%).
> >>
> >> I have glusterfs 3.4.2 on Ubuntu 12.04 on a another compute cluster,
> running
> >> fine since it was deployed.
> >> I had glusterfs 3.4.2 on Ubuntu 14.04 on this cluster, running fine for
> >> almost a year.
> >>
> >> After upgrading to 3.8.5, the problems (as described) started. I would
> like
> >> to use some of the new features of the newer versions (like bitrot),
> but the
> >> users can't run their compute jobs right now because the result files
> are
> >> garbled.
> >>
> >> There also seems to be a bug report with a smiliar problem: (but no
> progress)
> >> https://bugzilla.redhat.com/ show_bug.cgi?id=1370683
> >>
> >> For me, ALL servers are affected (not isolated to one or two servers)
> >>
> >> I also see messages like "INFO: task gpu_graphene_bv:4476 blocked for
> more
> >> than 120 seconds." in the syslog.
> >>
> >> For completeness (gv0 is the scratch volume, gv2 the slurm volume):
> >>
> >> [root at giant2: ~]# gluster v info
> >>
> >> Volume Name: gv0
> >> Type: Distributed-Replicate
> >> Volume ID: 993ec7c9-e4bc-44d0-b7c4-2d977e622e86
> >> Status: Started
> >> Snapshot Count: 0
> >> Number of Bricks: 6 x 2 = 12
> >> Transport-type: tcp
> >> Bricks:
> >> Brick1: giant1:/gluster/sdc/gv0
> >> Brick2: giant2:/gluster/sdc/gv0
> >> Brick3: giant3:/gluster/sdc/gv0
> >> Brick4: giant4:/gluster/sdc/gv0
> >> Brick5: giant5:/gluster/sdc/gv0
> >> Brick6: giant6:/gluster/sdc/gv0
> >> Brick7: giant1:/gluster/sdd/gv0
> >> Brick8: giant2:/gluster/sdd/gv0
> >> Brick9: giant3:/gluster/sdd/gv0
> >> Brick10: giant4:/gluster/sdd/gv0
> >> Brick11: giant5:/gluster/sdd/gv0
> >> Brick12: giant6:/gluster/sdd/gv0
> >> Options Reconfigured:
> >> auth.allow: X.X.X.*,127.0.0.1
> >> nfs.disable: on
> >>
> >> Volume Name: gv2
> >> Type: Replicate
> >> Volume ID: 30c78928-5f2c-4671-becc-8deaee1a7a8d
> >> Status: Started
> >> Snapshot Count: 0
> >> Number of Bricks: 1 x 2 = 2
> >> Transport-type: tcp
> >> Bricks:
> >> Brick1: giant1:/gluster/sdd/gv2
> >> Brick2: giant2:/gluster/sdd/gv2
> >> Options Reconfigured:
> >> auth.allow: X.X.X.*,127.0.0.1
> >> cluster.granular-entry-heal: on
> >> cluster.locking-scheme: granular
> >> nfs.disable: on
> >>
> >>
> >> 2016-11-30 0:10 GMT+01:00 Micha Ober < micha2k at gmail.com > :
> >>
> >>
> >>
> >> There also seems to be a bug report with a smiliar problem: (but no
> progress)
> >> https://bugzilla.redhat.com/sh ow_bug.cgi?id=1370683
> >>
> >> For me, ALL servers are affected (not isolated to one or two servers)
> >>
> >> I also see messages like "INFO: task gpu_graphene_bv:4476 blocked for
> more
> >> than 120 seconds." in the syslog.
> >>
> >> For completeness (gv0 is the scratch volume, gv2 the slurm volume):
> >>
> >> [root at giant2: ~]# gluster v info
> >>
> >> Volume Name: gv0
> >> Type: Distributed-Replicate
> >> Volume ID: 993ec7c9-e4bc-44d0-b7c4-2d977e622e86
> >> Status: Started
> >> Snapshot Count: 0
> >> Number of Bricks: 6 x 2 = 12
> >> Transport-type: tcp
> >> Bricks:
> >> Brick1: giant1:/gluster/sdc/gv0
> >> Brick2: giant2:/gluster/sdc/gv0
> >> Brick3: giant3:/gluster/sdc/gv0
> >> Brick4: giant4:/gluster/sdc/gv0
> >> Brick5: giant5:/gluster/sdc/gv0
> >> Brick6: giant6:/gluster/sdc/gv0
> >> Brick7: giant1:/gluster/sdd/gv0
> >> Brick8: giant2:/gluster/sdd/gv0
> >> Brick9: giant3:/gluster/sdd/gv0
> >> Brick10: giant4:/gluster/sdd/gv0
> >> Brick11: giant5:/gluster/sdd/gv0
> >> Brick12: giant6:/gluster/sdd/gv0
> >> Options Reconfigured:
> >> auth.allow: X.X.X.*,127.0.0.1
> >> nfs.disable: on
> >>
> >> Volume Name: gv2
> >> Type: Replicate
> >> Volume ID: 30c78928-5f2c-4671-becc-8deaee1a7a8d
> >> Status: Started
> >> Snapshot Count: 0
> >> Number of Bricks: 1 x 2 = 2
> >> Transport-type: tcp
> >> Bricks:
> >> Brick1: giant1:/gluster/sdd/gv2
> >> Brick2: giant2:/gluster/sdd/gv2
> >> Options Reconfigured:
> >> auth.allow: X.X.X.*,127.0.0.1
> >> cluster.granular-entry-heal: on
> >> cluster.locking-scheme: granular
> >> nfs.disable: on
> >>
> >>
> >> 2016-11-29 19:21 GMT+01:00 Micha Ober < micha2k at gmail.com > :
> >>
> >>
> >>
> >> I had opened another thread on this mailing list (Subject: "After
> upgrade
> >> from 3.4.2 to 3.8.5 - High CPU usage resulting in disconnects and
> >> split-brain").
> >>
> >> The title may be a bit misleading now, as I am no longer observing high
> CPU
> >> usage after upgrading to 3.8.6, but the disconnects are still happening
> and
> >> the number of files in split-brain is growing.
> >>
> >> Setup: 6 compute nodes, each serving as a glusterfs server and client,
> Ubuntu
> >> 14.04, two bricks per node, distribute-replicate
> >>
> >> I have two gluster volumes set up (one for scratch data, one for the
> slurm
> >> scheduler). Only the scratch data volume shows critical errors "[...]
> has
> >> not responded in the last 42 seconds, disconnecting.". So I can rule out
> >> network problems, the gigabit link between the nodes is not saturated at
> >> all. The disks are almost idle (<10%).
> >>
> >> I have glusterfs 3.4.2 on Ubuntu 12.04 on a another compute cluster,
> running
> >> fine since it was deployed.
> >> I had glusterfs 3.4.2 on Ubuntu 14.04 on this cluster, running fine for
> >> almost a year.
> >>
> >> After upgrading to 3.8.5, the problems (as described) started. I would
> like
> >> to use some of the new features of the newer versions (like bitrot),
> but the
> >> users can't run their compute jobs right now because the result files
> are
> >> garbled.
> >>
> >> 2016-11-29 18:53 GMT+01:00 Atin Mukherjee < amukherj at redhat.com > :
> >>
> >>
> >>
> >> Would you be able to share what is not working for you in 3.8.x
> (mention the
> >> exact version). 3.4 is quite old and falling back to an unsupported
> version
> >> doesn't look a feasible option.
> >>
> >> On Tue, 29 Nov 2016 at 17:01, Micha Ober < micha2k at gmail.com > wrote:
> >>
> >>
> >>
> >> Hi,
> >>
> >> I was using gluster 3.4 and upgraded to 3.8, but that version showed to
> be
> >> unusable for me. I now need to downgrade.
> >>
> >> I'm running Ubuntu 14.04. As upgrades of the op version are
> irreversible, I
> >> guess I have to delete all gluster volumes and re-create them with the
> >> downgraded version.
> >>
> >> 0. Backup data
> >> 1. Unmount all gluster volumes
> >> 2. apt-get purge glusterfs-server glusterfs-client
> >> 3. Remove PPA for 3.8
> >> 4. Add PPA for older version
> >> 5. apt-get install glusterfs-server glusterfs-client
> >> 6. Create volumes
> >>
> >> Is "purge" enough to delete all configuration files of the currently
> >> installed version or do I need to manually clear some residues before
> >> installing an older version?
> >>
> >> Thanks.
> >> _______________________________________________
> >> Gluster-users mailing list
> >> Gluster-users at gluster.org
> >> http://www.gluster.org/mailman /listinfo/gluster-users
> >> --
> >> - Atin (atinm)
> >>
> >>
> >>
> >>
> >>
> >> _______________________________________________
> >> Gluster-users mailing list Gluster-users at gluster.org
> >> http://www.gluster.org/mailman/listinfo/gluster-users
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >> --
> >> ~ Atin (atinm)
> >>
> >>
> >>
> >> _______________________________________________
> >> Gluster-users mailing list
> >> Gluster-users at gluster.org
> >> http://lists.gluster.org/mailman/listinfo/gluster-users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20170314/ba3f4912/attachment.html>


More information about the Gluster-users mailing list