[Gluster-users] Gluster trouble

Kaushal M kshlmster at gmail.com
Fri Jul 5 08:53:01 UTC 2013


On Wed, Jul 3, 2013 at 8:25 PM, Marcus Bointon
<marcus at synchromedia.co.uk> wrote:
> Back in March I posted about some gluster problems:
>
> http://gluster.org/pipermail/gluster-users/2013-March/035737.html
> http://gluster.org/pipermail/gluster-users/2013-March/035655.html
>
> I'm still in the same situation - a straightforward 2-node, 2-way AFR setup with each server mounting the single shared volume via NFS using gluster 3.3.0 (can't use 3.3.1 dues to its NFS issues) on 64-bit linux (ubuntu lucid). Gluster appears to be working, but won't mount on boot by any means I've tried, and it's still logging prodigious amounts of incomprehensible rubbish (to me!).
>
> gluster says everything is ok:
>
> gluster volume status
> Status of volume: shared
> Gluster process                                         Port    Online  Pid
> ------------------------------------------------------------------------------
> Brick 192.168.1.10:/var/shared                          24009   Y       3097
> Brick 192.168.1.11:/var/shared                          24009   Y       3020
> NFS Server on localhost                                 38467   Y       3103
> Self-heal Daemon on localhost                           N/A     Y       3109
> NFS Server on 192.168.1.11                              38467   Y       3057
> Self-heal Daemon on 192.168.1.11                        N/A     Y       3096
>
> (other node says the same thing with IPs the other way around)
>
> Yet the logs tell a different story.
>
> In syslog, this happens every second:
>
> Jul  3 00:17:29 web1 init: glusterd main process (14958) terminated with status 255
> Jul  3 00:17:29 web1 init: glusterd main process ended, respawning
>

This seems like the init system is trying to restart glusterd.
Glusterd is a daemon process which is spawned by the process launched
by init. Init might be thinking that the main process dying as
glusterd dying and try to restart it. But since glusterd is already
running you are getting the below logs.
What packages are you using and what distro is this on?

> In /var/log/glusterfs/etc-glusterfs-glusterd.vol.log I have lots of this:
>
> [2013-07-03 14:24:08.350429] I [glusterfsd.c:1666:main] 0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd version 3.3.0
> [2013-07-03 14:24:08.350592] E [glusterfsd.c:1296:glusterfs_pidfile_setup] 0-glusterfsd: pidfile /var/run/glusterd.pid lock error (Resource temporarily unavailable)
>
> In /var/log/glusterfs/glustershd.log, every minute I get hundreds of these:
>
> 2013-07-03 14:24:00.792751] I [afr-self-heal-common.c:2159:afr_self_heal_completion_cbk] 0-shared-replicate-0: background  meta-data self-heal completed on <gfid:16adce4d-1933-485f-8359-66c47c757cd3>
> [2013-07-03 14:24:00.794251] I [afr-common.c:1340:afr_launch_self_heal] 0-shared-replicate-0: background  meta-data self-heal triggered. path: <gfid:52bd33a0-0df8-408a-bf6f-c5b6a48c4bd3>, reason: lookup detected pending operations
> [2013-07-03 14:24:00.796411] I [afr-self-heal-common.c:2159:afr_self_heal_completion_cbk] 0-shared-replicate-0: background  meta-data self-heal completed on <gfid:52bd33a0-0df8-408a-bf6f-c5b6a48c4bd3>
>
> 'gluster volume heal shared info' says:
>
> Heal operation on volume shared has been successful
>
> Brick 192.168.1.10:/var/shared
> Number of entries: 335
> ...
>
> I'm not clear whether this means it has 335 files still to fix, or whether it's done so already.

This means there are 335 files still to fix.

> Both servers are logging the same kind of stuff. I'm sure all these are related since they happen at about the same rate.
>
> The lock error looks the most interesting, but I've no idea why that should happen. As before, I've tried deleting all traces of gluster, reinstalling and reconfiguring and putting all the data back on, but nothing changes.
>
> Here's the command I used to create the volume:
>
> gluster volume create shared replica 2 transport tcp 192.168.1.10:/var/shared 192.168.1.11:/var/shared
>
> Here's the volume file it created:
>
> +------------------------------------------------------------------------------+
>   1: volume shared-posix
>   2:     type storage/posix
>   3:     option directory /var/shared
>   4:     option volume-id 2600e26c-b6c4-448f-a6f6-ad27c14745a0
>   5: end-volume
>   6:
>   7: volume shared-access-control
>   8:     type features/access-control
>   9:     subvolumes shared-posix
>  10: end-volume
>  11:
>  12: volume shared-locks
>  13:     type features/locks
>  14:     subvolumes shared-access-control
>  15: end-volume
>  16:
>  17: volume shared-io-threads
>  18:     type performance/io-threads
>  19:     subvolumes shared-locks
>  20: end-volume
>  21:
>  22: volume shared-index
>  23:     type features/index
>  24:     option index-base /var/shared/.glusterfs/indices
>  25:     subvolumes shared-io-threads
>  26: end-volume
>  27:
>  28: volume shared-marker
>  29:     type features/marker
>  30:     option volume-uuid 2600e26c-b6c4-448f-a6f6-ad27c14745a0
>  31:     option timestamp-file /var/lib/glusterd/vols/shared/marker.tstamp
>  32:     option xtime off
>  33:     option quota off
>  34:     subvolumes shared-index
>  35: end-volume
>  36:
>  37: volume /var/shared
>  38:     type debug/io-stats
>  39:     option latency-measurement off
>  40:     option count-fop-hits off
>  41:     subvolumes shared-marker
>  42: end-volume
>  43:
>  44: volume shared-server
>  45:     type protocol/server
>  46:     option transport-type tcp
>  47:     option auth.login./var/shared.allow 94017411-d986-48e4-a7ac-47c1db14fba0
>  48:     option auth.login.94017411-d986-48e4-a7ac-47c1db14fba0.password 3929acf9-fcf1-4684-b271-07927d375c9b
>  49:     option auth.addr./var/shared.allow *
>  50:     subvolumes /var/shared
>  51: end-volume
>
> Despite all this, I've not seen gluster do anything visibly wrong - if I create a file on the shared volume it appears on the other node, checksums match, clients can read etc, but I don't want to be running on luck! It's all very troubling, and it's making a right mess of my new distributed logging system...
>
> Any ideas?
>
> Marcus
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users



More information about the Gluster-users mailing list