[Gluster-users] strange error hangs hangs any access to gluster mount
Jeff Darcy
jdarcy at redhat.com
Mon Mar 28 16:52:35 UTC 2011
On 03/28/2011 12:43 PM, Burnash, James wrote:
> I am receiving an error on a client trying to access a gluster mount
> (/pfs2, in this case).
>
> [2011-03-28 12:26:17.897887] I
> [dht-layout.c:588:dht_layout_normalize] pfs-ro1-dht: found anomalies
> in /. holes=1 overlaps=2
>
> This is seen on the client in the /var/log/glusterfs/pfs2.log, which
> is the mount point associated with that storage.
>
> All other clients accessing the same storage do not have the hanging
> symptom, and have no such entry in their logs.
>
> One possibly helpful note - this node worked fine until I upgraded
> the client from 3.1.1-1 to 3.1.3-1 on the x86_64 architecture,
> running CentOS 5.2. Even after I completely uninstalled GlusterFS
> from this node and reinstalled 3.1.1-1, the problem persisted.
>
> Here is the RPM info:
>
> root at jc1lnxsamm33:~# rpm -qa fuse fuse-2.7.4-8.el5.x86_64
> root at jc1lnxsamm33:~# rpm -qa "glusterfs*"
> glusterfs-fuse-3.1.1-1.x86_64 glusterfs-core-3.1.1-1.x86_64
> glusterfs-debuginfo-3.1.1-1.x86_64
>
> Servers are 4 Replicated-Distribute machines running CentOS 5.5 and
> GlusterFs 3.1.3-1.
>
> Volume Name: pfs-ro1 Type: Distributed-Replicate Status: Started
> Number of Bricks: 20 x 2 = 40 Transport-type: tcp Bricks: Brick1:
> jc1letgfs17-pfs1:/export/read-only/g01 Brick2:
> jc1letgfs18-pfs1:/export/read-only/g01 Brick3:
> jc1letgfs17-pfs1:/export/read-only/g02 Brick4:
> jc1letgfs18-pfs1:/export/read-only/g02 ... Brick35:
> jc1letgfs14-pfs1:/export/read-only/g08 Brick36:
> jc1letgfs15-pfs1:/export/read-only/g08 Brick37:
> jc1letgfs14-pfs1:/export/read-only/g09 Brick38:
> jc1letgfs15-pfs1:/export/read-only/g09 Brick39:
> jc1letgfs14-pfs1:/export/read-only/g10 Brick40:
> jc1letgfs15-pfs1:/export/read-only/g10 Options Reconfigured:
> performance.stat-prefetch: on performance.cache-size: 2GB
> network.ping-timeout: 10
>
> Any help greatly appreciated.
Can you execute the following command on each of the brick roots?
getfattr -d -e hex -n trusted.glusterfs.dht $brick_root
That should give a clearer picture of what the layouts look like, and
what those gaps/overlaps are. How they happened is a bit of another
story. I see this kind of thing pretty often, but I know it's because
of some Weird Stuff (tm) I do in CloudFS. I'm not aware of any bugs
etc. that would cause this in other contexts.
More information about the Gluster-users
mailing list