[Gluster-devel] ESTALE / "Stale NFS file handle"

Dan Parsons dparsons at nyip.net
Tue Mar 10 05:40:01 UTC 2009


I'm getting the below error messages in rc4. Like my previous email, there
doesn't seem to be any pattern as to which server/client it's happening on,
though the errors are occurring fairly frequently.
2009-03-09 17:32:26 E [unify.c:585:unify_lookup] unify: returning ESTALE for
/bio/data/fast-hmmsearch-all/tmpP986E__fast-hmmsearch-all_job/result.tigrfam.TIGR02622.hmmhits:
file count is 1
2009-03-09 17:32:26 E [unify.c:591:unify_lookup] unify:
/bio/data/fast-hmmsearch-all/tmpP986E__fast-hmmsearch-all_job/result.tigrfam.TIGR02622.hmmhits:
found on unify-switch-ns
2009-03-09 17:32:26 W [fuse-bridge.c:301:need_fresh_lookup] fuse-bridge:
revalidate of
/bio/data/fast-hmmsearch-all/tmpP986E__fast-hmmsearch-all_job/result.tigrfam.TIGR02622.hmmhits
failed (Stale NFS file handle)

2009-03-09 17:32:28 E [unify.c:360:unify_lookup_cbk] unify: child(dht0):
path(/bio/data/fast-hmmsearch-all/tmpP986E__fast-hmmsearch-all_job/result.tigrfam.TIGR01420.hmmhits):
No such file or directory
2009-03-09 17:32:28 E [unify.c:360:unify_lookup_cbk] unify:
child(unify-switch-ns):
path(/bio/data/fast-hmmsearch-all/tmpP986E__fast-hmmsearch-all_job/result.tigrfam.TIGR01420.hmmhits):
No such file or directory

As you can see, there are two separate sets of errors for two different
files, though both errors are troubling. This problem has persisted from rc2
to rc4, though I can't say for certain that it was introduced in rc2 (I
think it was there prior to that as well). No matching errors in server
logs.

Any suggestions? My configs are below. Thanks!

CLIENT CONFIG:

volume unify-switch-ns
   type protocol/client
   option transport-type tcp
   option remote-host 10.8.101.51
   option remote-subvolume posix-unify-switch-ns
end-volume

#volume distfs01-ns-readahead
#   type performance/read-ahead
#   option page-size 1MB
#   option page-count 8
#   subvolumes distfs01-ns-brick
#end-volume

#volume unify-switch-ns
#   type performance/write-behind
#   option block-size 1MB
#   option cache-size 3MB
#   subvolumes distfs01-ns-readahead
#end-volume

volume distfs01-unify
   type protocol/client
   option transport-type tcp
   option remote-host 10.8.101.51
   option remote-subvolume posix-unify
end-volume

volume distfs02-unify
   type protocol/client
   option transport-type tcp
   option remote-host 10.8.101.52
   option remote-subvolume posix-unify
end-volume

volume distfs03-unify
   type protocol/client
   option transport-type tcp
   option remote-host 10.8.101.53
   option remote-subvolume posix-unify
end-volume

volume distfs04-unify
   type protocol/client
   option transport-type tcp
   option remote-host 10.8.101.54
   option remote-subvolume posix-unify
end-volume

volume distfs01-stripe
   type protocol/client
   option transport-type tcp
   option remote-host 10.8.101.51
   option remote-subvolume posix-stripe
end-volume

volume distfs02-stripe
   type protocol/client
   option transport-type tcp
   option remote-host 10.8.101.52
   option remote-subvolume posix-stripe
end-volume

volume distfs03-stripe
   type protocol/client
   option transport-type tcp
   option remote-host 10.8.101.53
   option remote-subvolume posix-stripe
end-volume

volume distfs04-stripe
   type protocol/client
   option transport-type tcp
   option remote-host 10.8.101.54
   option remote-subvolume posix-stripe
end-volume

volume stripe0
type cluster/stripe
option block-size *.jar,*.pin:1MB,*:2MB
subvolumes distfs01-stripe distfs02-stripe distfs03-stripe distfs04-stripe
end-volume

volume dht0
type cluster/dht
# option lookup-unhashed yes
subvolumes distfs01-unify distfs02-unify distfs03-unify distfs04-unify
end-volume

volume unify
type cluster/unify
option namespace unify-switch-ns
option self-heal off
option scheduler switch
# send *.phr/psq/pnd etc to stripe0, send the rest to hash
# extensions have to be *.foo* and not simply *.foo or rsync's tmp file
naming will prevent files from being matched
option scheduler.switch.case
*.phr*:stripe0;*.psq*:stripe0;*.pnd*:stripe0;*.psd*:stripe0;*.pin*:stripe0;*.nsi*:stripe0;*.nin*:stripe0;*.nsd*:stripe0;*.nhr*:stripe0;*.nsq*:stripe0;*.tar*:stripe0;*.tar.gz*:stripe0;*.jar*:stripe0;*.img*:stripe0;*.perf*:stripe0;*.tgz*:stripe0;*.fasta*:stripe0;*.huge*:stripe0
subvolumes stripe0 dht0
end-volume

volume ioc
   type performance/io-cache
   subvolumes unify
   option cache-size 3000MB
option cache-timeout 3600
end-volume

volume filter
  type features/filter
  option fixed-uid 0
  option fixed-gid 900
  subvolumes ioc
end-volume




SERVER CONFIG:
volume posix-unify-brick
type storage/posix
option directory /distfs-storage-space/glusterfs/unify
# the below line is here to make the output of 'df' accurate, as both
volumes are served from the same local drive
option export-statfs-size off
end-volume

volume posix-stripe-brick
        type storage/posix
        option directory /distfs-storage-space/glusterfs/stripe
end-volume

volume posix-unify-switch-ns-brick
        type storage/posix
        option directory /distfs-storage-space/glusterfs/unify-switch-ns
end-volume

volume posix-unify
type performance/io-threads
option thread-count 4
subvolumes posix-unify-brick
end-volume

volume posix-stripe
type performance/io-threads
option thread-count 4
subvolumes posix-stripe-brick
end-volume

volume posix-unify-switch-ns
type performance/io-threads
option thread-count 2
subvolumes posix-unify-switch-ns-brick
end-volume

volume server
type protocol/server
option transport-type tcp
option auth.addr.posix-unify.allow 10.8.101.*,10.8.15.50
option auth.addr.posix-stripe.allow 10.8.101.*,10.8.15.50
option auth.addr.posix-unify-switch-ns.allow 10.8.101.*,10.8.15.50
subvolumes posix-unify posix-stripe posix-unify-switch-ns
end-volume
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-devel/attachments/20090309/776d7446/attachment-0003.html>


More information about the Gluster-devel mailing list