[Gluster-users] Redux: Stale NFS file handle

harry mangalam harry.mangalam at uci.edu
Fri Jan 4 18:50:30 UTC 2013


Following up on myself..

There is a RH bug report <https://bugzilla.redhat.com/show_bug.cgi?id=678544> 
that sounds like what I'm seeing, but it's associated with the NFS client.  
We're using the gluster native client, so it shouldn't be related, but it 
looks like there may be some NFS client code that was used in the gluster 
native client code as well (the cause of the 'Stale NFS handle' warning in a 
non-NFS setting). 

Could this be the case?  Could the pseudo-solution for the NFS case also work 
for the gluster native client case?  ie: mount -o acdirmin=[23] ...

It doesn't seem like 'acdirmin'  is a mount option for glusterfs but is there 
a gluster equivalent?

Also, this bug <https://bugzilla.redhat.com/show_bug.cgi?id=844584> that 
complains about the the same thing as I reference below has this notation from 
Amar:
---
..this issue ['Stale NFS handle'] can happen in the case where file is changed 
on server after it was being accessed from one node (changed as in, re-
created). as long as the application doesn't see any issue, this should be ok.
---

In our case, it's causing the SGE array job to fail (or at least it appears to 
be highly related):
=====
SGE error extract:

Shepherd error:
01/03/2013 20:03:03 [507:64196]: error: can't chdir to 
/gl/bio/krthornt/build_div/yak/line10_CY22B/prinses: No such file or directory
=====
glusterfs log on client (native gluster client):

FS file handle. Path: /bio/krthornt/build_div/yak/line10_CY22B/prinses 
(590802f1-7fba-4103-ba30-e4d415b9db36)
[2013-01-03 20:03:03.168229] W [client3_1-fops.c:2630:client3_1_lookup_cbk] 0-
gl-client-1: remote operation failed: Stale NFS file handle. Path: 
/bio/krthornt/build_div/yak/line10_CY22B/prinses (590802f1-7fba-4103-ba30-
e4d415b9db36)
[2013-01-03 20:03:03.168287] W [client3_1-fops.c:2630:client3_1_lookup_cbk] 0-
gl-client-2: remote operation failed: Stale NFS file handle. Path: 
/bio/krthornt/build_div/yak/line10_CY22B/prinses (590802f1-7fba-4103-ba30-
e4d415b9db36)

(and 13 more identical lines except for the timestamp (extending out to 
'20:03:17.202688').

so while we might not be losing data /per se/, this failure is definitely 
causing our cluster to lose jobs.

If this has not been reported previously, I'd be happy to file an official bug 
report.

hjm


On Thursday, January 03, 2013 05:56:10 PM harry mangalam wrote:
> From ~ an hour's googling and reading, it looks like this (not uncommon)
> bug/warning/error has not necessarily been associated with data loss, but we
> are finding that our gluster fs is interrupting our cluster jobs with the
> 'Stale NFS handle' Warnings like this (on the client):
> 
> [2013-01-03 12:30:59.149230] W [client3_1-fops.c:2630:client3_1_lookup_cbk]
> 0- gl-client-0: remote operation failed: Stale NFS file handle. Path:
> /bio/krthornt/build_div/yak/line06_CY08A/prinses (3b0aa7b2-bf7f-4b27-
> b515-32e94b1206e3)
> 
> (and 7 more, differing by the timestamp of <<1s).
> 
> The dir mentioned existed before the job was asked to read from it and
> shortly after the SGE failed, I checked that the glusterfs (/bio) was still
> mounted and that dir was still r/w.
> 
> We are getting these errors infrequently, but fairly regularly (a couple
> times a week, usually during a big array job that heavily reads from a
> particular dir) and I haven't seen any resolutions of the fault besides the
> vocabulary being corrected.  I know it's not nec an NFS problem, but I
> haven't seen a fix from the gluster folks.
> 
> Our glusterfs on this system is set up like this (over QDR/tcpoib)
> 
> $ gluster volume info gl
> 
> Volume Name: gl
> Type: Distribute
> Volume ID: 21f480f7-fc5a-4fd8-a084-3964634a9332
> Status: Started
> Number of Bricks: 8
> Transport-type: tcp,rdma
> Bricks:
> Brick1: bs2:/raid1
> Brick2: bs2:/raid2
> Brick3: bs3:/raid1
> Brick4: bs3:/raid2
> Brick5: bs4:/raid1
> Brick6: bs4:/raid2
> Brick7: bs1:/raid1
> Brick8: bs1:/raid2
> Options Reconfigured:
> auth.allow: 10.2.*.*,10.1.*.*
> performance.io-thread-count: 64
> performance.quick-read: on
> performance.io-cache: on
> nfs.disable: on
> performance.cache-size: 268435456
> performance.flush-behind: on
> performance.write-behind-window-size: 1024MB
> 
> and otherwise appears to be happy.
> 
> We were having a low-level problem with the RAID servers, where this
> LSI/3ware error was temporally close (~2m) to the gluster error:
> 
> LSI 3DM2 alert -- host: biostor4.oit.uci.edu
> Jan 03, 2013 03:32:09PM - Controller 6
> ERROR - Drive timeout detected: encl=1, slot=3
> 
> This error seemed to be related to construction around our data center and
> dust related with it.  We have had 10s of these LSI/3ware errors with no
> related gluster errors or apparent problems with the RAIDs.  No drives were
> ejected from the RAIDs and the errors did not repeat.  3ware explains:
> <http://cholla.mmto.org/computers/3ware/3dm2/en/3DM_2_OLH-8-6.html>
> ==============================
> 009h Drive timeout detected
> 
> The 3ware RAID controller has a sophisticated recovery mechanism to handle
> various types of failures of a disk drive. One such possible failure of a
> disk drive is a failure of a command that is pending from the 3ware RAID
> controller to complete within a reasonable amount of time. If the 3ware
> RAID controller detects this condition, it notifies the user, prior to
> entering the recovery phase, by displaying this AEN.
> 
> Possible causes of APORT time-outs include a bad or intermittent disk drive,
> power cable or interface cable.
> ================================
> We've checked into this and it doesn't seem to be related, but I thought I'd
> bring it up.
> 
> hjm
> 
> On Thursday, August 23, 2012 09:54:13 PM Joe Julian wrote:
> > *Bug 832694* <https://bugzilla.redhat.com/show_bug.cgi?id=832694>
> > -ESTALE error text should be reworded
> > 
> > On 08/23/2012 09:50 PM, Kaushal M wrote:
> > > The "Stale NFS file handle" message is the default string given by
> > > strerror() for errno ESTALE.
> > > Gluster uses ESTALE as errno to indicate that the file being referred
> > > to no longer exists, ie. the reference is stale.
> > > 
> > > - Kaushal
> > > 
> > > On Fri, Aug 24, 2012 at 7:03 AM, Jules Wang <lancelotds at 163.com
> > > 
> > > <mailto:lancelotds at 163.com>> wrote:
> > >     Hi, Jon ,
> > >     
> > >         I also met the same issue, and reported a
> > >     
> > >     bug(https://bugzilla.redhat.com/show_bug.cgi?id=851381)
> > >     <https://bugzilla.redhat.com/show_bug.cgi?id=851381>
> > >     
> > >         In the bug report, I share a simple way to reproduce the bug.
> > >         Have fun.
> > >     
> > >     Best Regards.
> > >     Jules Wang.
> > >     
> > >     At 2012-08-23 23:02:34,"Bùi Hùng Vie^.t" <buihungviet at gmail.com
> > >     
> > >     <mailto:buihungviet at gmail.com>> wrote:
> > >         Hi Jon,
> > >         I have no answer for you. Just want to share with you guys
> > >         that I met the same issue with this message. In my gluster
> > >         system, Gluster client log files have a lot of these messages.
> > >         I tried to ask and found nothing on the Web. Amazingly,
> > >         Gluster have been running for long time :)
> > >         
> > >         On Thu, Aug 23, 2012 at 8:43 PM, Jon Tegner <tegner at renget.se
> > >         
> > >         <mailto:tegner at renget.se>> wrote:
> > >             Hi, I'm a bit curious of error messages of the type
> > >             "remote operation failed: Stale NFS file handle". All
> > >             clients using the file system use Gluster Native Client,
> > >             so why should stale nfs file handle be reported?
> > >             
> > >             Regards,
> > >             
> > >             /jon
> > >             

---
Harry Mangalam - Research Computing, OIT, Rm 225 MSTB, UC Irvine
[m/c 2225] / 92697 Google Voice Multiplexer: (949) 478-4487
415 South Circle View Dr, Irvine, CA, 92697 [shipping]
MSTB Lat/Long: (33.642025,-117.844414) (paste into Google Maps)




More information about the Gluster-users mailing list