[Gluster-users] Remote operation failed: Stale NFS file handle

Neil Van Lysel van-lyse at cs.wisc.edu
Wed Oct 16 15:44:55 UTC 2013


Everything seems to be working great, but the logs are filled with these 
warnings. We can access the files that are reporting a stale nfs handle 
fine. What would these warnings be a symptom of? Warnings worry me so I 
don't want to ignore these errors even if it is log spam.


On 10/15/2013 06:19 PM, Joe Julian wrote:
> Note the " W ". Errors are " E ". If something's not working, however, 
> please elaborate. Those warnings may be a symptom, or just might be 
> log spam.
>
> On 10/15/2013 2:54 PM, Justin Dossey wrote:
>> I've seen these errors too on GlusterFS 3.3.1 nodes with 
>> glusterfs-fuse mounts.  It's particularly strange because we're not 
>> using NFS to mount the volumes.
>>
>>
>> On Tue, Oct 15, 2013 at 1:44 PM, Neil Van Lysel <van-lyse at cs.wisc.edu 
>> <mailto:van-lyse at cs.wisc.edu>> wrote:
>>
>>     Hello!
>>
>>     Many of our Gluster client nodes are seeing a lot of these errors
>>     in their log files:
>>
>>     [2013-10-15 06:48:59.467263] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-home-client-6:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (3cfbebf4-40e4-4300-aa6e-bd43b4310b94)
>>     [2013-10-15 06:48:59.467331] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-home-client-7:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (3cfbebf4-40e4-4300-aa6e-bd43b4310b94)
>>     [2013-10-15 06:48:59.470554] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-home-client-0:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (d662e7db-7864-4b18-b587-bdc5e8756076)
>>     [2013-10-15 06:48:59.470624] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-home-client-1:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (d662e7db-7864-4b18-b587-bdc5e8756076)
>>     [2013-10-15 06:49:04.537548] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-home-client-3:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (a4ea32e0-25f8-440d-b258-23430490624d)
>>     [2013-10-15 06:49:04.537651] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-home-client-2:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (a4ea32e0-25f8-440d-b258-23430490624d)
>>     [2013-10-15 06:49:14.380551] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-home-client-0:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (669a2d6b-2998-48b2-8f3f-93d5f65cdd87)
>>     [2013-10-15 06:49:14.380663] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-home-client-1:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (669a2d6b-2998-48b2-8f3f-93d5f65cdd87)
>>     [2013-10-15 06:49:14.386390] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-home-client-4:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (016aafa9-35ac-4f6f-90bd-b4ac5d435ad0)
>>     [2013-10-15 06:49:14.386471] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-home-client-5:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (016aafa9-35ac-4f6f-90bd-b4ac5d435ad0)
>>     [2013-10-15 18:28:10.630357] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-scratch-client-2:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (5d6153cc-64b3-4151-85cd-2646c33c6918)
>>     [2013-10-15 18:28:10.630425] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-scratch-client-3:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (5d6153cc-64b3-4151-85cd-2646c33c6918)
>>     [2013-10-15 18:28:10.636301] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-scratch-client-4:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (2f64b9fe-02a0-408b-9edb-0c5e5bf0ed0e)
>>     [2013-10-15 18:28:10.636377] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-scratch-client-5:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (2f64b9fe-02a0-408b-9edb-0c5e5bf0ed0e)
>>     [2013-10-15 18:28:10.638574] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-scratch-client-5:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (990de721-1fc9-461d-8412-8c17c23ebbbd)
>>     [2013-10-15 18:28:10.638647] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-scratch-client-4:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (990de721-1fc9-461d-8412-8c17c23ebbbd)
>>     [2013-10-15 18:28:10.645043] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-scratch-client-7:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (0d8d3c5a-d26e-4c15-a8d5-987a4033a6d0)
>>     [2013-10-15 18:28:10.645157] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-scratch-client-6:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (0d8d3c5a-d26e-4c15-a8d5-987a4033a6d0)
>>     [2013-10-15 18:28:10.648126] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-scratch-client-6:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (c1c84d57-f54d-4dc1-a5df-9be563da78fb)
>>     [2013-10-15 18:28:10.648276] W
>>     [client-rpc-fops.c:2624:client3_3_lookup_cbk] 0-scratch-client-7:
>>     remote operation failed: Stale NFS file handle. Path: /path
>>     (c1c84d57-f54d-4dc1-a5df-9be563da78fb)
>>
>>
>>     How can I resolve these errors?
>>
>>
>>     *gluster --version:
>>     glusterfs 3.4.0 built on Jul 25 2013 04:12:27
>>
>>
>>     *gluster volume info:
>>     Volume Name: scratch
>>     Type: Distributed-Replicate
>>     Volume ID: 198b9d77-96e6-4c7f-9f0c-3618cbcaa940
>>     Status: Started
>>     Number of Bricks: 4 x 2 = 8
>>     Transport-type: tcp
>>     Bricks:
>>     Brick1: 10.129.40.21:/data/glusterfs/brick1/scratch
>>     Brick2: 10.129.40.22:/data/glusterfs/brick1/scratch
>>     Brick3: 10.129.40.23:/data/glusterfs/brick1/scratch
>>     Brick4: 10.129.40.24:/data/glusterfs/brick1/scratch
>>     Brick5: 10.129.40.21:/data/glusterfs/brick2/scratch
>>     Brick6: 10.129.40.22:/data/glusterfs/brick2/scratch
>>     Brick7: 10.129.40.23:/data/glusterfs/brick2/scratch
>>     Brick8: 10.129.40.24:/data/glusterfs/brick2/scratch
>>     Options Reconfigured:
>>     features.quota: off
>>
>>     Volume Name: home
>>     Type: Distributed-Replicate
>>     Volume ID: 0d8ebafc-471e-4b16-a4a9-787ce8616225
>>     Status: Started
>>     Number of Bricks: 4 x 2 = 8
>>     Transport-type: tcp
>>     Bricks:
>>     Brick1: 10.129.40.21:/data/glusterfs/brick1/home
>>     Brick2: 10.129.40.22:/data/glusterfs/brick1/home
>>     Brick3: 10.129.40.23:/data/glusterfs/brick1/home
>>     Brick4: 10.129.40.24:/data/glusterfs/brick1/home
>>     Brick5: 10.129.40.21:/data/glusterfs/brick2/home
>>     Brick6: 10.129.40.22:/data/glusterfs/brick2/home
>>     Brick7: 10.129.40.23:/data/glusterfs/brick2/home
>>     Brick8: 10.129.40.24:/data/glusterfs/brick2/home
>>     Options Reconfigured:
>>     features.quota: off
>>
>>
>>     *gluster volume status:
>>     Status of volume: scratch
>>     Gluster process Port Online  Pid
>>     ------------------------------------------------------------------------------
>>     Brick 10.129.40.21:/data/glusterfs/brick1/scratch 49154 Y       7536
>>     Brick 10.129.40.22:/data/glusterfs/brick1/scratch 49154 Y       27976
>>     Brick 10.129.40.23:/data/glusterfs/brick1/scratch 49154 Y       7436
>>     Brick 10.129.40.24:/data/glusterfs/brick1/scratch 49154 Y       19773
>>     Brick 10.129.40.21:/data/glusterfs/brick2/scratch 49155 Y       7543
>>     Brick 10.129.40.22:/data/glusterfs/brick2/scratch 49155 Y       27982
>>     Brick 10.129.40.23:/data/glusterfs/brick2/scratch 49155 Y       7442
>>     Brick 10.129.40.24:/data/glusterfs/brick2/scratch 49155 Y       19778
>>     NFS Server on localhost 2049 Y       7564
>>     Self-heal Daemon on localhost N/A Y       7569
>>     NFS Server on 10.129.40.24  2049 Y       19788
>>     Self-heal Daemon on 10.129.40.24  N/A Y       19792
>>     NFS Server on 10.129.40.23  2049 Y       7464
>>     Self-heal Daemon on 10.129.40.23  N/A Y       7468
>>     NFS Server on 10.129.40.22  2049 Y       28004
>>     Self-heal Daemon on 10.129.40.22  N/A Y       28008
>>
>>     There are no active volume tasks
>>     Status of volume: home
>>     Gluster process Port Online  Pid
>>     ------------------------------------------------------------------------------
>>     Brick 10.129.40.21:/data/glusterfs/brick1/home  49152 Y       7549
>>     Brick 10.129.40.22:/data/glusterfs/brick1/home  49152 Y       27989
>>     Brick 10.129.40.23:/data/glusterfs/brick1/home  49152 Y       7449
>>     Brick 10.129.40.24:/data/glusterfs/brick1/home  49152 Y       19760
>>     Brick 10.129.40.21:/data/glusterfs/brick2/home  49153 Y       7554
>>     Brick 10.129.40.22:/data/glusterfs/brick2/home  49153 Y       27994
>>     Brick 10.129.40.23:/data/glusterfs/brick2/home  49153 Y       7454
>>     Brick 10.129.40.24:/data/glusterfs/brick2/home  49153 Y       19766
>>     NFS Server on localhost 2049 Y       7564
>>     Self-heal Daemon on localhost N/A Y       7569
>>     NFS Server on 10.129.40.24  2049 Y       19788
>>     Self-heal Daemon on 10.129.40.24  N/A Y       19792
>>     NFS Server on 10.129.40.22  2049 Y       28004
>>     Self-heal Daemon on 10.129.40.22  N/A Y       28008
>>     NFS Server on 10.129.40.23  2049 Y       7464
>>     Self-heal Daemon on 10.129.40.23  N/A Y       7468
>>
>>     There are no active volume tasks
>>
>>
>>     *The gluster volumes are mounted using the glusterfs-fuse package
>>     (glusterfs-fuse-3.4.0-3.el6.x86_64) on the clients like so:
>>     /sbin/mount.glusterfs 10.129.40.21:home /home
>>     /sbin/mount.glusterfs 10.129.40.21:scratch /scratch
>>
>>
>>     *Gluster packages on Gluster servers:
>>     glusterfs-server-3.4.0-3.el6.x86_64
>>     glusterfs-libs-3.4.0-8.el6.x86_64
>>     glusterfs-3.4.0-3.el6.x86_64
>>     glusterfs-geo-replication-3.4.0-3.el6.x86_64
>>     glusterfs-fuse-3.4.0-3.el6.x86_64
>>     glusterfs-rdma-3.4.0-3.el6.x86_64
>>
>>
>>     *Gluster packages on clients:
>>     glusterfs-fuse-3.4.0-3.el6.x86_64
>>     glusterfs-3.4.0-3.el6.x86_64
>>
>>
>>     All clients and servers are running the same OS and kernel:
>>
>>     *uname -a:
>>     Linux <hostname> 2.6.32-358.6.1.el6.x86_64 #1 SMP Tue Apr 23
>>     16:15:13 CDT 2013 x86_64 x86_64 x86_64 GNU/Linux
>>
>>     *cat /etc/redhat-release :
>>     Scientific Linux release 6.3 (Carbon)
>>
>>
>>     Thanks for your help,
>>
>>     Neil Van Lysel
>>     UNIX Systems Administrator
>>     Center for High Throughput Computing
>>     University of Wisconsin - Madison
>>
>>
>>
>>     _______________________________________________
>>     Gluster-users mailing list
>>     Gluster-users at gluster.org <mailto:Gluster-users at gluster.org>
>>     http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>
>>
>>
>>
>> -- 
>> Justin Dossey
>> CTO, PodOmatic
>>
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://supercolony.gluster.org/mailman/listinfo/gluster-users
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://supercolony.gluster.org/mailman/listinfo/gluster-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20131016/c031350d/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/pkcs7-signature
Size: 3528 bytes
Desc: S/MIME Cryptographic Signature
URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20131016/c031350d/attachment.p7s>


More information about the Gluster-users mailing list