[Gluster-users] Input/output error on FUSE log
Nithya Balachandran
nbalacha at redhat.com
Thu Jan 10 10:47:19 UTC 2019
I don't see write failures in the log but I do see fallocate failing with
EIO.
[2019-01-07 19:16:44.846187] W [MSGID: 109011]
[dht-layout.c:163:dht_layout_search] 0-gv1-dht: no subvolume for hash
(value) = 1285124113
[2019-01-07 19:16:44.846194] D [MSGID: 0]
[dht-helper.c:969:dht_subvol_get_hashed] 0-gv1-dht: No hashed subvolume for
path=/.shard/aa3ef10e-95e0-40d3-9464-133d72fa8a95.185
[2019-01-07 19:16:44.846200] D [MSGID: 0] [dht-common.c:7631:dht_mknod]
0-gv1-dht: no subvolume in layout for
path=/.shard/aa3ef10e-95e0-40d3-9464-133d72fa8a95.185 <--- *** *DHT
failed to find a hashed subvol* ***
[2019-01-07 19:16:44.846207] D [MSGID: 0] [dht-common.c:7712:dht_mknod]
0-stack-trace: stack-address: 0x7f6748006778, gv1-dht returned -1 error:
Input/output error [Input/output error]
[2019-01-07 19:16:44.846215] D [MSGID: 0]
[shard.c:3645:shard_common_mknod_cbk] 0-gv1-shard: mknod of shard 185
failed: Input/output error
[2019-01-07 19:16:44.846223] D [MSGID: 0]
[shard.c:720:shard_common_inode_write_failure_unwind] 0-stack-trace:
stack-address: 0x7f6748006778, gv1-shard returned -1 error: Input/output
error [Input/output error]
[2019-01-07 19:16:44.846234] D [MSGID: 0]
[defaults.c:1352:default_fallocate_cbk] 0-stack-trace: stack-address:
0x7f6748006778, gv1-quick-read returned -1 error: Input/output error
[Input/output error]
[2019-01-07 19:16:44.846244] D [MSGID: 0]
[defaults.c:1352:default_fallocate_cbk] 0-stack-trace: stack-address:
0x7f6748006778, gv1-open-behind returned -1 error: Input/output error
[Input/output error]
[2019-01-07 19:16:44.846254] D [MSGID: 0]
[md-cache.c:2715:mdc_fallocate_cbk] 0-stack-trace: stack-address:
0x7f6748006778, gv1-md-cache returned -1 error: Input/output error
[Input/output error]
[2019-01-07 19:16:44.846264] D [MSGID: 0]
[defaults.c:1352:default_fallocate_cbk] 0-stack-trace: stack-address:
0x7f6748006778, gv1-io-threads returned -1 error: Input/output error
[Input/output error]
[2019-01-07 19:16:44.846274] D [MSGID: 0]
[io-stats.c:2528:io_stats_fallocate_cbk] 0-stack-trace: stack-address:
0x7f6748006778, gv1 returned -1 error: Input/output error [Input/output
error]
[2019-01-07 19:16:44.846284] W [fuse-bridge.c:1441:fuse_err_cbk]
0-glusterfs-fuse: 1373: FALLOCATE() ERR => -1 (Input/output error)
[2019-01-07 19:16:44.846298] T [fuse-bridge.c:278:send_fuse_iov]
0-glusterfs-fuse: writev() result 16/16
Please get the xattrs on the .shard directory on each brick of the volume
so we can check if the layout is complete:
getfattr -e hex -m . -d <brick_root>/.shard
Thanks,
Nithya
On Thu, 10 Jan 2019 at 02:25, Matt Waymack <mwaymack at nsgdv.com> wrote:
> Has anyone any other ideas where to look? This is only affecting FUSE
> clients. SMB clients are unaffected by this problem.
>
>
>
> Thanks!
>
>
>
> *From:* gluster-users-bounces at gluster.org <
> gluster-users-bounces at gluster.org> *On Behalf Of *Matt Waymack
> *Sent:* Monday, January 7, 2019 1:19 PM
> *To:* Raghavendra Gowdappa <rgowdapp at redhat.com>
> *Cc:* gluster-users at gluster.org List <gluster-users at gluster.org>
> *Subject:* Re: [Gluster-users] Input/output error on FUSE log
>
>
>
> Attached are the logs from when a failure occurred with diagnostics set to
> trace.
>
>
>
> Thank you!
>
>
>
> *From:* Raghavendra Gowdappa <rgowdapp at redhat.com>
> *Sent:* Saturday, January 5, 2019 8:32 PM
> *To:* Matt Waymack <mwaymack at nsgdv.com>
> *Cc:* gluster-users at gluster.org List <gluster-users at gluster.org>
> *Subject:* Re: [Gluster-users] Input/output error on FUSE log
>
>
>
>
>
>
>
> On Sun, Jan 6, 2019 at 7:58 AM Raghavendra Gowdappa <rgowdapp at redhat.com>
> wrote:
>
>
>
>
>
> On Sun, Jan 6, 2019 at 4:19 AM Matt Waymack <mwaymack at nsgdv.com> wrote:
>
> Hi all,
>
>
>
> I'm having a problem writing to our volume. When writing files larger
> than about 2GB, I get an intermittent issue where the write will fail and
> return Input/Output error. This is also shown in the FUSE log of the
> client (this is affecting all clients). A snip of a client log is below:
>
> [2019-01-05 22:39:44.581371] W [fuse-bridge.c:2474:fuse_writev_cbk]
> 0-glusterfs-fuse: 51040978: WRITE => -1
> gfid=82a0b5c4-7ef3-43c2-ad86-41e16673d7c2 fd=0x7f949839a368 (Input/output
> error)
>
> [2019-01-05 22:39:44.598392] W [fuse-bridge.c:1441:fuse_err_cbk]
> 0-glusterfs-fuse: 51040979: FLUSH() ERR => -1 (Input/output error)
>
> [2019-01-05 22:39:47.420920] W [fuse-bridge.c:2474:fuse_writev_cbk]
> 0-glusterfs-fuse: 51041266: WRITE => -1
> gfid=0e8e1e13-97a5-478a-bc58-e81ddf3698a3 fd=0x7f949809b7f8 (Input/output
> error)
>
> [2019-01-05 22:39:47.433377] W [fuse-bridge.c:1441:fuse_err_cbk]
> 0-glusterfs-fuse: 51041267: FLUSH() ERR => -1 (Input/output error)
>
> [2019-01-05 22:39:50.441531] W [fuse-bridge.c:2474:fuse_writev_cbk]
> 0-glusterfs-fuse: 51041548: WRITE => -1
> gfid=0e8e1e13-97a5-478a-bc58-e81ddf3698a3 fd=0x7f949839a368 (Input/output
> error)
>
> [2019-01-05 22:39:50.451914] W [fuse-bridge.c:1441:fuse_err_cbk]
> 0-glusterfs-fuse: 51041549: FLUSH() ERR => -1 (Input/output error)
>
> The message "W [MSGID: 109011] [dht-layout.c:163:dht_layout_search]
> 0-gv1-dht: no subvolume for hash (value) = 1311504267" repeated 1721 times
> between [2019-01-05 22:39:33.906241] and [2019-01-05 22:39:44.598371]
>
> The message "E [MSGID: 101046] [dht-common.c:1502:dht_lookup_dir_cbk]
> 0-gv1-dht: dict is null" repeated 1714 times between [2019-01-05
> 22:39:33.925981] and [2019-01-05 22:39:50.451862]
>
> The message "W [MSGID: 109011] [dht-layout.c:163:dht_layout_search]
> 0-gv1-dht: no subvolume for hash (value) = 1137142622" repeated 1707 times
> between [2019-01-05 22:39:39.636552] and [2019-01-05 22:39:50.451895]
>
>
>
> This looks to be a DHT issue. Some questions:
>
> * Are all subvolumes of DHT up and client is connected to them?
> Particularly the subvolume which contains the file in question.
>
> * Can you get all extended attributes of parent directory of the file from
> all bricks?
>
> * set diagnostics.client-log-level to TRACE, capture these errors again
> and attach the client log file.
>
>
>
> I spoke a bit early. dht_writev doesn't search hashed subvolume as its
> already been looked up in lookup. So, these msgs looks to be of a different
> issue - not writev failure.
>
>
>
>
>
> This is intermittent for most files, but eventually if a file is large
> enough it will not write. The workflow is SFTP tot he client which then
> writes to the volume over FUSE. When files get to a certain point,w e can
> no longer write to them. The file sizes are different as well, so it's not
> like they all get to the same size and just stop either. I've ruled out a
> free space issue, our files at their largest are only a few hundred GB and
> we have tens of terrabytes free on each brick. We are also sharding at 1GB.
>
>
>
> I'm not sure where to go from here as the error seems vague and I can only
> see it on the client log. I'm not seeing these errors on the nodes
> themselves. This is also seen if I mount the volume via FUSE on any of the
> nodes as well and it is only reflected in the FUSE log.
>
>
>
> Here is the volume info:
>
> Volume Name: gv1
>
> Type: Distributed-Replicate
>
> Volume ID: 1472cc78-e2a0-4c3f-9571-dab840239b3c
>
> Status: Started
>
> Snapshot Count: 0
>
> Number of Bricks: 8 x (2 + 1) = 24
>
> Transport-type: tcp
>
> Bricks:
>
> Brick1: tpc-glus4:/exp/b1/gv1
>
> Brick2: tpc-glus2:/exp/b1/gv1
>
> Brick3: tpc-arbiter1:/exp/b1/gv1 (arbiter)
>
> Brick4: tpc-glus2:/exp/b2/gv1
>
> Brick5: tpc-glus4:/exp/b2/gv1
>
> Brick6: tpc-arbiter1:/exp/b2/gv1 (arbiter)
>
> Brick7: tpc-glus4:/exp/b3/gv1
>
> Brick8: tpc-glus2:/exp/b3/gv1
>
> Brick9: tpc-arbiter1:/exp/b3/gv1 (arbiter)
>
> Brick10: tpc-glus4:/exp/b4/gv1
>
> Brick11: tpc-glus2:/exp/b4/gv1
>
> Brick12: tpc-arbiter1:/exp/b4/gv1 (arbiter)
>
> Brick13: tpc-glus1:/exp/b5/gv1
>
> Brick14: tpc-glus3:/exp/b5/gv1
>
> Brick15: tpc-arbiter2:/exp/b5/gv1 (arbiter)
>
> Brick16: tpc-glus1:/exp/b6/gv1
>
> Brick17: tpc-glus3:/exp/b6/gv1
>
> Brick18: tpc-arbiter2:/exp/b6/gv1 (arbiter)
>
> Brick19: tpc-glus1:/exp/b7/gv1
>
> Brick20: tpc-glus3:/exp/b7/gv1
>
> Brick21: tpc-arbiter2:/exp/b7/gv1 (arbiter)
>
> Brick22: tpc-glus1:/exp/b8/gv1
>
> Brick23: tpc-glus3:/exp/b8/gv1
>
> Brick24: tpc-arbiter2:/exp/b8/gv1 (arbiter)
>
> Options Reconfigured:
>
> performance.cache-samba-metadata: on
>
> performance.cache-invalidation: off
>
> features.shard-block-size: 1000MB
>
> features.shard: on
>
> transport.address-family: inet
>
> nfs.disable: on
>
> cluster.lookup-optimize: on
>
>
>
> I'm a bit stumped on this, any help is appreciated. Thank you!
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190110/395e4d86/attachment.html>
More information about the Gluster-users
mailing list