[Gluster-users] Lot of EIO errors in disperse volume

Ankireddypalle Reddy areddy at commvault.com
Tue Jan 10 10:39:00 UTC 2017


Hi,
      We upgraded to GlusterFS 3.7.18 yesterday.  We see lot of failures in our applications. Most of the errors are EIO. The following log lines are commonly seen in the logs:

The message "W [MSGID: 122056] [ec-combine.c:873:ec_combine_check] 0-StoragePool-disperse-4: Mismatching xdata in answers of 'LOOKUP'" repeated 2 times between [2017-01-10 02:46:25.069809] and [2017-01-10 02:46:25.069835]
[2017-01-10 02:46:25.069852] W [MSGID: 122056] [ec-combine.c:873:ec_combine_check] 0-StoragePool-disperse-5: Mismatching xdata in answers of 'LOOKUP'
The message "W [MSGID: 122056] [ec-combine.c:873:ec_combine_check] 0-StoragePool-disperse-5: Mismatching xdata in answers of 'LOOKUP'" repeated 2 times between [2017-01-10 02:46:25.069852] and [2017-01-10 02:46:25.069873]
[2017-01-10 02:46:25.069910] W [MSGID: 122056] [ec-combine.c:873:ec_combine_check] 0-StoragePool-disperse-6: Mismatching xdata in answers of 'LOOKUP'
...
[2017-01-10 02:46:26.520774] I [MSGID: 109036] [dht-common.c:9076:dht_log_new_layout_for_dir_selfheal] 0-StoragePool-dht: Setting layout of /Folder_07.11.2016_23.02/CV_MAGNETIC/V_8854213/CHUNK_51334585 with [Subvol_name: StoragePool-disperse-0, Err: -1 , Start: 3221225466 , Stop: 3758096376 , Hash: 1 ], [Subvol_name: StoragePool-disperse-1, Err: -1 , Start: 3758096377 , Stop: 4294967295 , Hash: 1 ], [Subvol_name: StoragePool-disperse-2, Err: -1 , Start: 0 , Stop: 536870910 , Hash: 1 ], [Subvol_name: StoragePool-disperse-3, Err: -1 , Start: 536870911 , Stop: 1073741821 , Hash: 1 ], [Subvol_name: StoragePool-disperse-4, Err: -1 , Start: 1073741822 , Stop: 1610612732 , Hash: 1 ], [Subvol_name: StoragePool-disperse-5, Err: -1 , Start: 1610612733 , Stop: 2147483643 , Hash: 1 ], [Subvol_name: StoragePool-disperse-6, Err: -1 , Start: 2147483644 , Stop: 2684354554 , Hash: 1 ], [Subvol_name: StoragePool-disperse-7, Err: -1 , Start: 2684354555 , Stop: 3221225465 , Hash: 1 ],
[2017-01-10 02:46:26.522841] N [MSGID: 122031] [ec-generic.c:1130:ec_combine_xattrop] 0-StoragePool-disperse-3: Mismatching dictionary in answers of 'GF_FOP_XATTROP'
The message "N [MSGID: 122031] [ec-generic.c:1130:ec_combine_xattrop] 0-StoragePool-disperse-3: Mismatching dictionary in answers of 'GF_FOP_XATTROP'" repeated 2 times between [2017-01-10 02:46:26.522841] and [2017-01-10 02:46:26.522894]
[2017-01-10 02:46:26.522898] W [MSGID: 122040] [ec-common.c:919:ec_prepare_update_cbk] 0-StoragePool-disperse-3: Failed to get size and version [Input/output error]
[2017-01-10 02:46:26.523115] N [MSGID: 122031] [ec-generic.c:1130:ec_combine_xattrop] 0-StoragePool-disperse-6: Mismatching dictionary in answers of 'GF_FOP_XATTROP'
The message "N [MSGID: 122031] [ec-generic.c:1130:ec_combine_xattrop] 0-StoragePool-disperse-6: Mismatching dictionary in answers of 'GF_FOP_XATTROP'" repeated 2 times between [2017-01-10 02:46:26.523115] and [2017-01-10 02:46:26.523143]
[2017-01-10 02:46:26.523147] W [MSGID: 122040] [ec-common.c:919:ec_prepare_update_cbk] 0-StoragePool-disperse-6: Failed to get size and version [Input/output error]
[2017-01-10 02:46:26.523302] N [MSGID: 122031] [ec-generic.c:1130:ec_combine_xattrop] 0-StoragePool-disperse-2: Mismatching dictionary in answers of 'GF_FOP_XATTROP'
The message "N [MSGID: 122031] [ec-generic.c:1130:ec_combine_xattrop] 0-StoragePool-disperse-2: Mismatching dictionary in answers of 'GF_FOP_XATTROP'" repeated 2 times between [2017-01-10 02:46:26.523302] and [2017-01-10 02:46:26.523324]
[2017-01-10 02:46:26.523328] W [MSGID: 122040] [ec-common.c:919:ec_prepare_update_cbk] 0-StoragePool-disperse-2: Failed to get size and version [Input/output error]

[root at glusterfs3 Log_Files]# gluster --version
glusterfs 3.7.18 built on Dec  8 2016 06:34:26

[root at glusterfs3 Log_Files]# gluster volume info

Volume Name: StoragePool
Type: Distributed-Disperse
Volume ID: 149e976f-4e21-451c-bf0f-f5691208531f
Status: Started
Number of Bricks: 8 x (2 + 1) = 24
Transport-type: tcp
Bricks:
Brick1: glusterfs1sds:/ws/disk1/ws_brick
Brick2: glusterfs2sds:/ws/disk1/ws_brick
Brick3: glusterfs3sds:/ws/disk1/ws_brick
Brick4: glusterfs1sds:/ws/disk2/ws_brick
Brick5: glusterfs2sds:/ws/disk2/ws_brick
Brick6: glusterfs3sds:/ws/disk2/ws_brick
Brick7: glusterfs1sds:/ws/disk3/ws_brick
Brick8: glusterfs2sds:/ws/disk3/ws_brick
Brick9: glusterfs3sds:/ws/disk3/ws_brick
Brick10: glusterfs1sds:/ws/disk4/ws_brick
Brick11: glusterfs2sds:/ws/disk4/ws_brick
Brick12: glusterfs3sds:/ws/disk4/ws_brick
Brick13: glusterfs1sds:/ws/disk5/ws_brick
Brick14: glusterfs2sds:/ws/disk5/ws_brick
Brick15: glusterfs3sds:/ws/disk5/ws_brick
Brick16: glusterfs1sds:/ws/disk6/ws_brick
Brick17: glusterfs2sds:/ws/disk6/ws_brick
Brick18: glusterfs3sds:/ws/disk6/ws_brick
Brick19: glusterfs1sds:/ws/disk7/ws_brick
Brick20: glusterfs2sds:/ws/disk7/ws_brick
Brick21: glusterfs3sds:/ws/disk7/ws_brick
Brick22: glusterfs1sds:/ws/disk8/ws_brick
Brick23: glusterfs2sds:/ws/disk8/ws_brick
Brick24: glusterfs3sds:/ws/disk8/ws_brick
Options Reconfigured:
performance.readdir-ahead: on
diagnostics.client-log-level: INFO

Thanks and Regards,
Ram
***************************Legal Disclaimer***************************
"This communication may contain confidential and privileged material for the
sole use of the intended recipient. Any unauthorized review, use or distribution
by others is strictly prohibited. If you have received the message by mistake,
please advise the sender by reply email and delete the message. Thank you."
**********************************************************************
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20170110/982d939e/attachment.html>


More information about the Gluster-users mailing list