[Gluster-users] Possible error not being returned
Xavier Hernandez
xhernandez at datalab.es
Tue May 24 06:24:05 UTC 2016
The error "XDR decoding failed" also happened to me recently. There's a
bug report [1] and a patch for it [2] that should solve the problem.
Once communication failures are solved, random brick failures should
disappear and ec should behave correctly. Anyway I'll try to see why ec
is not able to detect these random failures.
The patch will be included in 3.7.12. Or you can apply it on top of
3.7.11 sources and compile if you want to try it.
Xavi
[1] https://bugzilla.redhat.com/show_bug.cgi?id=1331502
[2] http://review.gluster.org/14292/
On 23/05/16 15:37, Ankireddypalle Reddy wrote:
> Xavier,
> Please find below logs from ws-glus.lg where /ws/glus is the gluster mount point in question.
>
> 2016-05-18 21:13:00.477609] W [MSGID: 122002] [ec-common.c:71:ec_heal_report] 0-SDSStoragePool-disperse-2: Heal failed [Transport endpoint is not connected]
> [2016-05-18 21:13:00.556553] E [MSGID: 114030] [client-rpc-fops.c:3022:client3_3_readv_cbk] 0-SDSStoragePool-client-1: XDR decoding failed [Invalid argument]
> [2016-05-18 21:13:00.556588] W [MSGID: 114031] [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-SDSStoragePool-client-1: remote operation failed [Invalid argument]
> [2016-05-18 21:13:00.557754] W [MSGID: 122053] [ec-common.c:116:ec_check_status] 0-SDSStoragePool-disperse-0: Operation failed on some subvolumes (up=7, mask=7, remaining=0, good=5, bad=2)
> [2016-05-18 21:13:00.566626] W [MSGID: 122053] [ec-common.c:116:ec_check_status] 0-SDSStoragePool-disperse-0: Operation failed on some subvolumes (up=7, mask=5, remaining=0, good=5, bad=2)
> [2016-05-18 21:13:00.568694] W [MSGID: 122053] [ec-common.c:116:ec_check_status] 0-SDSStoragePool-disperse-0: Operation failed on some subvolumes (up=7, mask=5, remaining=0, good=5, bad=2)
> [2016-05-18 21:13:00.620563] W [MSGID: 122002] [ec-common.c:71:ec_heal_report] 0-SDSStoragePool-disperse-0: Heal failed [Transport endpoint is not connected]
> [2016-05-18 21:13:00.631860] W [MSGID: 122002] [ec-common.c:71:ec_heal_report] 0-SDSStoragePool-disperse-0: Heal failed [Transport endpoint is not connected]
> [2016-05-18 21:13:01.576095] E [MSGID: 114030] [client-rpc-fops.c:3022:client3_3_readv_cbk] 0-SDSStoragePool-client-13: XDR decoding failed [Invalid argument]
> [2016-05-18 21:13:01.576137] W [MSGID: 114031] [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-SDSStoragePool-client-13: remote operation failed [Invalid argument]
> [2016-05-18 21:13:01.576574] E [MSGID: 114030] [client-rpc-fops.c:3022:client3_3_readv_cbk] 0-SDSStoragePool-client-12: XDR decoding failed [Invalid argument]
> [2016-05-18 21:13:01.576598] W [MSGID: 114031] [client-rpc-fops.c:3050:client3_3_readv_cbk] 0-SDSStoragePool-client-12: remote operation failed [Invalid argument]
> [2016-05-18 21:13:01.576810] W [MSGID: 122053] [ec-common.c:116:ec_check_status] 0-SDSStoragePool-disperse-4: Operation failed on some subvolumes (up=7, mask=7, remaining=0, good=6, bad=1)
> [2016-05-18 21:13:01.582926] W [MSGID: 122053] [ec-common.c:116:ec_check_status] 0-SDSStoragePool-disperse-4: Operation failed on some subvolumes (up=7, mask=7, remaining=0, good=5, bad=2)
> [2016-05-18 21:13:01.590063] W [MSGID: 122053] [ec-common.c:116:ec_check_status] 0-SDSStoragePool-disperse-4: Operation failed on some subvolumes (up=7, mask=6, remaining=0, good=6, bad=1)
> [2016-05-18 21:13:01.590798] E [MSGID: 122034] [ec-common.c:461:ec_child_select] 0-SDSStoragePool-disperse-4: Insufficient available childs for this request (have 1, need 2)
> [2016-05-18 21:13:01.592769] W [MSGID: 122053] [ec-common.c:116:ec_check_status] 0-SDSStoragePool-disperse-4: Operation failed on some subvolumes (up=7, mask=6, remaining=0, good=6, bad=1)
> The message "W [MSGID: 122053] [ec-common.c:116:ec_check_status] 0-SDSStoragePool-disperse-4: Operation failed on some subvolumes (up=7, mask=6, remaining=0, good=6, bad=1)" repeated 3 times between [2016-05-18 21:13:01.592769] and [2016-05-18 21:13:01.598369]
> [2016-05-18 21:13:01.613843] I [MSGID: 122058] [ec-heal.c:2338:ec_heal_do] 0-SDSStoragePool-disperse-4: /Folder_05.11.2016_19.56/CV_MAGNETIC/V_22378/CHUNK_209062/SFILE_CONTAINER_020: name heal successful on 7
> [2016-05-18 21:13:01.699580] W [MSGID: 122002] [ec-common.c:71:ec_heal_report] 0-SDSStoragePool-disperse-2: Heal failed [Transport endpoint is not connected]
> [2016-05-18 21:13:01.833863] I [MSGID: 122058] [ec-heal.c:2338:ec_heal_do] 0-SDSStoragePool-disperse-4: /Folder_05.11.2016_19.56/CV_MAGNETIC/V_22378/CHUNK_209062/SFILE_CONTAINER_020: name heal successful on 7
> The message "I [MSGID: 122058] [ec-heal.c:2338:ec_heal_do] 0-SDSStoragePool-disperse-4: /Folder_05.11.2016_19.56/CV_MAGNETIC/V_22378/CHUNK_209062/SFILE_CONTAINER_020: name heal successful on 7" repeated 5 times between [2016-05-18 21:13:01.833863] and [2016-05-18 21:13:02.098833]
>
> The read from file /ws/glus/Folder_05.11.2016_19.56/CV_MAGNETIC/V_22378/CHUNK_209062/SFILE_CONTAINER_020 succeeded but the CRC verification failed for the data read.
>
> Thanks and Regards,
> Ram
>
> -----Original Message-----
> From: Xavier Hernandez [mailto:xhernandez at datalab.es]
> Sent: Monday, May 23, 2016 8:11 AM
> To: Ankireddypalle Reddy; gluster-users at gluster.org
> Subject: Re: [Gluster-users] Possible error not being returned
>
> In this case you should have more warnings/errors in your log files beside the ones related to EC. Can you post them ?
>
> EC tries to keep track of good and bad bricks, however if multiple internal operations are failing (specially if related with communications), maybe some special case happens and it's unable to determine that some brick is bad when it should.
>
> If that's the case, we need to know how this happens to try to solve it.
>
> Xavi
>
> On 23/05/16 11:22, Ankireddypalle Reddy wrote:
>> Xavier,
>> We are using disperse volume to save data being backed up by Commvault Simpana software. We are performing stress testing and have noticed that the issue happens when the 10G link is completely saturated consistently. The read would succeed but would return incorrect data. CRC checks fail on the returned data. The brick daemons are up and operational. To us it mostly appears to be communication issues. We have noticed lot of these issues when 1G NIC's were used. The error frequency went down drastically after moving the gluster traffic to 10G.
>>
>> Volume Name: SDSStoragePool
>> Type: Distributed-Disperse
>> Volume ID: c5ebb780-669f-4c31-9970-e12dae1f473c
>> Status: Started
>> Number of Bricks: 8 x (2 + 1) = 24
>> Transport-type: tcp
>> Bricks:
>> Brick1: cvltpbba1sds:/ws/disk1/ws_brick
>> Brick2: cvltpbba3sds:/ws/disk1/ws_brick
>> Brick3: cvltpbba4sds:/ws/disk1/ws_brick
>> Brick4: cvltpbba1sds:/ws/disk2/ws_brick
>> Brick5: cvltpbba3sds:/ws/disk2/ws_brick
>> Brick6: cvltpbba4sds:/ws/disk2/ws_brick
>> Brick7: cvltpbba1sds:/ws/disk3/ws_brick
>> Brick8: cvltpbba3sds:/ws/disk3/ws_brick
>> Brick9: cvltpbba4sds:/ws/disk3/ws_brick
>> Brick10: cvltpbba1sds:/ws/disk4/ws_brick
>> Brick11: cvltpbba3sds:/ws/disk4/ws_brick
>> Brick12: cvltpbba4sds:/ws/disk4/ws_brick
>> Brick13: cvltpbba1sds:/ws/disk5/ws_brick
>> Brick14: cvltpbba3sds:/ws/disk5/ws_brick
>> Brick15: cvltpbba4sds:/ws/disk5/ws_brick
>> Brick16: cvltpbba1sds:/ws/disk6/ws_brick
>> Brick17: cvltpbba3sds:/ws/disk6/ws_brick
>> Brick18: cvltpbba4sds:/ws/disk6/ws_brick
>> Brick19: cvltpbba1sds:/ws/disk7/ws_brick
>> Brick20: cvltpbba3sds:/ws/disk7/ws_brick
>> Brick21: cvltpbba4sds:/ws/disk7/ws_brick
>> Brick22: cvltpbba1sds:/ws/disk8/ws_brick
>> Brick23: cvltpbba3sds:/ws/disk8/ws_brick
>> Brick24: cvltpbba4sds:/ws/disk8/ws_brick Options Reconfigured:
>> performance.readdir-ahead: on
>>
>> Thanks and Regards,
>> Ram
>>
>> -----Original Message-----
>> From: Xavier Hernandez [mailto:xhernandez at datalab.es]
>> Sent: Monday, May 23, 2016 3:22 AM
>> To: Ankireddypalle Reddy; gluster-users at gluster.org
>> Subject: Re: [Gluster-users] Possible error not being returned
>>
>> It's possible that the operation that failed is an internal one made by disperse itself or any other translator, so this error is not reported to the application.
>>
>> The read issued by the application will only fail if anything fails while processing the read itself. If everything goes well, the read will succeed and it should contain healthy data.
>>
>> What configuration are you using ? (gluster volume info) What are you doing exactly ? (workload) Why is one brick down/damaged ? are you doing tests ? how are you doing them ?
>>
>> Best regards,
>>
>> Xavi
>>
>> On 20/05/16 16:54, Ankireddypalle Reddy wrote:
>>> Hi,
>>>
>>> Did anyone get a chance to check this. We are intermittently
>>> receiving corrupted data in read operations because of this.
>>>
>>>
>>>
>>> Thanks and Regards,
>>>
>>> Ram
>>>
>>>
>>>
>>> *From:*gluster-users-bounces at gluster.org
>>> [mailto:gluster-users-bounces at gluster.org] *On Behalf Of
>>> *Ankireddypalle Reddy
>>> *Sent:* Thursday, May 19, 2016 3:59 PM
>>> *To:* gluster-users at gluster.org
>>> *Subject:* [Gluster-users] Possible error not being returned
>>>
>>>
>>>
>>> Hi,
>>>
>>> A disperse volume was configured on servers with limited
>>> network bandwidth. Some of the read operations failed with error
>>>
>>>
>>>
>>> [2016-05-16 18:38:36.035559] E [MSGID: 122034]
>>> [ec-common.c:461:ec_child_select] 0-SDSStoragePool-disperse-2:
>>> Insufficient available childs for this request (have 1, need 2)
>>>
>>> [2016-05-16 18:38:36.035713] W [fuse-bridge.c:2213:fuse_readv_cbk]
>>> 0-glusterfs-fuse: 155121179: READ => -1 (Input/output error)
>>>
>>>
>>>
>>> For some read operations just the following error was logged but the
>>> I/O did not fail.
>>>
>>> [2016-05-16 18:42:45.401570] E [MSGID: 122034]
>>> [ec-common.c:461:ec_child_select] 0-SDSStoragePool-disperse-3:
>>> Insufficient available childs for this request (have 1, need 2)
>>>
>>> [2016-05-16 18:42:45.402054] W [MSGID: 122053]
>>> [ec-common.c:116:ec_check_status] 0-SDSStoragePool-disperse-3:
>>> Operation failed on some subvolumes (up=7, mask=6, remaining=0,
>>> good=6, bad=1)
>>>
>>>
>>>
>>> We are receiving corrupted data in the read operation when the error
>>> is logged but the read call did not return any error.
>>>
>>>
>>>
>>> Thanks and Regards,
>>>
>>> Ram
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> ***************************Legal
>>> Disclaimer***************************
>>>
>>> "This communication may contain confidential and privileged material
>>> for the
>>>
>>> sole use of the intended recipient. Any unauthorized review, use or
>>> distribution
>>>
>>> by others is strictly prohibited. If you have received the message by
>>> mistake,
>>>
>>> please advise the sender by reply email and delete the message. Thank you."
>>>
>>> *********************************************************************
>>> *
>>>
>>>
>>> ***************************Legal
>>> Disclaimer***************************
>>> "This communication may contain confidential and privileged material
>>> for the sole use of the intended recipient. Any unauthorized review,
>>> use or distribution by others is strictly prohibited. If you have
>>> received the message by mistake, please advise the sender by reply email and delete the message. Thank you."
>>> *********************************************************************
>>> *
>>>
>>>
>>>
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>
>>
>>
>>
>> ***************************Legal Disclaimer***************************
>> "This communication may contain confidential and privileged material
>> for the sole use of the intended recipient. Any unauthorized review,
>> use or distribution by others is strictly prohibited. If you have
>> received the message by mistake, please advise the sender by reply email and delete the message. Thank you."
>> **********************************************************************
>>
>
>
>
> ***************************Legal Disclaimer***************************
> "This communication may contain confidential and privileged material for the
> sole use of the intended recipient. Any unauthorized review, use or distribution
> by others is strictly prohibited. If you have received the message by mistake,
> please advise the sender by reply email and delete the message. Thank you."
> **********************************************************************
>
More information about the Gluster-users
mailing list