[Gluster-users] Corrupted File readable via FUSE?
David Spisla
spisla80 at gmail.com
Wed Feb 6 15:36:24 UTC 2019
Hello Raghavendra,
I can not give you the output of the gluster commands because I repaired
the system already. But beside of this this errors occurs randomly. I am
sure that only one copy of the file was corrupted because it is part of a
test and I corrupt one copy of the file manually on brick level and after
this I check if it is still readable. During this conversation the error
occurs again.
Here is the Log from brick of node1:
[2019-02-06 14:15:09.524638] E [MSGID: 115070]
> [server-rpc-fops_v2.c:1503:server4_open_cbk] 0-
> archive1-server: 32: OPEN /data/file1.txt
> (23b623cb-7256-4fe6-85b0-1026b1531a86), client: CTX_
>
> ID:e3871169-af62-44aa-a990-fa4248283c08-GRAPH_ID:0-PID:31830-HOST:fs-lrunning-c1-n1-PC_NAME:ar
> chive1-client-0-RECON_NO:-0, error-xlator: archive1-bitrot-stub
> [Input/output error]
> [2019-02-06 14:15:09.535587] E [MSGID: 115070]
> [server-rpc-fops_v2.c:1503:server4_open_cbk] 0-
> archive1-server: 56: OPEN /data/file1.txt
> (23b623cb-7256-4fe6-85b0-1026b1531a86), client: CTX_
>
> ID:e3871169-af62-44aa-a990-fa4248283c08-GRAPH_ID:0-PID:31830-HOST:fs-lrunning-c1-n1-PC_NAME:ar
> chive1-client-0-RECON_NO:-0, error-xlator: archive1-bitrot-stub
> [Input/output error]
> The message "E [MSGID: 116020]
> [bit-rot-stub.c:647:br_stub_check_bad_object] 0-archive1-bitrot
> -stub: 23b623cb-7256-4fe6-85b0-1026b1531a86 is a bad object. Returning"
> repeated 2 times betwe
> en [2019-02-06 14:15:09.524599] and [2019-02-06 14:15:09.549409]
> [2019-02-06 14:15:09.549427] E [MSGID: 115070]
> [server-rpc-fops_v2.c:1503:server4_open_cbk] 0-
> archive1-server: 70: OPEN /data/file1.txt
> (23b623cb-7256-4fe6-85b0-1026b1531a86), client: CTX_
>
> ID:e3871169-af62-44aa-a990-fa4248283c08-GRAPH_ID:0-PID:31830-HOST:fs-lrunning-c1-n1-PC_NAME:ar
> chive1-client-0-RECON_NO:-0, error-xlator: archive1-bitrot-stub
> [Input/output error]
> [2019-02-06 14:15:09.561450] I [MSGID: 115036]
> [server.c:469:server_rpc_notify] 0-archive1-ser
> ver: disconnecting connection from
> CTX_ID:e3871169-af62-44aa-a990-fa4248283c08-GRAPH_ID:0-PID:
> 31830-HOST:fs-lrunning-c1-n1-PC_NAME:archive1-client-0-RECON_NO:-0
> [2019-02-06 14:15:09.561568] I [MSGID: 101055]
> [client_t.c:435:gf_client_unref] 0-archive1-ser
> ver: Shutting down connection
> CTX_ID:e3871169-af62-44aa-a990-fa4248283c08-GRAPH_ID:0-PID:31830-HOST:fs-lrunning-c1-n1-PC_NAME:archive1-client-0-RECON_NO:-0
> [2019-02-06 14:15:10.188406] I [glusterfsd-mgmt.c:58:mgmt_cbk_spec]
> 0-mgmt: Volume file changed
> [2019-02-06 14:15:10.201029] I [glusterfsd-mgmt.c:2005:mgmt_getspec_cbk]
> 0-glusterfs: No change in volfile,continuing
> [2019-02-06 14:15:10.514721] I [glusterfsd-mgmt.c:58:mgmt_cbk_spec]
> 0-mgmt: Volume file changed
> [2019-02-06 14:15:10.526216] I [glusterfsd-mgmt.c:2005:mgmt_getspec_cbk]
> 0-glusterfs: No change in volfile,continuing
> The message "E [MSGID: 101191]
> [event-epoll.c:671:event_dispatch_epoll_worker] 0-epoll: Failed to dispatch
> handler" repeated 79 times between [2019-02-06 14:15:09.499105] and
> [2019-02-06 14:15:10.682592]
> [2019-02-06 14:15:10.684204] E [MSGID: 116020]
> [bit-rot-stub.c:647:br_stub_check_bad_object] 0-archive1-bitrot-stub:
> 23b623cb-7256-4fe6-85b0-1026b1531a86 is a bad object. Returning
> [2019-02-06 14:15:10.684262] E [MSGID: 115070]
> [server-rpc-fops_v2.c:1503:server4_open_cbk] 0-archive1-server: 2146148:
> OPEN /data/file1.txt (23b623cb-7256-4fe6-85b0-1026b1531a86), client:
> CTX_ID:0545b52c-2843-4833-a5fc-b11e062a72d3-GRAPH_ID:0-PID:2458-HOST:fs-lrunning-c1-n1-PC_NAME:archive1-client-0-RECON_NO:-3,
> error-xlator: archive1-bitrot-stub [Input/output error]
> [2019-02-06 14:15:10.684949] E [MSGID: 116020]
> [bit-rot-stub.c:647:br_stub_check_bad_object] 0-archive1-bitrot-stub:
> 23b623cb-7256-4fe6-85b0-1026b1531a86 is a bad object. Returning
> [2019-02-06 14:15:10.684982] E [MSGID: 115070]
> [server-rpc-fops_v2.c:1503:server4_open_cbk] 0-archive1-server: 2146149:
> OPEN /data/file1.txt (23b623cb-7256-4fe6-85b0-1026b1531a86), client:
> CTX_ID:0545b52c-2843-4833-a5fc-b11e062a72d3-GRAPH_ID:0-PID:2458-HOST:fs-lrunning-c1-n1-PC_NAME:archive1-client-0-RECON_NO:-3,
> error-xlator: archive1-bitrot-stub [Input/output error]
> [2019-02-06 14:15:10.686566] E [MSGID: 116020]
> [bit-rot-stub.c:647:br_stub_check_bad_object] 0-archive1-bitrot-stub:
> 23b623cb-7256-4fe6-85b0-1026b1531a86 is a bad object. Returning
> [2019-02-06 14:15:10.686600] E [MSGID: 115070]
> [server-rpc-fops_v2.c:1503:server4_open_cbk] 0-archive1-server: 2146150:
> OPEN /data/file1.txt (23b623cb-7256-4fe6-85b0-1026b1531a86), client:
> CTX_ID:0545b52c-2843-4833-a5fc-b11e062a72d3-GRAPH_ID:0-PID:2458-HOST:fs-lrunning-c1-n1-PC_NAME:archive1-client-0-RECON_NO:-3,
> error-xlator: archive1-bitrot-stub [Input/output error]
> [2019-02-06 14:15:11.189361] I [glusterfsd-mgmt.c:58:mgmt_cbk_spec]
> 0-mgmt: Volume file changed
> [2019-02-06 14:15:11.207835] I [glusterfsd-mgmt.c:58:mgmt_cbk_spec]
> 0-mgmt: Volume file changed
> [2019-02-06 14:15:11.220763] I [glusterfsd-mgmt.c:58:mgmt_cbk_spec]
> 0-mgmt: Volume file changed
>
>
One can see that there is bitrot file on brick of node1. This seems to be
correct.
Here the Log of the FUSE Mount Node1:
[2019-02-06 14:15:10.684387] E [MSGID: 114031]
> [client-rpc-fops_v2.c:281:client4_0_open_cbk] 0-archive1-client-0: remote
> operation failed. Path: /data/file1.txt
> (23b623cb-7256-4fe6-85b0-1026b1531a86) [Input/output error]
> [2019-02-06 14:15:10.684556] W [dict.c:761:dict_ref]
> (-->/usr/lib64/glusterfs/5.3/xlator/performance/quick-read.so(+0x7329)
> [0x7feff4bde329]
> -->/usr/lib64/glusterfs/5.3/xlator/performance/io-cache.so(+0xaaf5)
> [0x7feff4defaf5] -->/usr/lib64/libglusterfs.so.0(dict_ref+0x58)
> [0x7feffcecf218] ) 0-dict: dict is NULL [Invalid argument]
> [2019-02-06 14:15:10.685122] E [MSGID: 114031]
> [client-rpc-fops_v2.c:281:client4_0_open_cbk] 0-archive1-client-0: remote
> operation failed. Path: /data/file1.txt
> (23b623cb-7256-4fe6-85b0-1026b1531a86) [Input/output error]
> [2019-02-06 14:15:10.685127] E [MSGID: 108009]
> [afr-open.c:220:afr_openfd_fix_open_cbk] 0-archive1-replicate-0: Failed to
> open /data/file1.txt on subvolume archive1-client-0 [Input/output error]
> [2019-02-06 14:15:10.686207] W [fuse-bridge.c:2371:fuse_readv_cbk]
> 0-glusterfs-fuse: 4623583: READ => -1
> gfid=23b623cb-7256-4fe6-85b0-1026b1531a86 fd=0x7fefa8c5d618 (Transport
> endpoint is not connected)
> [2019-02-06 14:15:10.686306] W [dict.c:761:dict_ref]
> (-->/usr/lib64/glusterfs/5.3/xlator/performance/quick-read.so(+0x7329)
> [0x7feff4bde329]
> -->/usr/lib64/glusterfs/5.3/xlator/performance/io-cache.so(+0xaaf5)
> [0x7feff4defaf5] -->/usr/lib64/libglusterfs.so.0(dict_ref+0x58)
> [0x7feffcecf218] ) 0-dict: dict is NULL [Invalid argument]
> [2019-02-06 14:15:10.686690] E [MSGID: 114031]
> [client-rpc-fops_v2.c:281:client4_0_open_cbk] 0-archive1-client-0: remote
> operation failed. Path: /data/file1.txt
> (23b623cb-7256-4fe6-85b0-1026b1531a86) [Input/output error]
> [2019-02-06 14:15:10.686714] E [MSGID: 108009]
> [afr-open.c:220:afr_openfd_fix_open_cbk] 0-archive1-replicate-0: Failed to
> open /data/file1.txt on subvolume archive1-client-0 [Input/output error]
> [2019-02-06 14:15:10.686877] W [fuse-bridge.c:2371:fuse_readv_cbk]
> 0-glusterfs-fuse: 4623584: READ => -1
> gfid=23b623cb-7256-4fe6-85b0-1026b1531a86 fd=0x7fefa8c5d618 (Transport
> endpoint is not connected)
> [2019-02-06 14:15:10.687500] W [MSGID: 114028]
> [client-lk.c:347:delete_granted_locks_owner] 0-archive1-client-0: fdctx not
> valid [Invalid argument]
>
>
One can see an "Input/output error" because of the corrupted file from
brick of node1. At this time the *brick on node 2 was really down* but on
Node 3, 4 they were up. So still 2 good copies are reachable. Or not?
The Log of the bricks from node 3,4 has no entry for this "file1.txt". It
seems to be that the Client Stack did no requests to this bricks.
Example Log of brick from node 3:
[2019-02-06 14:15:09.561650] E [MSGID: 101191]
> [event-epoll.c:671:event_dispatch_epoll_worker] 0-epoll: Failed to dispatch
> handler
> [2019-02-06 14:15:10.220218] I [glusterfsd-mgmt.c:58:mgmt_cbk_spec]
> 0-mgmt: Volume file changed
> [2019-02-06 14:15:10.236379] I [glusterfsd-mgmt.c:2005:mgmt_getspec_cbk]
> 0-glusterfs: No change in volfile,continuing
> [2019-02-06 14:15:10.541472] I [glusterfsd-mgmt.c:58:mgmt_cbk_spec]
> 0-mgmt: Volume file changed
> [2019-02-06 14:15:10.556125] I [glusterfsd-mgmt.c:2005:mgmt_getspec_cbk]
> 0-glusterfs: No change in volfile,continuing
> [2019-02-06 14:15:11.248253] I [glusterfsd-mgmt.c:58:mgmt_cbk_spec]
> 0-mgmt: Volume file changed
> [2019-02-06 14:15:11.264428] I [glusterfsd-mgmt.c:58:mgmt_cbk_spec]
> 0-mgmt: Volume file changed
> [2019-02-06 14:15:11.277016] I [glusterfsd-mgmt.c:58:mgmt_cbk_spec]
> 0-mgmt: Volume file changed
>
>
Is there a hidden quorum active? I have a 4-way Replica Volume, so 2 of 4
Copies are good and reachable
Regards
David
Am Di., 5. Feb. 2019 um 21:06 Uhr schrieb FNU Raghavendra Manjunath <
rabhat at redhat.com>:
>
> Hi David,
>
> Do you have any bricks down? Can you please share the output of the
> following commands and also the logs of the server and the client nodes?
>
> 1) gluster volume info
> 2) gluster volume status
> 3) gluster volume bitrot <volume name> scrub status
>
> Few more questions
>
> 1) How many copies of the file were corrupted? (All? Or Just one?)
>
> 2 things I am trying to understand
>
> A) IIUC, if only one copy is corrupted, then the replication module from
> the gluster client should serve the data from the
> remaining good copy
> B) If all the copies were corrupted (or say more than quorum copies were
> corrupted which means 2 in case of 3 way replication)
> then there will be an error to the application. But the error to be
> reported should 'Input/Output Error'. Not 'Transport endpoint not connected'
> 'Transport endpoint not connected' error usually comes when a brick
> where the operation is being directed to is not connected to the client.
>
>
>
> Regards,
> Raghavendra
>
> On Mon, Feb 4, 2019 at 6:02 AM David Spisla <spisla80 at gmail.com> wrote:
>
>> Hello Amar,
>> sounds good. Until now this patch is only merged into master. I think it
>> should be part of the next v5.x patch release!
>>
>> Regards
>> David
>>
>> Am Mo., 4. Feb. 2019 um 09:58 Uhr schrieb Amar Tumballi Suryanarayan <
>> atumball at redhat.com>:
>>
>>> Hi David,
>>>
>>> I guess https://review.gluster.org/#/c/glusterfs/+/21996/ helps to fix
>>> the issue. I will leave it to Raghavendra Bhat to reconfirm.
>>>
>>> Regards,
>>> Amar
>>>
>>> On Fri, Feb 1, 2019 at 8:45 PM David Spisla <spisla80 at gmail.com> wrote:
>>>
>>>> Hello Gluster Community,
>>>> I have got a 4 Node Cluster with a Replica 4 Volume, so each node has a
>>>> brick with a copy of a file. Now I tried out the bitrot functionality and
>>>> corrupt the copy on the brick of node1. After this I scrub ondemand and the
>>>> file is marked correctly as corrupted.
>>>>
>>>> No I try to read that file from FUSE on node1 (with corrupt copy):
>>>> $ cat file1.txt
>>>> cat: file1.txt: Transport endpoint is not connected
>>>> FUSE log says:
>>>>
>>>> *[2019-02-01 15:02:19.191984] E [MSGID: 114031]
>>>> [client-rpc-fops_v2.c:281:client4_0_open_cbk] 0-archive1-client-0: remote
>>>> operation failed. Path: /data/file1.txt
>>>> (b432c1d6-ece2-42f2-8749-b11e058c4be3) [Input/output error]*
>>>> [2019-02-01 15:02:19.192269] W [dict.c:761:dict_ref]
>>>> (-->/usr/lib64/glusterfs/5.3/xlator/performance/quick-read.so(+0x7329)
>>>> [0x7fc642471329]
>>>> -->/usr/lib64/glusterfs/5.3/xlator/performance/io-cache.so(+0xaaf5)
>>>> [0x7fc642682af5] -->/usr/lib64/libglusterfs.so.0(dict_ref+0x58)
>>>> [0x7fc64a78d218] ) 0-dict: dict is NULL [Invalid argument]
>>>> [2019-02-01 15:02:19.192714] E [MSGID: 108009]
>>>> [afr-open.c:220:afr_openfd_fix_open_cbk] 0-archive1-replicate-0: Failed to
>>>> open /data/file1.txt on subvolume archive1-client-0 [Input/output error]
>>>> *[2019-02-01 15:02:19.193009] W [fuse-bridge.c:2371:fuse_readv_cbk]
>>>> 0-glusterfs-fuse: 147733: READ => -1
>>>> gfid=b432c1d6-ece2-42f2-8749-b11e058c4be3 fd=0x7fc60408bbb8 (Transport
>>>> endpoint is not connected)*
>>>> [2019-02-01 15:02:19.193653] W [MSGID: 114028]
>>>> [client-lk.c:347:delete_granted_locks_owner] 0-archive1-client-0: fdctx not
>>>> valid [Invalid argument]
>>>>
>>>> And from FUSE on node2 (with heal copy):
>>>> $ cat file1.txt
>>>> file1
>>>>
>>>> It seems to be that node1 wants to get the file from its own brick, but
>>>> the copy there is broken. Node2 gets the file from its own brick with a
>>>> heal copy, so reading the file succeed.
>>>> But I am wondering myself because sometimes reading the file from node1
>>>> with the broken copy succeed
>>>>
>>>> What is the expected behaviour here? Is it possibly to read files with
>>>> a corrupted copy from any client access?
>>>>
>>>> Regards
>>>> David Spisla
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users at gluster.org
>>>> https://lists.gluster.org/mailman/listinfo/gluster-users
>>>
>>>
>>>
>>> --
>>> Amar Tumballi (amarts)
>>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20190206/0ffb10c6/attachment.html>
More information about the Gluster-users
mailing list