[Gluster-users] One error/warning message after upgrade 5.11 -> 6.8

Hu Bert revirii at googlemail.com
Mon Jun 8 09:48:22 UTC 2020


Hi Strahil,

thx for your answer, but i assume that your approach won't help. It
seems like that this behaviour is permanent; e.g. a log entry like
this:

[2020-06-08 09:40:03.948269] E [MSGID: 113001]
[posix-metadata.c:234:posix_store_mdata_xattr] 0-persistent-posix:
file: /gluster/md3/persistent/.glusterfs/38/30/38306ef8-6588-40cf-8be3-c0a022714612:
gfid: 38306ef8-6588-40cf-8be3-c0a022714612 key:trusted.glusterfs.mdata
 [No such file or directory]
[2020-06-08 09:40:03.948333] E [MSGID: 113114]
[posix-metadata.c:433:posix_set_mdata_xattr_legacy_files]
0-persistent-posix: gfid: 38306ef8-6588-40cf-8be3-c0a022714612
key:trusted.glusterfs.mdata  [No such file or directory]
[2020-06-08 09:40:03.948422] I [MSGID: 115060]
[server-rpc-fops.c:938:_gf_server_log_setxattr_failure]
0-persistent-server: 14193413: SETXATTR
/images/generated/207/039/2070391/484x425r.jpg
(38306ef8-6588-40cf-8be3-c0a022714612) ==> set-ctime-mdata, client:
CTX_ID:b738017c-20a3-4547-afba-5b8933d8e6e5-GRAPH_ID:0-PID:1078-HOST:pepe-PC_NAME:persistent-client-2-RECON_NO:-1,
error-xlator: persistent-posix

tells me that an error (ctime-mdata) is found and fixed. And this is
happening over and over again. A couple of minutes ago i wanted to
begin with what you suggested and called 'gluster volume heal
persistent info' and suddenly saw:

Brick gluster1:/gluster/md3/persistent
Status: Connected
Number of entries: 0

Brick gluster2:/gluster/md3/persistent
Status: Connected
Number of entries: 0

Brick gluster3:/gluster/md3/persistent
Status: Connected
Number of entries: 0

I thought 'wtf...'; the heal-count was 0 as well; but the next call
~15s later showed this again:

Brick gluster1:/gluster/md3/persistent
Number of entries: 31

Brick gluster2:/gluster/md3/persistent
Number of entries: 27

Brick gluster3:/gluster/md3/persistent
Number of entries: 4

For me it looks like the 'error found -> heal it' process works as it
should, but due to the permanent errors (log file entries) the heal
count of zero is almost impossible to read.

Well, one could deactivate features.ctime as this seems to be the
reason (as the log entries suggest), but i don't know if that is
reasonable, i.e. if this feature is needed.


Best regards,
Hubert

Am Mo., 8. Juni 2020 um 11:22 Uhr schrieb Strahil Nikolov
<hunter86_bg at yahoo.com>:
>
> Hi Hubert,
>
> Here is one idea:
> Using 'gluster volume  heal VOL  info' can provide  the gfids of files pending heal.
> Once  you have them, you can find the inode of each file via 'ls  -li /gluster/brick/.gfid/<first_two_characters_of_gfid>/<next_two_characters>/gfid
>
> Then you can search the brick with find for that inode number (don't forget the 'ionice' to reduce the pressure).
>
> Once  you have the list of files, stat them via the FUSE client and check if they got healed.
>
> I fully agree that you need to first heal the golumes  before proceeding further or you might get into a nasty situation.
>
> Best Regards,
> Strahil Nikolov
>
>
> На 8 юни 2020 г. 8:30:57 GMT+03:00, Hu Bert <revirii at googlemail.com> написа:
> >Good morning,
> >
> >i just wanted to update the version from 6.8 to 6.9 on our replicate 3
> >system (formerly was version 5.11), and i see tons of these messages:
> >
> >[2020-06-08 05:25:55.192301] E [MSGID: 113001]
> >[posix-metadata.c:234:posix_store_mdata_xattr] 0-persistent-posix:
> >file:
> >/gluster/md3/persistent/.glusterfs/43/31/43312aba-75c6-42c2-855c-e0db66d7748f:
> >gfid: 43312aba-75c6-42c2-855c-e0db66d7748f key:trusted.glusterfs.mdata
> > [No such file or directory]
> >[2020-06-08 05:25:55.192375] E [MSGID: 113114]
> >[posix-metadata.c:433:posix_set_mdata_xattr_legacy_files]
> >0-persistent-posix: gfid: 43312aba-75c6-42c2-855c-e0db66d7748f
> >key:trusted.glusterfs.mdata  [No such file or directory]
> >[2020-06-08 05:25:55.192426] I [MSGID: 115060]
> >[server-rpc-fops.c:938:_gf_server_log_setxattr_failure]
> >0-persistent-server: 13382741: SETXATTR
> ><gfid:43312aba-75c6-42c2-855c-e0db66d7748f>
> >(43312aba-75c6-42c2-855c-e0db66d7748f) ==> set-ctime-mdata, client:
> >CTX_ID:e223ca30-6c30-4a40-ae98-a418143ce548-GRAPH_ID:0-PID:1006-HOST:sam-PC_NAME:persistent-client-2-RECON_NO:-1,
> >error-xlator: persistent-posix
> >
> >Still the ctime-message. And a lot of these messages:
> >
> >[2020-06-08 05:25:53.016606] W [MSGID: 101159]
> >[inode.c:1330:__inode_unlink] 0-inode:
> >7043eed7-dbd7-4277-976f-d467349c1361/21194684.jpg: dentry not found in
> >839512f0-75de-414f-993d-1c35892f8560
> >
> >Well... the problem is: the volume seems to be in a permanent heal
> >status:
> >
> >Gathering count of entries to be healed on volume persistent has been
> >successful
> >Brick gluster1:/gluster/md3/persistent
> >Number of entries: 31
> >Brick gluster2:/gluster/md3/persistent
> >Number of entries: 6
> >Brick gluster3:/gluster/md3/persistent
> >Number of entries: 5
> >
> >a bit later:
> >Gathering count of entries to be healed on volume persistent has been
> >successful
> >Brick gluster1:/gluster/md3/persistent
> >Number of entries: 100
> >Brick gluster2:/gluster/md3/persistent
> >Number of entries: 74
> >Brick gluster3:/gluster/md3/persistent
> >Number of entries: 1
> >
> >The number of entries never reach 0-0-0; i already updated one of the
> >systems from 6.8 to 6.9, but updating the other 2 when heal isn't zero
> >doesn't seem to be a good idea. Well... any idea?
> >
> >
> >Best regards,
> >Hubert
> >
> >Am Fr., 8. Mai 2020 um 21:47 Uhr schrieb Strahil Nikolov
> ><hunter86_bg at yahoo.com>:
> >>
> >> On April 21, 2020 8:00:32 PM GMT+03:00, Amar Tumballi
> ><amar at kadalu.io> wrote:
> >> >There seems to be a burst of issues when people upgraded to 5.x or
> >6.x
> >> >from
> >> >3.12 (Thanks to you and Strahil, who have reported most of them).
> >> >
> >> >Latest update from Strahil is that if files are copied fresh on 7.5
> >> >series,
> >> >there are no issues.
> >> >
> >> >We are in process of identifying the patch, and also provide an
> >option
> >> >to
> >> >disable 'acl' for testing. Will update once we identify the issue.
> >> >
> >> >Regards,
> >> >Amar
> >> >
> >> >
> >> >
> >> >On Sat, Apr 11, 2020 at 11:10 AM Hu Bert <revirii at googlemail.com>
> >> >wrote:
> >> >
> >> >> Hi,
> >> >>
> >> >> no one has seen such messages?
> >> >>
> >> >> Regards,
> >> >> Hubert
> >> >>
> >> >> Am Mo., 6. Apr. 2020 um 06:13 Uhr schrieb Hu Bert
> >> ><revirii at googlemail.com
> >> >> >:
> >> >> >
> >> >> > Hello,
> >> >> >
> >> >> > i just upgraded my servers and clients from 5.11 to 6.8; besides
> >> >one
> >> >> > connection problem to the gluster download server everything
> >went
> >> >> > fine.
> >> >> >
> >> >> > On the 3 gluster servers i mount the 2 volumes as well, and only
> >> >there
> >> >> > (and not on all the other clients) there are some messages in
> >the
> >> >log
> >> >> > file of both mount logs:
> >> >> >
> >> >> > [2020-04-06 04:10:53.552561] W [MSGID: 114031]
> >> >> > [client-rpc-fops_v2.c:851:client4_0_setxattr_cbk]
> >> >> > 0-persistent-client-2: remote operation failed [Permission
> >denied]
> >> >> > [2020-04-06 04:10:53.552635] W [MSGID: 114031]
> >> >> > [client-rpc-fops_v2.c:851:client4_0_setxattr_cbk]
> >> >> > 0-persistent-client-1: remote operation failed [Permission
> >denied]
> >> >> > [2020-04-06 04:10:53.552639] W [MSGID: 114031]
> >> >> > [client-rpc-fops_v2.c:851:client4_0_setxattr_cbk]
> >> >> > 0-persistent-client-0: remote operation failed [Permission
> >denied]
> >> >> > [2020-04-06 04:10:53.553226] E [MSGID: 148002]
> >> >> > [utime.c:146:gf_utime_set_mdata_setxattr_cbk]
> >0-persistent-utime:
> >> >dict
> >> >> > set of key for set-ctime-mdata failed [Permission denied]
> >> >> > The message "W [MSGID: 114031]
> >> >> > [client-rpc-fops_v2.c:851:client4_0_setxattr_cbk]
> >> >> > 0-persistent-client-2: remote operation failed [Permission
> >denied]"
> >> >> > repeated 4 times between [2020-04-06 04:10:53.552561] and
> >> >[2020-04-06
> >> >> > 04:10:53.745542]
> >> >> > The message "W [MSGID: 114031]
> >> >> > [client-rpc-fops_v2.c:851:client4_0_setxattr_cbk]
> >> >> > 0-persistent-client-1: remote operation failed [Permission
> >denied]"
> >> >> > repeated 4 times between [2020-04-06 04:10:53.552635] and
> >> >[2020-04-06
> >> >> > 04:10:53.745610]
> >> >> > The message "W [MSGID: 114031]
> >> >> > [client-rpc-fops_v2.c:851:client4_0_setxattr_cbk]
> >> >> > 0-persistent-client-0: remote operation failed [Permission
> >denied]"
> >> >> > repeated 4 times between [2020-04-06 04:10:53.552639] and
> >> >[2020-04-06
> >> >> > 04:10:53.745632]
> >> >> > The message "E [MSGID: 148002]
> >> >> > [utime.c:146:gf_utime_set_mdata_setxattr_cbk]
> >0-persistent-utime:
> >> >dict
> >> >> > set of key for set-ctime-mdata failed [Permission denied]"
> >repeated
> >> >4
> >> >> > times between [2020-04-06 04:10:53.553226] and [2020-04-06
> >> >> > 04:10:53.746080]
> >> >> >
> >> >> > Anything to worry about?
> >> >> >
> >> >> >
> >> >> > Regards,
> >> >> > Hubert
> >> >> ________
> >> >>
> >> >>
> >> >>
> >> >> Community Meeting Calendar:
> >> >>
> >> >> Schedule -
> >> >> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
> >> >> Bridge: https://bluejeans.com/441850968
> >> >>
> >> >> Gluster-users mailing list
> >> >> Gluster-users at gluster.org
> >> >> https://lists.gluster.org/mailman/listinfo/gluster-users
> >> >>
> >>
> >> Hi,
> >>
> >> Can you provide the xfs_info for the bricks from the volume ?
> >>
> >> I have a theory that I want to confirm or reject.
> >>
> >> Best Regards,
> >> Strahil Nikolov


More information about the Gluster-users mailing list