[Gluster-users] posix set mdata failed, No ctime

Kotresh Hiremath Ravishankar khiremat at redhat.com
Sat Sep 22 04:44:43 UTC 2018


You can ignore this error. It is fixed and should be available in next
4.1.x release.



On Sat, 22 Sep 2018, 07:07 Pedro Costa, <pedro at pmc.digital> wrote:

> Forgot to mention, I’m running all VM’s with 16.04.1-Ubuntu,  Kernel
> 4.15.0-1023-azure #24
>
>
>
>
>
> *From:* Pedro Costa
> *Sent:* 21 September 2018 10:16
> *To:* 'gluster-users at gluster.org' <gluster-users at gluster.org>
> *Subject:* posix set mdata failed, No ctime
>
>
>
> Hi,
>
>
>
> I have a replicate x3 volume with the following config:
>
>
>
> ```
>
> Volume Name: gvol1
>
> Type: Replicate
>
> Volume ID: 384acec2-5b5f-40da-bf0e-5c53d12b3ae2
>
> Status: Started
>
> Snapshot Count: 0
>
> Number of Bricks: 1 x 3 = 3
>
> Transport-type: tcp
>
> Bricks:
>
> Brick1: vm0:/srv/brick1/gvol1
>
> Brick2: vm1:/srv/brick1/gvol1
>
> Brick3: vm2:/srv/brick1/gvol1
>
> Options Reconfigured:
>
> storage.ctime: on
>
> features.utime: on
>
> storage.fips-mode-rchecksum: on
>
> performance.client-io-threads: off
>
> nfs.disable: on
>
> transport.address-family: inet
>
> ```
>
>
>
> This volume was actually created on v3.8, but as since been upgraded
> (version by version) to v4.1.4 and it’s working fine (for the most part):
>
>
>
> ```
>
> Client connections for volume gvol1
>
> ----------------------------------------------
>
> Brick : vm0:/srv/brick1/gvol1
>
> Clients connected : 6
>
> Hostname                                               BytesRead
> BytesWritten       OpVersion
>
> --------                                               ---------
> ------------       ---------
>
> 10.X.0.5:49143                                          2096520
> 2480212           40100
>
> 10.X.0.6:49141                                            14000
> 12812           40100
>
> 10.X.0.4:49134                                           258324
> 333456           40100
>
> 10.X.0.4:49141                                        565372566
> 1643447105           40100
>
> 10.X.0.5:49145                                        491262003
> 291782440           40100
>
> 10.X.0.6:49139                                        482629418
> 328228888           40100
>
> ----------------------------------------------
>
> Brick : vm1:/srv/brick1/gvol1
>
> Clients connected : 6
>
> Hostname                                               BytesRead
> BytesWritten       OpVersion
>
> --------                                               ---------
> ------------       ---------
>
> 10.X.0.6:49146                                           658516
> 508904           40100
>
> 10.X.0.5:49133                                          4142848
> 7139858           40100
>
> 10.X.0.4:49138                                             4088
> 3696           40100
>
> 10.X.0.4:49140                                        471405874
> 284488736           40100
>
> 10.X.0.5:49140                                        585193563
> 1670630439           40100
>
> 10.X.0.6:49138                                        482407454
> 330274812           40100
>
> ----------------------------------------------
>
> Brick : vm2:/srv/brick1/gvol1
>
> Clients connected : 6
>
> Hostname                                               BytesRead
> BytesWritten       OpVersion
>
> --------                                               ---------
> ------------       ---------
>
> 10.X.0.6:49133                                          1789624
> 4340938           40100
>
> 10.X.0.5:49137                                          3010064
> 3005184           40100
>
> 10.X.0.4:49143                                             4268
> 3744           40100
>
> 10.X.0.4:49139                                        471328402
> 283798376           40100
>
> 10.X.0.5:49139                                        491404443
> 293342568           40100
>
> 10.X.0.6:49140                                        561683906
> 830511730           40100
>
> ----------------------------------------------
>
> ```
>
>
>
> I’m now getting a lot of errors on the brick log file, like:
>
>
>
> `The message "W [MSGID: 113117] [posix-metadata.c:627:posix_set_ctime]
> 0-gvol1-posix: posix set mdata failed, No ctime :
> /srv/brick1/gvol1/.glusterfs/18/d0/18d04927-1ec0-4779-8c5b-7ebb82e4a614
> gfid:18d04927-1ec0-4779-8c5b-7ebb82e4a614 [Function not implemented]"
> repeated 2 times between [2018-09-21 08:21:52.480797] and [2018-09-21
> 08:22:07.529625]`
>
>
>
> For different files but the most common is a file that the Node.js
> application that runs on top of the gluster via a fuse client (glusterfs)
> stats every 5s for changes,
> https://nodejs.org/api/fs.html#fs_fs_stat_path_options_callback
>
>
>
> I think this is also related to another issue, when reading the file it
> returns an empty result (not always), as the app reports:
>
>
>
> `2018-09-21 08:22:00 | [vm0] [nobody] sync hosts: invalid
> applications.json, response was empty.`
>
>
>
> Doing `gluster volume heal gvol1 info` yields 0 for all bricks.
>
>
>
> Should I be concerned about the warning, is this a known issue? If not,
> what could be causing the empty file return sometimes? Could they be
> related?
>
>
>
> The application that is running on top of the cluster build and spawns
> other node.js applications with mostly small files, do you have any
> optimization tips for it?
>
>
>
> FYI it is a slightly modified version of
> https://github.com/totaljs/superadmin to run as a web farm.
>
>
>
> Thank you so much for any help in advance,
>
>
>
> Kind Regards,
>
> Pedro Maia Costa
>
>
>
>
>
>
>
>
>
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20180922/cbc023c1/attachment.html>


More information about the Gluster-users mailing list