[Gluster-users] Strange Logs
Artem Russakovskii
archon810 at gmail.com
Fri Feb 14 21:17:19 UTC 2020
I've been seeing the same thing happen, and in our case, it's because of
running a script that checks gluster from time to time (
https://github.com/jtopjian/scripts/blob/master/gluster/gluster-status.sh
in our case).
Do you have a job that runs and periodically checks for gluster health?
Sincerely,
Artem
--
Founder, Android Police <http://www.androidpolice.com>, APK Mirror
<http://www.apkmirror.com/>, Illogical Robot LLC
beerpla.net | @ArtemR <http://twitter.com/ArtemR>
On Fri, Feb 14, 2020 at 3:10 AM Christian Reiss <email at christian-reiss.de>
wrote:
> Hey folks,
>
> my logs are constantly (every few secs, continuously) swamped with
>
> [2020-02-14 11:05:20.258542] I [MSGID: 114046]
> [client-handshake.c:1106:client_setvolume_cbk] 0-ssd_storage-client-0:
> Connected to ssd_storage-client-0, attached to remote volume
> '/gluster_bricks/node01.company.com/gluster'.
> [2020-02-14 11:05:20.258559] I [MSGID: 108005]
> [afr-common.c:5280:__afr_handle_child_up_event]
> 0-ssd_storage-replicate-0: Subvolume 'ssd_storage-client-0' came back
> up; going online.
> [2020-02-14 11:05:20.258920] I [rpc-clnt.c:1963:rpc_clnt_reconfig]
> 0-ssd_storage-client-2: changing port to 49152 (from 0)
> [2020-02-14 11:05:20.259132] I [socket.c:864:__socket_shutdown]
> 0-ssd_storage-client-2: intentional socket shutdown(11)
> [2020-02-14 11:05:20.260010] I [MSGID: 114057]
> [client-handshake.c:1376:select_server_supported_programs]
> 0-ssd_storage-client-1: Using Program GlusterFS 4.x v1, Num (1298437),
> Version (400)
> [2020-02-14 11:05:20.261077] I [MSGID: 114046]
> [client-handshake.c:1106:client_setvolume_cbk] 0-ssd_storage-client-1:
> Connected to ssd_storage-client-1, attached to remote volume
> '/gluster_bricks/node02.company.com/gluster'.
> [2020-02-14 11:05:20.261089] I [MSGID: 108002]
> [afr-common.c:5647:afr_notify] 0-ssd_storage-replicate-0: Client-quorum
> is met
> [2020-02-14 11:05:20.262005] I [MSGID: 114057]
> [client-handshake.c:1376:select_server_supported_programs]
> 0-ssd_storage-client-2: Using Program GlusterFS 4.x v1, Num (1298437),
> Version (400)
> [2020-02-14 11:05:20.262685] I [MSGID: 114046]
> [client-handshake.c:1106:client_setvolume_cbk] 0-ssd_storage-client-2:
> Connected to ssd_storage-client-2, attached to remote volume
> '/gluster_bricks/node03.company.com/gluster'.
> [2020-02-14 11:05:20.263909] I [MSGID: 108031]
> [afr-common.c:2580:afr_local_discovery_cbk] 0-ssd_storage-replicate-0:
> selecting local read_child ssd_storage-client-0
> [2020-02-14 11:05:20.264124] I [MSGID: 104041]
> [glfs-resolve.c:954:__glfs_active_subvol] 0-ssd_storage: switched to
> graph 6e6f6465-3031-2e64-632d-6475732e6461 (0)
> [2020-02-14 11:05:22.407851] I [MSGID: 114007]
> [client.c:2478:client_check_remote_host] 0-ssd_storage-snapd-client:
> Remote host is not set. Assuming the volfile server as remote host
> [Invalid argument]
> [2020-02-14 11:05:22.409711] I [MSGID: 104045] [glfs-master.c:80:notify]
> 0-gfapi: New graph 6e6f6465-3031-2e64-632d-6475732e6461 (0) coming up
> [2020-02-14 11:05:22.409738] I [MSGID: 114020] [client.c:2436:notify]
> 0-ssd_storage-client-0: parent translators are ready, attempting connect
> on transport
> [2020-02-14 11:05:22.412949] I [MSGID: 114020] [client.c:2436:notify]
> 0-ssd_storage-client-1: parent translators are ready, attempting connect
> on transport
> [2020-02-14 11:05:22.413130] I [rpc-clnt.c:1963:rpc_clnt_reconfig]
> 0-ssd_storage-client-0: changing port to 49152 (from 0)
> [2020-02-14 11:05:22.413154] I [socket.c:864:__socket_shutdown]
> 0-ssd_storage-client-0: intentional socket shutdown(10)
> [2020-02-14 11:05:22.415534] I [MSGID: 114020] [client.c:2436:notify]
> 0-ssd_storage-client-2: parent translators are ready, attempting connect
> on transport
> [2020-02-14 11:05:22.417836] I [MSGID: 114057]
> [client-handshake.c:1376:select_server_supported_programs]
> 0-ssd_storage-client-0: Using Program GlusterFS 4.x v1, Num (1298437),
> Version (400)
> [2020-02-14 11:05:22.418036] I [rpc-clnt.c:1963:rpc_clnt_reconfig]
> 0-ssd_storage-client-1: changing port to 49152 (from 0)
> [2020-02-14 11:05:22.418095] I [socket.c:864:__socket_shutdown]
> 0-ssd_storage-client-1: intentional socket shutdown(12)
> [2020-02-14 11:05:22.420029] I [MSGID: 114020] [client.c:2436:notify]
> 0-ssd_storage-snapd-client: parent translators are ready, attempting
> connect on transport
> [2020-02-14 11:05:22.420533] E [MSGID: 101075]
> [common-utils.c:505:gf_resolve_ip6] 0-resolver: getaddrinfo failed
> (family:2) (Name or service not known)
> [2020-02-14 11:05:22.420545] E
> [name.c:266:af_inet_client_get_remote_sockaddr]
> 0-ssd_storage-snapd-client: DNS resolution failed on host
> /var/run/glusterd.socket
> Final graph:
>
> +------------------------------------------------------------------------------+
> 1: volume ssd_storage-client-0
> 2: type protocol/client
> 3: option opversion 70000
> 4: option clnt-lk-version 1
> 5: option volfile-checksum 0
> 6: option volfile-key ssd_storage
> 7: option client-version 7.0
> 8: option process-name gfapi.glfsheal
> 9: option process-uuid
>
> CTX_ID:50cec79e-6028-4e6f-b8ed-dda9db36b2d0-GRAPH_ID:0-PID:24926-HOST:node01.company.com-PC_NAME:ssd_storage-client-0-RECON_NO:-0
> 10: option fops-version 1298437
> 11: option ping-timeout 42
> 12: option remote-host node01.company.com
> 13: option remote-subvolume /gluster_bricks/
> node01.company.com/gluster
> 14: option transport-type socket
> 15: option transport.address-family inet
> 16: option username 96bcf4d4-932f-4654-86c3-470a081d5021
> 17: option password 069e7ee9-b17d-4228-a612-b0f33588a9ec
> 18: option transport.socket.ssl-enabled off
> 19: option transport.tcp-user-timeout 0
> 20: option transport.socket.keepalive-time 20
> 21: option transport.socket.keepalive-interval 2
> 22: option transport.socket.keepalive-count 9
> 23: option send-gids true
> 24: end-volume
> 25:
> 26: volume ssd_storage-client-1
> 27: type protocol/client
> 28: option ping-timeout 42
> 29: option remote-host node02.company.com
> 30: option remote-subvolume /gluster_bricks/
> node02.company.com/gluster
> 31: option transport-type socket
> 32: option transport.address-family inet
> 33: option username 96bcf4d4-932f-4654-86c3-470a081d5021
> 34: option password 069e7ee9-b17d-4228-a612-b0f33588a9ec
> 35: option transport.socket.ssl-enabled off
> 36: option transport.tcp-user-timeout 0
> 37: option transport.socket.keepalive-time 20
> 38: option transport.socket.keepalive-interval 2
> 39: option transport.socket.keepalive-count 9
> 40: option send-gids true
> 41: end-volume
> 42:
> 43: volume ssd_storage-client-2
> 44: type protocol/client
> 45: option ping-timeout 42
> 46: option remote-host node03.company.com
> 47: option remote-subvolume /gluster_bricks/
> node03.company.com/gluster
> 48: option transport-type socket
> 49: option transport.address-family inet
> 50: option username 96bcf4d4-932f-4654-86c3-470a081d5021
> 51: option password 069e7ee9-b17d-4228-a612-b0f33588a9ec
> 52: option transport.socket.ssl-enabled off
> 53: option transport.tcp-user-timeout 0
> 54: option transport.socket.keepalive-time 20
> 55: option transport.socket.keepalive-interval 2
> 56: option transport.socket.keepalive-count 9
> 57: option send-gids true
> 58: end-volume
> 59:
> 60: volume ssd_storage-replicate-0
> 61: type cluster/replicate
> 62: option background-self-heal-count 0
> 63: option afr-pending-xattr
> ssd_storage-client-0,ssd_storage-client-1,ssd_storage-client-2
> 64: option metadata-self-heal on
> 65: option data-self-heal on
> 66: option entry-self-heal on
> 67: option data-self-heal-algorithm full
> 68: option use-compound-fops off
> 69: subvolumes ssd_storage-client-0 ssd_storage-client-1
> ssd_storage-client-2
> 70: end-volume
> 71:
> 72: volume ssd_storage-dht
> 73: type cluster/distribute
> 74: option readdir-optimize on
> 75: option lock-migration off
> 76: option force-migration off
> 77: subvolumes ssd_storage-replicate-0
> 78: end-volume
> 79:
> 80: volume ssd_storage-utime
> 81: type features/utime
> 82: option noatime on
> 83: subvolumes ssd_storage-dht
> 84: end-volume
> 85:
> 86: volume ssd_storage-write-behind
> 87: type performance/write-behind
> 88: subvolumes ssd_storage-utime
> 89: end-volume
> 90:
> 91: volume ssd_storage-read-ahead
> 92: type performance/read-ahead
> 93: subvolumes ssd_storage-write-behind
> 94: end-volume
> 95:
> 96: volume ssd_storage-readdir-ahead
> 97: type performance/readdir-ahead
> 98: option parallel-readdir off
> 99: option rda-request-size 131072
> 100: option rda-cache-limit 10MB
> 101: subvolumes ssd_storage-read-ahead
> 102: end-volume
> 103:
> 104: volume ssd_storage-io-cache
> 105: type performance/io-cache
> 106: subvolumes ssd_storage-readdir-ahead
> 107: end-volume
> 108:
> 109: volume ssd_storage-open-behind
> 110: type performance/open-behind
> 111: subvolumes ssd_storage-io-cache
> 112: end-volume
> 113:
> 114: volume ssd_storage-quick-read
> 115: type performance/quick-read
> 116: subvolumes ssd_storage-open-behind
> 117: end-volume
> 118:
> 119: volume ssd_storage-md-cache
> 120: type performance/md-cache
> 121: subvolumes ssd_storage-quick-read
> 122: end-volume
> 123:
> 124: volume ssd_storage-snapd-client
> 125: type protocol/client
> 126: option remote-host /var/run/glusterd.socket
> 127: option ping-timeout 42
> 128: option remote-subvolume snapd-ssd_storage
> 129: option transport-type socket
> 130: option transport.address-family inet
> 131: option username 96bcf4d4-932f-4654-86c3-470a081d5021
> 132: option password 069e7ee9-b17d-4228-a612-b0f33588a9ec
> 133: option transport.socket.ssl-enabled off
> 134: option transport.tcp-user-timeout 0
> 135: option transport.socket.keepalive-time 20
> 136: option transport.socket.keepalive-interval 2
> 137: option transport.socket.keepalive-count 9
> 138: option send-gids true
> 139: end-volume
> 140:
> 141: volume ssd_storage-snapview-client
> 142: type features/snapview-client
> 143: option snapshot-directory .snaps
> 144: option show-snapshot-directory on
> 145: subvolumes ssd_storage-md-cache ssd_storage-snapd-client
> 146: end-volume
> 147:
> 148: volume ssd_storage
> 149: type debug/io-stats
> 150: option log-level INFO
> 151: option threads 16
> 152: option latency-measurement off
> 153: option count-fop-hits off
> 154: option global-threading off
> 155: subvolumes ssd_storage-snapview-client
> 156: end-volume
> 157:
> 158: volume meta-autoload
> 159: type meta
> 160: subvolumes ssd_storage
> 161: end-volume
> 162:
>
> +------------------------------------------------------------------------------+
> [2020-02-14 11:05:22.421366] I [MSGID: 114046]
> [client-handshake.c:1106:client_setvolume_cbk] 0-ssd_storage-client-0:
> Connected to ssd_storage-client-0, attached to remote volume
> '/gluster_bricks/node01.company.com/gluster'.
> [2020-02-14 11:05:22.421379] I [MSGID: 108005]
> [afr-common.c:5280:__afr_handle_child_up_event]
> 0-ssd_storage-replicate-0: Subvolume 'ssd_storage-client-0' came back
> up; going online.
> [2020-02-14 11:05:22.421669] I [rpc-clnt.c:1963:rpc_clnt_reconfig]
> 0-ssd_storage-client-2: changing port to 49152 (from 0)
> [2020-02-14 11:05:22.421686] I [socket.c:864:__socket_shutdown]
> 0-ssd_storage-client-2: intentional socket shutdown(11)
> [2020-02-14 11:05:22.422460] I [MSGID: 114057]
> [client-handshake.c:1376:select_server_supported_programs]
> 0-ssd_storage-client-1: Using Program GlusterFS 4.x v1, Num (1298437),
> Version (400)
> [2020-02-14 11:05:22.423377] I [MSGID: 114046]
> [client-handshake.c:1106:client_setvolume_cbk] 0-ssd_storage-client-1:
> Connected to ssd_storage-client-1, attached to remote volume
> '/gluster_bricks/node02.company.com/gluster'.
> [2020-02-14 11:05:22.423391] I [MSGID: 108002]
> [afr-common.c:5647:afr_notify] 0-ssd_storage-replicate-0: Client-quorum
> is met
> [2020-02-14 11:05:22.424586] I [MSGID: 114057]
> [client-handshake.c:1376:select_server_supported_programs]
> 0-ssd_storage-client-2: Using Program GlusterFS 4.x v1, Num (1298437),
> Version (400)
> [2020-02-14 11:05:22.425323] I [MSGID: 114046]
> [client-handshake.c:1106:client_setvolume_cbk] 0-ssd_storage-client-2:
> Connected to ssd_storage-client-2, attached to remote volume
> '/gluster_bricks/node03.company.com/gluster'.
> [2020-02-14 11:05:22.426613] I [MSGID: 108031]
> [afr-common.c:2580:afr_local_discovery_cbk] 0-ssd_storage-replicate-0:
> selecting local read_child ssd_storage-client-0
> [2020-02-14 11:05:22.426758] I [MSGID: 104041]
> [glfs-resolve.c:954:__glfs_active_subvol] 0-ssd_storage: switched to
> graph 6e6f6465-3031-2e64-632d-6475732e6461 (0)
>
>
> Can you guys make any sense out of this? 5 unsynced entries remain.
>
> --
> with kind regards,
> mit freundlichen Gruessen,
>
> Christian Reiss
>
> ________
>
> Community Meeting Calendar:
>
> APAC Schedule -
> Every 2nd and 4th Tuesday at 11:30 AM IST
> Bridge: https://bluejeans.com/441850968
>
> NA/EMEA Schedule -
> Every 1st and 3rd Tuesday at 01:00 PM EDT
> Bridge: https://bluejeans.com/441850968
>
> Gluster-users mailing list
> Gluster-users at gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20200214/9a810e98/attachment.html>
More information about the Gluster-users
mailing list