<div dir="ltr"><div dir="ltr">Sorry! I should have been more specific. I over-looked the option:<div><br></div><div>---</div><div><div>[root@localhost ~]# gluster volume set demo1 tcp-user-timeout 42</div><div>volume set: failed: option : tcp-user-timeout does not exist</div><div>Did you mean tcp-user-timeout?</div><div>[root@localhost ~]# gluster volume set demo1 <b>client.tcp-user-timeout</b> 42<br></div><div>volume set: success</div><div>[root@localhost ~]# gluster volume set demo1 <b>server.tcp-user-timeout</b> 42</div><div>volume set: success</div><div>----</div></div><div>Looks like you need to set the option specifically on client and server. </div><div><br></div><div>Regards,</div></div></div><br><div class="gmail_quote"><div dir="ltr">On Wed, Oct 3, 2018 at 11:32 AM Dmitry Melekhov &lt;<a href="mailto:dm@belkam.com">dm@belkam.com</a>&gt; wrote:<br></div><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
  
    
  
  <div text="#000000" bgcolor="#FFFFFF">
    <div class="m_-4573436687730019044moz-cite-prefix"><br>
      It doesn&#39;t work for some reason:<br>
      <br>
       gluster volume set pool tcp-user-timeout 42<br>
      volume set: failed: option : tcp-user-timeout does not exist<br>
      Did you mean tcp-user-timeout?<br>
      <br>
      <br>
      4.1.5.<br>
      <br>
      <br>
      <br>
      03.10.2018 08:30, Dmitry Melekhov пишет:<br>
    </div>
    <blockquote type="cite">
      <div class="m_-4573436687730019044moz-cite-prefix">02.10.2018 12:59, Amar Tumballi
        пишет:<br>
      </div>
      <blockquote type="cite">
        <div dir="ltr">
          <div dir="ltr">
            <div dir="ltr">Recently, in one of the situation, we found
              that locks were not freed up due to not getting TCP
              timeout..
              <div><br>
              </div>
              <div>Can you try the option like below and let us know?</div>
              <div><br>
              </div>
              <div>`gluster volume set $volname tcp-user-timeout 42` </div>
              <div><br>
              </div>
              <div>(ref: <a href="https://review.gluster.org/21170/" target="_blank">https://review.gluster.org/21170/</a>
                )</div>
              <div><br>
              </div>
              <div>
                <div>Regards,</div>
                <div>Amar</div>
                <div><br>
                </div>
              </div>
            </div>
          </div>
        </div>
      </blockquote>
      <br>
      Thank you, we&#39;ll try this.<br>
      <br>
      <blockquote type="cite"><br>
        <div class="gmail_quote">
          <div dir="ltr">On Tue, Oct 2, 2018 at 10:40 AM Dmitry Melekhov
            &lt;<a href="mailto:dm@belkam.com" target="_blank">dm@belkam.com</a>&gt; wrote:<br>
          </div>
          <blockquote class="gmail_quote">
            <div>
              <div class="m_-4573436687730019044m_-998520419163836936m_5347744416915771342moz-cite-prefix">01.10.2018
                23:09, Danny Lee пишет:<br>
              </div>
              <blockquote type="cite">
                <div dir="auto">Ran into this issue too with 4.1.5 with
                  an arbiter setup.  Also could not run a statedump due
                  to &quot;Segmentation fault&quot;.
                  <div dir="auto"><br>
                  </div>
                  <div dir="auto">Tried with 3.12.13 and had issues with
                    locked files as well.  We were able to do a
                    statedump and found that some of our files were
                    &quot;BLOCKED&quot; (xlator.features.locks.vol-locks.inode). 
                    Attached part of statedump.</div>
                  <div dir="auto"><br>
                  </div>
                  <div dir="auto">Also tried clearing the locks using
                    clear-locks, which did remove the lock, but as soon
                    as I tried to cat the file, it got locked again and
                    the cat process hung.</div>
                </div>
              </blockquote>
              <br>
              I created issue in bugzilla, can&#39;t find it though :-(<br>
              Looks like there is no activity after I sent all logs...<br>
              <br>
              <br>
              <blockquote type="cite"><br>
                <div class="gmail_quote">
                  <div dir="ltr">On Wed, Aug 29, 2018, 3:13 AM Dmitry
                    Melekhov &lt;<a href="mailto:dm@belkam.com" rel="noreferrer" target="_blank">dm@belkam.com</a>&gt;
                    wrote:<br>
                  </div>
                  <blockquote class="gmail_quote">
                    <div>
                      <div class="m_-4573436687730019044m_-998520419163836936m_5347744416915771342m_7076642256643114549m_4002192505348597522moz-cite-prefix">28.08.2018
                        10:43, Amar Tumballi пишет:<br>
                      </div>
                      <blockquote type="cite">
                        <div dir="ltr"><br>
                          <div class="gmail_extra"><br>
                            <div class="gmail_quote">On Tue, Aug 28,
                              2018 at 11:24 AM, Dmitry Melekhov <span dir="ltr">&lt;<a href="mailto:dm@belkam.com" rel="noreferrer noreferrer" target="_blank">dm@belkam.com</a>&gt;</span>
                              wrote:<br>
                              <blockquote class="gmail_quote">Hello!<br>
                                <br>
                                <br>
                                Yesterday we hit something like this on
                                4.1.2<br>
                                <br>
                                Centos 7.5.<br>
                                <br>
                                <br>
                                Volume is replicated - two bricks and
                                one arbiter.<br>
                                <br>
                                <br>
                                We rebooted arbiter, waited for heal
                                end,  and tried to live migrate VM to
                                another node ( we run VMs on gluster
                                nodes ):<br>
                                <br>
                                <br>
                                [2018-08-27 09:56:22.085411] I [MSGID:
                                115029]
                                [server-handshake.c:763:server_setvolume]
                                0-pool-server: accepted client from
CTX_ID:b55f4a90-e241-48ce-bd4d-268c8a956f4a-GRAPH_ID:0-PID:8887-HOST:son-PC_NAME:pool-<br>
                                client-6-RECON_NO:-0 (version: 4.1.2)<br>
                                [2018-08-27 09:56:22.107609] I [MSGID:
                                115036] [server.c:483:server_rpc_notify]
                                0-pool-server: disconnecting connection
                                from
CTX_ID:b55f4a90-e241-48ce-bd4d-268c8a956f4a-GRAPH_ID:0-PID:8887-HOST:son-PC_NAME:pool-<br>
                                client-6-RECON_NO:-0<br>
                                [2018-08-27 09:56:22.107747] I [MSGID:
                                101055] [client_t.c:444:gf_client_unref]
                                0-pool-server: Shutting down connection
CTX_ID:b55f4a90-e241-48ce-bd4d-268c8a956f4a-GRAPH_ID:0-PID:8887-HOST:son-PC_NAME:pool-clien<br>
                                t-6-RECON_NO:-0<br>
                                [2018-08-27 09:58:37.905829] I [MSGID:
                                115036] [server.c:483:server_rpc_notify]
                                0-pool-server: disconnecting connection
                                from
CTX_ID:c3eb6cfc-2ef9-470a-89d1-a87170d00da5-GRAPH_ID:0-PID:30292-HOST:father-PC_NAME:p<br>
                                ool-client-6-RECON_NO:-0<br>
                                [2018-08-27 09:58:37.905926] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=28c831d8bc550000}<br>
                                [2018-08-27 09:58:37.905959] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=2870a7d6bc550000}<br>
                                [2018-08-27 09:58:37.905979] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=2880a7d6bc550000}<br>
                                [2018-08-27 09:58:37.905997] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=28f031d8bc550000}<br>
                                [2018-08-27 09:58:37.906016] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=28b07dd5bc550000}<br>
                                [2018-08-27 09:58:37.906034] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=28e0a7d6bc550000}<br>
                                [2018-08-27 09:58:37.906056] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=28b845d8bc550000}<br>
                                [2018-08-27 09:58:37.906079] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=2858a7d8bc550000}<br>
                                [2018-08-27 09:58:37.906098] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=2868a8d7bc550000}<br>
                                [2018-08-27 09:58:37.906121] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=28f80bd7bc550000}<br>
                                ...<br>
                                <br>
                                [2018-08-27 09:58:37.907375] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=28a8cdd6bc550000}<br>
                                [2018-08-27 09:58:37.907393] W
                                [inodelk.c:610:pl_inodelk_log_cleanup]
                                0-pool-server: releasing lock on
                                12172afe-f0a4-4e10-bc0f-c5e4e0d9f318
                                held by {client=0x7ffb58035bc0,
                                pid=30292 lk-owner=2880cdd6bc550000}<br>
                                [2018-08-27 09:58:37.907476] I
                                [socket.c:3837:socket_submit_reply]
                                0-tcp.pool-server: not connected
                                (priv-&gt;connected = -1)<br>
                                [2018-08-27 09:58:37.907520] E
                                [rpcsvc.c:1378:rpcsvc_submit_generic]
                                0-rpc-service: failed to submit message
                                (XID: 0xcb88cb, Program: GlusterFS 4.x
                                v1, ProgVers: 400, Proc: 30) to
                                rpc-transport (tcp.pool-server)<br>
                                [2018-08-27 09:58:37.910727] E
                                [server.c:137:server_submit_reply]
                                (--&gt;/usr/lib64/glusterfs/4.1.2/xlator/debug/io-stats.so(+0x20084)
                                [0x7ffb64379084]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0x605<br>
                                ba) [0x7ffb5fddf5ba]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0xafce)
                                [0x7ffb5fd89fce] ) 0-: Reply submission
                                failed<br>
                                [2018-08-27 09:58:37.910814] E
                                [rpcsvc.c:1378:rpcsvc_submit_generic]
                                0-rpc-service: failed to submit message
                                (XID: 0xcb88ce, Program: GlusterFS 4.x
                                v1, ProgVers: 400, Proc: 30) to
                                rpc-transport (tcp.pool-server)<br>
                                [2018-08-27 09:58:37.910861] E
                                [server.c:137:server_submit_reply]
                                (--&gt;/usr/lib64/glusterfs/4.1.2/xlator/debug/io-stats.so(+0x20084)
                                [0x7ffb64379084]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0x605<br>
                                ba) [0x7ffb5fddf5ba]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0xafce)
                                [0x7ffb5fd89fce] ) 0-: Reply submission
                                failed<br>
                                [2018-08-27 09:58:37.910904] E
                                [rpcsvc.c:1378:rpcsvc_submit_generic]
                                0-rpc-service: failed to submit message
                                (XID: 0xcb88cf, Program: GlusterFS 4.x
                                v1, ProgVers: 400, Proc: 30) to
                                rpc-transport (tcp.pool-server)<br>
                                [2018-08-27 09:58:37.910940] E
                                [server.c:137:server_submit_reply]
                                (--&gt;/usr/lib64/glusterfs/4.1.2/xlator/debug/io-stats.so(+0x20084)
                                [0x7ffb64379084]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0x605<br>
                                ba) [0x7ffb5fddf5ba]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0xafce)
                                [0x7ffb5fd89fce] ) 0-: Reply submission
                                failed<br>
                                [2018-08-27 09:58:37.910979] E
                                [rpcsvc.c:1378:rpcsvc_submit_generic]
                                0-rpc-service: failed to submit message
                                (XID: 0xcb88d1, Program: GlusterFS 4.x
                                v1, ProgVers: 400, Proc: 30) to
                                rpc-transport (tcp.pool-server)<br>
                                [2018-08-27 09:58:37.911012] E
                                [server.c:137:server_submit_reply]
                                (--&gt;/usr/lib64/glusterfs/4.1.2/xlator/debug/io-stats.so(+0x20084)
                                [0x7ffb64379084]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0x605<br>
                                ba) [0x7ffb5fddf5ba]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0xafce)
                                [0x7ffb5fd89fce] ) 0-: Reply submission
                                failed<br>
                                [2018-08-27 09:58:37.911050] E
                                [rpcsvc.c:1378:rpcsvc_submit_generic]
                                0-rpc-service: failed to submit message
                                (XID: 0xcb88d8, Program: GlusterFS 4.x
                                v1, ProgVers: 400, Proc: 30) to
                                rpc-transport (tcp.pool-server)<br>
                                [2018-08-27 09:58:37.911083] E
                                [server.c:137:server_submit_reply]
                                (--&gt;/usr/lib64/glusterfs/4.1.2/xlator/debug/io-stats.so(+0x20084)
                                [0x7ffb64379084]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0x605<br>
                                ba) [0x7ffb5fddf5ba]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0xafce)
                                [0x7ffb5fd89fce] ) 0-: Reply submission
                                failed<br>
                                [2018-08-27 09:58:37.916217] E
                                [server.c:137:server_submit_reply]
                                (--&gt;/usr/lib64/glusterfs/4.1.2/xlator/debug/io-stats.so(+0x20084)
                                [0x7ffb64379084]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0x605<br>
                                ba) [0x7ffb5fddf5ba]
                                --&gt;/usr/lib64/glusterfs/4.1.2/xlator/protocol/server.so(+0xafce)
                                [0x7ffb5fd89fce] ) 0-: Reply submission
                                failed<br>
                                [2018-08-27 09:58:37.916520] I [MSGID:
                                115013]
                                [server-helpers.c:286:do_fd_cleanup]
                                0-pool-server: fd cleanup on
                                /balamak.img<br>
                                <br>
                                <br>
                                after this I/O on  /balamak.img was
                                blocked.<br>
                                <br>
                                <br>
                                Only solution we found was to reboot all
                                3 nodes.<br>
                                <br>
                                <br>
                                Is there any bug report in bugzilla we
                                can add logs?<br>
                                <br>
                              </blockquote>
                              <div><br>
                              </div>
                              <div>Not aware of such bugs!</div>
                              <div> </div>
                              <blockquote class="gmail_quote"> Is it
                                possible to turn of these locks?<br>
                                <br>
                              </blockquote>
                              <div><br>
                              </div>
                              <div>Not sure, will get back on this one!</div>
                            </div>
                          </div>
                        </div>
                      </blockquote>
                      <br>
                      <br>
                      btw, found this link<br>
                      <a class="m_-4573436687730019044m_-998520419163836936m_5347744416915771342m_7076642256643114549m_4002192505348597522moz-txt-link-freetext" href="https://docs.gluster.org/en/v3/Troubleshooting/troubleshooting-filelocks/" rel="noreferrer noreferrer" target="_blank">https://docs.gluster.org/en/v3/Troubleshooting/troubleshooting-filelocks/</a><br>
                      <br>
                      tried on another (test) cluster:<br>
                      <br>
                       [root@marduk ~]# gluster volume statedump pool<br>
                      Segmentation fault (core dumped)<br>
                       <br>
                      <br>
                      4.1.2 too...<br>
                      <br>
                      something is wrong here.<br>
                      <br>
                      <br>
                      <blockquote type="cite">
                        <div dir="ltr">
                          <div class="gmail_extra">
                            <div class="gmail_quote">
                              <div> </div>
                              <blockquote class="gmail_quote"> Thank
                                you!<br>
                                <br>
                                <br>
                                <br>
                                <br>
_______________________________________________<br>
                                Gluster-users mailing list<br>
                                <a href="mailto:Gluster-users@gluster.org" rel="noreferrer noreferrer" target="_blank">Gluster-users@gluster.org</a><br>
                                <a href="https://lists.gluster.org/mailman/listinfo/gluster-users" rel="noreferrer noreferrer noreferrer" target="_blank">https://lists.gluster.org/mailman/listinfo/gluster-users</a></blockquote>
                            </div>
                            <br>
                            <br>
                            <div><br>
                            </div>
                            -- <br>
                            <div class="m_-4573436687730019044m_-998520419163836936m_5347744416915771342m_7076642256643114549m_4002192505348597522gmail_signature" data-smartmail="gmail_signature">
                              <div dir="ltr">
                                <div>
                                  <div dir="ltr">
                                    <div>Amar Tumballi (amarts)<br>
                                    </div>
                                  </div>
                                </div>
                              </div>
                            </div>
                          </div>
                        </div>
                      </blockquote>
                      <p><br>
                      </p>
                    </div>
                    _______________________________________________<br>
                    Gluster-users mailing list<br>
                    <a href="mailto:Gluster-users@gluster.org" rel="noreferrer noreferrer" target="_blank">Gluster-users@gluster.org</a><br>
                    <a href="https://lists.gluster.org/mailman/listinfo/gluster-users" rel="noreferrer noreferrer noreferrer" target="_blank">https://lists.gluster.org/mailman/listinfo/gluster-users</a></blockquote>
                </div>
              </blockquote>
              <p><br>
              </p>
            </div>
            _______________________________________________<br>
            Gluster-users mailing list<br>
            <a href="mailto:Gluster-users@gluster.org" target="_blank">Gluster-users@gluster.org</a><br>
            <a href="https://lists.gluster.org/mailman/listinfo/gluster-users" rel="noreferrer" target="_blank">https://lists.gluster.org/mailman/listinfo/gluster-users</a></blockquote>
        </div>
        <br>
        <div><br>
        </div>
        -- <br>
        <div dir="ltr" class="m_-4573436687730019044m_-998520419163836936gmail_signature" data-smartmail="gmail_signature">
          <div dir="ltr">
            <div>
              <div dir="ltr">
                <div>Amar Tumballi (amarts)<br>
                </div>
              </div>
            </div>
          </div>
        </div>
      </blockquote>
      <p><br>
      </p>
    </blockquote>
    <p><br>
    </p>
  </div>

_______________________________________________<br>
Gluster-users mailing list<br>
<a href="mailto:Gluster-users@gluster.org" target="_blank">Gluster-users@gluster.org</a><br>
<a href="https://lists.gluster.org/mailman/listinfo/gluster-users" rel="noreferrer" target="_blank">https://lists.gluster.org/mailman/listinfo/gluster-users</a></blockquote></div><br clear="all"><div><br></div>-- <br><div dir="ltr" class="gmail_signature" data-smartmail="gmail_signature"><div dir="ltr"><div><div dir="ltr"><div>Amar Tumballi (amarts)<br></div></div></div></div></div>