<html>
  <head>
    <meta http-equiv="Content-Type" content="text/html; charset=UTF-8">
  </head>
  <body text="#000000" bgcolor="#FFFFFF">
    <p>Resend as I did not reply to the list earlier.  TBird responded
      to the poster and not the list.<br>
    </p>
    <div class="moz-cite-prefix">On 12/27/18 11:46 AM, Brett Holcomb
      wrote:<br>
    </div>
    <blockquote type="cite"
      cite="mid:a09be52c-2001-a599-fbdc-64f8131d40ba@l1049h.com">
      <meta http-equiv="Content-Type" content="text/html; charset=UTF-8">
      <p>Thank you. I appreciate the help  Here is the information.  Let
        me know if you need anything else.  I'm fairly new to gluster.<br>
      </p>
      <p>Gluster version is 5.2</p>
      <p>1. gluster v info<br>
      </p>
      <p>Volume Name: projects<br>
        Type: Distributed-Replicate<br>
        Volume ID: 5aac71aa-feaa-44e9-a4f9-cb4dd6e0fdc3<br>
        Status: Started<br>
        Snapshot Count: 0<br>
        Number of Bricks: 2 x 3 = 6<br>
        Transport-type: tcp<br>
        Bricks:<br>
        Brick1: gfssrv1:/srv/gfs01/Projects<br>
        Brick2: gfssrv2:/srv/gfs01/Projects<br>
        Brick3: gfssrv3:/srv/gfs01/Projects<br>
        Brick4: gfssrv4:/srv/gfs01/Projects<br>
        Brick5: gfssrv5:/srv/gfs01/Projects<br>
        Brick6: gfssrv6:/srv/gfs01/Projects<br>
        Options Reconfigured:<br>
        cluster.self-heal-daemon: enable<br>
        performance.quick-read: off<br>
        performance.parallel-readdir: off<br>
        performance.readdir-ahead: off<br>
        performance.write-behind: off<br>
        performance.read-ahead: off<br>
        performance.client-io-threads: off<br>
        nfs.disable: on<br>
        transport.address-family: inet<br>
        server.allow-insecure: on<br>
        storage.build-pgfid: on<br>
        changelog.changelog: on<br>
        changelog.capture-del-path: on<br>
        <br>
      </p>
      <p>2.  gluster v status<br>
      </p>
      <p>Status of volume: projects<br>
        Gluster process                             TCP Port  RDMA Port 
        Online  Pid<br>
------------------------------------------------------------------------------<br>
        Brick gfssrv1:/srv/gfs01/Projects           49154     0         
        Y       7213 <br>
        Brick gfssrv2:/srv/gfs01/Projects           49154     0         
        Y       6932 <br>
        Brick gfssrv3:/srv/gfs01/Projects           49154     0         
        Y       6920 <br>
        Brick gfssrv4:/srv/gfs01/Projects           49154     0         
        Y       6732 <br>
        Brick gfssrv5:/srv/gfs01/Projects           49154     0         
        Y       6950 <br>
        Brick gfssrv6:/srv/gfs01/Projects           49154     0         
        Y       6879 <br>
        Self-heal Daemon on localhost               N/A       N/A       
        Y       11484<br>
        Self-heal Daemon on gfssrv2                 N/A       N/A       
        Y       10366<br>
        Self-heal Daemon on gfssrv4                 N/A       N/A       
        Y       9872 <br>
        Self-heal Daemon on srv-1-gfs3.corp.l1049h.<br>
        net                                         N/A       N/A       
        Y       9892 <br>
        Self-heal Daemon on gfssrv6                 N/A       N/A       
        Y       10372<br>
        Self-heal Daemon on gfssrv5                 N/A       N/A       
        Y       10761<br>
         <br>
        Task Status of Volume projects<br>
------------------------------------------------------------------------------<br>
        There are no active volume tasks</p>
      <p>3. I've given the summary since the actual list for two volumes
        is around 5900 entries.</p>
      <p>Brick gfssrv1:/srv/gfs01/Projects<br>
        Status: Connected<br>
        Total Number of entries: 85<br>
        Number of entries in heal pending: 85<br>
        Number of entries in split-brain: 0<br>
        Number of entries possibly healing: 0<br>
        <br>
        Brick gfssrv2:/srv/gfs01/Projects<br>
        Status: Connected<br>
        Total Number of entries: 0<br>
        Number of entries in heal pending: 0<br>
        Number of entries in split-brain: 0<br>
        Number of entries possibly healing: 0<br>
        <br>
        Brick gfssrv3:/srv/gfs01/Projects<br>
        Status: Connected<br>
        Total Number of entries: 0<br>
        Number of entries in heal pending: 0<br>
        Number of entries in split-brain: 0<br>
        Number of entries possibly healing: 0<br>
        <br>
        Brick gfssrv4:/srv/gfs01/Projects<br>
        Status: Connected<br>
        Total Number of entries: 0<br>
        Number of entries in heal pending: 0<br>
        Number of entries in split-brain: 0<br>
        Number of entries possibly healing: 0<br>
      </p>
      <p>Brick gfssrv5:/srv/gfs01/Projects<br>
        Status: Connected<br>
        Total Number of entries: 58854<br>
        Number of entries in heal pending: 58854<br>
        Number of entries in split-brain: 0<br>
        Number of entries possibly healing: 0<br>
        <br>
        Brick gfssrv6:/srv/gfs01/Projects<br>
        Status: Connected<br>
        Total Number of entries: 58854<br>
        Number of entries in heal pending: 58854<br>
        Number of entries in split-brain: 0<br>
        Number of entries possibly healing: 0<br>
      </p>
      <div class="moz-cite-prefix">On 12/27/18 3:09 AM, Ashish Pandey
        wrote:<br>
      </div>
      <blockquote type="cite"
        cite="mid:1851464190.54195617.1545898152060.JavaMail.zimbra@redhat.com">
        <meta http-equiv="content-type" content="text/html;
          charset=UTF-8">
        <div style="font-family: times new roman, new york, times,
          serif;&#xA; font-size: 12pt; color: #000000">
          <div>Hi Brett,<br>
          </div>
          <div><br>
          </div>
          <div>Could you please tell us more about the setup?<br>
          </div>
          <div><br>
          </div>
          <div>1 - Gluster v info<br>
          </div>
          <div>2 - gluster v status<br>
          </div>
          <div>3 - gluster v heal &lt;volname&gt; info<br>
          </div>
          <div><br>
          </div>
          <div>These are the very basic information to start with
            debugging or suggesting any workaround.</div>
          <div>It should always be included when asking such questions
            on mailing list so that people can reply sooner. <br>
          </div>
          <div><br>
          </div>
          <div><br>
          </div>
          <div>Note: Please hide IP address/hostname or any other
            information you don't want world to see.<br>
          </div>
          <div><br>
          </div>
          <div>---<br>
          </div>
          <div>Ashish<br>
          </div>
          <div><br>
          </div>
          <hr id="zwchr">
          <div
style="color:#000;font-weight:normal;font-style:normal;text-decoration:none;font-family:Helvetica,Arial,sans-serif;font-size:12pt;"
            data-mce-style="color: #000; font-weight: normal;
            font-style:&#xA; normal; text-decoration: none;
            font-family:&#xA; Helvetica,Arial,sans-serif; font-size:
            12pt;"><b>From: </b>"Brett Holcomb" <a
              class="moz-txt-link-rfc2396E"
              href="mailto:biholcomb@l1049h.com" moz-do-not-send="true">&lt;biholcomb@l1049h.com&gt;</a><br>
            <b>To: </b><a class="moz-txt-link-abbreviated"
              href="mailto:gluster-users@gluster.org"
              moz-do-not-send="true">gluster-users@gluster.org</a><br>
            <b>Sent: </b>Thursday, December 27, 2018 12:19:15 AM<br>
            <b>Subject: </b>Re: [Gluster-users] Self Heal Confusion<br>
            <div><br>
            </div>
            <p>Still no change in the heals pending.  I found this
              reference, <a class="moz-txt-link-freetext"
href="https://archive.fosdem.org/2017/schedule/event/glusterselinux/attachments/slides/1876/export/events/attachments/glusterselinux/slides/1876/fosdem.pdf"
                target="_blank"
data-mce-href="https://archive.fosdem.org/2017/schedule/event/glusterselinux/attachments/slides/1876/export/events/attachments/glusterselinux/slides/1876/fosdem.pdf"
                moz-do-not-send="true">https://archive.fosdem.org/2017/schedule/event/glusterselinux/attachments/slides/1876/export/events/attachments/glusterselinux/slides/1876/fosdem.pdf</a>,
              which mentions the default SELinux context for a brick and
              that internal operations such as self-heal, rebalance
              should be ignored. but they do not elaborate on what
              ignore means - is it just not doing self-heal or something
              else.</p>
            <p>I did set SELinux to permissive and nothing changed. 
              I'll try setting the bricks to the context mentioned in
              this pdf and see what happens.</p>
            <p><br>
            </p>
            <div class="moz-cite-prefix">On 12/20/18 8:26 PM, John
              Strunk wrote:<br>
            </div>
            <blockquote
cite="mid:CAMLs-gRPSPvjEAvZNZLyNYvruEGvs-v2WL6m=un5v00pXXcMnw@mail.gmail.com">
              <div dir="ltr">Assuming your bricks are up... yes, the
                heal count should be decreasing.
                <div><br>
                </div>
                <div>There is/was a bug wherein self-heal would stop
                  healing but would still be running. I don't know
                  whether your version is affected, but the remedy is to
                  just restart the self-heal daemon.</div>
                <div>Force start one of the volumes that has heals
                  pending. The bricks are already running, but it will
                  cause shd to restart and, assuming this is the
                  problem, healing should begin...</div>
                <div><br>
                </div>
                <div>$ gluster vol start my-pending-heal-vol force</div>
                <div><br>
                </div>
                <div>Others could better comment on the status of the
                  bug.</div>
                <div><br>
                </div>
                <div>-John</div>
                <div><br>
                </div>
              </div>
              <br>
              <div class="gmail_quote">
                <div dir="ltr">On Thu, Dec 20, 2018 at 5:45 PM Brett
                  Holcomb &lt;<a href="mailto:biholcomb@l1049h.com"
                    target="_blank"
                    data-mce-href="mailto:biholcomb@l1049h.com"
                    moz-do-not-send="true">biholcomb@l1049h.com</a>&gt;
                  wrote:<br>
                </div>
                <blockquote class="gmail_quote" style="margin:0px
                  0px&#xA; 0px&#xA; 0.8ex;border-left:1px solid&#xA;
                  rgb(204,204,204);padding-left:1ex"
                  data-mce-style="margin: 0px 0px 0px&#xA; 0.8ex;&#xA;
                  border-left: 1px solid #cccccc; padding-left: 1ex;">I
                  have one volume that has 85 pending entries in healing
                  and two more <br>
                  volumes with 58,854 entries in healing pending.  These
                  numbers are from <br>
                  the volume heal info summary command.  They have
                  stayed constant for two <br>
                  days now.  I've read the gluster docs and many more. 
                  The Gluster docs <br>
                  just give some commands and non gluster docs basically
                  repeat that.  <br>
                  Given that it appears no self-healing is going on for
                  my volume I am <br>
                  confused as to why.<br>
                  <br>
                  1.  If a self-heal deamon is listed on a host (all of
                  mine show one with <br>
                  a volume status command) can I assume it's enabled and
                  running?<br>
                  <br>
                  2.  I assume the volume that has all the self-heals
                  pending has some <br>
                  serious issues even though I can access the files and
                  directories on <br>
                  it.  If self-heal is running shouldn't the numbers be
                  decreasing?<br>
                  <br>
                  It appears to me self-heal is not working properly so
                  how to I get it to <br>
                  start working or should I delete the volume and start
                  over?<br>
                  <br>
                  I'm running gluster 5.2 on Centos 7 latest and
                  updated.<br>
                  <br>
                  Thank you.<br>
                  <br>
                  <br>
                  _______________________________________________<br>
                  Gluster-users mailing list<br>
                  <a href="mailto:Gluster-users@gluster.org"
                    target="_blank"
                    data-mce-href="mailto:Gluster-users@gluster.org"
                    moz-do-not-send="true">Gluster-users@gluster.org</a><br>
                  <a
                    href="https://lists.gluster.org/mailman/listinfo/gluster-users"
                    rel="noreferrer" target="_blank"
                    data-mce-href="https://lists.gluster.org/mailman/listinfo/gluster-users"
                    moz-do-not-send="true">https://lists.gluster.org/mailman/listinfo/gluster-users</a><br
                    data-mce-bogus="1">
                </blockquote>
              </div>
            </blockquote>
            <br>
            _______________________________________________<br>
            Gluster-users mailing list<br>
            <a class="moz-txt-link-abbreviated"
              href="mailto:Gluster-users@gluster.org"
              moz-do-not-send="true">Gluster-users@gluster.org</a><br>
            <a class="moz-txt-link-freetext"
              href="https://lists.gluster.org/mailman/listinfo/gluster-users"
              moz-do-not-send="true">https://lists.gluster.org/mailman/listinfo/gluster-users</a></div>
          <div><br>
          </div>
        </div>
      </blockquote>
    </blockquote>
  </body>
</html>