[Gluster-users] GlusterFS-3.7.14 released
Pranith Kumar Karampuri
pkarampu at redhat.com
Thu Aug 11 13:00:05 UTC 2016
On Thu, Aug 11, 2016 at 4:29 PM, Serkan Çoban <cobanserkan at gmail.com> wrote:
> I can wait for the patch to complete, please inform me when you ready.
> If it will take too much time to solve the crawl issue I can test
> without it too...
>
I don't know the Root cause for the problem, so I am not sure by when it
will be ready. Let me build the rpms, I have a meeting now for around an
hour. I will start building rpms after that.
>
> Serkan
>
> On Thu, Aug 11, 2016 at 5:52 AM, Pranith Kumar Karampuri
> <pkarampu at redhat.com> wrote:
> >
> >
> > On Wed, Aug 10, 2016 at 1:58 PM, Serkan Çoban <cobanserkan at gmail.com>
> wrote:
> >>
> >> Hi,
> >>
> >> Any progress about the patch?
> >
> >
> > hi Serkan,
> > While testing the patch by myself, I am seeing that it is taking
> more
> > than one crawl to complete heals even when there are no directory
> > hierarchies. It is faster than before but it shouldn't take more than 1
> > crawl to complete the heal because all the files exist already. I am
> > investigating why that is the case now. If you want to test things out
> > without this patch I will give you rpms today. Otherwise we need to find
> > until we find RCA for this crawl problem. Let me know your decision. If
> you
> > are okay with testing progressive versions of this feature, that would be
> > great. We can compare how each patch improved the performance.
> >
> > Pranith
> >
> >>
> >>
> >> On Thu, Aug 4, 2016 at 10:16 AM, Pranith Kumar Karampuri
> >> <pkarampu at redhat.com> wrote:
> >> >
> >> >
> >> > On Thu, Aug 4, 2016 at 11:30 AM, Serkan Çoban <cobanserkan at gmail.com>
> >> > wrote:
> >> >>
> >> >> Thanks Pranith,
> >> >> I am waiting for RPMs to show, I will do the tests as soon as
> possible
> >> >> and inform you.
> >> >
> >> >
> >> > I guess on 3.7.x the RPMs are not automatically built. Let me find how
> >> > it
> >> > can be done. I will inform you after finding that out. Give me a day.
> >> >
> >> >>
> >> >>
> >> >> On Wed, Aug 3, 2016 at 11:19 PM, Pranith Kumar Karampuri
> >> >> <pkarampu at redhat.com> wrote:
> >> >> >
> >> >> >
> >> >> > On Thu, Aug 4, 2016 at 1:47 AM, Pranith Kumar Karampuri
> >> >> > <pkarampu at redhat.com> wrote:
> >> >> >>
> >> >> >>
> >> >> >>
> >> >> >> On Thu, Aug 4, 2016 at 12:51 AM, Serkan Çoban
> >> >> >> <cobanserkan at gmail.com>
> >> >> >> wrote:
> >> >> >>>
> >> >> >>> I use rpms for installation. Redhat/Centos 6.8.
> >> >> >>
> >> >> >>
> >> >> >> http://review.gluster.org/#/c/15084 is the patch. In some time
> the
> >> >> >> rpms
> >> >> >> will be built actually.
> >> >> >
> >> >> >
> >> >> > In the same URL above it will actually post the rpms for
> >> >> > fedora/el6/el7
> >> >> > at
> >> >> > the end of the page.
> >> >> >
> >> >> >>
> >> >> >>
> >> >> >> Use gluster volume set <volname> disperse.shd-max-threads
> >> >> >> <num-threads
> >> >> >> (range: 1-64)>
> >> >> >>
> >> >> >> While testing this I thought of ways to decrease the number of
> >> >> >> crawls
> >> >> >> as
> >> >> >> well. But they are a bit involved. Try to create same set of data
> >> >> >> and
> >> >> >> see
> >> >> >> what is the time it takes to complete heals using number of
> threads
> >> >> >> as
> >> >> >> you
> >> >> >> increase the number of parallel heals from 1 to 64.
> >> >> >>
> >> >> >>>
> >> >> >>> On Wed, Aug 3, 2016 at 10:16 PM, Pranith Kumar Karampuri
> >> >> >>> <pkarampu at redhat.com> wrote:
> >> >> >>> >
> >> >> >>> >
> >> >> >>> > On Thu, Aug 4, 2016 at 12:45 AM, Serkan Çoban
> >> >> >>> > <cobanserkan at gmail.com>
> >> >> >>> > wrote:
> >> >> >>> >>
> >> >> >>> >> I prefer 3.7 if it is ok for you. Can you also provide build
> >> >> >>> >> instructions?
> >> >> >>> >
> >> >> >>> >
> >> >> >>> > 3.7 should be fine. Do you use rpms/debs/anything-else?
> >> >> >>> >
> >> >> >>> >>
> >> >> >>> >>
> >> >> >>> >> On Wed, Aug 3, 2016 at 10:12 PM, Pranith Kumar Karampuri
> >> >> >>> >> <pkarampu at redhat.com> wrote:
> >> >> >>> >> >
> >> >> >>> >> >
> >> >> >>> >> > On Thu, Aug 4, 2016 at 12:37 AM, Serkan Çoban
> >> >> >>> >> > <cobanserkan at gmail.com>
> >> >> >>> >> > wrote:
> >> >> >>> >> >>
> >> >> >>> >> >> Yes, but I can create 2+1(or 8+2) ec using two servers
> right?
> >> >> >>> >> >> I
> >> >> >>> >> >> have
> >> >> >>> >> >> 26 disks on each server.
> >> >> >>> >> >
> >> >> >>> >> >
> >> >> >>> >> > On which release-branch do you want the patch? I am testing
> it
> >> >> >>> >> > on
> >> >> >>> >> > master-branch now.
> >> >> >>> >> >
> >> >> >>> >> >>
> >> >> >>> >> >>
> >> >> >>> >> >> On Wed, Aug 3, 2016 at 9:59 PM, Pranith Kumar Karampuri
> >> >> >>> >> >> <pkarampu at redhat.com> wrote:
> >> >> >>> >> >> >
> >> >> >>> >> >> >
> >> >> >>> >> >> > On Thu, Aug 4, 2016 at 12:23 AM, Serkan Çoban
> >> >> >>> >> >> > <cobanserkan at gmail.com>
> >> >> >>> >> >> > wrote:
> >> >> >>> >> >> >>
> >> >> >>> >> >> >> I have two of my storage servers free, I think I can use
> >> >> >>> >> >> >> them
> >> >> >>> >> >> >> for
> >> >> >>> >> >> >> testing. Is two server testing environment ok for you?
> >> >> >>> >> >> >
> >> >> >>> >> >> >
> >> >> >>> >> >> > I think it would be better if you have at least 3. You
> can
> >> >> >>> >> >> > test
> >> >> >>> >> >> > it
> >> >> >>> >> >> > with
> >> >> >>> >> >> > 2+1
> >> >> >>> >> >> > ec configuration.
> >> >> >>> >> >> >
> >> >> >>> >> >> >>
> >> >> >>> >> >> >>
> >> >> >>> >> >> >> On Wed, Aug 3, 2016 at 9:44 PM, Pranith Kumar Karampuri
> >> >> >>> >> >> >> <pkarampu at redhat.com> wrote:
> >> >> >>> >> >> >> >
> >> >> >>> >> >> >> >
> >> >> >>> >> >> >> > On Wed, Aug 3, 2016 at 6:01 PM, Serkan Çoban
> >> >> >>> >> >> >> > <cobanserkan at gmail.com>
> >> >> >>> >> >> >> > wrote:
> >> >> >>> >> >> >> >>
> >> >> >>> >> >> >> >> Hi,
> >> >> >>> >> >> >> >>
> >> >> >>> >> >> >> >> May I ask if multi-threaded self heal for distributed
> >> >> >>> >> >> >> >> disperse
> >> >> >>> >> >> >> >> volumes
> >> >> >>> >> >> >> >> implemented in this release?
> >> >> >>> >> >> >> >
> >> >> >>> >> >> >> >
> >> >> >>> >> >> >> > Serkan,
> >> >> >>> >> >> >> > At the moment I am a bit busy with different
> >> >> >>> >> >> >> > work,
> >> >> >>> >> >> >> > Is
> >> >> >>> >> >> >> > it
> >> >> >>> >> >> >> > possible
> >> >> >>> >> >> >> > for you to help test the feature if I provide a patch?
> >> >> >>> >> >> >> > Actually
> >> >> >>> >> >> >> > the
> >> >> >>> >> >> >> > patch
> >> >> >>> >> >> >> > should be small. Testing is where lot of time will be
> >> >> >>> >> >> >> > spent
> >> >> >>> >> >> >> > on.
> >> >> >>> >> >> >> >
> >> >> >>> >> >> >> >>
> >> >> >>> >> >> >> >>
> >> >> >>> >> >> >> >> Thanks,
> >> >> >>> >> >> >> >> Serkan
> >> >> >>> >> >> >> >>
> >> >> >>> >> >> >> >> On Tue, Aug 2, 2016 at 5:30 PM, David Gossage
> >> >> >>> >> >> >> >> <dgossage at carouselchecks.com> wrote:
> >> >> >>> >> >> >> >> > On Tue, Aug 2, 2016 at 6:01 AM, Lindsay Mathieson
> >> >> >>> >> >> >> >> > <lindsay.mathieson at gmail.com> wrote:
> >> >> >>> >> >> >> >> >>
> >> >> >>> >> >> >> >> >> On 2/08/2016 5:07 PM, Kaushal M wrote:
> >> >> >>> >> >> >> >> >>>
> >> >> >>> >> >> >> >> >>> GlusterFS-3.7.14 has been released. This is a
> >> >> >>> >> >> >> >> >>> regular
> >> >> >>> >> >> >> >> >>> minor
> >> >> >>> >> >> >> >> >>> release.
> >> >> >>> >> >> >> >> >>> The release-notes are available at
> >> >> >>> >> >> >> >> >>>
> >> >> >>> >> >> >> >> >>>
> >> >> >>> >> >> >> >> >>>
> >> >> >>> >> >> >> >> >>>
> >> >> >>> >> >> >> >> >>>
> >> >> >>> >> >> >> >> >>>
> >> >> >>> >> >> >> >> >>>
> >> >> >>> >> >> >> >> >>>
> >> >> >>> >> >> >> >> >>> https://github.com/gluster/
> glusterfs/blob/release-3.7/doc/release-notes/3.7.14.md
> >> >> >>> >> >> >> >> >>
> >> >> >>> >> >> >> >> >>
> >> >> >>> >> >> >> >> >> Thanks Kaushal, I'll check it out
> >> >> >>> >> >> >> >> >>
> >> >> >>> >> >> >> >> >
> >> >> >>> >> >> >> >> > So far on my test box its working as expected. At
> >> >> >>> >> >> >> >> > least
> >> >> >>> >> >> >> >> > the
> >> >> >>> >> >> >> >> > issues
> >> >> >>> >> >> >> >> > that
> >> >> >>> >> >> >> >> > prevented it from running as before have
> disappeared.
> >> >> >>> >> >> >> >> > Will
> >> >> >>> >> >> >> >> > need
> >> >> >>> >> >> >> >> > to
> >> >> >>> >> >> >> >> > see
> >> >> >>> >> >> >> >> > how
> >> >> >>> >> >> >> >> > my test VM behaves after a few days.
> >> >> >>> >> >> >> >> >
> >> >> >>> >> >> >> >> >
> >> >> >>> >> >> >> >> >
> >> >> >>> >> >> >> >> >> --
> >> >> >>> >> >> >> >> >> Lindsay Mathieson
> >> >> >>> >> >> >> >> >>
> >> >> >>> >> >> >> >> >> _______________________________________________
> >> >> >>> >> >> >> >> >> Gluster-users mailing list
> >> >> >>> >> >> >> >> >> Gluster-users at gluster.org
> >> >> >>> >> >> >> >> >>
> >> >> >>> >> >> >> >> >> http://www.gluster.org/
> mailman/listinfo/gluster-users
> >> >> >>> >> >> >> >> >
> >> >> >>> >> >> >> >> >
> >> >> >>> >> >> >> >> >
> >> >> >>> >> >> >> >> > _______________________________________________
> >> >> >>> >> >> >> >> > Gluster-users mailing list
> >> >> >>> >> >> >> >> > Gluster-users at gluster.org
> >> >> >>> >> >> >> >> > http://www.gluster.org/
> mailman/listinfo/gluster-users
> >> >> >>> >> >> >> >> _______________________________________________
> >> >> >>> >> >> >> >> Gluster-users mailing list
> >> >> >>> >> >> >> >> Gluster-users at gluster.org
> >> >> >>> >> >> >> >> http://www.gluster.org/
> mailman/listinfo/gluster-users
> >> >> >>> >> >> >> >
> >> >> >>> >> >> >> >
> >> >> >>> >> >> >> >
> >> >> >>> >> >> >> >
> >> >> >>> >> >> >> > --
> >> >> >>> >> >> >> > Pranith
> >> >> >>> >> >> >
> >> >> >>> >> >> >
> >> >> >>> >> >> >
> >> >> >>> >> >> >
> >> >> >>> >> >> > --
> >> >> >>> >> >> > Pranith
> >> >> >>> >> >
> >> >> >>> >> >
> >> >> >>> >> >
> >> >> >>> >> >
> >> >> >>> >> > --
> >> >> >>> >> > Pranith
> >> >> >>> >
> >> >> >>> >
> >> >> >>> >
> >> >> >>> >
> >> >> >>> > --
> >> >> >>> > Pranith
> >> >> >>
> >> >> >>
> >> >> >>
> >> >> >>
> >> >> >> --
> >> >> >> Pranith
> >> >> >
> >> >> >
> >> >> >
> >> >> >
> >> >> > --
> >> >> > Pranith
> >> >
> >> >
> >> >
> >> >
> >> > --
> >> > Pranith
> >
> >
> >
> >
> > --
> > Pranith
>
--
Pranith
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160811/bbf2d742/attachment.html>
More information about the Gluster-users
mailing list