[Gluster-Maintainers] glusterfs-3.12.7 released

Jiffin Tony Thottan jthottan at redhat.com
Thu Mar 22 08:08:56 UTC 2018



On Thursday 22 March 2018 01:07 PM, Atin Mukherjee wrote:
>
>
> On Thu, Mar 22, 2018 at 12:38 PM, Jiffin Tony Thottan 
> <jthottan at redhat.com <mailto:jthottan at redhat.com>> wrote:
>
>
>
>     On Thursday 22 March 2018 12:29 PM, Jiffin Tony Thottan wrote:
>>
>>
>>
>>     On Wednesday 21 March 2018 09:06 AM, Atin Mukherjee wrote:
>>>
>>>
>>>     On Wed, Mar 21, 2018 at 12:18 AM, Shyam Ranganathan
>>>     <srangana at redhat.com <mailto:srangana at redhat.com>> wrote:
>>>
>>>         On 03/20/2018 01:10 PM, Jiffin Thottan wrote:
>>>         > Hi Shyam,
>>>         >
>>>         > Actually I planned to do the release on March 8th(posted
>>>         the release note on that day). But it didn't happen.
>>>         > I didn't merge any patches post sending the release
>>>         note(blocker bug had some merge conflict issue on that so I
>>>         skipped AFAIR).
>>>         > I performed 3.12.7 tagging yesterday and ran the build job
>>>         today.
>>>         >
>>>         > Can u please provide a suggestion here ? Do I need to
>>>         perform a 3.12.7-1 for the blocker bug ?
>>>
>>>         I see that the bug is marked against the tracker, but is not a
>>>         regression or an issue that is serious enough that it cannot
>>>         wait for
>>>         the next minor release.
>>>
>>>         Copied Atin to the mail, who opened that issue for his
>>>         comments. If he
>>>         agrees, let's get this moving and get the fix into the next
>>>         minor release.
>>>
>>>
>>>     Even though it's not a regression and a day 1 bug with brick
>>>     multiplexing, the issue is severe enough to consider this to be
>>>     fixed *asap* . In this scenario, if you're running a multi node
>>>     cluster with brick multiplexing enabled and one node down and
>>>     there're some volume operations performed and post that when the
>>>     node comes back, brick processes fail to come up.
>>
>>     Issue is impact only with glusterd, whether any other component
>>     needs this fix?
>
>     Sorry I meant brick multiplexing not glusterd
>     --
>     Jiffin
>
>>     If it is issue not report from upstream user/community, I prefer
>>     to take it for next release.
>
>
> IMO, assessment of an issue should be done based on its merit, not 
> based on where it originates from. It might be a fair question to ask 
> that "do we have users who have brick multiplexing enabled" and based 
> on that take a call to fix it immediately or as part of next update 
> but at the same time, you're still exposing a known problem with out 
> flagging a warning that don't use brick multiplexing till this bug is 
> fixed.

I have not yet sent the announcement mail for the release nor sent 
release notes to https://docs.gluster.org/en.  I can mention about it 
over there
--
Jiffin


>
>>
>>     Regards,
>>     Jiffin
>>
>>>
>>>         >
>>>         > --
>>>         > Regards,
>>>         > Jiffin
>>>         >
>>>         >
>>>         >
>>>         >
>>>         > ----- Original Message -----
>>>         > From: "Shyam Ranganathan" <srangana at redhat.com
>>>         <mailto:srangana at redhat.com>>
>>>         > To: jenkins at build.gluster.org
>>>         <mailto:jenkins at build.gluster.org>, packaging at gluster.org
>>>         <mailto:packaging at gluster.org>, maintainers at gluster.org
>>>         <mailto:maintainers at gluster.org>
>>>         > Sent: Tuesday, March 20, 2018 9:06:57 PM
>>>         > Subject: Re: [Gluster-Maintainers] glusterfs-3.12.7 released
>>>         >
>>>         > On 03/20/2018 11:19 AM, jenkins at build.gluster.org
>>>         <mailto:jenkins at build.gluster.org> wrote:
>>>         >> SRC:
>>>         https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.tar.gz
>>>         <https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.tar.gz>
>>>         >> HASH:
>>>         https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.sha512sum
>>>         <https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.sha512sum>
>>>         >>
>>>         >> This release is made off jenkins-release-47
>>>         >
>>>         > Jiffin, there are about 6 patches ready in the 3.12 queue,
>>>         that are not
>>>         > merged for this release, why?
>>>         >
>>>         https://review.gluster.org/#/projects/glusterfs,dashboards/dashboard:3-12-dashboard
>>>         <https://review.gluster.org/#/projects/glusterfs,dashboards/dashboard:3-12-dashboard>
>>>         >
>>>         > The tracker bug for 3.12.7 calls out
>>>         > https://bugzilla.redhat.com/show_bug.cgi?id=1543708
>>>         <https://bugzilla.redhat.com/show_bug.cgi?id=1543708> as a
>>>         blocker, and
>>>         > has a patch, which is not merged.
>>>         >
>>>         > Was this some test packaging job?
>>>         >
>>>         >
>>>         >
>>>         >
>>>         >>
>>>         >>
>>>         >>
>>>         >> _______________________________________________
>>>         >> maintainers mailing list
>>>         >> maintainers at gluster.org <mailto:maintainers at gluster.org>
>>>         >> http://lists.gluster.org/mailman/listinfo/maintainers
>>>         <http://lists.gluster.org/mailman/listinfo/maintainers>
>>>         >>
>>>         > _______________________________________________
>>>         > maintainers mailing list
>>>         > maintainers at gluster.org <mailto:maintainers at gluster.org>
>>>         > http://lists.gluster.org/mailman/listinfo/maintainers
>>>         <http://lists.gluster.org/mailman/listinfo/maintainers>
>>>         >
>>>
>>>
>>
>>
>>
>>     _______________________________________________
>>     maintainers mailing list
>>     maintainers at gluster.org <mailto:maintainers at gluster.org>
>>     http://lists.gluster.org/mailman/listinfo/maintainers
>>     <http://lists.gluster.org/mailman/listinfo/maintainers>
>
>
>     _______________________________________________
>     maintainers mailing list
>     maintainers at gluster.org <mailto:maintainers at gluster.org>
>     http://lists.gluster.org/mailman/listinfo/maintainers
>     <http://lists.gluster.org/mailman/listinfo/maintainers>
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/maintainers/attachments/20180322/1e6a3f65/attachment.html>


More information about the maintainers mailing list