[Gluster-infra] [Gluster-devel] bug-1432542-mpx-restart-crash.t failing

Karthik Subrahmanya ksubrahm at redhat.com
Mon Jul 9 09:13:15 UTC 2018


Hi Deepshikha,

Are you looking into this failure? I can still see this happening for all
the regression runs.

Thanks & Regards,
Karthik

On Sun, Jul 8, 2018 at 7:18 AM Atin Mukherjee <amukherj at redhat.com> wrote:

>
> https://build.gluster.org/job/regression-test-with-multiplex/794/display/redirect
> has the same test failing. Is the reason of the failure different given
> this is on jenkins?
>
> On Sat, 7 Jul 2018 at 19:12, Deepshikha Khandelwal <dkhandel at redhat.com>
> wrote:
>
>> Hi folks,
>>
>> The issue[1] has been resolved. Now the softserve instance will be
>> having 2GB RAM i.e. same as that of the Jenkins builder's sizing
>> configurations.
>>
>> [1] https://github.com/gluster/softserve/issues/40
>>
>> Thanks,
>> Deepshikha Khandelwal
>>
>> On Fri, Jul 6, 2018 at 6:14 PM, Karthik Subrahmanya <ksubrahm at redhat.com>
>> wrote:
>> >
>> >
>> > On Fri 6 Jul, 2018, 5:18 PM Deepshikha Khandelwal, <dkhandel at redhat.com
>> >
>> > wrote:
>> >>
>> >> Hi Poornima/Karthik,
>> >>
>> >> We've looked into the memory error that this softserve instance have
>> >> showed up. These machine instances have 1GB RAM which is not in the
>> >> case with the Jenkins builder. It's 2GB RAM there.
>> >>
>> >> We've created the issue [1] and will solve it sooner.
>> >
>> > Great. Thanks for the update.
>> >>
>> >>
>> >> Sorry for the inconvenience.
>> >>
>> >> [1] https://github.com/gluster/softserve/issues/40
>> >>
>> >> Thanks,
>> >> Deepshikha Khandelwal
>> >>
>> >> On Fri, Jul 6, 2018 at 3:44 PM, Karthik Subrahmanya <
>> ksubrahm at redhat.com>
>> >> wrote:
>> >> > Thanks Poornima for the analysis.
>> >> > Can someone work on fixing this please?
>> >> >
>> >> > ~Karthik
>> >> >
>> >> > On Fri, Jul 6, 2018 at 3:17 PM Poornima Gurusiddaiah
>> >> > <pgurusid at redhat.com>
>> >> > wrote:
>> >> >>
>> >> >> The same test case is failing for my patch as well [1]. I requested
>> for
>> >> >> a
>> >> >> regression system and tried to reproduce it.
>> >> >> From my analysis, the brick process (mutiplexed) is consuming a lot
>> of
>> >> >> memory, and is being OOM killed. The regression has 1GB ram and the
>> >> >> process
>> >> >> is consuming more than 1GB. 1GB for 120 bricks is acceptable
>> >> >> considering
>> >> >> there is 1000 threads in that brick process.
>> >> >> Ways to fix:
>> >> >> - Increase the regression system RAM size OR
>> >> >> - Decrease the number of volumes in the test case.
>> >> >>
>> >> >> But what is strange is why the test passes sometimes for some
>> patches.
>> >> >> There could be some bug/? in memory consumption.
>> >> >>
>> >> >> Regards,
>> >> >> Poornima
>> >> >>
>> >> >>
>> >> >> On Fri, Jul 6, 2018 at 2:11 PM, Karthik Subrahmanya
>> >> >> <ksubrahm at redhat.com>
>> >> >> wrote:
>> >> >>>
>> >> >>> Hi,
>> >> >>>
>> >> >>> $subject is failing on centos regression for most of the patches
>> with
>> >> >>> timeout error.
>> >> >>>
>> >> >>> 07:32:34
>> >> >>>
>> >> >>>
>> ================================================================================
>> >> >>> 07:32:34 [07:33:05] Running tests in file
>> >> >>> ./tests/bugs/core/bug-1432542-mpx-restart-crash.t
>> >> >>> 07:32:34 Timeout set is 300, default 200
>> >> >>> 07:37:34 ./tests/bugs/core/bug-1432542-mpx-restart-crash.t timed
>> out
>> >> >>> after 300 seconds
>> >> >>> 07:37:34 ./tests/bugs/core/bug-1432542-mpx-restart-crash.t: bad
>> status
>> >> >>> 124
>> >> >>> 07:37:34
>> >> >>> 07:37:34        *********************************
>> >> >>> 07:37:34        *       REGRESSION FAILED       *
>> >> >>> 07:37:34        * Retrying failed tests in case *
>> >> >>> 07:37:34        * we got some spurious failures *
>> >> >>> 07:37:34        *********************************
>> >> >>> 07:37:34
>> >> >>> 07:42:34 ./tests/bugs/core/bug-1432542-mpx-restart-crash.t timed
>> out
>> >> >>> after 300 seconds
>> >> >>> 07:42:34 End of test
>> ./tests/bugs/core/bug-1432542-mpx-restart-crash.t
>> >> >>> 07:42:34
>> >> >>>
>> >> >>>
>> ================================================================================
>> >> >>>
>> >> >>> Can anyone take a look?
>> >> >>>
>> >> >>> Thanks,
>> >> >>> Karthik
>> >> >>>
>> >> >>>
>> >> >>>
>> >> >>> _______________________________________________
>> >> >>> Gluster-devel mailing list
>> >> >>> Gluster-devel at gluster.org
>> >> >>> https://lists.gluster.org/mailman/listinfo/gluster-devel
>> >> >>
>> >> >>
>> >> >
>> >> > _______________________________________________
>> >> > Gluster-infra mailing list
>> >> > Gluster-infra at gluster.org
>> >> > https://lists.gluster.org/mailman/listinfo/gluster-infra
>> _______________________________________________
>> Gluster-devel mailing list
>> Gluster-devel at gluster.org
>> https://lists.gluster.org/mailman/listinfo/gluster-devel
>>
> --
> - Atin (atinm)
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.gluster.org/pipermail/gluster-infra/attachments/20180709/284aa3dc/attachment.html>


More information about the Gluster-infra mailing list