[Gluster-devel] 3.7 pending patches
Raghavendra Gowdappa
rgowdapp at redhat.com
Thu Jan 28 15:18:49 UTC 2016
+ Sakshi
----- Original Message -----
> From: "Raghavendra Gowdappa" <rgowdapp at redhat.com>
> To: "Pranith Kumar Karampuri" <pkarampu at redhat.com>
> Cc: "Venky Shankar" <vshankar at redhat.com>, "Gluster Devel" <gluster-devel at gluster.org>, "Vijay Bellur"
> <vbellur at redhat.com>, "Nithya Balachandran" <nbalacha at redhat.com>
> Sent: Thursday, January 28, 2016 8:47:51 PM
> Subject: Re: 3.7 pending patches
>
>
>
> ----- Original Message -----
> > From: "Pranith Kumar Karampuri" <pkarampu at redhat.com>
> > To: "Venky Shankar" <vshankar at redhat.com>, "Gluster Devel"
> > <gluster-devel at gluster.org>
> > Cc: "Vijay Bellur" <vbellur at redhat.com>, "Raghavendra Gowdappa"
> > <rgowdapp at redhat.com>, "Nithya Balachandran"
> > <nbalacha at redhat.com>
> > Sent: Thursday, January 28, 2016 8:29:16 PM
> > Subject: Re: 3.7 pending patches
> >
> >
> >
> > On 01/28/2016 07:05 PM, Venky Shankar wrote:
> > > Hey folks,
> > >
> > > I just merged patch #13302 (and it's 3.7 equivalent) which fixes a
> > > scrubber
> > > crash.
> > > This was causing other patches to fail regression.
> > >
> > > Requesting a rebase of patches (especially 3.7 pending) that were blocked
> > > due to
> > > this.
> > Thanks a lot for this venky, kotresh, Emmanuel. I re-triggered the builds.
> >
> > I observed the following crash in one of the runs for
> > https://build.gluster.org/job/rackspace-regression-2GB-triggered/17819/console
> > (3.7):
> > (gdb) bt
> > #0 0x000000000040ecff in glusterfs_rebalance_event_notify_cbk (
> > req=0x7f0e58006dbc, iov=0x7f0e6cadb5d0, count=1,
> > myframe=0x7f0e58003a7c)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/glusterfsd/src/glusterfsd-mgmt.c:1812
> > #1 0x00007f0e79a1274b in saved_frames_unwind (saved_frames=0x19ffe70)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:366
> > #2 0x00007f0e79a127ea in saved_frames_destroy (frames=0x19ffe70)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:383
> > #3 0x00007f0e79a12c41 in rpc_clnt_connection_cleanup (conn=0x19fea20)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:532
> > #4 0x00007f0e79a136cb in rpc_clnt_notify (trans=0x19fee70,
> > mydata=0x19fea20,
> > event=RPC_TRANSPORT_DISCONNECT, data=0x19fee70)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:854
> > #5 0x00007f0e79a0fb76 in rpc_transport_notify (this=0x19fee70,
> > event=RPC_TRANSPORT_DISCONNECT, data=0x19fee70)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-transport.c:546
> > #6 0x00007f0e6f1fd621 in socket_event_poll_err (this=0x19fee70)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-t---Type
> > <return> to continue, or q <return> to quit---
> > ransport/socket/src/socket.c:1151
> > #7 0x00007f0e6f20234c in socket_event_handler (fd=9, idx=1,
> > data=0x19fee70,
> > poll_in=1, poll_out=0, poll_err=24)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2356
> > #8 0x00007f0e79cc386c in event_dispatch_epoll_handler
> > (event_pool=0x19c3c90,
> > event=0x7f0e6cadbe70)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/libglusterfs/src/event-epoll.c:575
> > #9 0x00007f0e79cc3c5a in event_dispatch_epoll_worker (data=0x7f0e68014970)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/libglusterfs/src/event-epoll.c:678
> > #10 0x00007f0e78f2aa51 in start_thread () from ./lib64/libpthread.so.0
> > #11 0x00007f0e7889493d in clone () from ./lib64/libc.so.6
> > (gdb) fr 0
> > #0 0x000000000040ecff in glusterfs_rebalance_event_notify_cbk (
> > req=0x7f0e58006dbc, iov=0x7f0e6cadb5d0, count=1,
> > myframe=0x7f0e58003a7c)
> > at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/glusterfsd/src/glusterfsd-mgmt.c:1812
> > 1812 in
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/glusterfsd/src/glusterfsd-mgmt.c
> > (gdb) info locals
> > rsp = {op_ret = 0, op_errno = 0, dict = {dict_len = 0, dict_val = 0x0}}
> > frame = 0x7f0e58003a7c
> > ctx = 0x0
> > ret = 0
> > __FUNCTION__ = "glusterfs_rebalance_event_notify_cbk"
> > (gdb) p frame->this
> > $1 = (xlator_t *) 0x3a600000000000
> > (gdb) p frame->this->name
> > Cannot access memory at address 0x3a600000000000
>
> There is a patch by sakshi on master at:
> http://review.gluster.org/#/c/13262/
>
> Its blocked on netbsd regression failures.
>
> @sakshi,
>
> Can you figure out what is the issue with netbsd regression failure? Also,
> can you send a backport for 3.7.
>
> regards,
> Raghavendra.
>
> >
> > Pranith
> > >
> > > Thanks,
> > >
> > > Venky
> >
> >
More information about the Gluster-devel
mailing list