[Gluster-devel] 3.7 pending patches

Raghavendra Gowdappa rgowdapp at redhat.com
Thu Jan 28 15:18:49 UTC 2016


+ Sakshi

----- Original Message -----
> From: "Raghavendra Gowdappa" <rgowdapp at redhat.com>
> To: "Pranith Kumar Karampuri" <pkarampu at redhat.com>
> Cc: "Venky Shankar" <vshankar at redhat.com>, "Gluster Devel" <gluster-devel at gluster.org>, "Vijay Bellur"
> <vbellur at redhat.com>, "Nithya Balachandran" <nbalacha at redhat.com>
> Sent: Thursday, January 28, 2016 8:47:51 PM
> Subject: Re: 3.7 pending patches
> 
> 
> 
> ----- Original Message -----
> > From: "Pranith Kumar Karampuri" <pkarampu at redhat.com>
> > To: "Venky Shankar" <vshankar at redhat.com>, "Gluster Devel"
> > <gluster-devel at gluster.org>
> > Cc: "Vijay Bellur" <vbellur at redhat.com>, "Raghavendra Gowdappa"
> > <rgowdapp at redhat.com>, "Nithya Balachandran"
> > <nbalacha at redhat.com>
> > Sent: Thursday, January 28, 2016 8:29:16 PM
> > Subject: Re: 3.7 pending patches
> > 
> > 
> > 
> > On 01/28/2016 07:05 PM, Venky Shankar wrote:
> > > Hey folks,
> > >
> > > I just merged patch #13302 (and it's 3.7 equivalent) which fixes a
> > > scrubber
> > > crash.
> > > This was causing other patches to fail regression.
> > >
> > > Requesting a rebase of patches (especially 3.7 pending) that were blocked
> > > due to
> > > this.
> > Thanks a lot for this venky, kotresh, Emmanuel. I re-triggered the builds.
> > 
> > I observed the following crash in one of the runs for
> > https://build.gluster.org/job/rackspace-regression-2GB-triggered/17819/console
> > (3.7):
> > (gdb) bt
> > #0  0x000000000040ecff in glusterfs_rebalance_event_notify_cbk (
> >      req=0x7f0e58006dbc, iov=0x7f0e6cadb5d0, count=1,
> > myframe=0x7f0e58003a7c)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/glusterfsd/src/glusterfsd-mgmt.c:1812
> > #1  0x00007f0e79a1274b in saved_frames_unwind (saved_frames=0x19ffe70)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:366
> > #2  0x00007f0e79a127ea in saved_frames_destroy (frames=0x19ffe70)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:383
> > #3  0x00007f0e79a12c41 in rpc_clnt_connection_cleanup (conn=0x19fea20)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:532
> > #4  0x00007f0e79a136cb in rpc_clnt_notify (trans=0x19fee70,
> > mydata=0x19fea20,
> >      event=RPC_TRANSPORT_DISCONNECT, data=0x19fee70)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:854
> > #5  0x00007f0e79a0fb76 in rpc_transport_notify (this=0x19fee70,
> >      event=RPC_TRANSPORT_DISCONNECT, data=0x19fee70)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-transport.c:546
> > #6  0x00007f0e6f1fd621 in socket_event_poll_err (this=0x19fee70)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-t---Type
> > <return> to continue, or q <return> to quit---
> > ransport/socket/src/socket.c:1151
> > #7  0x00007f0e6f20234c in socket_event_handler (fd=9, idx=1,
> > data=0x19fee70,
> >      poll_in=1, poll_out=0, poll_err=24)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2356
> > #8  0x00007f0e79cc386c in event_dispatch_epoll_handler
> > (event_pool=0x19c3c90,
> >      event=0x7f0e6cadbe70)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/libglusterfs/src/event-epoll.c:575
> > #9  0x00007f0e79cc3c5a in event_dispatch_epoll_worker (data=0x7f0e68014970)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/libglusterfs/src/event-epoll.c:678
> > #10 0x00007f0e78f2aa51 in start_thread () from ./lib64/libpthread.so.0
> > #11 0x00007f0e7889493d in clone () from ./lib64/libc.so.6
> > (gdb) fr 0
> > #0  0x000000000040ecff in glusterfs_rebalance_event_notify_cbk (
> >      req=0x7f0e58006dbc, iov=0x7f0e6cadb5d0, count=1,
> > myframe=0x7f0e58003a7c)
> >      at
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/glusterfsd/src/glusterfsd-mgmt.c:1812
> > 1812    in
> > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/glusterfsd/src/glusterfsd-mgmt.c
> > (gdb) info locals
> > rsp = {op_ret = 0, op_errno = 0, dict = {dict_len = 0, dict_val = 0x0}}
> > frame = 0x7f0e58003a7c
> > ctx = 0x0
> > ret = 0
> > __FUNCTION__ = "glusterfs_rebalance_event_notify_cbk"
> > (gdb) p frame->this
> > $1 = (xlator_t *) 0x3a600000000000
> > (gdb) p frame->this->name
> > Cannot access memory at address 0x3a600000000000
> 
> There is a patch by sakshi on master at:
> http://review.gluster.org/#/c/13262/
> 
> Its blocked on netbsd regression failures.
> 
> @sakshi,
> 
> Can you figure out what is the issue with netbsd regression failure? Also,
> can you send a backport for 3.7.
> 
> regards,
> Raghavendra.
> 
> > 
> > Pranith
> > >
> > > Thanks,
> > >
> > >                  Venky
> > 
> >


More information about the Gluster-devel mailing list