----- Original Message ----- > From: "Pranith Kumar Karampuri" <pkarampu@xxxxxxxxxx> > To: "Venky Shankar" <vshankar@xxxxxxxxxx>, "Gluster Devel" <gluster-devel@xxxxxxxxxxx> > Cc: "Vijay Bellur" <vbellur@xxxxxxxxxx>, "Raghavendra Gowdappa" <rgowdapp@xxxxxxxxxx>, "Nithya Balachandran" > <nbalacha@xxxxxxxxxx> > Sent: Thursday, January 28, 2016 8:29:16 PM > Subject: Re: 3.7 pending patches > > > > On 01/28/2016 07:05 PM, Venky Shankar wrote: > > Hey folks, > > > > I just merged patch #13302 (and it's 3.7 equivalent) which fixes a scrubber > > crash. > > This was causing other patches to fail regression. > > > > Requesting a rebase of patches (especially 3.7 pending) that were blocked > > due to > > this. > Thanks a lot for this venky, kotresh, Emmanuel. I re-triggered the builds. > > I observed the following crash in one of the runs for > https://build.gluster.org/job/rackspace-regression-2GB-triggered/17819/console > (3.7): > (gdb) bt > #0 0x000000000040ecff in glusterfs_rebalance_event_notify_cbk ( > req=0x7f0e58006dbc, iov=0x7f0e6cadb5d0, count=1, > myframe=0x7f0e58003a7c) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/glusterfsd/src/glusterfsd-mgmt.c:1812 > #1 0x00007f0e79a1274b in saved_frames_unwind (saved_frames=0x19ffe70) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:366 > #2 0x00007f0e79a127ea in saved_frames_destroy (frames=0x19ffe70) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:383 > #3 0x00007f0e79a12c41 in rpc_clnt_connection_cleanup (conn=0x19fea20) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:532 > #4 0x00007f0e79a136cb in rpc_clnt_notify (trans=0x19fee70, > mydata=0x19fea20, > event=RPC_TRANSPORT_DISCONNECT, data=0x19fee70) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-clnt.c:854 > #5 0x00007f0e79a0fb76 in rpc_transport_notify (this=0x19fee70, > event=RPC_TRANSPORT_DISCONNECT, data=0x19fee70) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-lib/src/rpc-transport.c:546 > #6 0x00007f0e6f1fd621 in socket_event_poll_err (this=0x19fee70) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-t---Type > <return> to continue, or q <return> to quit--- > ransport/socket/src/socket.c:1151 > #7 0x00007f0e6f20234c in socket_event_handler (fd=9, idx=1, > data=0x19fee70, > poll_in=1, poll_out=0, poll_err=24) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/rpc/rpc-transport/socket/src/socket.c:2356 > #8 0x00007f0e79cc386c in event_dispatch_epoll_handler > (event_pool=0x19c3c90, > event=0x7f0e6cadbe70) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/libglusterfs/src/event-epoll.c:575 > #9 0x00007f0e79cc3c5a in event_dispatch_epoll_worker (data=0x7f0e68014970) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/libglusterfs/src/event-epoll.c:678 > #10 0x00007f0e78f2aa51 in start_thread () from ./lib64/libpthread.so.0 > #11 0x00007f0e7889493d in clone () from ./lib64/libc.so.6 > (gdb) fr 0 > #0 0x000000000040ecff in glusterfs_rebalance_event_notify_cbk ( > req=0x7f0e58006dbc, iov=0x7f0e6cadb5d0, count=1, > myframe=0x7f0e58003a7c) > at > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/glusterfsd/src/glusterfsd-mgmt.c:1812 > 1812 in > /home/jenkins/root/workspace/rackspace-regression-2GB-triggered/glusterfsd/src/glusterfsd-mgmt.c > (gdb) info locals > rsp = {op_ret = 0, op_errno = 0, dict = {dict_len = 0, dict_val = 0x0}} > frame = 0x7f0e58003a7c > ctx = 0x0 > ret = 0 > __FUNCTION__ = "glusterfs_rebalance_event_notify_cbk" > (gdb) p frame->this > $1 = (xlator_t *) 0x3a600000000000 > (gdb) p frame->this->name > Cannot access memory at address 0x3a600000000000 There is a patch by sakshi on master at: http://review.gluster.org/#/c/13262/ Its blocked on netbsd regression failures. @sakshi, Can you figure out what is the issue with netbsd regression failure? Also, can you send a backport for 3.7. regards, Raghavendra. > > Pranith > > > > Thanks, > > > > Venky > > _______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-devel