Re: iscsi_trx going into D state

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




----- Original Message -----
> From: "Robert LeBlanc" <robert@xxxxxxxxxxxxx>
> To: "Doug Ledford" <dledford@xxxxxxxxxx>
> Cc: "Nicholas A. Bellinger" <nab@xxxxxxxxxxxxxxx>, "Zhu Lingshan" <lszhu@xxxxxxxx>, "linux-rdma"
> <linux-rdma@xxxxxxxxxxxxxxx>, linux-scsi@xxxxxxxxxxxxxxx, "Sagi Grimberg" <sagi@xxxxxxxxxxx>, "Christoph Hellwig"
> <hch@xxxxxx>
> Sent: Tuesday, January 3, 2017 7:11:40 PM
> Subject: Re: iscsi_trx going into D state
> 
> With the last patch it is getting hung up on wait_for_completion in
> target_wait_for_sess_cmds. I don't know what t_state or fabric state
> mean. To me it looks like a queue is not being emptied, but it would
> help if someone confirmed this and has some pointers on how to
> properly flush them when the communication is interrupted.
> 
> [  222.989134] Starting iscsit_close_connection.
> [  222.993555] Calling flush_workqueue.
> [  222.997703] Returned from flush_workqueue.
> [  223.005802] isert_wait_conn calling ib_close_qp/ib_drain_qp.
> [  223.011892] isert_wait_conn finished ib_close_qp/ib_drain_qp.
> [  223.018063] isert_wait_conn calling isert_put_unsol_pending_cmds.
> [  223.024574] isert_wait_conn returned from
> isert_put_unsol_pending_cmds.
> [  223.031582] isert_wait_conn calling isert_wait4cmds.
> [  223.036942] isert_wait4cmds calling
> target_sess_cmd_list_set_waiting.
> [  223.043789] isert_wait4cmds returned from
> target_sess_cmd_list_set_waiting.
> [  223.051135] isert_wait4cmds calling target_wait_for_sess_cmds.
> [  223.057362] Waiting for se_cmd: ffff887ebf88bd00 t_state: 6, fabric
> state: 29
> [  223.064893] target_wait_for_sess_cmds calling
> spin_unlock_irqrestore.
> [  223.071748] target_wait_for_sess_cmds calling wait_for_completion.
> [  224.997636] Calling wait_for_common.
> [  225.001936] Starting __wait_for_common.
> [  225.006226] Calling do_wait_for_common.
> 
> Thanks
> ----------------
> Robert LeBlanc
> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> 
> 
> On Tue, Jan 3, 2017 at 1:07 PM, Robert LeBlanc <robert@xxxxxxxxxxxxx> wrote:
> > With this patch I'm not seeing the __ib_drain_sq backtraces, but I'm
> > still seeing the previous backtraces.
> >
> > diff --git a/drivers/infiniband/ulp/isert/ib_isert.c
> > b/drivers/infiniband/ulp/isert/ib_isert.c
> > index 6dd43f6..1e53502 100644
> > --- a/drivers/infiniband/ulp/isert/ib_isert.c
> > +++ b/drivers/infiniband/ulp/isert/ib_isert.c
> > @@ -2595,7 +2595,7 @@ static void isert_wait_conn(struct iscsi_conn *conn)
> >        isert_conn_terminate(isert_conn);
> >        mutex_unlock(&isert_conn->mutex);
> >
> > -       ib_drain_qp(isert_conn->qp);
> > +       ib_close_qp(isert_conn->qp);
> >        isert_put_unsol_pending_cmds(conn);
> >        isert_wait4cmds(conn);
> >        isert_wait4logout(isert_conn);
> >
> > I was thinking that if the connection was brought down uncleanly then
> > there may be messages(??) it the send queue that would never be
> > consumed by the application, so it would never drain and would have to
> > be forcibly emptied. Maybe there is something stuck in the command
> > queue as well?
> >
> > [(support-1.0) root@prv-0-13-roberttest ~]# ps aux | grep " D "
> > root     15426  0.0  0.0      0     0 ?        D    12:48   0:00 [iscsi_np]
> > root     15429  0.0  0.0      0     0 ?        D    12:48   0:00
> > [iscsi_ttx]
> > root     16077  0.0  0.0 112656  2216 pts/0    S+   12:55   0:00 grep
> > --color=auto  D
> > [(support-1.0) root@prv-0-13-roberttest ~]# cat /proc/15426/stack
> > [<ffffffff815304d1>] iscsit_stop_session+0x1b1/0x1c0
> > [<ffffffff81521c62>] iscsi_check_for_session_reinstatement+0x1e2/0x270
> > [<ffffffff81524660>] iscsi_target_check_for_existing_instances+0x30/0x40
> > [<ffffffff815247a8>] iscsi_target_do_login+0x138/0x630
> > [<ffffffff815259be>] iscsi_target_start_negotiation+0x4e/0xa0
> > [<ffffffff8152355e>] __iscsi_target_login_thread+0x83e/0xf20
> > [<ffffffff81523c64>] iscsi_target_login_thread+0x24/0x30
> > [<ffffffff810a3059>] kthread+0xd9/0xf0
> > [<ffffffff817732d5>] ret_from_fork+0x25/0x30
> > [<ffffffffffffffff>] 0xffffffffffffffff
> > [(support-1.0) root@prv-0-13-roberttest ~]# cat /proc/15429/stack
> > [<ffffffff8150c689>] target_wait_for_sess_cmds+0x49/0x190
> > [<ffffffffa0705744>] isert_wait_conn+0x1a4/0x2d0 [ib_isert]
> > [<ffffffff815309b7>] iscsit_close_connection+0x157/0x860
> > [<ffffffff8151f10b>] iscsit_take_action_for_connection_exit+0x7b/0xf0
> > [<ffffffff81530150>] iscsi_target_tx_thread+0x150/0x1d0
> > [<ffffffff810a3059>] kthread+0xd9/0xf0
> > [<ffffffff817732d5>] ret_from_fork+0x25/0x30
> > [<ffffffffffffffff>] 0xffffffffffffffff
> > ----------------
> > Robert LeBlanc
> > PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> >
> >
> > On Fri, Dec 30, 2016 at 4:07 PM, Robert LeBlanc <robert@xxxxxxxxxxxxx>
> > wrote:
> >> I decided to try something completely different... Running the stock
> >> CentOS 3.10 kernel and OFED 3.4 on both hosts, I'm not seeing the hung
> >> processes and the tests complete successfully. The same seems to be
> >> true for the target on 4.9 and the initiator on 3.10.
> >>
> >> However, with the target on 3.10 and the initiator on 4.9, I get this
> >> on the target:
> >>
> >> [(support-1.0) root@prv-0-13-roberttest ~]# ps aux | grep " D "
> >> root     14791  0.0  0.0      0     0 ?        D    15:08   0:00
> >> [iscsi_np]
> >> root     14795  0.0  0.0      0     0 ?        D    15:08   0:00
> >> [iscsi_trx]
> >> root     14852  0.0  0.0 112648   976 pts/0    S+   15:11   0:00 grep
> >> --color=auto  D
> >> [(support-1.0) root@prv-0-13-roberttest ~]# uname -a
> >> Linux prv-0-13-roberttest.betterservers.com 3.10.0-327.36.3.el7.x86_64
> >> #1 SMP Mon Oct 24 16:09:20 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
> >> [(support-1.0) root@prv-0-13-roberttest ~]# cat /proc/14791/stack
> >> [<ffffffffa09dde48>] iscsit_stop_session+0x1c8/0x1e0 [iscsi_target_mod]
> >> [<ffffffffa09ceefa>] iscsi_check_for_session_reinstatement+0x1ea/0x280
> >> [iscsi_target_mod]
> >> [<ffffffffa09d19f5>]
> >> iscsi_target_check_for_existing_instances+0x35/0x40 [iscsi_target_mod]
> >> [<ffffffffa09d1b41>] iscsi_target_do_login+0x141/0x670 [iscsi_target_mod]
> >> [<ffffffffa09d2f4c>] iscsi_target_start_negotiation+0x1c/0xb0
> >> [iscsi_target_mod]
> >> [<ffffffffa09d0c6f>] iscsi_target_login_thread+0xadf/0x1050
> >> [iscsi_target_mod]
> >> [<ffffffff810a5b8f>] kthread+0xcf/0xe0
> >> [<ffffffff81646a98>] ret_from_fork+0x58/0x90
> >> [<ffffffffffffffff>] 0xffffffffffffffff
> >> [(support-1.0) root@prv-0-13-roberttest ~]# cat /proc/14795/stack
> >> [<ffffffffa0801469>] isert_wait4flush+0x79/0xc0 [ib_isert]
> >> [<ffffffffa080150b>] isert_wait_conn+0x5b/0x2d0 [ib_isert]
> >> [<ffffffffa09ddfbd>] iscsit_close_connection+0x15d/0x820
> >> [iscsi_target_mod]
> >> [<ffffffffa09cc183>] iscsit_take_action_for_connection_exit+0x83/0x110
> >> [iscsi_target_mod]
> >> [<ffffffffa09dccb7>] iscsi_target_rx_thread+0x1e7/0xf80 [iscsi_target_mod]
> >> [<ffffffff810a5b8f>] kthread+0xcf/0xe0
> >> [<ffffffff81646a98>] ret_from_fork+0x58/0x90
> >> [<ffffffffffffffff>] 0xffffffffffffffff
> >>
> >> [  345.970157] iSCSI Login timeout on Network Portal 0.0.0.0:3260
> >> [  483.850714] INFO: task iscsi_np:14791 blocked for more than 120
> >> seconds.
> >> [  483.857467] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs"
> >> disables this message.
> >> [  483.865326] iscsi_np        D 0000000000000000     0 14791      2
> >> 0x00000004
> >> [  483.872460]  ffff886e3b117be0 0000000000000046 ffff887ede579700
> >> ffff886e3b117fd8
> >> [  483.879983]  ffff886e3b117fd8 ffff886e3b117fd8 ffff887ede579700
> >> ffff883ef7898160
> >> [  483.887500]  ffff883ef7898168 7fffffffffffffff ffff887ede579700
> >> 0000000000000000
> >> [  483.895025] Call Trace:
> >> [  483.897505]  [<ffffffff8163bb39>] schedule+0x29/0x70
> >> [  483.902496]  [<ffffffff81639829>] schedule_timeout+0x209/0x2d0
> >> [  483.908355]  [<ffffffff812fc60b>] ? simple_strtoull+0x3b/0x70
> >> [  483.914128]  [<ffffffff8163bf06>] wait_for_completion+0x116/0x170
> >> [  483.920253]  [<ffffffff810b8940>] ? wake_up_state+0x20/0x20
> >> [  483.925847]  [<ffffffffa09dde48>] iscsit_stop_session+0x1c8/0x1e0
> >> [iscsi_target_mod]
> >> [  483.933612]  [<ffffffffa09ceefa>]
> >> iscsi_check_for_session_reinstatement+0x1ea/0x280 [iscsi_target_mod]
> >> [  483.942944]  [<ffffffffa09d19f5>]
> >> iscsi_target_check_for_existing_instances+0x35/0x40 [iscsi_target_mod]
> >> [  483.953304]  [<ffffffffa09d1b41>] iscsi_target_do_login+0x141/0x670
> >> [iscsi_target_mod]
> >> [  483.961988]  [<ffffffffa09d2f4c>]
> >> iscsi_target_start_negotiation+0x1c/0xb0 [iscsi_target_mod]
> >> [  483.971278]  [<ffffffffa09d0c6f>]
> >> iscsi_target_login_thread+0xadf/0x1050 [iscsi_target_mod]
> >> [  483.980346]  [<ffffffff8163b401>] ? __schedule+0x1f1/0x900
> >> [  483.986525]  [<ffffffffa09d0190>] ?
> >> iscsi_target_login_sess_out+0x250/0x250 [iscsi_target_mod]
> >> [  483.995816]  [<ffffffff810a5b8f>] kthread+0xcf/0xe0
> >> [  484.001403]  [<ffffffff810a5ac0>] ? kthread_create_on_node+0x140/0x140
> >> [  484.008608]  [<ffffffff81646a98>] ret_from_fork+0x58/0x90
> >> [  484.014672]  [<ffffffff810a5ac0>] ? kthread_create_on_node+0x140/0x140
> >> [  484.021896] INFO: task iscsi_trx:14795 blocked for more than 120
> >> seconds.
> >> [  484.029349] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs"
> >> disables this message.
> >> [  484.037849] iscsi_trx       D ffff887ee64f8000     0 14795      2
> >> 0x00000004
> >> [  484.045598]  ffff886e391bfbe0 0000000000000046 ffff887ed715c500
> >> ffff886e391bffd8
> >> [  484.053753]  ffff886e391bffd8 ffff886e391bffd8 ffff887ed715c500
> >> ffff887ee64f91d0
> >> [  484.061891]  ffff887ee64f91d8 7fffffffffffffff ffff887ed715c500
> >> ffff887ee64f8000
> >> [  484.070049] Call Trace:
> >> [  484.073174]  [<ffffffff8163bb39>] schedule+0x29/0x70
> >> [  484.078797]  [<ffffffff81639829>] schedule_timeout+0x209/0x2d0
> >> [  484.085290]  [<ffffffffa0672125>] ? cm_alloc_msg+0x115/0x180 [ib_cm]
> >> [  484.092252]  [<ffffffff8163bf06>] wait_for_completion+0x116/0x170
> >> [  484.098960]  [<ffffffff810b8940>] ? wake_up_state+0x20/0x20
> >> [  484.105132]  [<ffffffffa0801469>] isert_wait4flush+0x79/0xc0 [ib_isert]
> >> [  484.112369]  [<ffffffffa080150b>] isert_wait_conn+0x5b/0x2d0 [ib_isert]
> >> [  484.119566]  [<ffffffffa09ddfbd>]
> >> iscsit_close_connection+0x15d/0x820 [iscsi_target_mod]
> >> [  484.128239]  [<ffffffff8163ca67>] ?
> >> wait_for_completion_interruptible+0x167/0x1d0
> >> [  484.136341]  [<ffffffffa09dcad0>] ?
> >> iscsi_target_tx_thread+0x200/0x200 [iscsi_target_mod]
> >> [  484.145135]  [<ffffffffa09cc183>]
> >> iscsit_take_action_for_connection_exit+0x83/0x110 [iscsi_target_mod]
> >> [  484.155067]  [<ffffffffa09dccb7>]
> >> iscsi_target_rx_thread+0x1e7/0xf80 [iscsi_target_mod]
> >> [  484.163700]  [<ffffffff81013588>] ? __switch_to+0xf8/0x4b0
> >> [  484.169774]  [<ffffffffa09dcad0>] ?
> >> iscsi_target_tx_thread+0x200/0x200 [iscsi_target_mod]
> >> [  484.178530]  [<ffffffff810a5b8f>] kthread+0xcf/0xe0
> >> [  484.183991]  [<ffffffff810a5ac0>] ? kthread_create_on_node+0x140/0x140
> >> [  484.191106]  [<ffffffff81646a98>] ret_from_fork+0x58/0x90
> >> [  484.197096]  [<ffffffff810a5ac0>] ? kthread_create_on_node+0x140/0x140
> >>
> >> I think there are two bugs here. Something in 4.9 iser (initiator) is
> >> causing a shutdown of the session when limited to 10 Gb. This second
> >> is in isert (target) where when a session isn't cleanly closed, it
> >> gets hung on cleaning up the session. It seems that bug #1 triggers
> >> bug #2 much easier than on Infiniband.
> >>
> >> I hope this is useful.
> >>
> >> ----------------
> >> Robert LeBlanc
> >> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> >>
> >>
> >> On Thu, Dec 29, 2016 at 4:57 PM, Robert LeBlanc <robert@xxxxxxxxxxxxx>
> >> wrote:
> >>> OK, I've drilled down a little more and
> >>>
> >>> timeout = action(timeout);
> >>>
> >>> in do_wait_for_common() in kernel/sched/completion.c is not returning.
> >>> I'll have to see if I can make more progress tomorrow.
> >>> ----------------
> >>> Robert LeBlanc
> >>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> >>>
> >>>
> >>> On Thu, Dec 29, 2016 at 2:23 PM, Robert LeBlanc <robert@xxxxxxxxxxxxx>
> >>> wrote:
> >>>> I know most people are ignoring this thread by now, but I hope someone
> >>>> is still reading and can offer some ideas.
> >>>>
> >>>> It looks like ib_drain_qp_done() is not being called the first time
> >>>> that __ib_drain_sq() is called from iscsit_close_connection(). I tried
> >>>> to debug wait_for_completion() and friends, but they are called by too
> >>>> many things and I don't know how to filter out what I'm looking for.
> >>>> My next idea is to copy the completion functions here so that I can
> >>>> add debug to only that path. I feel like I'm inching closer to the
> >>>> problem, stumbling around in the dark.
> >>>>
> >>>> [Thu Dec 29 14:02:03 2016] Starting iscsit_close_connection.
> >>>> [Thu Dec 29 14:02:03 2016] isert_wait_conn calling ib_drain_qp.
> >>>> [Thu Dec 29 14:02:03 2016] ib_drain_qp calling ib_drain_sq.
> >>>> [Thu Dec 29 14:02:03 2016] ib_drain_sq calling __ib_drain_sq.
> >>>> [Thu Dec 29 14:02:03 2016] Setting up drain callback.
> >>>> [Thu Dec 29 14:02:03 2016] Starting init_completion.
> >>>> [Thu Dec 29 14:02:03 2016] Calling ib_modify_qp.
> >>>> [Thu Dec 29 14:02:03 2016] Calling ib_post_send.
> >>>> [Thu Dec 29 14:02:03 2016] Calling wait_for_completion.
> >>>> [Thu Dec 29 14:02:03 2016] &sdrain.done->done = 0.
> >>>>
> >>>> Gets "stuck" here...
> >>>>
> >>>> [Thu Dec 29 14:02:20 2016] iSCSI Login timeout on Network Portal
> >>>> 0.0.0.0:3260
> >>>> [Thu Dec 29 14:02:37 2016] ib_drain_qp calling ib_drain_sq.
> >>>> [Thu Dec 29 14:02:37 2016] ib_drain_sq calling __ib_drain_sq.
> >>>> [Thu Dec 29 14:02:37 2016] Setting up drain callback.
> >>>> [Thu Dec 29 14:02:37 2016] Starting init_completion.
> >>>> [Thu Dec 29 14:02:37 2016] Calling ib_modify_qp.
> >>>> [Thu Dec 29 14:02:37 2016] Calling ib_post_send.
> >>>> [Thu Dec 29 14:02:37 2016] Calling wait_for_completion.
> >>>> [Thu Dec 29 14:02:37 2016] ib_drain_qp_done going to call complete.
> >>>> [Thu Dec 29 14:02:38 2016] &sdrain.done->done = 1.
> >>>> [Thu Dec 29 14:02:38 2016] Returned from wait_for_completion.
> >>>> [Thu Dec 29 14:02:38 2016] ib_drain_qp_done going to call complete.
> >>>>
> >>>> Next time ib_drain_qp is called, ib_drain_qp_done gets called...
> >>>>
> >>>> [Thu Dec 29 14:02:55 2016] ib_drain_qp calling ib_drain_sq.
> >>>> [Thu Dec 29 14:02:55 2016] ib_drain_sq calling __ib_drain_sq.
> >>>> [Thu Dec 29 14:02:55 2016] Setting up drain callback.
> >>>> [Thu Dec 29 14:02:55 2016] Starting init_completion.
> >>>> [Thu Dec 29 14:02:55 2016] Calling ib_modify_qp.
> >>>> [Thu Dec 29 14:02:55 2016] Calling ib_post_send.
> >>>> [Thu Dec 29 14:02:55 2016] Calling wait_for_completion.
> >>>> [Thu Dec 29 14:02:55 2016] ib_drain_qp_done going to call complete.
> >>>> [Thu Dec 29 14:02:55 2016] &sdrain.done->done = 1.
> >>>> [Thu Dec 29 14:02:55 2016] Returned from wait_for_completion.
> >>>> [Thu Dec 29 14:02:55 2016] ib_drain_qp_done going to call complete.
> >>>> ----------------
> >>>> Robert LeBlanc
> >>>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> >>>>
> >>>>
> >>>> On Wed, Dec 28, 2016 at 1:58 PM, Robert LeBlanc <robert@xxxxxxxxxxxxx>
> >>>> wrote:
> >>>>> Good news! I found a 10 Gb switch laying around and put it in place of
> >>>>> the Linux router. I'm getting the same failure with the switch, so it
> >>>>> is not something funky with the Linux router and easier to replicate.
> >>>>> ----------------
> >>>>> Robert LeBlanc
> >>>>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> >>>>>
> >>>>>
> >>>>> On Wed, Dec 28, 2016 at 1:39 PM, Robert LeBlanc <robert@xxxxxxxxxxxxx>
> >>>>> wrote:
> >>>>>> OK, here is some more info. This is a diagram of my current set up.
> >>>>>>
> >>>>>>                 +----------------+
> >>>>>>                 |  Linux Router  |
> >>>>>>                 |   ConnectX-3   |
> >>>>>>                 | port 1  port 2 |
> >>>>>>                 +----------------+
> >>>>>>                      /      \
> >>>>>> +---------------+   /        \   +---------------+
> >>>>>> |    Host 1     |  / A      A \  |    Host 2     |
> >>>>>> | ConnectX-4-LX | /            \ | ConnectX-4-LX |
> >>>>>> |        Port 1 |-              -| Port 1        |
> >>>>>> |        Port 2 |----------------| Port 2        |
> >>>>>> +---------------+        B       +---------------+
> >>>>>>
> >>>>>> The Linux router has the ConnectX-3 (not PRO) card in Ethernet mode
> >>>>>> and is using a breakout cable (port 1 only) to connect to the
> >>>>>> ConnectX-4-LX cards at 10 Gb as path 'A'. The second port of the
> >>>>>> ConnectX-4-LX cards are connected directly at 25 Gb as path 'B'.
> >>>>>>
> >>>>>> Running Iser and RoCE on path 'B' seems to run just fine.
> >>>>>>
> >>>>>> Running Iser and RoCE on path 'A' has issues when the Linux router is
> >>>>>> operating as a bridge or a router. Some small operations like mkfs
> >>>>>> seem to work just fine, but fio causes iser to want to log out and we
> >>>>>> get D state. I can run ib_send_bw 'all' tests through path 'A' and
> >>>>>> don't see a problem. It does seem to be load related, though. I have
> >>>>>> been trying to run
> >>>>>>
> >>>>>> echo "3" > /proc/sys/vm/drop_caches; fio --rw=read --bs=4K --size=1G
> >>>>>> --numjobs=40 --name=worker.matt --group_reporting
> >>>>>>
> >>>>>> If I reduce the number of jobs to 10 or less, it seems to work
> >>>>>> although I may see some of the debug messages I added in, it doesn't
> >>>>>> seem to completely hang and cause the logout lockup.
> >>>>>>
> >>>>>> Steps to reproduce:
> >>>>>> 1. 4.9 kernel
> >>>>>> 2. Bridge ports 1 & 2 on the Linux router
> >>>>>> 3. Configure port 1 on Host 1 & 2 on the same subnet
> >>>>>> 4. Create large ramdisk in targetcli and export from Host 1
> >>>>>> 5. Login from Host 2
> >>>>>> 6. Create EXT4 file system on imported disk
> >>>>>> 7. Mount and cd into mount
> >>>>>> 8. Run fio: echo "3" > /proc/sys/vm/drop_caches; fio --rw=read --bs=4K
> >>>>>> --size=1G --numjobs=40 --name=worker.matt --group_reporting
> >>>>>> 9. After some time, the fio process will report the file system is
> >>>>>> read only and the iscsi processes will be in D state on Host 1
> >>>>>>
> >>>>>> It does seem the problem is in iser and not specific to the generic
> >>>>>> RDMA stack.
> >>>>>>
> >>>>>> I'll keep digging and reporting back.
> >>>>>> ----------------
> >>>>>> Robert LeBlanc
> >>>>>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> >>>>>>
> >>>>>>
> >>>>>> On Tue, Dec 27, 2016 at 1:58 PM, Robert LeBlanc <robert@xxxxxxxxxxxxx>
> >>>>>> wrote:
> >>>>>>> I realized that I did not set the default RoCE mode to v2 and the
> >>>>>>> client is on a different subnet, probably why I'm seeing the -110
> >>>>>>> error. Iser should not go into D state because of this and should
> >>>>>>> handle this gracefully, but may provide an easy way to replicate the
> >>>>>>> issue.
> >>>>>>> ----------------
> >>>>>>> Robert LeBlanc
> >>>>>>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> >>>>>>>
> >>>>>>>
> >>>>>>> On Tue, Dec 27, 2016 at 1:22 PM, Robert LeBlanc
> >>>>>>> <robert@xxxxxxxxxxxxx> wrote:
> >>>>>>>> I looked at this code and it is quiet above my ability. I created
> >>>>>>>> this
> >>>>>>>> patch, but I don't know how to interrogate the queue to see how many
> >>>>>>>> items there are. If you can give me some more direction on what to
> >>>>>>>> try, I can keep fumbling around with this until someone smarter than
> >>>>>>>> me can figure it out. This is now a blocker for me so I'm going to
> >>>>>>>> beat my head on this until it is fixed.
> >>>>>>>>
> >>>>>>>> Thanks for being patient with me.
> >>>>>>>>
> >>>>>>>> diff --git a/drivers/infiniband/core/verbs.c
> >>>>>>>> b/drivers/infiniband/core/verbs.c
> >>>>>>>> index 8368764..9e5bd4b 100644
> >>>>>>>> --- a/drivers/infiniband/core/verbs.c
> >>>>>>>> +++ b/drivers/infiniband/core/verbs.c
> >>>>>>>> @@ -1954,22 +1954,27 @@ static void __ib_drain_sq(struct ib_qp *qp)
> >>>>>>>>                 return;
> >>>>>>>>         }
> >>>>>>>>
> >>>>>>>> +       printk("Setting up drain callback.");
> >>>>>>>>         swr.wr_cqe = &sdrain.cqe;
> >>>>>>>>         sdrain.cqe.done = ib_drain_qp_done;
> >>>>>>>> +       printk("Starting init_completion.");
> >>>>>>>>         init_completion(&sdrain.done);
> >>>>>>>>
> >>>>>>>> +       printk("Calling ib_modify_qp.");
> >>>>>>>>         ret = ib_modify_qp(qp, &attr, IB_QP_STATE);
> >>>>>>>>         if (ret) {
> >>>>>>>>                 WARN_ONCE(ret, "failed to drain send queue: %d\n",
> >>>>>>>>                 ret);
> >>>>>>>>                 return;
> >>>>>>>>         }
> >>>>>>>>
> >>>>>>>> +       printk("Calling ib_post_send.");
> >>>>>>>>         ret = ib_post_send(qp, &swr, &bad_swr);
> >>>>>>>>         if (ret) {
> >>>>>>>>                 WARN_ONCE(ret, "failed to drain send queue: %d\n",
> >>>>>>>>                 ret);
> >>>>>>>>                 return;
> >>>>>>>>         }
> >>>>>>>>
> >>>>>>>> +       printk("Starting wait_for_completion.");
> >>>>>>>>         wait_for_completion(&sdrain.done);
> >>>>>>>>  }
> >>>>>>>>
> >>>>>>>> I get the same processes in D state (and same backtrace) and this is
> >>>>>>>> what shows up in dmesg:
> >>>>>>>>
> >>>>>>>> [  920.317401] isert: isert_rdma_accept: rdma_accept() failed with:
> >>>>>>>> -110
> >>>>>>>> [  920.325554] ------------[ cut here ]------------
> >>>>>>>> [  920.330188] WARNING: CPU: 11 PID: 705 at
> >>>>>>>> drivers/infiniband/core/verbs.c:303 ib_dealloc_pd+0x58/0xa0
> >>>>>>>> [ib_core]
> >>>>>>>> [  920.340210] Modules linked in: target_core_user target_core_pscsi
> >>>>>>>> target_core_file target_core_iblock 8021q garp mrp rpcrdma sunrpc
> >>>>>>>> ib_isert ib_iser ib_srpt ib_srp scsi_transport_srp ib_ipoib
> >>>>>>>> iptable_filter rdma_ucm i
> >>>>>>>> b_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm sb_edac edac_core
> >>>>>>>> x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm ext4
> >>>>>>>> ipmi_devintf irqbypass crct10dif_pclmul crc32_pclmul
> >>>>>>>> ghash_clmulni_intel aesni_intel jbd2 lr
> >>>>>>>> w gf128mul glue_helper mbcache iTCO_wdt ablk_helper mei_me
> >>>>>>>> iTCO_vendor_support cryptd joydev sg mei i2c_i801 lpc_ich pcspkr
> >>>>>>>> mfd_core ioatdma shpchp i2c_smbus ipmi_si wmi ipmi_msghandler
> >>>>>>>> acpi_power_meter acpi_pad ip_table
> >>>>>>>> s xfs libcrc32c raid1 mlx4_en mlx4_ib mlx5_ib sd_mod ib_core ast
> >>>>>>>> drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops ttm
> >>>>>>>> mlx5_core igb mlx4_core
> >>>>>>>> [  920.412347]  ahci ptp drm libahci pps_core libata dca
> >>>>>>>> i2c_algo_bit
> >>>>>>>> be2iscsi bnx2i cnic uio qla4xxx iscsi_boot_sysfs
> >>>>>>>> [  920.421744] CPU: 11 PID: 705 Comm: kworker/11:2 Not tainted
> >>>>>>>> 4.9.0+ #3
> >>>>>>>> [  920.428199] Hardware name: Supermicro
> >>>>>>>> SYS-6028TP-HTFR/X10DRT-PIBF,
> >>>>>>>> BIOS 1.1 08/03/2015
> >>>>>>>> [  920.436126] Workqueue: ib_cm cm_work_handler [ib_cm]
> >>>>>>>> [  920.441113]  ffffc90032a03a40 ffffffff8134d45f 0000000000000000
> >>>>>>>> 0000000000000000
> >>>>>>>> [  920.448583]  ffffc90032a03a80 ffffffff81083371 0000012fa04a1c4a
> >>>>>>>> ffff883f5e886e80
> >>>>>>>> [  920.456073]  ffff887f1eaa4400 ffff887f1eaa5800 ffffc90032a03b08
> >>>>>>>> 00000000ffffff92
> >>>>>>>> [  920.463535] Call Trace:
> >>>>>>>> [  920.465993]  [<ffffffff8134d45f>] dump_stack+0x63/0x84
> >>>>>>>> [  920.471144]  [<ffffffff81083371>] __warn+0xd1/0xf0
> >>>>>>>> [  920.475941]  [<ffffffff810834ad>] warn_slowpath_null+0x1d/0x20
> >>>>>>>> [  920.481790]  [<ffffffffa026cf58>] ib_dealloc_pd+0x58/0xa0
> >>>>>>>> [ib_core]
> >>>>>>>> [  920.488072]  [<ffffffffa0695000>] isert_device_put+0x50/0xc0
> >>>>>>>> [ib_isert]
> >>>>>>>> [  920.494693]  [<ffffffffa069838e>]
> >>>>>>>> isert_connect_request+0x68e/0xd40
> >>>>>>>> [ib_isert]
> >>>>>>>> [  920.501924]  [<ffffffffa0699683>] isert_cma_handler+0xe3/0x3b0
> >>>>>>>> [ib_isert]
> >>>>>>>> [  920.508725]  [<ffffffffa042c5d6>] ? cma_new_conn_id+0x276/0x4b0
> >>>>>>>> [rdma_cm]
> >>>>>>>> [  920.515521]  [<ffffffffa0427050>] cma_listen_handler+0x20/0x30
> >>>>>>>> [rdma_cm]
> >>>>>>>> [  920.522227]  [<ffffffffa042ca05>] cma_req_handler+0x1f5/0x4c0
> >>>>>>>> [rdma_cm]
> >>>>>>>> [  920.528851]  [<ffffffffa03fb0f5>] cm_process_work+0x25/0xf0
> >>>>>>>> [ib_cm]
> >>>>>>>> [  920.535125]  [<ffffffffa03fba94>] cm_req_handler+0x8d4/0xc70
> >>>>>>>> [ib_cm]
> >>>>>>>> [  920.541485]  [<ffffffffa03fc1ce>] cm_work_handler+0x1ce/0x1648
> >>>>>>>> [ib_cm]
> >>>>>>>> [  920.548021]  [<ffffffff8109cc02>] process_one_work+0x152/0x400
> >>>>>>>> [  920.553861]  [<ffffffff8109d4f5>] worker_thread+0x125/0x4b0
> >>>>>>>> [  920.559443]  [<ffffffff8109d3d0>] ? rescuer_thread+0x380/0x380
> >>>>>>>> [  920.565284]  [<ffffffff810a3059>] kthread+0xd9/0xf0
> >>>>>>>> [  920.570178]  [<ffffffff810a2f80>] ? kthread_park+0x60/0x60
> >>>>>>>> [  920.576389]  [<ffffffff817732d5>] ret_from_fork+0x25/0x30
> >>>>>>>> [  920.582473] ---[ end trace 1f5a1831f9d2d964 ]---
> >>>>>>>> [  920.587907] ------------[ cut here ]------------
> >>>>>>>> [  920.593213] WARNING: CPU: 11 PID: 705 at
> >>>>>>>> drivers/infiniband/core/cq.c:189 ib_free_cq+0x97/0xc0 [ib_core]
> >>>>>>>> [  920.603383] Modules linked in: target_core_user target_core_pscsi
> >>>>>>>> target_core_file target_core_iblock 8021q garp mrp rpcrdma sunrpc
> >>>>>>>> ib_isert ib_iser ib_srpt ib_srp scsi_transport_srp ib_ipoib
> >>>>>>>> iptable_filter rdma_ucm i
> >>>>>>>> b_ucm ib_uverbs ib_umad rdma_cm ib_cm iw_cm sb_edac edac_core
> >>>>>>>> x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm ext4
> >>>>>>>> ipmi_devintf irqbypass crct10dif_pclmul crc32_pclmul
> >>>>>>>> ghash_clmulni_intel aesni_intel jbd2 lr
> >>>>>>>> w gf128mul glue_helper mbcache iTCO_wdt ablk_helper mei_me
> >>>>>>>> iTCO_vendor_support cryptd joydev sg mei i2c_i801 lpc_ich pcspkr
> >>>>>>>> mfd_core ioatdma shpchp i2c_smbus ipmi_si wmi ipmi_msghandler
> >>>>>>>> acpi_power_meter acpi_pad ip_table
> >>>>>>>> s xfs libcrc32c raid1 mlx4_en mlx4_ib mlx5_ib sd_mod ib_core ast
> >>>>>>>> drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops ttm
> >>>>>>>> mlx5_core igb mlx4_core
> >>>>>>>> [  920.679694]  ahci ptp drm libahci pps_core libata dca
> >>>>>>>> i2c_algo_bit
> >>>>>>>> be2iscsi bnx2i cnic uio qla4xxx iscsi_boot_sysfs
> >>>>>>>> [  920.690579] CPU: 11 PID: 705 Comm: kworker/11:2 Tainted: G
> >>>>>>>> W
> >>>>>>>>       4.9.0+ #3
> >>>>>>>> [  920.699008] Hardware name: Supermicro
> >>>>>>>> SYS-6028TP-HTFR/X10DRT-PIBF,
> >>>>>>>> BIOS 1.1 08/03/2015
> >>>>>>>> [  920.707701] Workqueue: ib_cm cm_work_handler [ib_cm]
> >>>>>>>> [  920.713438]  ffffc90032a03a18 ffffffff8134d45f 0000000000000000
> >>>>>>>> 0000000000000000
> >>>>>>>> [  920.721648]  ffffc90032a03a58 ffffffff81083371 000000bd5e886e80
> >>>>>>>> ffff887f1eaa6800
> >>>>>>>> [  920.729850]  ffff883f5e886e20 ffff883f5e886e18 ffffc90032a03b08
> >>>>>>>> 00000000ffffff92
> >>>>>>>> [  920.738026] Call Trace:
> >>>>>>>> [  920.741188]  [<ffffffff8134d45f>] dump_stack+0x63/0x84
> >>>>>>>> [  920.747027]  [<ffffffff81083371>] __warn+0xd1/0xf0
> >>>>>>>> [  920.752488]  [<ffffffff810834ad>] warn_slowpath_null+0x1d/0x20
> >>>>>>>> [  920.758989]  [<ffffffffa026e037>] ib_free_cq+0x97/0xc0 [ib_core]
> >>>>>>>> [  920.765649]  [<ffffffffa0694f88>]
> >>>>>>>> isert_free_comps.isra.26+0x38/0x60 [ib_isert]
> >>>>>>>> [  920.773609]  [<ffffffffa069500d>] isert_device_put+0x5d/0xc0
> >>>>>>>> [ib_isert]
> >>>>>>>> [  920.780868]  [<ffffffffa069838e>]
> >>>>>>>> isert_connect_request+0x68e/0xd40
> >>>>>>>> [ib_isert]
> >>>>>>>> [  920.788734]  [<ffffffffa0699683>] isert_cma_handler+0xe3/0x3b0
> >>>>>>>> [ib_isert]
> >>>>>>>> [  920.796157]  [<ffffffffa042c5d6>] ? cma_new_conn_id+0x276/0x4b0
> >>>>>>>> [rdma_cm]
> >>>>>>>> [  920.803586]  [<ffffffffa0427050>] cma_listen_handler+0x20/0x30
> >>>>>>>> [rdma_cm]
> >>>>>>>> [  920.810916]  [<ffffffffa042ca05>] cma_req_handler+0x1f5/0x4c0
> >>>>>>>> [rdma_cm]
> >>>>>>>> [  920.818167]  [<ffffffffa03fb0f5>] cm_process_work+0x25/0xf0
> >>>>>>>> [ib_cm]
> >>>>>>>> [  920.825063]  [<ffffffffa03fba94>] cm_req_handler+0x8d4/0xc70
> >>>>>>>> [ib_cm]
> >>>>>>>> [  920.832051]  [<ffffffffa03fc1ce>] cm_work_handler+0x1ce/0x1648
> >>>>>>>> [ib_cm]
> >>>>>>>> [  920.839208]  [<ffffffff8109cc02>] process_one_work+0x152/0x400
> >>>>>>>> [  920.845669]  [<ffffffff8109d4f5>] worker_thread+0x125/0x4b0
> >>>>>>>> [  920.851880]  [<ffffffff8109d3d0>] ? rescuer_thread+0x380/0x380
> >>>>>>>> [  920.858352]  [<ffffffff810a3059>] kthread+0xd9/0xf0
> >>>>>>>> [  920.863857]  [<ffffffff810a2f80>] ? kthread_park+0x60/0x60
> >>>>>>>> [  920.869975]  [<ffffffff817732d5>] ret_from_fork+0x25/0x30
> >>>>>>>> [  920.876006] ---[ end trace 1f5a1831f9d2d965 ]---
> >>>>>>>> [  920.884335] isert: isert_cma_handler: failed handle connect
> >>>>>>>> request -110
> >>>>>>>> [ 1639.592451] Setting up drain callback.
> >>>>>>>> [ 1639.596073] Starting init_completion.
> >>>>>>>> [ 1639.600683] Calling ib_modify_qp.
> >>>>>>>> [ 1639.602616] Calling ib_post_send.
> >>>>>>>> [ 1639.606550] Starting wait_for_completion.
> >>>>>>>> [ 1656.976015] iSCSI Login timeout on Network Portal 0.0.0.0:3260
> >>>>>>>> [ 1674.254027] Setting up drain callback.
> >>>>>>>> [ 1674.257634] Starting init_completion.
> >>>>>>>> [ 1674.262107] Calling ib_modify_qp.
> >>>>>>>> [ 1674.264011] Calling ib_post_send.
> >>>>>>>> [ 1674.267969] Starting wait_for_completion.
> >>>>>>>> [ 1691.583888] Setting up drain callback.
> >>>>>>>> [ 1691.588490] Starting init_completion.
> >>>>>>>> [ 1691.590677] Calling ib_modify_qp.
> >>>>>>>> [ 1691.594766] Calling ib_post_send.
> >>>>>>>> [ 1691.596607] Starting wait_for_completion.
> >>>>>>>> [ 1708.913356] Setting up drain callback.
> >>>>>>>> [ 1708.915658] Starting init_completion.
> >>>>>>>> [ 1708.920152] Calling ib_modify_qp.
> >>>>>>>> [ 1708.922041] Calling ib_post_send.
> >>>>>>>> [ 1708.926048] Starting wait_for_completion.
> >>>>>>>> [ 1726.244365] Setting up drain callback.
> >>>>>>>> [ 1726.248973] Starting init_completion.
> >>>>>>>> [ 1726.251165] Calling ib_modify_qp.
> >>>>>>>> [ 1726.255189] Calling ib_post_send.
> >>>>>>>> [ 1726.257031] Starting wait_for_completion.
> >>>>>>>> [ 1743.574751] Setting up drain callback.
> >>>>>>>> [ 1743.577044] Starting init_completion.
> >>>>>>>> [ 1743.581496] Calling ib_modify_qp.
> >>>>>>>> [ 1743.583404] Calling ib_post_send.
> >>>>>>>> [ 1743.587346] Starting wait_for_completion.
> >>>>>>>> [ 1760.904470] Setting up drain callback.
> >>>>>>>> [ 1760.908991] Starting init_completion.
> >>>>>>>> [ 1760.911206] Calling ib_modify_qp.
> >>>>>>>> [ 1760.915214] Calling ib_post_send.
> >>>>>>>> [ 1760.917062] Starting wait_for_completion.
> >>>>>>>> [ 1778.230821] Setting up drain callback.
> >>>>>>>> [ 1778.233116] Starting init_completion.
> >>>>>>>> [ 1778.237510] Calling ib_modify_qp.
> >>>>>>>> [ 1778.239413] Calling ib_post_send.
> >>>>>>>> .... [keeps repeating]
> >>>>>>>> ----------------
> >>>>>>>> Robert LeBlanc
> >>>>>>>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> On Thu, Dec 22, 2016 at 12:15 PM, Doug Ledford <dledford@xxxxxxxxxx>
> >>>>>>>> wrote:
> >>>>>>>>> On 12/21/2016 6:39 PM, Robert LeBlanc wrote:
> >>>>>>>>>> I hit a new backtrace today, hopefully it adds something.
> >>>>>>>>>>
> >>>>>>>>>> # cat /proc/19659/stack
> >>>>>>>>>> [<ffffffff815304d1>] iscsit_stop_session+0x1b1/0x1c0
> >>>>>>>>>> [<ffffffff81521c62>]
> >>>>>>>>>> iscsi_check_for_session_reinstatement+0x1e2/0x270
> >>>>>>>>>> [<ffffffff81524660>]
> >>>>>>>>>> iscsi_target_check_for_existing_instances+0x30/0x40
> >>>>>>>>>> [<ffffffff815247a8>] iscsi_target_do_login+0x138/0x630
> >>>>>>>>>> [<ffffffff815259be>] iscsi_target_start_negotiation+0x4e/0xa0
> >>>>>>>>>> [<ffffffff8152355e>] __iscsi_target_login_thread+0x83e/0xf20
> >>>>>>>>>> [<ffffffff81523c64>] iscsi_target_login_thread+0x24/0x30
> >>>>>>>>>> [<ffffffff810a3059>] kthread+0xd9/0xf0
> >>>>>>>>>> [<ffffffff817732d5>] ret_from_fork+0x25/0x30
> >>>>>>>>>> [<ffffffffffffffff>] 0xffffffffffffffff
> >>>>>>>>>>
> >>>>>>>>>> # cat /proc/21342/stack
> >>>>>>>>>> [<ffffffffa0292b10>] __ib_drain_sq+0x190/0x1c0 [ib_core]
> >>>>>>>>>> [<ffffffffa0292b65>] ib_drain_sq+0x25/0x30 [ib_core]
> >>>>>>>>>> [<ffffffffa0292d72>] ib_drain_qp+0x12/0x30 [ib_core]
> >>>>>>>>>> [<ffffffffa062c5ff>] isert_wait_conn+0x5f/0x2d0 [ib_isert]
> >>>>>>>>>> [<ffffffff815309b7>] iscsit_close_connection+0x157/0x860
> >>>>>>>>>> [<ffffffff8151f10b>]
> >>>>>>>>>> iscsit_take_action_for_connection_exit+0x7b/0xf0
> >>>>>>>>>> [<ffffffff81530265>] iscsi_target_rx_thread+0x95/0xa0
> >>>>>>>>>> [<ffffffff810a3059>] kthread+0xd9/0xf0
> >>>>>>>>>> [<ffffffff817732d5>] ret_from_fork+0x25/0x30
> >>>>>>>>>> [<ffffffffffffffff>] 0xffffffffffffffff
> >>>>>>>>>>
> >>>>>>>>>> # ps aux | grep iscsi | grep D
> >>>>>>>>>> root     19659  0.0  0.0      0     0 ?        D    16:12   0:00
> >>>>>>>>>> [iscsi_np]
> >>>>>>>>>> root     21342  0.0  0.0      0     0 ?        D    16:29   0:00
> >>>>>>>>>> [iscsi_trx]
> >>>>>>>>>> ----------------
> >>>>>>>>>> Robert LeBlanc
> >>>>>>>>>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
> >>>>>>>>>
> >>>>>>>>> That looks suspiciously like the __ib_drain_sq is stuck forever
> >>>>>>>>> waiting
> >>>>>>>>> on a completion that never comes.
> >>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> On Thu, Dec 15, 2016 at 1:38 PM, Robert LeBlanc
> >>>>>>>>>> <robert@xxxxxxxxxxxxx> wrote:
> >>>>>>>>>>> Nicholas,
> >>>>>>>>>>>
> >>>>>>>>>>> I've found that the kernels I used were not able to be inspected
> >>>>>>>>>>> using
> >>>>>>>>>>> crash and I could not build the debug info for them. So I built a
> >>>>>>>>>>> new
> >>>>>>>>>>> 4.9 kernel and verified that I could inspect the crash. It is
> >>>>>>>>>>> located
> >>>>>>>>>>> at [1].
> >>>>>>>>>>>
> >>>>>>>>>>> [1] http://mirrors.betterservers.com/trace/crash2.tar.xz
> >>>>>>>>>>> ----------------
> >>>>>>>>>>> Robert LeBlanc
> >>>>>>>>>>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62
> >>>>>>>>>>> B9F1
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> On Mon, Dec 12, 2016 at 4:57 PM, Robert LeBlanc
> >>>>>>>>>>> <robert@xxxxxxxxxxxxx> wrote:
> >>>>>>>>>>>> Nicholas,
> >>>>>>>>>>>>
> >>>>>>>>>>>> After lots of set backs and having to give up trying to get
> >>>>>>>>>>>> kernel
> >>>>>>>>>>>> dumps on our "production" systems, I've been able to work out
> >>>>>>>>>>>> the
> >>>>>>>>>>>> issues we had with kdump and replicate the issue on my dev
> >>>>>>>>>>>> boxes. I
> >>>>>>>>>>>> have dumps from 4.4.30 and 4.9-rc8 (makedumpfile would not dump,
> >>>>>>>>>>>> so it
> >>>>>>>>>>>> is a straight copy of /proc/vmcore from the crash kernel). In
> >>>>>>>>>>>> each
> >>>>>>>>>>>> crash directory, I put a details.txt file that has the process
> >>>>>>>>>>>> IDs
> >>>>>>>>>>>> that were having problems and a brief description of the set-up
> >>>>>>>>>>>> at the
> >>>>>>>>>>>> time. This was mostly replicated by starting fio and pulling the
> >>>>>>>>>>>> Infiniband cable until fio gave up. This hardware also has
> >>>>>>>>>>>> Mellanox
> >>>>>>>>>>>> ConnectX4-LX cards and I also replicated the issue over RoCE
> >>>>>>>>>>>> using 4.9
> >>>>>>>>>>>> since it has the drivers in-box. Please let me know if you need
> >>>>>>>>>>>> more
> >>>>>>>>>>>> info, I can test much faster now. The cores/kernels/modules are
> >>>>>>>>>>>> located at [1].
> >>>>>>>>>>>>
> >>>>>>>>>>>> [1] http://mirrors.betterservers.com/trace/crash.tar.xz
> >>>>>>>>>>>>
> >>>>>>>>>>>> Thanks,
> >>>>>>>>>>>> Robert
> >>>>>>>>>>>> ----------------
> >>>>>>>>>>>> Robert LeBlanc
> >>>>>>>>>>>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62
> >>>>>>>>>>>> B9F1
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> On Fri, Nov 4, 2016 at 3:57 PM, Robert LeBlanc
> >>>>>>>>>>>> <robert@xxxxxxxxxxxxx> wrote:
> >>>>>>>>>>>>> We hit this yesterday, this time it was on the tx thread (the
> >>>>>>>>>>>>> other
> >>>>>>>>>>>>> ones before seem to be on the rx thread). We weren't able to
> >>>>>>>>>>>>> get a
> >>>>>>>>>>>>> kernel dump on this. We'll try to get one next time.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> # ps axuw | grep "D.*iscs[i]"
> >>>>>>>>>>>>> root     12383  0.0  0.0      0     0 ?        D    Nov03
> >>>>>>>>>>>>> 0:04 [iscsi_np]
> >>>>>>>>>>>>> root     23016  0.0  0.0      0     0 ?        D    Nov03
> >>>>>>>>>>>>> 0:00 [iscsi_ttx]
> >>>>>>>>>>>>> root     23018  0.0  0.0      0     0 ?        D    Nov03
> >>>>>>>>>>>>> 0:00 [iscsi_ttx]
> >>>>>>>>>>>>> # cat /proc/12383/stack
> >>>>>>>>>>>>> [<ffffffff814f24af>] iscsit_stop_session+0x19f/0x1d0
> >>>>>>>>>>>>> [<ffffffff814e3c66>]
> >>>>>>>>>>>>> iscsi_check_for_session_reinstatement+0x1e6/0x270
> >>>>>>>>>>>>> [<ffffffff814e6620>]
> >>>>>>>>>>>>> iscsi_target_check_for_existing_instances+0x30/0x40
> >>>>>>>>>>>>> [<ffffffff814e6770>] iscsi_target_do_login+0x140/0x640
> >>>>>>>>>>>>> [<ffffffff814e7b0c>] iscsi_target_start_negotiation+0x1c/0xb0
> >>>>>>>>>>>>> [<ffffffff814e585b>] iscsi_target_login_thread+0xa9b/0xfc0
> >>>>>>>>>>>>> [<ffffffff8109d7c8>] kthread+0xd8/0xf0
> >>>>>>>>>>>>> [<ffffffff81721a8f>] ret_from_fork+0x3f/0x70
> >>>>>>>>>>>>> [<ffffffffffffffff>] 0xffffffffffffffff
> >>>>>>>>>>>>> # cat /proc/23016/stack
> >>>>>>>>>>>>> [<ffffffff814ce0d9>] target_wait_for_sess_cmds+0x49/0x1a0
> >>>>>>>>>>>>> [<ffffffffa058b92b>] isert_wait_conn+0x1ab/0x2f0 [ib_isert]
> >>>>>>>>>>>>> [<ffffffff814f2642>] iscsit_close_connection+0x162/0x870
> >>>>>>>>>>>>> [<ffffffff814e110f>]
> >>>>>>>>>>>>> iscsit_take_action_for_connection_exit+0x7f/0x100
> >>>>>>>>>>>>> [<ffffffff814f122a>] iscsi_target_tx_thread+0x1aa/0x1d0
> >>>>>>>>>>>>> [<ffffffff8109d7c8>] kthread+0xd8/0xf0
> >>>>>>>>>>>>> [<ffffffff81721a8f>] ret_from_fork+0x3f/0x70
> >>>>>>>>>>>>> [<ffffffffffffffff>] 0xffffffffffffffff
> >>>>>>>>>>>>> # cat /proc/23018/stack
> >>>>>>>>>>>>> [<ffffffff814ce0d9>] target_wait_for_sess_cmds+0x49/0x1a0
> >>>>>>>>>>>>> [<ffffffffa058b92b>] isert_wait_conn+0x1ab/0x2f0 [ib_isert]
> >>>>>>>>>>>>> [<ffffffff814f2642>] iscsit_close_connection+0x162/0x870
> >>>>>>>>>>>>> [<ffffffff814e110f>]
> >>>>>>>>>>>>> iscsit_take_action_for_connection_exit+0x7f/0x100
> >>>>>>>>>>>>> [<ffffffff814f122a>] iscsi_target_tx_thread+0x1aa/0x1d0
> >>>>>>>>>>>>> [<ffffffff8109d7c8>] kthread+0xd8/0xf0
> >>>>>>>>>>>>> [<ffffffff81721a8f>] ret_from_fork+0x3f/0x70
> >>>>>>>>>>>>> [<ffffffffffffffff>] 0xffffffffffffffff
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> From dmesg:
> >>>>>>>>>>>>> [  394.476332] INFO: rcu_sched self-detected stall on CPU
> >>>>>>>>>>>>> [  394.476334]  20-...: (23976 ticks this GP)
> >>>>>>>>>>>>> idle=edd/140000000000001/0 softirq=292/292 fqs=18788
> >>>>>>>>>>>>> [  394.476336]   (t=24003 jiffies g=3146 c=3145 q=0)
> >>>>>>>>>>>>> [  394.476337] Task dump for CPU 20:
> >>>>>>>>>>>>> [  394.476338] kworker/u68:2   R  running task        0 12906
> >>>>>>>>>>>>> 2 0x00000008
> >>>>>>>>>>>>> [  394.476345] Workqueue: isert_comp_wq isert_cq_work
> >>>>>>>>>>>>> [ib_isert]
> >>>>>>>>>>>>> [  394.476346]  ffff883f2fe38000 00000000f805705e
> >>>>>>>>>>>>> ffff883f7fd03da8
> >>>>>>>>>>>>> ffffffff810ac8ff
> >>>>>>>>>>>>> [  394.476347]  0000000000000014 ffffffff81adb680
> >>>>>>>>>>>>> ffff883f7fd03dc0
> >>>>>>>>>>>>> ffffffff810af239
> >>>>>>>>>>>>> [  394.476348]  0000000000000015 ffff883f7fd03df0
> >>>>>>>>>>>>> ffffffff810e1cd0
> >>>>>>>>>>>>> ffff883f7fd17b80
> >>>>>>>>>>>>> [  394.476348] Call Trace:
> >>>>>>>>>>>>> [  394.476354]  <IRQ>  [<ffffffff810ac8ff>]
> >>>>>>>>>>>>> sched_show_task+0xaf/0x110
> >>>>>>>>>>>>> [  394.476355]  [<ffffffff810af239>] dump_cpu_task+0x39/0x40
> >>>>>>>>>>>>> [  394.476357]  [<ffffffff810e1cd0>]
> >>>>>>>>>>>>> rcu_dump_cpu_stacks+0x80/0xb0
> >>>>>>>>>>>>> [  394.476359]  [<ffffffff810e6100>]
> >>>>>>>>>>>>> rcu_check_callbacks+0x540/0x820
> >>>>>>>>>>>>> [  394.476360]  [<ffffffff810afe11>] ?
> >>>>>>>>>>>>> account_system_time+0x81/0x110
> >>>>>>>>>>>>> [  394.476363]  [<ffffffff810faa60>] ?
> >>>>>>>>>>>>> tick_sched_do_timer+0x50/0x50
> >>>>>>>>>>>>> [  394.476364]  [<ffffffff810eb599>]
> >>>>>>>>>>>>> update_process_times+0x39/0x60
> >>>>>>>>>>>>> [  394.476365]  [<ffffffff810fa815>]
> >>>>>>>>>>>>> tick_sched_handle.isra.17+0x25/0x60
> >>>>>>>>>>>>> [  394.476366]  [<ffffffff810faa9d>] tick_sched_timer+0x3d/0x70
> >>>>>>>>>>>>> [  394.476368]  [<ffffffff810ec182>]
> >>>>>>>>>>>>> __hrtimer_run_queues+0x102/0x290
> >>>>>>>>>>>>> [  394.476369]  [<ffffffff810ec668>]
> >>>>>>>>>>>>> hrtimer_interrupt+0xa8/0x1a0
> >>>>>>>>>>>>> [  394.476372]  [<ffffffff81052c65>]
> >>>>>>>>>>>>> local_apic_timer_interrupt+0x35/0x60
> >>>>>>>>>>>>> [  394.476374]  [<ffffffff8172423d>]
> >>>>>>>>>>>>> smp_apic_timer_interrupt+0x3d/0x50
> >>>>>>>>>>>>> [  394.476376]  [<ffffffff817224f7>]
> >>>>>>>>>>>>> apic_timer_interrupt+0x87/0x90
> >>>>>>>>>>>>> [  394.476379]  <EOI>  [<ffffffff810d71be>] ?
> >>>>>>>>>>>>> console_unlock+0x41e/0x4e0
> >>>>>>>>>>>>> [  394.476380]  [<ffffffff810d757c>] vprintk_emit+0x2fc/0x500
> >>>>>>>>>>>>> [  394.476382]  [<ffffffff810d78ff>] vprintk_default+0x1f/0x30
> >>>>>>>>>>>>> [  394.476384]  [<ffffffff81174dde>] printk+0x5d/0x74
> >>>>>>>>>>>>> [  394.476388]  [<ffffffff814bce21>]
> >>>>>>>>>>>>> transport_lookup_cmd_lun+0x1d1/0x200
> >>>>>>>>>>>>> [  394.476390]  [<ffffffff814ee8c0>]
> >>>>>>>>>>>>> iscsit_setup_scsi_cmd+0x230/0x540
> >>>>>>>>>>>>> [  394.476392]  [<ffffffffa058dbf3>]
> >>>>>>>>>>>>> isert_rx_do_work+0x3f3/0x7f0 [ib_isert]
> >>>>>>>>>>>>> [  394.476394]  [<ffffffffa058e174>] isert_cq_work+0x184/0x770
> >>>>>>>>>>>>> [ib_isert]
> >>>>>>>>>>>>> [  394.476396]  [<ffffffff8109740f>]
> >>>>>>>>>>>>> process_one_work+0x14f/0x400
> >>>>>>>>>>>>> [  394.476397]  [<ffffffff81097c84>] worker_thread+0x114/0x470
> >>>>>>>>>>>>> [  394.476398]  [<ffffffff8171d32a>] ? __schedule+0x34a/0x7f0
> >>>>>>>>>>>>> [  394.476399]  [<ffffffff81097b70>] ?
> >>>>>>>>>>>>> rescuer_thread+0x310/0x310
> >>>>>>>>>>>>> [  394.476400]  [<ffffffff8109d7c8>] kthread+0xd8/0xf0
> >>>>>>>>>>>>> [  394.476402]  [<ffffffff8109d6f0>] ? kthread_park+0x60/0x60
> >>>>>>>>>>>>> [  394.476403]  [<ffffffff81721a8f>] ret_from_fork+0x3f/0x70
> >>>>>>>>>>>>> [  394.476404]  [<ffffffff8109d6f0>] ? kthread_park+0x60/0x60
> >>>>>>>>>>>>> [  405.716632] Unexpected ret: -104 send data 360
> >>>>>>>>>>>>> [  405.721711] tx_data returned -32, expecting 360.
> >>>>>>>>>>>>> ----------------
> >>>>>>>>>>>>> Robert LeBlanc
> >>>>>>>>>>>>> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62
> >>>>>>>>>>>>> B9F1
> >>>>>>>>>
> >>>>>>>>> When you combine this trace with the newest one, it really makes me
> >>>>>>>>> thing there is something of a bad interaction between the new drain
> >>>>>>>>> cq
> >>>>>>>>> API and the iser/isert implementation to use said API.  Sagi,
> >>>>>>>>> Christoph?
> >>>>>>>>>
> >>>>>>>>> --
> >>>>>>>>> Doug Ledford <dledford@xxxxxxxxxx>
> >>>>>>>>>     GPG Key ID: B826A3330E572FDD
> >>>>>>>>>     Key fingerprint = AE6B 1BDA 122B 23B4 265B  1274 B826 A333 0E57
> >>>>>>>>>     2FDD
> >>>>>>>>>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-scsi" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 

Hello Robert

I am going to have some time next week to set this up and see if I can help troubleshoot at all
Can you share a summary (direct email to me) of your latest tested configuration and module parameters in use that is still failing.
I looked back in the notes, can you confirm that with back-to-back direct connect the issue does not happen.

I have a setup with 2 MLX4 cards, so the intention is to use 1 as the server and 1 as the client. 
I have a FDR switch is will use between them.

Thanks
Laurence
--
To unsubscribe from this list: send the line "unsubscribe linux-scsi" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Index of Archives]     [SCSI Target Devel]     [Linux SCSI Target Infrastructure]     [Kernel Newbies]     [IDE]     [Security]     [Git]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Linux ATA RAID]     [Linux IIO]     [Samba]     [Device Mapper]

  Powered by Linux