Re: [PATCH for-rc] rdma-core/mad: Improve handling of timed out WRs of mad agent

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On Mon, Jul 29, 2024 at 12:27 PM Leon Romanovsky <leon@xxxxxxxxxx> wrote:
>
> On Mon, Jul 29, 2024 at 12:08:52PM +0530, Saravanan Vajravel wrote:
> > On Mon, Jul 29, 2024 at 11:40 AM Leon Romanovsky <leon@xxxxxxxxxx> wrote:
> > >
> > > On Mon, Jul 29, 2024 at 09:27:35AM +0530, Saravanan Vajravel wrote:
> > > > On Sun, Jul 28, 2024 at 1:08 PM Leon Romanovsky <leon@xxxxxxxxxx> wrote:
> > > > >
> > > > > On Mon, Jul 22, 2024 at 04:33:25PM +0530, Saravanan Vajravel wrote:
> > > > > > Current timeout handler of mad agent aquires/releases mad_agent_priv
> > > > > > lock for every timed out WRs. This causes heavy locking contention
> > > > > > when higher no. of WRs are to be handled inside timeout handler.
> > > > > >
> > > > > > This leads to softlockup with below trace in some use cases where
> > > > > > rdma-cm path is used to establish connection between peer nodes
> > > > > >
> > > > > > Trace:
> > > > > > -----
> > > > > >  BUG: soft lockup - CPU#4 stuck for 26s! [kworker/u128:3:19767]
> > > > > >  CPU: 4 PID: 19767 Comm: kworker/u128:3 Kdump: loaded Tainted: G OE
> > > > > >      -------  ---  5.14.0-427.13.1.el9_4.x86_64 #1
> > > > > >  Hardware name: Dell Inc. PowerEdge R740/01YM03, BIOS 2.4.8 11/26/2019
> > > > > >  Workqueue: ib_mad1 timeout_sends [ib_core]
> > > > > >  RIP: 0010:__do_softirq+0x78/0x2ac
> > > > > >  RSP: 0018:ffffb253449e4f98 EFLAGS: 00000246
> > > > > >  RAX: 00000000ffffffff RBX: 0000000000000000 RCX: 000000000000001f
> > > > > >  RDX: 000000000000001d RSI: 000000003d1879ab RDI: fff363b66fd3a86b
> > > > > >  RBP: ffffb253604cbcd8 R08: 0000009065635f3b R09: 0000000000000000
> > > > > >  R10: 0000000000000040 R11: ffffb253449e4ff8 R12: 0000000000000000
> > > > > >  R13: 0000000000000000 R14: 0000000000000000 R15: 0000000000000040
> > > > > >  FS:  0000000000000000(0000) GS:ffff8caa1fc80000(0000) knlGS:0000000000000000
> > > > > >  CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> > > > > >  CR2: 00007fd9ec9db900 CR3: 0000000891934006 CR4: 00000000007706e0
> > > > > >  DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
> > > > > >  DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
> > > > > >  PKRU: 55555554
> > > > > >  Call Trace:
> > > > > >   <IRQ>
> > > > > >   ? show_trace_log_lvl+0x1c4/0x2df
> > > > > >   ? show_trace_log_lvl+0x1c4/0x2df
> > > > > >   ? __irq_exit_rcu+0xa1/0xc0
> > > > > >   ? watchdog_timer_fn+0x1b2/0x210
> > > > > >   ? __pfx_watchdog_timer_fn+0x10/0x10
> > > > > >   ? __hrtimer_run_queues+0x127/0x2c0
> > > > > >   ? hrtimer_interrupt+0xfc/0x210
> > > > > >   ? __sysvec_apic_timer_interrupt+0x5c/0x110
> > > > > >   ? sysvec_apic_timer_interrupt+0x37/0x90
> > > > > >   ? asm_sysvec_apic_timer_interrupt+0x16/0x20
> > > > > >   ? __do_softirq+0x78/0x2ac
> > > > > >   ? __do_softirq+0x60/0x2ac
> > > > > >   __irq_exit_rcu+0xa1/0xc0
> > > > > >   sysvec_call_function_single+0x72/0x90
> > > > > >   </IRQ>
> > > > > >   <TASK>
> > > > > >   asm_sysvec_call_function_single+0x16/0x20
> > > > > >  RIP: 0010:_raw_spin_unlock_irq+0x14/0x30
> > > > > >  RSP: 0018:ffffb253604cbd88 EFLAGS: 00000247
> > > > > >  RAX: 000000000001960d RBX: 0000000000000002 RCX: ffff8cad2a064800
> > > > > >  RDX: 000000008020001b RSI: 0000000000000001 RDI: ffff8cad5d39f66c
> > > > > >  RBP: ffff8cad5d39f600 R08: 0000000000000001 R09: 0000000000000000
> > > > > >  R10: ffff8caa443e0c00 R11: ffffb253604cbcd8 R12: ffff8cacb8682538
> > > > > >  R13: 0000000000000005 R14: ffffb253604cbd90 R15: ffff8cad5d39f66c
> > > > > >   cm_process_send_error+0x122/0x1d0 [ib_cm]
> > > > > >   timeout_sends+0x1dd/0x270 [ib_core]
> > > > > >   process_one_work+0x1e2/0x3b0
> > > > > >   ? __pfx_worker_thread+0x10/0x10
> > > > > >   worker_thread+0x50/0x3a0
> > > > > >   ? __pfx_worker_thread+0x10/0x10
> > > > > >   kthread+0xdd/0x100
> > > > > >   ? __pfx_kthread+0x10/0x10
> > > > > >   ret_from_fork+0x29/0x50
> > > > > >   </TASK>
> > > > > >
> > > > > > Simplified timeout handler by creating local list of timed out WRs
> > > > > > and invoke send handler post creating the list. The new method acquires/
> > > > > > releases lock once to fetch the list and hence helps to reduce locking
> > > > > > contetiong when processing higher no. of WRs
> > > > > >
> > > > > > Signed-off-by: Saravanan Vajravel <saravanan.vajravel@xxxxxxxxxxxx>
> > > > > > ---
> > > > > >  drivers/infiniband/core/mad.c | 14 ++++++++------
> > > > > >  1 file changed, 8 insertions(+), 6 deletions(-)
> > > > > >
> > > > > > diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c
> > > > > > index 674344eb8e2f..58befbaaf0ad 100644
> > > > > > --- a/drivers/infiniband/core/mad.c
> > > > > > +++ b/drivers/infiniband/core/mad.c
> > > > > > @@ -2616,14 +2616,16 @@ static int retry_send(struct ib_mad_send_wr_private *mad_send_wr)
> > > > > >
> > > > > >  static void timeout_sends(struct work_struct *work)
> > > > > >  {
> > > > > > +     struct ib_mad_send_wr_private *mad_send_wr, *n;
> > > > > >       struct ib_mad_agent_private *mad_agent_priv;
> > > > > > -     struct ib_mad_send_wr_private *mad_send_wr;
> > > > > >       struct ib_mad_send_wc mad_send_wc;
> > > > > > +     struct list_head local_list;
> > > > > >       unsigned long flags, delay;
> > > > > >
> > > > > >       mad_agent_priv = container_of(work, struct ib_mad_agent_private,
> > > > > >                                     timed_work.work);
> > > > > >       mad_send_wc.vendor_err = 0;
> > > > > > +     INIT_LIST_HEAD(&local_list);
> > > > > >
> > > > > >       spin_lock_irqsave(&mad_agent_priv->lock, flags);
> > > > > >       while (!list_empty(&mad_agent_priv->wait_list)) {
> > > > > > @@ -2641,13 +2643,16 @@ static void timeout_sends(struct work_struct *work)
> > > > > >                       break;
> > > > > >               }
> > > > > >
> > > > > > -             list_del(&mad_send_wr->agent_list);
> > > > > > +             list_del_init(&mad_send_wr->agent_list);
> > > > > >               if (mad_send_wr->status == IB_WC_SUCCESS &&
> > > > > >                   !retry_send(mad_send_wr))
> > > > > >                       continue;
> > > > > >
> > > > > > -             spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
> > > > > > +             list_add_tail(&mad_send_wr->agent_list, &local_list);
> > > > > > +     }
> > > > > > +     spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
> > > > > >
> > > > > > +     list_for_each_entry_safe(mad_send_wr, n, &local_list, agent_list) {
> > > > > >               if (mad_send_wr->status == IB_WC_SUCCESS)
> > > > > >                       mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR;
> > > > > >               else
> > > > > > @@ -2655,11 +2660,8 @@ static void timeout_sends(struct work_struct *work)
> > > > > >               mad_send_wc.send_buf = &mad_send_wr->send_buf;
> > > > > >               mad_agent_priv->agent.send_handler(&mad_agent_priv->agent,
> > > > > >                                                  &mad_send_wc);
> > > > >
> > > > > I understand the problem, but I'm not sure that this is safe to do. How
> > > > > can we be sure that this is safe to call the send_handler on entry in
> > > > > wait_list without the locking?
> > > > >
> > > > > Thanks
> > > >
> > > > Per existing implementation, the send_handler is invoked without locking.
> > > > I didn't change that design. Existing implementation was acquiring and
> > > > releasing the lock for every Work Request (WR) that it handles inside
> > > > timeout_handler.
> > > > I improved it in such a way that once lock is acquired to fetch list
> > > > of WR to be handled
> > > > and process all fetched WRs outside the scope of lock. In both implementations,
> > > > send_handler is always called without lock. This patch aims to reduce locking
> > > > contention
> > >
> > > Indeed send_handler is called without lock, but not on the wait_list.
> > > I've asked explicitly about operations on the wait_list.
> > >
> > > Thanks
> >
> > The entry in wait_list is removed under the scope of mad_agent_priv->lock.
> > It is then added to local_list declared inside timeout_sends() handler
> > under the same
> > lock. Once all such entries are removed and added to the local_list,
> > this timeout_handler
> > processes each entry from the local_list and it invokes send_handler.
> > This local_list
> > doesn't need any locking as there is no possibility of race condition.
> >
> > Existing implementation also removed entry from wait_list and invoked
> > send_handler.
>
> Yes, and what is the difference between the current and proposed code?
>   2649                 spin_unlock_irqrestore(&mad_agent_priv->lock, flags);  <----- unlock
>   2650                                                                      
>   2651                 if (mad_send_wr->status == IB_WC_SUCCESS)          
>   2652                         mad_send_wc.status = IB_WC_RESP_TIMEOUT_ERR;
>   2653                 else                                              
>   2654                         mad_send_wc.status = mad_send_wr->status;
>   2655                 mad_send_wc.send_buf = &mad_send_wr->send_buf;  
>   2656                 mad_agent_priv->agent.send_handler(&mad_agent_priv->agent,  <-- performed without lock
>   2657                                                    &mad_send_wc);        
>   2658                                                                        
>   2659                 deref_mad_agent(mad_agent_priv);                      
>   2660                 spin_lock_irqsave(&mad_agent_priv->lock, flags);      <---   ## lock() again
>   2661         }                                                            
>   2662         spin_unlock_irqrestore(&mad_agent_priv->lock, flags);     <-- ## unlock.
>
> > So operation on the wait_list should not change in the new patch
> >
> > Thanks,

Please check your snippet of code above. I corrected it with ## tag.
It is acquiring / releasing / again acquiring / again releasing the lock 
for handling one WR. So the existing implementation has heavy locking
contention. It will  be surfacing as an issue only when higher no. of WRs
are handled inside time_out handler().

Thanks.

> >
> > > > >
> > > > > > -
> > > > > >               deref_mad_agent(mad_agent_priv);
> > > > > > -             spin_lock_irqsave(&mad_agent_priv->lock, flags);
> > > > > >       }
> > > > > > -     spin_unlock_irqrestore(&mad_agent_priv->lock, flags);
> > > > > >  }
> > > > > >
> > > > > >  /*
> > > > > > --
> > > > > > 2.39.3
> > > > > >
> > >
> > >
>
>

Attachment: smime.p7s
Description: S/MIME Cryptographic Signature


[Index of Archives]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Photo]     [Yosemite News]     [Yosemite Photos]     [Linux Kernel]     [Linux SCSI]     [XFree86]

  Powered by Linux