Re: RCU: rcu stall issues and an approach to the fix

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Jul 23, 2021 at 12:20:41PM -0400, donghai qiao wrote:
> On Thu, Jul 22, 2021 at 11:49 PM Paul E. McKenney <paulmck@xxxxxxxxxx>
> wrote:
> 
> > On Fri, Jul 23, 2021 at 08:29:53AM +0800, Boqun Feng wrote:
> > > On Thu, Jul 22, 2021 at 04:08:06PM -0400, donghai qiao wrote:
> > > > RCU experts,
> > > >
> > > > When you reply, please also keep me CC'ed.
> > > >
> > > > The problem of RCU stall might be an old problem and it can happen
> > quite often.
> > > > As I have observed, when the problem occurs,  at least one CPU in the
> > system
> > > > on which its rdp->gp_seq falls behind others by 4 (qs).
> > > >
> > > > e.g.  On CPU 0, rdp->gp_seq = 0x13889d, but on other CPUs, their
> > > > rdp->gp_seq = 0x1388a1.
> > > >
> > > > Because RCU stall issues can last a long period of time, the number of
> > callbacks
> > > > in the list rdp->cblist of all CPUs can accumulate to thousands. In
> > > > the worst case,
> > > > it triggers panic.
> > > >
> > > > When looking into the problem further, I'd think the problem is
> > related to the
> > > > Linux scheduler. When the RCU core detects the stall on a CPU,
> > rcu_gp_kthread
> > > > would send a rescheduling request via send_IPI to that CPU to try to
> > force a
> > > > context switch to make some progress. However, at least one situation
> > can fail
> > > > this effort, which is when the CPU is running a user thread and it is
> > the only
> > > > user thread in the rq, then this attempted context switching will not
> > happen
> > > > immediately. In particular if the system is also configured with
> > NOHZ_FULL for
> > >
> > > Correct me if I'm wrong, if a CPU is solely running a user thread, how
> > > can that CPU stall RCU? Because you need to be in a RCU read-side
> > > critical section to stall RCU. Or the problem you're talking here is
> > > about *recovering* from RCU stall?
> 
> In response to Boqun's question, the crashdumps I analyzed were configured
> with this :
> 
> CONFIG_PREEMPT_RCU=n
> CONFIG_PREEMPT_COUNT=n
> CONFIG_PROVE_RCU=n
> 
> Because these configurations were not enabled, the compiler generated empty
> binary code for functions rcu_read_lock() and rcu_read_unlock() which
> delimit rcu read-side critical sections. And the crashdump showed both
> functions have no binary code in the kernel module and I am pretty sure.

Agreed, that is expected behavior.

> In the first place I thought this kernel might be built the wrong way,
> but later I found other sources that said this was ok.  That's why when
> CPUs enter or leave rcu critical section, the rcu core
> is not informed.

If RCU core was informed every time that a CPU entered or left an RCU
read-side critical section, performance and scalability would be
abysmal.  So yes, this interaction is very arms-length.

> When the current grace period is closed, rcu_gp_kthread will open a new
> period for all. This will be reflected from every
> CPU's rdp->gp_seq. Every CPU is responsible to update its own gp when a
> progress is made. So when a cpu is running
> a user thread whilst a new period is open, it can not update its rcu unless
> a context switch occurs or upon a sched tick.
> But if a CPU is configured as NOHZ, this will be a problem to RCU, so rcu
> stall will happen.

Except that if a CPU is running in nohz_full mode, each transition from
kernel to user execution must invoke rcu_user_enter() and each transition
back must invoke rcu_user_exit().  These update RCU's per-CPU state, which
allows RCU's grace-period kthread ("rcu_sched" in this configuration)
to detect even momentary nohz_full usermode execution.

You can check this in your crash dump by looking at the offending CPU's
rcu_data structure's ->dynticks field and comparing to the activities
of rcu_user_enter().

> When RCU detects that qs is stalled on a CPU, it tries to force a context
> switch to make progress on that CPU. This is
> done through a resched IPI. But this can not always succeed depending on
> the scheduler.   A while ago, this code
> process the resched IPI:
> 
> void scheduler_ipi(void)
> {
>         ...
>         if (llist_empty(&this_rq()->wake_list) && !got_nohz_idle_kick())
>                 return;
>         ...
>         irq_enter();
>         sched_ttwu_pending();
>         ...
>         if (unlikely(got_nohz_idle_kick())) {
>                 this_rq()->idle_balance = 1;
>                 raise_softirq_irqoff(SCHED_SOFTIRQ);
>         }
>         irq_exit();
> }
> 
> As you can see the function returns from the first "if statement" before it
> can issue a SCHED_SOFTIRQ. Later this
> code has been changed, but similar check/optimization remains in many
> places in the scheduler. The things I try to
> fix are those that resched_cpu fails to do.

???  Current mainline has this instead:

static __always_inline void scheduler_ipi(void)
{
	/*
	 * Fold TIF_NEED_RESCHED into the preempt_count; anybody setting
	 * TIF_NEED_RESCHED remotely (for the first time) will also send
	 * this IPI.
	 */
	preempt_fold_need_resched();
}

Combined with the activities of resched_curr(), which is invoked
from resched_cpu(), this should force a call to the scheduler on
the return path from this IPI.

So what kernel version are you using?

Recent kernels have logic to enable the tick on nohz_full CPUs that are
slow to supply RCU with a quiescent state, but this should happen only
when such CPUs are spinning in kernel mode.  Again, usermode execution
is dealt with by rcu_user_enter().

> Hope this explains it.
> Donghai
> 
> 
> > Excellent point, Boqun!
> >
> > Donghai, have you tried reproducing this using a kernel built with
> > CONFIG_RCU_EQS_DEBUG=y?
> >
> 
> I can give this configuration a try. Will let you know the results.

This should help detect any missing rcu_user_enter() or rcu_user_exit()
calls.

							Thanx, Paul

> Thanks.
> Donghai
> 
> 
> >
> >                                                         Thanx, Paul
> >
> > > Regards,
> > > Boqun
> > >
> > > > the CPU and as long as the user thread is running, the forced context
> > > > switch will
> > > > never happen unless the user thread volunteers to yield the CPU. I
> > think this
> > > > should be one of the major root causes of these RCU stall issues. Even
> > if
> > > > NOHZ_FULL is not configured, there will be at least 1 tick delay which
> > can
> > > > affect the realtime kernel, by the way.
> > > >
> > > > But it seems not a good idea to craft a fix from the scheduler side
> > because
> > > > this has to invalidate some existing scheduling optimizations. The
> > current
> > > > scheduler is deliberately optimized to avoid such context switching.
> > So my
> > > > question is why the RCU core cannot effectively update qs for the
> > stalled CPU
> > > > when it detects that the stalled CPU is running a user thread?  The
> > reason
> > > > is pretty obvious because when a CPU is running a user thread, it must
> > not
> > > > be in any kernel read-side critical sections. So it should be safe to
> > close
> > > > its current RCU grace period on this CPU. Also, with this approach we
> > can make
> > > > RCU work more efficiently than the approach of context switch which
> > needs to
> > > > go through an IPI interrupt and the destination CPU needs to wake up
> > its
> > > > ksoftirqd or wait for the next scheduling cycle.
> > > >
> > > > If my suggested approach makes sense, I can go ahead to fix it that
> > way.
> > > >
> > > > Thanks
> > > > Donghai
> >



[Index of Archives]     [Linux Samsung SoC]     [Linux Rockchip SoC]     [Linux Actions SoC]     [Linux for Synopsys ARC Processors]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]


  Powered by Linux