On Wed, Mar 03, 2021 at 04:00:16PM -0800, paulmck@xxxxxxxxxx wrote: > From: "Paul E. McKenney" <paulmck@xxxxxxxxxx> > > After interrupts have enabled at boot but before some random point > in early_initcall() processing, softirq processing is unreliable. > If softirq sees a need to push softirq-handler invocation to ksoftirqd > during this time, then those handlers can be delayed until the ksoftirqd > kthreads have been spawned, which happens at some random point in the > early_initcall() processing. In many cases, this delay is just fine. > However, if the boot sequence blocks waiting for a wakeup from a softirq > handler, this delay will result in a silent-hang deadlock. > > This commit therefore prevents these hangs by ensuring that the tick > stays active until after the ksoftirqd kthreads have been spawned. > This change causes the tick to eventually drain the backlog of delayed > softirq handlers, breaking this deadlock. > > Signed-off-by: Paul E. McKenney <paulmck@xxxxxxxxxx> > --- > kernel/rcu/tree_plugin.h | 11 +++++++++++ > 1 file changed, 11 insertions(+) > > diff --git a/kernel/rcu/tree_plugin.h b/kernel/rcu/tree_plugin.h > index 2d60377..36212de 100644 > --- a/kernel/rcu/tree_plugin.h > +++ b/kernel/rcu/tree_plugin.h > @@ -1255,6 +1255,11 @@ static void rcu_prepare_kthreads(int cpu) > */ > int rcu_needs_cpu(u64 basemono, u64 *nextevt) > { > + /* Through early_initcall(), need tick for softirq handlers. */ > + if (!IS_ENABLED(CONFIG_HZ_PERIODIC) && !this_cpu_ksoftirqd()) { > + *nextevt = 1; > + return 1; > + } > *nextevt = KTIME_MAX; > return !rcu_segcblist_empty(&this_cpu_ptr(&rcu_data)->cblist) && > !rcu_segcblist_is_offloaded(&this_cpu_ptr(&rcu_data)->cblist); > @@ -1350,6 +1355,12 @@ int rcu_needs_cpu(u64 basemono, u64 *nextevt) > > lockdep_assert_irqs_disabled(); > > + /* Through early_initcall(), need tick for softirq handlers. */ > + if (!IS_ENABLED(CONFIG_HZ_PERIODIC) && !this_cpu_ksoftirqd()) { > + *nextevt = 1; > + return 1; > + } > + > /* If no non-offloaded callbacks, RCU doesn't need the CPU. */ > if (rcu_segcblist_empty(&rdp->cblist) || > rcu_segcblist_is_offloaded(&this_cpu_ptr(&rcu_data)->cblist)) { I suspect rcutiny should be concerned as well? In fact this patch doesn't look necessary because can_stop_idle_tick() refuse to stop the tick when softirqs are pending. Thanks.