On Sat, Oct 22, 2022 at 08:45:25PM +0800, Zhen Lei wrote: > In some extreme cases, such as the I/O pressure test, the CPU usage may > be 100%, causing RCU stall. In this case, the printed information about > current is not useful. Displays the number and usage of hard interrupts, > soft interrupts, and context switches that are generated within half of > the CPU stall timeout, can help us make a general judgment. In other > cases, we can preliminarily determine whether an infinite loop occurs > when local_irq, local_bh or preempt is disabled. > > For example: > rcu: INFO: rcu_preempt self-detected stall on CPU > rcu: 0-....: (1250 ticks this GP) <omitted> > rcu: hardirqs softirqs csw/system > rcu: number: 624 45 0 > rcu: cputime: 69 1 2425 ==> 2500(ms) > > The example above shows that the number of hard and soft interrupts is > small, there is zero context switching, and the system takes up a lot of > time. We can quickly conclude that the current task is infinitely looped > with preempt_disable(). > > The impact on system performance is negligible because snapshot is > recorded only one time after 1/2 CPU stall timeout. > > Signed-off-by: Zhen Lei <thunder.leizhen@xxxxxxxxxx> > --- > kernel/rcu/tree.c | 16 ++++++++++++++++ > kernel/rcu/tree.h | 11 +++++++++++ > kernel/rcu/tree_stall.h | 28 ++++++++++++++++++++++++++++ > 3 files changed, 55 insertions(+) > > diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c > index 6bb8e72bc8151ef..56c49a3117e7a81 100644 > --- a/kernel/rcu/tree.c > +++ b/kernel/rcu/tree.c > @@ -931,6 +931,22 @@ static int rcu_implicit_dynticks_qs(struct rcu_data *rdp) > rdp->rcu_iw_gp_seq = rnp->gp_seq; > irq_work_queue_on(&rdp->rcu_iw, rdp->cpu); > } > + > + if (rdp->snap_record.gp_seq != rdp->gp_seq) { > + u64 *cpustat; > + struct rcu_snap_record *r; > + > + cpustat = kcpustat_cpu(rdp->cpu).cpustat; > + > + r = &rdp->snap_record; > + r->cputime_irq = cpustat[CPUTIME_IRQ]; > + r->cputime_softirq = cpustat[CPUTIME_SOFTIRQ]; > + r->cputime_system = cpustat[CPUTIME_SYSTEM]; > + r->nr_hardirqs = kstat_cpu_irqs_sum(rdp->cpu); > + r->nr_softirqs = kstat_cpu_softirqs_sum(rdp->cpu); > + r->nr_csw = nr_context_switches_cpu(rdp->cpu); > + r->gp_seq = rdp->gp_seq; This needs to be optional. Yes, it is normally rarely executed, but people who don't want the additional information should not pay the price for it. > + } > } > > return 0; > diff --git a/kernel/rcu/tree.h b/kernel/rcu/tree.h > index d4a97e40ea9c3e2..fb3121d15cca6f8 100644 > --- a/kernel/rcu/tree.h > +++ b/kernel/rcu/tree.h > @@ -158,6 +158,16 @@ union rcu_noqs { > u16 s; /* Set of bits, aggregate OR here. */ > }; > > +struct rcu_snap_record { > + unsigned long gp_seq; > + u64 cputime_irq; > + u64 cputime_softirq; > + u64 cputime_system; > + unsigned int nr_hardirqs; > + unsigned int nr_softirqs; > + unsigned long long nr_csw; > +}; Please add a comment saying what this is and what it is used for. > + > /* Per-CPU data for read-copy update. */ > struct rcu_data { > /* 1) quiescent-state and grace-period handling : */ > @@ -262,6 +272,7 @@ struct rcu_data { > short rcu_onl_gp_flags; /* ->gp_flags at last online. */ > unsigned long last_fqs_resched; /* Time of last rcu_resched(). */ > unsigned long last_sched_clock; /* Jiffies of last rcu_sched_clock_irq(). */ > + struct rcu_snap_record snap_record; And here as well, please. > int cpu; > }; > diff --git a/kernel/rcu/tree_stall.h b/kernel/rcu/tree_stall.h > index 5653560573e22d6..f8c9d0284d116a8 100644 > --- a/kernel/rcu/tree_stall.h > +++ b/kernel/rcu/tree_stall.h > @@ -428,6 +428,32 @@ static bool rcu_is_rcuc_kthread_starving(struct rcu_data *rdp, unsigned long *jp > return j > 2 * HZ; > } > > +static void print_cpu_stat_info(int cpu) > +{ > + u64 *cpustat; > + unsigned long half_timeout; > + struct rcu_snap_record *r; Let's please follow convention and call it "rsrp" rather than just "r". > + struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); > + > + r = &rdp->snap_record; > + if (r->gp_seq != rdp->gp_seq) > + return; > + > + cpustat = kcpustat_cpu(cpu).cpustat; > + half_timeout = rcu_jiffies_till_stall_check() / 2; > + > + pr_err(" hardirqs softirqs csw/system\n"); > + pr_err(" number: %8d %10d %12lld\n", > + kstat_cpu_irqs_sum(cpu) - r->nr_hardirqs, > + kstat_cpu_softirqs_sum(cpu) - r->nr_softirqs, > + nr_context_switches_cpu(cpu) - r->nr_csw); > + pr_err("cputime: %8lld %10lld %12lld ==> %lld(ms)\n", > + div_u64(cpustat[CPUTIME_IRQ] - r->cputime_irq, NSEC_PER_MSEC), > + div_u64(cpustat[CPUTIME_SOFTIRQ] - r->cputime_softirq, NSEC_PER_MSEC), > + div_u64(cpustat[CPUTIME_SYSTEM] - r->cputime_system, NSEC_PER_MSEC), > + jiffies64_to_msecs(half_timeout)); > +} > + > /* > * Print out diagnostic information for the specified stalled CPU. > * > @@ -484,6 +510,8 @@ static void print_cpu_stall_info(int cpu) > data_race(rcu_state.n_force_qs) - rcu_state.n_force_qs_gpstart, > rcuc_starved ? buf : "", > falsepositive ? " (false positive?)" : ""); > + > + print_cpu_stat_info(cpu); Again, please make this conditional. One way to do that is with a Kconfig option. Another is with a kernel boot parameter, as is done wtih module_param() elsewhere in tree_stall.h. Or if the parsing needs to be fancy (it shouldn't) using kernel_param_ops as is done in tree.c. Distros tend to like kernel boot parameters, while people dealing with large numbers of devices tend to like Kconfig options. Choose wisely. ;-) Please make this initially default-off. If enough people enable it for long enough, we can later switch it to default-on and maybe even later to unconditional. But let's start carefully. Thanx, Paul > } > > /* Complain about starvation of grace-period kthread. */ > -- > 2.25.1 >