On Fri 2023-05-19 10:18:38, Douglas Anderson wrote: > Implement a hardlockup detector that doesn't doesn't need any extra > arch-specific support code to detect lockups. Instead of using > something arch-specific we will use the buddy system, where each CPU > watches out for another one. Specifically, each CPU will use its > softlockup hrtimer to check that the next CPU is processing hrtimer > interrupts by verifying that a counter is increasing. > > --- a/kernel/watchdog.c > +++ b/kernel/watchdog.c > @@ -85,7 +85,7 @@ __setup("nmi_watchdog=", hardlockup_panic_setup); > > #endif /* CONFIG_HARDLOCKUP_DETECTOR */ > > -#if defined(CONFIG_HARDLOCKUP_DETECTOR_PERF) > +#if defined(CONFIG_HARDLOCKUP_DETECTOR_COUNTS_HRTIMER) > > static DEFINE_PER_CPU(atomic_t, hrtimer_interrupts); > static DEFINE_PER_CPU(int, hrtimer_interrupts_saved); > @@ -106,6 +106,14 @@ notrace void arch_touch_nmi_watchdog(void) > } > EXPORT_SYMBOL(arch_touch_nmi_watchdog); > > +void watchdog_hardlockup_touch_cpu(unsigned int cpu) > +{ > + per_cpu(watchdog_hardlockup_touched, cpu) = true; > + > + /* Match with smp_rmb() in watchdog_hardlockup_check() */ > + smp_wmb(); It is great that you described where the related barrier is. Another important information is what exactly is synchronized. And I am actually not sure what we are synchronizing here. My understanding is that a write barrier should synchronize related writes, for example: X = ...; /* Make sure that X is modified before Y */ smp_wmb(); Y = ...; And the related read barrier should synchronize the related reads, for example: if (test(Y)) { /* * Make sure that we use the updated X when * we saw the updated Y. */ smp_rmb(); do_something(X); } IMHO, we do not need any barrier here because we have only one variable "watchdog_hardlockup_touched" here. watchdog_hardlockup_check() will either see the updated value or not. But it does not synchronize it against any other variables or values. > +} > + > static bool is_hardlockup(unsigned int cpu) > { > int hrint = atomic_read(&per_cpu(hrtimer_interrupts, cpu)); > @@ -443,11 +454,15 @@ static enum hrtimer_restart watchdog_timer_fn(struct hrtimer *hrtimer) > struct pt_regs *regs = get_irq_regs(); > int duration; > int softlockup_all_cpu_backtrace = sysctl_softlockup_all_cpu_backtrace; > + unsigned long hrtimer_interrupts; > > if (!watchdog_enabled) > return HRTIMER_NORESTART; > > - watchdog_hardlockup_kick(); > + hrtimer_interrupts = watchdog_hardlockup_kick(); > + > + /* test for hardlockups */ I would omit the comment. It is not valid when perf detector is used. And checking the buddy is clear from the function name. > + watchdog_buddy_check_hardlockup(hrtimer_interrupts); I would personally move this into watchdog_hardlockup_kick(). watchdog_timer_fn() is already complex enough. And checking the buddy when kicking a CPU makes sense. Also I would not pass "hrtimer_interrupts". I guess that it is just an optimization. It is an extra churn in the code. IMHO, is is not wort it. This code does not need to be super optimized. > + > > /* kick the softlockup detector */ > if (completion_done(this_cpu_ptr(&softlockup_completion))) { > diff --git a/kernel/watchdog_buddy.c b/kernel/watchdog_buddy.c > new file mode 100644 > index 000000000000..fee45af2e5bd > --- /dev/null > +++ b/kernel/watchdog_buddy.c > @@ -0,0 +1,93 @@ > +// SPDX-License-Identifier: GPL-2.0 > + > +#include <linux/cpu.h> > +#include <linux/cpumask.h> > +#include <linux/kernel.h> > +#include <linux/nmi.h> > +#include <linux/percpu-defs.h> > + > +static cpumask_t __read_mostly watchdog_cpus; > + > +static unsigned int watchdog_next_cpu(unsigned int cpu) > +{ > + cpumask_t cpus = watchdog_cpus; A copy should be done by cpumask_copy(). But the question is why a copy would be needed. When called from watchdog_buddy_check_hardlockup(), this function is not sychronized against the CPU hotplug. And even the copying would be racy. IMHO, the copy does not help much and we do not need it. The only important this is that this function would return a valid CPU. And I think that it is guarantted because CPU0 could not be disabled. > + unsigned int next_cpu; > + > + next_cpu = cpumask_next(cpu, &cpus); > + if (next_cpu >= nr_cpu_ids) > + next_cpu = cpumask_first(&cpus); > + > + if (next_cpu == cpu) > + return nr_cpu_ids; >> + return next_cpu; > +} > + > +int __init watchdog_hardlockup_probe(void) > +{ > + return 0; > +} > + > +void watchdog_hardlockup_enable(unsigned int cpu) > +{ > + unsigned int next_cpu; > + > + /* > + * The new CPU will be marked online before the hrtimer interrupt > + * gets a chance to run on it. If another CPU tests for a > + * hardlockup on the new CPU before it has run its the hrtimer > + * interrupt, it will get a false positive. Touch the watchdog on > + * the new CPU to delay the check for at least 3 sampling periods > + * to guarantee one hrtimer has run on the new CPU. > + */ > + watchdog_hardlockup_touch_cpu(cpu); > + > + /* > + * We are going to check the next CPU. Our watchdog_hrtimer > + * need not be zero if the CPU has already been online earlier. > + * Touch the watchdog on the next CPU to avoid false positive > + * if we try to check it in less then 3 interrupts. > + */ > + next_cpu = watchdog_next_cpu(cpu); > + if (next_cpu < nr_cpu_ids) > + watchdog_hardlockup_touch_cpu(next_cpu); Thinking loudly: This feels racy when many CPUs are enabled/disabled in parallel. I am not 100% sure it it can happen though. My understanding is that it can't happen because the CPU hotplug is serialized by cpu_add_remove_lock. So, this seems to work after all. > + > + cpumask_set_cpu(cpu, &watchdog_cpus); > +} > + > --- a/lib/Kconfig.debug > +++ b/lib/Kconfig.debug > @@ -1035,10 +1035,55 @@ config BOOTPARAM_SOFTLOCKUP_PANIC > > Say N if unsure. > > -config HARDLOCKUP_DETECTOR_PERF > +# Both the "perf" and "buddy" hardlockup detectors count hrtimer > +# interrupts. This config enables functions managing this common code. > +config HARDLOCKUP_DETECTOR_COUNTS_HRTIMER > bool > select SOFTLOCKUP_DETECTOR > > +config HARDLOCKUP_DETECTOR_PERF > + bool > + depends on HAVE_HARDLOCKUP_DETECTOR_PERF > + select HARDLOCKUP_DETECTOR_COUNTS_HRTIMER > + > +config HARDLOCKUP_DETECTOR_BUDDY > + bool > + depends on SMP > + select HARDLOCKUP_DETECTOR_COUNTS_HRTIMER > + > +# For hardlockup detectors you can have one directly provided by the arch > +# or use a "non-arch" one. If you're using a "non-arch" one that is > +# further divided the perf hardlockup detector (which, confusingly, needs > +# arch-provided perf support) and the buddy hardlockup detector (which just > +# needs SMP). In either case, using the "non-arch" code conflicts with > +# the NMI watchdog code (which is sometimes used directly and sometimes used > +# by the arch-provided hardlockup detector). > +config HAVE_HARDLOCKUP_DETECTOR_NON_ARCH > + bool > + depends on (HAVE_HARDLOCKUP_DETECTOR_PERF || SMP) && !HAVE_NMI_WATCHDOG > + default y > + > +config HARDLOCKUP_DETECTOR_PREFER_BUDDY > + bool "Prefer the buddy CPU hardlockup detector" > + depends on HAVE_HARDLOCKUP_DETECTOR_NON_ARCH && HAVE_HARDLOCKUP_DETECTOR_PERF && SMP Huh, I have big troubles to scratch my head around this check: HAVE_HARDLOCKUP_DETECTOR_NON_ARCH depends on HAVE_HARDLOCKUP_DETECTOR_PERF and SMP and this depends on HAVE_HARDLOCKUP_DETECTOR_NON_ARCH and again on HAVE_HARDLOCKUP_DETECTOR_PERF and SMP. > + help > + Say Y here to prefer the buddy hardlockup detector over the perf one. > + > + With the buddy detector, each CPU uses its softlockup hrtimer > + to check that the next CPU is processing hrtimer interrupts by > + verifying that a counter is increasing. > + > + This hardlockup detector is useful on systems that don't have > + an arch-specific hardlockup detector or if resources needed > + for the hardlockup detector are better used for other things. > + > +# This will select the appropriate non-arch hardlockdup detector > +config HARDLOCKUP_DETECTOR_NON_ARCH > + bool > + depends on HAVE_HARDLOCKUP_DETECTOR_NON_ARCH > + select HARDLOCKUP_DETECTOR_BUDDY if !HAVE_HARDLOCKUP_DETECTOR_PERF || HARDLOCKUP_DETECTOR_PREFER_BUDDY > + select HARDLOCKUP_DETECTOR_PERF if HAVE_HARDLOCKUP_DETECTOR_PERF && !HARDLOCKUP_DETECTOR_PREFER_BUDDY > + > # > # Enables a timestamp based low pass filter to compensate for perf based > # hard lockup detection which runs too fast due to turbo modes. > @@ -1053,9 +1098,10 @@ config HARDLOCKUP_CHECK_TIMESTAMP > config HARDLOCKUP_DETECTOR > bool "Detect Hard Lockups" > depends on DEBUG_KERNEL && !S390 Is there any reason why S390 could not or do not want to use the buddy hardlockup detector. > - depends on HAVE_HARDLOCKUP_DETECTOR_PERF || HAVE_HARDLOCKUP_DETECTOR_ARCH > + depends on HAVE_HARDLOCKUP_DETECTOR_NON_ARCH || HAVE_HARDLOCKUP_DETECTOR_ARCH > select LOCKUP_DETECTOR > - select HARDLOCKUP_DETECTOR_PERF if HAVE_HARDLOCKUP_DETECTOR_PERF > + select HARDLOCKUP_DETECTOR_NON_ARCH if HAVE_HARDLOCKUP_DETECTOR_NON_ARCH Anyway, the configuration of the hard lockup detectors is insane and this patchset makes it even worse, especially the new HARDLOCKUP_DETECTOR_NON_ARCH stuff. It seems that sparc, powerpc and s390 are somehow special. Do you still have in mind how they are distinguished using the Kconfig variables? For example, I am pretty confused by the meaning of HAVE_NMI_WATCHDOG. And sparc has its own variant of watchdog_hardlockup_enable()/disable(). It means that it is arch-specific. Does it work with the 13th patch which made watchdog_hardlockup_enable()/disable() to be watchdog-hardlockup-type specific? Is is somehow related to HAVE_NMI_WATCHDOG? Does this replace the entire watchdog only only the enable part? I think that we need to make this more straightforward. But I first need to understand the existing maze of config variables. Best Regards, Petr