On Fri, Jan 24, 2025 at 06:48:40PM +0100, Uladzislau Rezki wrote: > On Fri, Jan 24, 2025 at 09:36:07AM -0800, Paul E. McKenney wrote: > > On Fri, Jan 24, 2025 at 06:21:30PM +0100, Uladzislau Rezki wrote: > > > On Fri, Jan 24, 2025 at 07:45:23AM -0800, Paul E. McKenney wrote: > > > > On Fri, Jan 24, 2025 at 12:41:38PM +0100, Uladzislau Rezki wrote: > > > > > On Thu, Jan 23, 2025 at 12:29:45PM -0800, Paul E. McKenney wrote: > > > > > > On Thu, Jan 23, 2025 at 07:58:26PM +0100, Uladzislau Rezki (Sony) wrote: > > > > > > > This configuration specifies the maximum number of CPUs which > > > > > > > is set to 8. The problem is that it can not be overwritten for > > > > > > > something higher. > > > > > > > > > > > > > > Remove that configuration for TREE05, so it is possible to run > > > > > > > the torture test on as many CPUs as many system has. > > > > > > > > > > > > > > Signed-off-by: Uladzislau Rezki (Sony) <urezki@xxxxxxxxx> > > > > > > > > > > > > You should be able to override this on the kvm.sh command line by > > > > > > specifying "--kconfig CONFIG_NR_CPUS=128" or whatever number you wish. > > > > > > For example, see the torture.sh querying the system's number of CPUs > > > > > > and then specifying it to a number of tests. > > > > > > > > > > > > Or am I missing something here? > > > > > > > > > > > It took me a while to understand what happens. Apparently there is this > > > > > 8 CPUs limitation. Yes, i can do it manually by passing --kconfig but > > > > > you need to know about that. I have not expected that. > > > > > > > > > > Therefore i removed it from the configuration because i have not found > > > > > a good explanation why we need. It is confusing instead :) > > > > > > > > Right now, if I do a run with --configs "TREE10 14*CFLIST", this will > > > > make use of 20 systems with 80 CPUs each. If you remove that line from > > > > TREE05, won't each instance of TREE05 consume a full system, for a total > > > > of 33 systems? Yes, I could use "--kconfig CONFIG_NR_CPUS=8" on the > > > > command line, but that would affect all the scenarios, not just TREE05. > > > > Including (say) TINY01, where I believe that it would cause kvm.sh > > > > to complain about a Kconfig conflict. > > > > > > > > Hence me not being in favor of this change. ;-) > > > > > > > > Is there another way to make things work for both situations? > > > > > > > OK, i see. Well. I will just go with --kconfig CONFIG_NR_CPUS=foo if i > > > need more CPUs for TREE05. > > > > > > I will not resist, we just drop this patch :) > > > > Thank you! > > > > The bug you are chasing happens when a given synchonize_rcu() interacts > > with RCU readers, correct? > > > Below one: > > <snip> > /* > * RCU torture fake writer kthread. Repeatedly calls sync, with a random > * delay between calls. > */ > static int > rcu_torture_fakewriter(void *arg) > { > ... > <snip> > > > In rcutorture, only the rcu_torture_writer() call to synchronize_rcu() > > interacts with rcu_torture_reader(). So my guess is that running > > many small TREE05 guest OSes would reproduce this bug more quickly. > > So instead of this: > > > > --kconfig CONFIG_NR_CPUS=128 > > > > Do this: > > > > --configs "16*TREE05" > > > > Or maybe even this: > > > > --configs "16*TREE05" --kconfig CONFIG_NR_CPUS=4 > Thanks for input. > > > > > Thoughts? > > > If you mean below splat: No, instead the one reported by cheung wall <zzqq0103.hey@xxxxxxxxx>. > <snip> > [ 32.107748] ============================= > [ 32.108512] WARNING: suspicious RCU usage > [ 32.109232] 6.12.0-rc4-dirty #66 Not tainted > [ 32.110058] ----------------------------- > [ 32.110817] kernel/events/core.c:13962 RCU-list traversed in non-reader section!! > [ 32.111221] kworker/u34:2 (251) used greatest stack depth: 12112 bytes left > [ 32.112125] > [ 32.112125] other info that might help us debug this: > [ 32.112125] > [ 32.112130] > [ 32.112130] rcu_scheduler_active = 2, debug_locks = 1 > [ 32.116039] 3 locks held by cpuhp/1/20: > [ 32.116758] #0: ffffffff93a6a750 (cpu_hotplug_lock){++++}-{0:0}, at: cpuhp_thread_fun+0x50/0x220 > [ 32.118410] #1: ffffffff93a6ce00 (cpuhp_state-down){+.+.}-{0:0}, at: cpuhp_thread_fun+0x50/0x220 > [ 32.120091] #2: ffffffff93b7eb68 (pmus_lock){+.+.}-{3:3}, at: perf_event_exit_cpu_context+0x32/0x2d0 > [ 32.121723] > [ 32.121723] stack backtrace: > [ 32.122413] CPU: 1 UID: 0 PID: 20 Comm: cpuhp/1 Not tainted 6.12.0-rc4-dirty #66 > [ 32.123666] Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.13.0-1ubuntu1.1 04/01/2014 > [ 32.125302] Call Trace: > [ 32.125769] <TASK> > [ 32.126148] dump_stack_lvl+0x83/0xa0 > [ 32.126823] lockdep_rcu_suspicious+0x113/0x180 > [ 32.127652] perf_event_exit_cpu_context+0x2c4/0x2d0 > [ 32.128593] ? __pfx_perf_event_exit_cpu+0x10/0x10 > [ 32.129489] perf_event_exit_cpu+0x9/0x10 > [ 32.130243] cpuhp_invoke_callback+0x187/0x6e0 > [ 32.131065] ? cpuhp_thread_fun+0x50/0x220 > [ 32.131800] cpuhp_thread_fun+0x185/0x220 > [ 32.132560] ? __pfx_smpboot_thread_fn+0x10/0x10 [ 32.133394] smpboot_thread_fn+0xd8/0x1d0 > [ 32.134050] kthread+0xd0/0x100 > [ 32.134592] ? __pfx_kthread+0x10/0x10 > [ 32.135270] ret_from_fork+0x2f/0x50 > [ 32.135896] ? __pfx_kthread+0x10/0x10 > [ 32.136610] ret_from_fork_asm+0x1a/0x30 > [ 32.137356] </TASK> > [ 32.140997] smpboot: CPU 1 is now offline > <snip> > > I reproduced that using: > > +rcutorture.nfakewriters=128 > +rcutorture.gp_sync=1 > +rcupdate.rcu_expedited=0 > +rcupdate.rcu_normal=1 > +rcutree.rcu_normal_wake_from_gp=1 > <snip> > > The test script: > > for (( i=0; i<$LOOPS; i++ )); do > tools/testing/selftests/rcutorture/bin/kvm.sh --cpus 64 --configs \ > '100*TREE05' --memory 20G --bootargs 'rcutorture.fwd_progress=1' > echo "Done $i" > done > > i.e. with more nfakewriters. Right, and large nfakewriters would help push the synchronize_rcu() wakeups off of the grace-period kthread. > If you mean the one that has recently reported, i am not able to > reproduce it anyhow :) Using larger numbers of smaller rcutorture guest OSes might help to reproduce it. Maybe as small as three CPUs each. ;-) Thanx, Paul