On Sun, 12 Sep 2021, Finn Thain wrote:
... I've now done as you did, that is, diff --git a/arch/m68k/kernel/irq.c b/arch/m68k/kernel/irq.c index 9ab4f550342e..b46d8a57f4da 100644 --- a/arch/m68k/kernel/irq.c +++ b/arch/m68k/kernel/irq.c @@ -20,10 +20,13 @@ asmlinkage void do_IRQ(int irq, struct pt_regs *regs) { struct pt_regs *oldregs = set_irq_regs(regs); + unsigned long flags; + local_irq_save(flags); irq_enter(); generic_handle_irq(irq); irq_exit(); + local_irq_restore(flags); set_irq_regs(oldregs); } There may be a better way to achieve that. If the final IPL can be found in regs then it doesn't need to be saved again. I haven't looked for a possible entropy pool improvement from correct locking in random.c -- it would not surprise me if there was one. But none of this explains the panics I saw so I went looking for potential race conditions in the irq_enter_rcu() and irq_exit_rcu() code. I haven't found the bug yet.
Turns out that the panic bug was not affected by that patch... running --mmap -1 --mmap-odirect --mmap-bytes 100% -t 60 --timestamp --no-rand-seed --times stress-ng: 17:06:09.62 info: [1241] setting to a 60 second run per stressor stress-ng: 17:06:09.62 info: [1241] dispatching hogs: 1 mmap [ 807.270000] Kernel panic - not syncing: Aiee, killing interrupt handler! [ 807.270000] CPU: 0 PID: 1243 Comm: stress-ng Not tainted 5.14.0-multi-00002-g69f953866c7e #2 [ 807.270000] Stack from 00bcbde4: [ 807.270000] 00bcbde4 00488d85 00488d85 000c0000 00bcbe00 003f3708 00488d85 00bcbe20 [ 807.270000] 003f270e 000c0000 418004fc 00bca000 009f8a80 00bca000 00a06fc0 00bcbe5c [ 807.270000] 000317f6 0048098b 00000009 418004fc 00bca000 00000000 07408000 00000009 [ 807.270000] 00000008 00bcbf38 00a06fc0 00000006 00000000 00000001 00bcbe6c 000319ac [ 807.270000] 00000009 01438a20 00bcbeb8 0003acf0 00000009 0000000f 0000000e c043c000 [ 807.270000] 00000000 07408000 00000003 00bcbf98 efb2c944 efb2b8a8 00039afa 00bca000 [ 807.270000] Call Trace: [<000c0000>] insert_vmap_area.constprop.91+0xbc/0x15a [ 807.270000] [<003f3708>] dump_stack+0x10/0x16 [ 807.270000] [<003f270e>] panic+0xba/0x2bc [ 807.270000] [<000c0000>] insert_vmap_area.constprop.91+0xbc/0x15a [ 807.270000] [<000317f6>] do_exit+0x87e/0x9d6 [ 807.270000] [<000319ac>] do_group_exit+0x28/0xb6 [ 807.270000] [<0003acf0>] get_signal+0x126/0x720 [ 807.270000] [<00039afa>] send_signal+0xde/0x16e [ 807.270000] [<00004f70>] do_notify_resume+0x38/0x61c [ 807.270000] [<0003abaa>] force_sig_fault_to_task+0x36/0x3a [ 807.270000] [<0003abc6>] force_sig_fault+0x18/0x1c [ 807.270000] [<000074f4>] send_fault_sig+0x44/0xc6 [ 807.270000] [<00006a62>] buserr_c+0x2c8/0x6a2 [ 807.270000] [<00002cfc>] do_signal_return+0x10/0x1a [ 807.270000] [<0018800e>] ext4_htree_fill_tree+0x7c/0x32a [ 807.270000] [<0010800a>] d_absolute_path+0x18/0x6a [ 807.270000] [ 807.270000] ---[ end Kernel panic - not syncing: Aiee, killing interrupt handler! ]--- On the Quadra 630, the panic almost completely disappeared when I enabled the relevant CONFIG_DEBUG_* options. After about 7 hours of stress testing I got this: [23982.680000] list_add corruption. next->prev should be prev (00b51e98), but was 00bb22d8. (next=00b75cd0). [23982.690000] kernel BUG at lib/list_debug.c:25! [23982.700000] *** TRAP #7 *** FORMAT=0 [23982.710000] Current process id is 15489 [23982.720000] BAD KERNEL TRAP: 00000000 [23982.740000] Modules linked in: [23982.750000] PC: [<00261e62>] __list_add_valid+0x62/0xc0 [23982.760000] SR: 2000 SP: e2fb938b a2: 00bcba80 [23982.770000] d0: 00000022 d1: 00000002 d2: 008c4e40 d3: 00b7a9c0 [23982.780000] d4: 00b51e98 d5: 000da3c0 a0: 00067f00 a1: 00b51d2c [23982.790000] Process stress-ng (pid: 15489, task=35ee07ca) [23982.800000] Frame format=0 [23982.810000] Stack from 00b51e80: [23982.810000] 004cbab9 004ea3a1 00000019 004ea34f 00b51e98 00bb22d8 00b75cd0 008c4e38 [23982.810000] 00b51ecc 000da3f2 008c4e40 00b51e98 00b75cd0 00b51e5c 000f5d40 00b75cd0 [23982.810000] 00b7a9c0 00bb22d0 00b7a9c0 00b51f04 000dc346 00b51e5c 008c4e38 00b7a9c0 [23982.810000] c4c97000 00000000 c4c96000 00102073 00b14960 c4c97000 00b51e5c 00b75c94 [23982.810000] 00000001 00b51f24 000d5628 00b51e5c 00b75c94 00102070 00000000 00b75c94 [23982.810000] 00b75c94 00b51f3c 000d5728 00b14960 00b75c94 c4c97000 00000000 00b51f78 [23982.830000] Call Trace: [<000da3f2>] anon_vma_chain_link+0x32/0x80 [23982.840000] [<000f5d40>] kmem_cache_alloc+0x0/0x200 [23982.850000] [<000dc346>] anon_vma_clone+0xc6/0x180 [23982.860000] [<00102073>] cdev_get+0x33/0x80 [23982.870000] [<000d5628>] __split_vma+0x68/0x140 [23982.880000] [<00102070>] cdev_get+0x30/0x80 [23982.890000] [<000d5728>] split_vma+0x28/0x40 [23982.900000] [<000d83ba>] mprotect_fixup+0x13a/0x200 [23982.910000] [<00102070>] cdev_get+0x30/0x80 [23982.920000] [<000d8280>] mprotect_fixup+0x0/0x200 [23982.930000] [<000d85b2>] sys_mprotect+0x132/0x1c0 [23982.940000] [<00102070>] cdev_get+0x30/0x80 [23982.950000] [<00001000>] kernel_pg_dir+0x0/0x1000 [23982.960000] [<000071df>] flush_icache_range+0x1f/0x40 [23982.970000] [<00002ca4>] syscall+0x8/0xc [23982.980000] [<00001000>] kernel_pg_dir+0x0/0x1000 [23982.990000] [<00001000>] kernel_pg_dir+0x0/0x1000 [23983.000000] [<00002000>] _start+0x0/0x40 [23983.010000] [<0018800e>] ext4_ext_remove_space+0x20e/0x1540 [23983.030000] [23983.040000] Code: 4879 004e a3a1 4879 004c bab9 4e93 4e47 <b089> 6704 b088 661c 2f08 2f2e 000c 2f00 4879 004e a404 47f9 0043 d16c 4e93 4878 [23983.060000] Disabling lock debugging due to kernel taint I am still unable to reproduce this in Aranym or QEMU. (Though I did find a QEMU bug in the attempt.) I suppose list pointer corruption could have resulted in the above panic had it gone undetected. So it's tempting to blame the panic on bad DRAM -- especially if this anon_vma_chain struct always gets placed at the same physical address (?)