The patch titled Subject: dump_stack: avoid potential deadlocks has been added to the -mm tree. Its filename is dump_stack-avoid-potential-deadlocks.patch This patch should soon appear at http://ozlabs.org/~akpm/mmots/broken-out/dump_stack-avoid-potential-deadlocks.patch and later at http://ozlabs.org/~akpm/mmotm/broken-out/dump_stack-avoid-potential-deadlocks.patch Before you just go and hit "reply", please: a) Consider who else should be cc'ed b) Prefer to cc a suitable mailing list as well c) Ideally: find the original patch on the mailing list and do a reply-to-all to that, adding suitable additional cc's *** Remember to use Documentation/SubmitChecklist when testing your code *** The -mm tree is included into linux-next and is updated there every 3-4 working days ------------------------------------------------------ From: Eric Dumazet <edumazet@xxxxxxxxxx> Subject: dump_stack: avoid potential deadlocks Some servers experienced fatal deadlocks because of a combination of bugs, leading to multiple cpus calling dump_stack(). The checksumming bug was fixed in commit 34ae6a1aa054 ("ipv6: update skb->csum when CE mark is propagated"). The second problem is a faulty locking in dump_stack() CPU1 runs in process context and calls dump_stack(), grabs dump_lock. CPU2 receives a TCP packet under softirq, grabs socket spinlock, and call dump_stack() from netdev_rx_csum_fault(). dump_stack() spins on atomic_cmpxchg(&dump_lock, -1, 2), since dump_lock is owned by CPU1 While dumping its stack, CPU1 is interrupted by a softirq, and happens to process a packet for the TCP socket locked by CPU2. CPU1 spins forever in spin_lock() : deadlock Stack trace on CPU1 looked like : [306295.402231] NMI backtrace for cpu 1 [306295.402238] RIP: 0010:[<ffffffffa9804e95>] [<ffffffffa9804e95>] _raw_spin_lock+0x25/0x30 ... [306295.402255] Stack: [306295.402256] ffff88407f023cb0 ffffffffa99cbdc3 ffff88407f023ca0 ffff88012f496bb0 [306295.402266] ffffffffaa4dc1f0 ffff8820d94f0dc0 000000000000000a ffffffffaa4b4280 [306295.402275] ffff88407f023ce0 ffffffffa98a21d0 ffff88407f023cc0 ffff88407f023ca0 [306295.402284] Call Trace: [306295.402286] <IRQ> [306295.402288] [306295.402291] [<ffffffffa99cbdc3>] tcp_v6_rcv+0x243/0x620 [306295.402304] [<ffffffffa99c38af>] ip6_input_finish+0x11f/0x330 [306295.402309] [<ffffffffa99c3b38>] ip6_input+0x38/0x40 [306295.402313] [<ffffffffa99c3b7c>] ip6_rcv_finish+0x3c/0x90 [306295.402318] [<ffffffffa99c3e79>] ipv6_rcv+0x2a9/0x500 [306295.402323] [<ffffffffa989a1a1>] process_backlog+0x461/0xaa0 [306295.402332] [<ffffffffa9899a57>] net_rx_action+0x147/0x430 [306295.402337] [<ffffffffa980cca7>] __do_softirq+0x167/0x2d0 [306295.402341] [<ffffffffa9d912dc>] call_softirq+0x1c/0x30 [306295.402345] [<ffffffffa980687f>] do_softirq+0x3f/0x80 [306295.402350] [<ffffffffa980cf7e>] irq_exit+0x6e/0xc0 [306295.402355] [<ffffffffa980a5b5>] smp_call_function_single_interrupt+0x35/0x40 [306295.402360] [<ffffffffa9d90bfa>] call_function_single_interrupt+0x6a/0x70 [306295.402361] <EOI> [306295.402364] [306295.402376] [<ffffffffa9a3fa32>] printk+0x4d/0x4f [306295.402390] [<ffffffffa99e1726>] printk_address+0x31/0x33 [306295.402395] [<ffffffffa99e175b>] print_trace_address+0x33/0x3c [306295.402408] [<ffffffffa99e165b>] print_context_stack+0x7f/0x119 [306295.402412] [<ffffffffa99e07f1>] dump_trace+0x26b/0x28e [306295.402417] [<ffffffffa99e17b3>] show_trace_log_lvl+0x4f/0x5c [306295.402421] [<ffffffffa99e0a14>] show_stack_log_lvl+0x104/0x113 [306295.402425] [<ffffffffa99e1819>] show_stack+0x42/0x44 [306295.402429] [<ffffffffa9ba350a>] dump_stack+0x46/0x58 [306295.402434] [<ffffffffa9d0957d>] netdev_rx_csum_fault+0x38/0x3c [306295.402439] [<ffffffffa9988e0e>] __skb_checksum_complete_head+0x6e/0x80 [306295.402444] [<ffffffffa9988e31>] __skb_checksum_complete+0x11/0x20 [306295.402449] [<ffffffffa98acf65>] tcp_rcv_established+0x2bd5/0x2fd0 [306295.402468] [<ffffffffa99cb69c>] tcp_v6_do_rcv+0x13c/0x620 [306295.402477] [<ffffffffa9984be5>] sk_backlog_rcv+0x15/0x30 [306295.402482] [<ffffffffa98931e2>] release_sock+0xd2/0x150 [306295.402486] [<ffffffffa98a51b1>] tcp_recvmsg+0x1c1/0xfc0 [306295.402491] [<ffffffffa98b637d>] inet_recvmsg+0x7d/0x90 [306295.402495] [<ffffffffa9891fcf>] sock_recvmsg+0xaf/0xe0 [306295.402505] [<ffffffffa9892611>] ___sys_recvmsg+0x111/0x3b0 [306295.402528] [<ffffffffa9892f5c>] SyS_recvmsg+0x5c/0xb0 [306295.402532] [<ffffffffa9d8fba2>] system_call_fastpath+0x16/0x1b Fixes: b58d977432c8 ("dump_stack: serialize the output from dump_stack()") Signed-off-by: Eric Dumazet <edumazet@xxxxxxxxxx> Cc: Alex Thorlton <athorlton@xxxxxxx> Cc: <stable@xxxxxxxxxxxxxxx> Signed-off-by: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> --- lib/dump_stack.c | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff -puN lib/dump_stack.c~dump_stack-avoid-potential-deadlocks lib/dump_stack.c --- a/lib/dump_stack.c~dump_stack-avoid-potential-deadlocks +++ a/lib/dump_stack.c @@ -25,6 +25,7 @@ static atomic_t dump_lock = ATOMIC_INIT( asmlinkage __visible void dump_stack(void) { + unsigned long flags; int was_locked; int old; int cpu; @@ -33,9 +34,8 @@ asmlinkage __visible void dump_stack(voi * Permit this cpu to perform nested stack dumps while serialising * against other CPUs */ - preempt_disable(); - retry: + local_irq_save(flags); cpu = smp_processor_id(); old = atomic_cmpxchg(&dump_lock, -1, cpu); if (old == -1) { @@ -43,6 +43,7 @@ retry: } else if (old == cpu) { was_locked = 1; } else { + local_irq_restore(flags); cpu_relax(); goto retry; } @@ -52,7 +53,7 @@ retry: if (!was_locked) atomic_set(&dump_lock, -1); - preempt_enable(); + local_irq_restore(flags); } #else asmlinkage __visible void dump_stack(void) _ Patches currently in -mm which might be from edumazet@xxxxxxxxxx are dump_stack-avoid-potential-deadlocks.patch -- To unsubscribe from this list: send the line "unsubscribe mm-commits" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html