On Mon, 23 Nov 2020 13:08:55 +0200 Leon Romanovsky <leon@xxxxxxxxxx> wrote: > [ 10.028024] Chain exists of: > [ 10.028025] console_owner --> target_list_lock --> _xmit_ETHER#2 Note, the problem is that we have a location that grabs the xmit_lock while holding target_list_lock (and possibly console_owner). > [ 10.028028] > [ 10.028028] Possible interrupt unsafe locking scenario: > [ 10.028029] > [ 10.028029] CPU0 CPU1 > [ 10.028030] ---- ---- > [ 10.028030] lock(_xmit_ETHER#2); > [ 10.028032] local_irq_disable(); > [ 10.028032] lock(console_owner); > [ 10.028034] lock(target_list_lock); > [ 10.028035] <Interrupt> > [ 10.028035] lock(console_owner); > [ 10.028036] > [ 10.028037] *** DEADLOCK *** > [ 10.028037] > [ 10.028107] the dependencies between the lock to be acquired > [ 10.028107] and HARDIRQ-irq-unsafe lock: > [ 10.028108] -> (_xmit_ETHER#2){+.-.}-{2:2} ops: 217 { > [ 10.028110] HARDIRQ-ON-W at: > [ 10.028111] __lock_acquire+0x8bc/0x1a94 > [ 10.028111] lock_acquire.part.0+0x170/0x360 > [ 10.028112] lock_acquire+0x68/0x8c > [ 10.028113] _raw_spin_trylock+0x80/0xd0 > [ 10.028113] virtnet_poll+0xac/0x360 xmit_lock is taken in virtnet_poll() (via virtnet_poll_cleantx()). This is called from the softirq, and interrupts are not disabled. > [ 10.028114] net_rx_action+0x1b0/0x4e0 > [ 10.028115] __do_softirq+0x1f4/0x638 > [ 10.028115] do_softirq+0xb8/0xcc > [ 10.028116] __local_bh_enable_ip+0x18c/0x200 > [ 10.028116] virtnet_napi_enable+0xc0/0xd4 > [ 10.028117] virtnet_open+0x98/0x1c0 > [ 10.028118] __dev_open+0x12c/0x200 > [ 10.028118] __dev_change_flags+0x1a0/0x220 > [ 10.028119] dev_change_flags+0x2c/0x70 > [ 10.028119] do_setlink+0x214/0xe20 > [ 10.028120] __rtnl_newlink+0x514/0x820 > [ 10.028120] rtnl_newlink+0x58/0x84 > [ 10.028121] rtnetlink_rcv_msg+0x184/0x4b4 > [ 10.028122] netlink_rcv_skb+0x60/0x124 > [ 10.028122] rtnetlink_rcv+0x20/0x30 > [ 10.028123] netlink_unicast+0x1b4/0x270 > [ 10.028124] netlink_sendmsg+0x1f0/0x400 > [ 10.028124] sock_sendmsg+0x5c/0x70 > [ 10.028125] ____sys_sendmsg+0x24c/0x280 > [ 10.028125] ___sys_sendmsg+0x88/0xd0 > [ 10.028126] __sys_sendmsg+0x70/0xd0 > [ 10.028127] __arm64_sys_sendmsg+0x2c/0x40 > [ 10.028128] el0_svc_common.constprop.0+0x84/0x200 > [ 10.028128] do_el0_svc+0x2c/0x90 > [ 10.028129] el0_svc+0x18/0x50 > [ 10.028129] el0_sync_handler+0xe0/0x350 > [ 10.028130] el0_sync+0x158/0x180 [..] > [ 10.028171] ... key at: [<ffff80001312aef8>] netdev_xmit_lock_key+0x10/0x390 > [ 10.028171] ... acquired at: > [ 10.028172] __lock_acquire+0x134c/0x1a94 > [ 10.028172] lock_acquire.part.0+0x170/0x360 > [ 10.028173] lock_acquire+0x68/0x8c > [ 10.028173] _raw_spin_lock+0x64/0x90 > [ 10.028174] virtnet_poll_tx+0x84/0x120 > [ 10.028174] netpoll_poll_dev+0x12c/0x350 > [ 10.028175] netpoll_send_skb+0x39c/0x400 > [ 10.028175] netpoll_send_udp+0x2b8/0x440 > [ 10.028176] write_msg+0xfc/0x120 [netconsole] > [ 10.028176] console_unlock+0x3ec/0x6a4 The above shows the problem. We have: console_unlock() (which holds the console_owner lock) write_msg() (which holds the target_list_lock) Then we write_msg() calls: netpoll_send_udp() { netpoll_send_skb() { netpoll_poll_dev() { virtnet_poll_tx() (which takes the xmit_lock!) DEADLOCK! In netpoll_send_skb() I see this: /* tickle device maybe there is some cleanup */ netpoll_poll_dev(np->dev); Which looks to me that it will call some code that should only be used in softirq context. It's called with locks held that are taken in interrupt context, and any locks that are taken in netpoll_poll_dev() must always be taken with interrupts disabled. That is, if xmit_lock is taken within netpoll_poll_dev(), then it must always be taken with interrupts disabled. Otherwise you can have the deadlock that lockdep reported. -- Steve > [ 10.028177] register_console+0x17c/0x2f4 > [ 10.028178] init_netconsole+0x20c/0x1000 [netconsole] > [ 10.028178] do_one_initcall+0x8c/0x480 > [ 10.028179] do_init_module+0x60/0x270 > [ 10.028179] load_module+0x21f8/0x2734 > [ 10.028180] __do_sys_finit_module+0xbc/0x12c > [ 10.028180] __arm64_sys_finit_module+0x28/0x34 > [ 10.028181] el0_svc_common.constprop.0+0x84/0x200 > [ 10.028181] do_el0_svc+0x2c/0x90 > [ 10.028182] el0_svc+0x18/0x50 > [ 10.028182] el0_sync_handler+0xe0/0x350 > [ 10.028183] el0_sync+0x158/0x180 > [ 10.028183] > [ 10.028183] > [ 10.028184] stack backtrace: > [ 10.028185] CPU: 14 PID: 638 Comm: modprobe Not tainted 5.10.0-rc4_for_upstream_min_debug_2020_11_22_19_37 #1 > [ 10.028186] Hardware name: QEMU KVM Virtual Machine, BIOS 0.0.0 02/06/2015 > [ 10.028186] Call trace: > [ 10.028186] dump_backtrace+0x0/0x1d0 > [ 10.028187] show_stack+0x20/0x3c > [ 10.028187] dump_stack+0xec/0x138 > [ 10.028188] check_irq_usage+0x6b8/0x6cc > [ 10.028188] __lock_acquire+0x134c/0x1a94 > [ 10.028189] lock_acquire.part.0+0x170/0x360 > [ 10.028189] lock_acquire+0x68/0x8c > [ 10.028190] _raw_spin_lock+0x64/0x90 > [ 10.028191] virtnet_poll_tx+0x84/0x120 > [ 10.028191] netpoll_poll_dev+0x12c/0x350 > [ 10.028192] netpoll_send_skb+0x39c/0x400 > [ 10.028192] netpoll_send_udp+0x2b8/0x440 > [ 10.028193] write_msg+0xfc/0x120 [netconsole] > [ 10.028193] console_unlock+0x3ec/0x6a4 > [ 10.028194] register_console+0x17c/0x2f4 > [ 10.028194] init_netconsole+0x20c/0x1000 [netconsole] > [ 10.028195] do_one_initcall+0x8c/0x480 > [ 10.028195] do_init_module+0x60/0x270 > [ 10.028196] load_module+0x21f8/0x2734 > [ 10.028197] __do_sys_finit_module+0xbc/0x12c > [ 10.028197] __arm64_sys_finit_module+0x28/0x34 > [ 10.028198] el0_svc_common.constprop.0+0x84/0x200 > [ 10.028198] do_el0_svc+0x2c/0x90 > [ 10.028199] el0_svc+0x18/0x50 > [ 10.028199] el0_sync_handler+0xe0/0x350 > [ 10.028200] el0_sync+0x158/0x180 > [ 10.073569] random: crng init done > [ 10.073964] printk: console [netcon0] enabled > [ 10.074704] random: 7 urandom warning(s) missed due to ratelimiting > [ 10.075340] netconsole: network logging started > _______________________________________________ Virtualization mailing list Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx https://lists.linuxfoundation.org/mailman/listinfo/virtualization