On 26/10/2023 23:06, Andrii Nakryiko wrote: > On Wed, Oct 25, 2023 at 3:28 PM Eduard Zingerman <eddyz87@xxxxxxxxx> wrote: >> >> On Wed, 2023-10-25 at 15:30 -0300, Arnaldo Carvalho de Melo wrote: >>> Em Wed, Oct 25, 2023 at 03:12:49PM -0300, Arnaldo Carvalho de Melo escreveu: >>>> But I guess the acks/reviews + my tests are enough to merge this as-is, >>>> thanks for your work on this! >>> >>> Ok, its in the 'next' branch so that it can go thru: >>> >>> https://github.com/libbpf/libbpf/actions/workflows/pahole.yml >>> >>> But the previous days are all failures, probably something else is >>> preventing this test from succeeding? Andrii? >> >> It looks like the latest run succeeded, while a number of previous >> runs got locked up for some reason. All using the same kernel >> checkpoint commit. I know how to setup local github runner, >> so I can try to replicate this by forking the repo, >> redirecting CI to my machine and executing it several times. >> Will do this over the weekend, need to work on some verifier >> bugs first. >> > > BPF selftests are extremely unreliable under slow Github runners, > unfortunately. Kernel either crashes or locks up very frequently. It > has nothing to do with libbpf and we don't seem to see this in BPF CI > due to having much faster runners there. > > I'm not sure what to do about this apart from trying to identify a > selftest that causes lock up (extremely time consuming endeavor) or > just wait till libbpf CI will be privileged enough to gain its own > fast AWS-based worker :) > > But it seems like the last scheduled run succeeded, I think you are good. > Perhaps related, I've starting seeing regular RCU stalls with bpf-next based kernels when running selftests that seem to have the kprobe_multi_link_prog_run tests as a common factor. The system goes out to lunch for a while and sometimes returns depending on which CPUs are hit. In the same run I've also just run into a test hang after the send_signal_sched_switch test completes; judging by the stack traces and the alphabetical order it looks like setget_sockopt.c is the culprit there. Seems to be a socket locking issue. dmesg output follows for both issues.. [ 4151.256451] rcu: INFO: rcu_preempt detected stalls on CPUs/tasks: [ 4151.273835] rcu: 2-....: (4 ticks this GP) idle=2d34/1/0x4000000000000000 softirq=194674/194674 fqs=3185 [ 4151.293115] rcu: (detected by 0, t=60043 jiffies, g=201485, q=5607 ncpus=8) [ 4151.309961] Sending NMI from CPU 0 to CPUs 2: [ 4151.322250] NMI backtrace for cpu 2 [ 4151.322295] CPU: 2 PID: 1101 Comm: systemd-journal Kdump: loaded Tainted: G O 6.6.0-rc7+ #159 [ 4151.322327] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.5.1 06/16/2021 [ 4151.322362] RIP: 0010:fprobe_handler+0x2d/0x2b0 [ 4151.322430] Code: fa 55 48 89 e5 41 57 49 89 cf 41 56 41 55 49 89 f5 41 54 49 89 fc 53 48 89 d3 48 83 ec 08 48 85 d2 74 0d f6 82 c8 00 00 00 01 <0f> 85 4f 01 00 00 65 4c 8b 34 25 80 27 03 00 49 8b b6 d0 13 00 00 [ 4151.322463] RSP: 0018:ffffc900001208a0 EFLAGS: 00000046 [ 4151.322525] RAX: ffff88811336b380 RBX: ffff88810614ca40 RCX: ffffc900001208e0 [ 4151.322558] RDX: ffff88810614ca40 RSI: ffffffffb4666b5d RDI: ffffffffb4584884 [ 4151.322585] RBP: ffffc900001208d0 R08: 0000000000000000 R09: 0000000000000000 [ 4151.322619] R10: 0000000000000000 R11: 0000000000000000 R12: ffffffffb4584884 [ 4151.322658] R13: ffffffffb4666b5d R14: ffffffffffffffef R15: ffffc900001208e0 [ 4151.322694] FS: 00007f7ec44b4a40(0000) GS:ffff888237c80000(0000) knlGS:0000000000000000 [ 4151.322725] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 [ 4151.322757] CR2: 00007f7ebb9462d0 CR3: 0000000101b10003 CR4: 0000000000770ee0 [ 4151.322789] PKRU: 55555554 [ 4151.322818] Call Trace: [ 4151.322856] <NMI> [ 4151.322926] ? show_regs+0x69/0x80 [ 4151.323045] ? nmi_cpu_backtrace+0xa0/0x110 [ 4151.323271] ? nmi_cpu_backtrace_handler+0x15/0x20 [ 4151.323356] ? nmi_handle+0x64/0x160 [ 4151.323526] ? fprobe_handler+0x2d/0x2b0 [ 4151.323687] ? default_do_nmi+0x6e/0x180 [ 4151.323855] ? exc_nmi+0x151/0x1d0 [ 4151.324015] ? end_repeat_nmi+0x16/0x67 [ 4151.324114] ? kprobe_multi_link_prog_run.isra.0+0x9d/0x130 [ 4151.324172] ? __rcu_read_unlock+0x4/0x40 [ 4151.324459] ? kprobe_multi_link_prog_run.isra.0+0x9d/0x130 [ 4151.324509] ? __rcu_read_unlock+0x4/0x40 [ 4151.324585] ? fprobe_handler+0x2d/0x2b0 [ 4151.324755] ? fprobe_handler+0x2d/0x2b0 [ 4151.324958] ? fprobe_handler+0x2d/0x2b0 [ 4151.325102] </NMI> [ 4151.325130] <IRQ> [ 4151.325749] ? __rcu_read_unlock+0x9/0x40 [ 4151.325918] ? __rcu_read_unlock+0x9/0x40 [ 4151.325996] ? kprobe_multi_link_prog_run.isra.0+0x9d/0x130 [ 4151.326097] ? __rcu_read_unlock+0x9/0x40 [ 4151.326146] ? kprobe_multi_link_prog_run.isra.0+0x9d/0x130 [ 4151.326224] ? __pfx_copy_from_kernel_nofault_allowed+0x10/0x10 [ 4151.326360] ? copy_from_kernel_nofault+0x26/0xf0 [ 4151.326460] ? kprobe_multi_link_handler+0x28/0x40 [ 4151.326542] ? copy_from_kernel_nofault_allowed+0x4/0x50 [ 4151.326615] ? fprobe_handler+0x12f/0x2b0 [ 4151.326696] ? __pfx_irq_enter_rcu+0x10/0x10 [ 4151.327030] ? __pfx_irq_enter_rcu+0x10/0x10 [ 4151.327282] ? __pfx_irq_enter_rcu+0x10/0x10 [ 4151.327357] ? copy_from_kernel_nofault_allowed+0x9/0x50 [ 4151.327545] ? copy_from_kernel_nofault_allowed+0x9/0x50 [ 4151.327631] ? copy_from_kernel_nofault+0x26/0xf0 [ 4151.327737] ? copy_from_kernel_nofault_allowed+0x9/0x50 [ 4151.327786] ? copy_from_kernel_nofault+0x26/0xf0 [ 4151.327833] ? irq_enter_rcu+0x4/0x80 [ 4151.327884] ? __pfx_irq_enter_rcu+0x10/0x10 [ 4151.327934] ? sysvec_irq_work+0x4f/0xe0 [ 4151.328011] ? get_entry_ip+0x32/0x80 [ 4151.328191] ? kprobe_multi_link_handler+0x19/0x40 [ 4151.328271] ? irq_enter_rcu+0x4/0x80 [ 4151.328349] ? fprobe_handler+0x12f/0x2b0 [ 4151.328904] ? early_xen_iret_patch+0xc/0xc [ 4151.329053] ? irq_enter_rcu+0x9/0x80 [ 4151.329230] ? irq_enter_rcu+0x9/0x80 [ 4151.329308] ? sysvec_irq_work+0x4f/0xe0 [ 4151.329415] ? irq_enter_rcu+0x9/0x80 [ 4151.329465] ? sysvec_irq_work+0x4f/0xe0 [ 4151.329599] ? asm_sysvec_irq_work+0x1f/0x30 [ 4151.330062] ? rcu_read_unlock_special+0xd5/0x190 [ 4151.330211] ? rcu_read_unlock_special+0xcf/0x190 [ 4151.330419] ? __rcu_read_unlock+0x33/0x40 [ 4151.330495] ? uncharge_batch+0xe8/0x140 [ 4151.330618] ? __mem_cgroup_uncharge+0x70/0x90 [ 4151.330860] ? __folio_put+0x36/0x80 [ 4151.330964] ? free_page_and_swap_cache+0x87/0x90 [ 4151.331064] ? tlb_remove_table_rcu+0x2c/0x50 [ 4151.331189] ? rcu_do_batch+0x1bf/0x550 [ 4151.331264] ? rcu_do_batch+0x159/0x550 [ 4151.331614] ? rcu_core+0x183/0x370 [ 4151.331794] ? rcu_core_si+0x12/0x20 [ 4151.331870] ? __do_softirq+0x10e/0x327 [ 4151.332187] ? __irq_exit_rcu+0xa8/0x110 [ 4151.332295] ? irq_exit_rcu+0x12/0x20 [ 4151.332374] ? sysvec_apic_timer_interrupt+0xb0/0xe0 [ 4151.332450] </IRQ> [ 4151.332479] <TASK> [ 4151.332595] ? asm_sysvec_apic_timer_interrupt+0x1f/0x30 [ 4151.333055] ? _raw_spin_unlock_irqrestore+0x21/0x50 [ 4151.333227] ? __wake_up_common_lock+0x8a/0xc0 [ 4151.333644] ? __wake_up+0x17/0x20 [ 4151.333717] ? fsnotify_insert_event+0x121/0x1a0 [ 4151.334004] ? inotify_handle_inode_event+0x158/0x240 [ 4151.334336] ? fsnotify_handle_inode_event.isra.0+0x6f/0xe0 [ 4151.334410] ? send_to_group+0x2ad/0x320 [ 4151.334771] ? fsnotify+0x25c/0x5a0 [ 4151.335385] ? __fsnotify_parent+0x21b/0x330 [ 4151.335698] ? __fsnotify_parent+0x9/0x330 [ 4151.335963] ? notify_change+0x42f/0x4c0 [ 4151.336018] ? __fsnotify_parent+0x9/0x330 [ 4151.336070] ? notify_change+0x42f/0x4c0 [ 4151.336173] ? notify_change+0x9/0x4c0 [ 4151.336474] ? do_truncate+0x87/0xe0 [ 4151.336528] ? do_truncate+0x87/0xe0 [ 4151.336960] ? do_sys_ftruncate+0x1af/0x210 [ 4151.337209] ? __x64_sys_ftruncate+0x1e/0x30 [ 4151.337285] ? do_syscall_64+0x3e/0x90 [ 4151.337382] ? entry_SYSCALL_64_after_hwframe+0x6e/0xd8 [ 4151.337984] </TASK> [ 4151.338018] INFO: NMI handler (nmi_cpu_backtrace_handler) took too long to run: 15.770 msecs [ 4154.063929] rcu: INFO: rcu_preempt detected expedited stalls on CPUs/tasks: { } 72585 jiffies s: 2429 root: 0x0/. [ 4200.980354] 8021q: 802.1Q VLAN Support v1.8 [ 4200.982157] 8021q: adding VLAN 0 to HW filter on device ens3 [ 4254.123516] test_progs[47486] is installing a program with bpf_probe_write_user helper that may corrupt user memory! [ 4254.123530] test_progs[47486] is installing a program with bpf_probe_write_user helper that may corrupt user memory! This second issue causes a hang in selftest execution as described above after send_signal_sched_switch [ 4263.776211] ------------[ cut here ]------------ [ 4263.781022] Voluntary context switch within RCU read-side critical section! [ 4263.781036] WARNING: CPU: 2 PID: 47486 at kernel/rcu/tree_plugin.h:320 rcu_note_context_switch+0x2dd/0x320 [ 4263.783303] Modules linked in: sch_fq 8021q garp mrp dummy tun ipip tunnel4 ip_tunnel bpf_preload bpf_testmod(O) veth cls_bpf sch_ingress ipt_REJECT xt_comment xt_owner nft_compat nft_fib_inet nft_fib_ipv4 nft_fib_ipv6 nft_fib nft_reject_inet nf_reject_ipv4 nf_reject_ipv6 nft_reject nft_ct nft_chain_nat nf_nat ip_set cuse vfat fat intel_rapl_msr intel_rapl_common bochs kvm_amd drm_vram_helper ccp drm_ttm_helper ttm kvm drm_kms_helper irqbypass drm joydev pcspkr i2c_piix4 sch_fq_codel xfs iscsi_tcp libiscsi_tcp libiscsi nvme_tcp nvme_fabrics nvme nvme_core sd_mod t10_pi crc64_rocksoft_generic crc64_rocksoft sg virtio_net net_failover failover virtio_scsi crct10dif_pclmul crc32_pclmul ghash_clmulni_intel sha512_ssse3 ata_generic pata_acpi ata_piix aesni_intel crypto_simd cryptd virtio_pci virtio_pci_legacy_dev serio_raw virtio_pci_modern_dev libata floppy qemu_fw_cfg dm_multipath sunrpc dm_mirror dm_region_hash dm_log dm_mod scsi_transport_iscsi ipmi_devintf ipmi_msghandler fuse [last unloaded: bpf_testmod(O)] [ 4263.795254] CPU: 2 PID: 47486 Comm: new_name Kdump: loaded Tainted: G O 6.6.0-rc7+ #159 [ 4263.796602] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.5.1 06/16/2021 [ 4263.797804] RIP: 0010:rcu_note_context_switch+0x2dd/0x320 [ 4263.798772] Code: 08 f0 83 44 24 fc 00 48 89 de 4c 89 f7 e8 5b 98 ff ff e9 f0 fd ff ff 48 c7 c7 d0 00 d3 b5 c6 05 96 1b 3b 02 01 e8 c3 19 f4 ff <0f> 0b e9 83 fd ff ff a9 ff ff ff 7f 0f 84 3f fe ff ff 65 48 8b 3c [ 4263.801495] RSP: 0018:ffffc9000f4f7b68 EFLAGS: 00010082 [ 4263.802493] RAX: 0000000000000000 RBX: ffff888237cb3e00 RCX: 0000000000000000 [ 4263.803676] RDX: 0000000000000003 RSI: ffffffffb5cdb4f1 RDI: 00000000ffffffff [ 4263.804842] RBP: ffffc9000f4f7b88 R08: 0000000000000000 R09: ffffc9000f4f79d0 [ 4263.805989] R10: 0000000000000003 R11: ffffffffb67f4908 R12: 0000000000000000 [ 4263.807158] R13: ffff8881003acd40 R14: 0000000000000001 R15: 00000000000000e1 [ 4263.808334] FS: 00007fd0577417c0(0000) GS:ffff888237c80000(0000) knlGS:0000000000000000 [ 4263.809625] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 [ 4263.810690] CR2: 0000000006954568 CR3: 000000010ecb2004 CR4: 0000000000770ee0 [ 4263.811903] PKRU: 55555554 [ 4263.812656] Call Trace: [ 4263.813401] <TASK> [ 4263.814095] ? show_regs+0x69/0x80 [ 4263.814897] ? __warn+0x8d/0x150 [ 4263.815713] ? rcu_note_context_switch+0x2dd/0x320 [ 4263.816686] ? report_bug+0x196/0x1c0 [ 4263.817549] ? srso_alias_return_thunk+0x5/0x7f [ 4263.818503] ? irq_work_queue+0x13/0x60 [ 4263.819387] ? handle_bug+0x45/0x80 [ 4263.820219] ? exc_invalid_op+0x1c/0x70 [ 4263.821089] ? asm_exc_invalid_op+0x1f/0x30 [ 4263.821994] ? rcu_note_context_switch+0x2dd/0x320 [ 4263.822961] ? rcu_note_context_switch+0x2dd/0x320 [ 4263.823916] __schedule+0x60/0x770 [ 4263.824765] ? __raw_spin_lock_irqsave+0x23/0x60 [ 4263.825720] ? srso_alias_return_thunk+0x5/0x7f [ 4263.826662] ? _raw_spin_unlock_irqrestore+0x2b/0x50 [ 4263.827617] schedule+0x6a/0xf0 [ 4263.828383] __lock_sock+0x7d/0xc0 [ 4263.829178] ? __pfx_autoremove_wake_function+0x10/0x10 [ 4263.830145] lock_sock_nested+0x54/0x60 [ 4263.830982] do_ip_setsockopt+0x10b4/0x11a0 [ 4263.831859] ? srso_alias_return_thunk+0x5/0x7f [ 4263.832766] ? srso_alias_return_thunk+0x5/0x7f [ 4263.833655] sol_ip_sockopt+0x3c/0x80 [ 4263.834444] __bpf_setsockopt+0x5f/0xa0 [ 4263.835222] bpf_sock_ops_setsockopt+0x19/0x30 [ 4263.836064] bpf_prog_e2095e3611d57c9f_bpf_test_sockopt_int+0xbc/0x127 [ 4263.837110] bpf_prog_493685a3bae00bbd_bpf_test_ip_sockopt+0x49/0x4f [ 4263.838124] bpf_prog_e80acd4ed535c5e2_skops_sockopt+0x433/0xe95 [ 4263.839102] ? migrate_disable+0x7b/0x90 [ 4263.839865] __cgroup_bpf_run_filter_sock_ops+0x91/0x140 [ 4263.840769] __inet_listen_sk+0x106/0x130 [ 4263.841529] ? lock_sock_nested+0x43/0x60 [ 4263.842259] inet_listen+0x4d/0x70 [ 4263.842935] __sys_listen+0x75/0xc0 [ 4263.843624] __x64_sys_listen+0x18/0x20 [ 4263.844309] do_syscall_64+0x3e/0x90 [ 4263.844969] entry_SYSCALL_64_after_hwframe+0x6e/0xd8 [ 4263.845788] RIP: 0033:0x7fd156f9c82b [ 4263.846461] Code: f0 ff ff 73 01 c3 48 8b 0d 52 46 38 00 f7 d8 64 89 01 48 83 c8 ff c3 0f 1f 84 00 00 00 00 00 f3 0f 1e fa b8 32 00 00 00 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 8b 0d 25 46 38 00 f7 d8 64 89 01 48 [ 4263.848837] RSP: 002b:00007ffda4b93ed8 EFLAGS: 00000246 ORIG_RAX: 0000000000000032 [ 4263.849892] RAX: ffffffffffffffda RBX: 00000000068d85b0 RCX: 00007fd156f9c82b [ 4263.850912] RDX: 0000000000000010 RSI: 0000000000000001 RDI: 000000000000002c [ 4263.851932] RBP: 00007ffda4b93f20 R08: 0000000000000010 R09: 0000000000000000 [ 4263.852953] R10: 00007ffda4b93eb0 R11: 0000000000000246 R12: 0000000000411fa0 [ 4263.853975] R13: 00007ffda4b94370 R14: 0000000000000000 R15: 0000000000000000 [ 4263.855011] </TASK> [ 4263.855564] ---[ end trace 0000000000000000 ]--- [ 4323.847899] rcu: INFO: rcu_preempt detected stalls on CPUs/tasks: [ 4323.849648] rcu: Tasks blocked on level-0 rcu_node (CPUs 0-7): P47486/1:b..l [ 4323.850749] rcu: (detected by 4, t=60004 jiffies, g=212293, q=44343 ncpus=8) [ 4323.851852] task:new_name state:D stack:0 pid:47486 ppid:47485 flags:0x00000000 [ 4323.853085] Call Trace: [ 4323.853711] <TASK> [ 4323.854296] __schedule+0x265/0x770 [ 4323.855023] ? __raw_spin_lock_irqsave+0x23/0x60 [ 4323.855861] schedule+0x6a/0xf0 [ 4323.856547] __lock_sock+0x7d/0xc0 [ 4323.857271] ? __pfx_autoremove_wake_function+0x10/0x10 [ 4323.858168] lock_sock_nested+0x54/0x60 [ 4323.858923] do_ip_setsockopt+0x10b4/0x11a0 [ 4323.859710] ? srso_alias_return_thunk+0x5/0x7f [ 4323.860545] ? srso_alias_return_thunk+0x5/0x7f [ 4323.861373] sol_ip_sockopt+0x3c/0x80 [ 4323.862093] __bpf_setsockopt+0x5f/0xa0 [ 4323.862837] bpf_sock_ops_setsockopt+0x19/0x30 [ 4323.863643] bpf_prog_e2095e3611d57c9f_bpf_test_sockopt_int+0xbc/0x127 [ 4323.864668] bpf_prog_493685a3bae00bbd_bpf_test_ip_sockopt+0x49/0x4f [ 4323.865678] bpf_prog_e80acd4ed535c5e2_skops_sockopt+0x433/0xe95 [ 4323.866664] ? migrate_disable+0x7b/0x90 [ 4323.867441] __cgroup_bpf_run_filter_sock_ops+0x91/0x140 [ 4323.868356] __inet_listen_sk+0x106/0x130 [ 4323.869117] ? lock_sock_nested+0x43/0x60 [ 4323.869885] inet_listen+0x4d/0x70 [ 4323.870595] __sys_listen+0x75/0xc0 [ 4323.871300] __x64_sys_listen+0x18/0x20 [ 4323.872042] do_syscall_64+0x3e/0x90 [ 4323.872761] entry_SYSCALL_64_after_hwframe+0x6e/0xd8 [ 4323.873637] RIP: 0033:0x7fd156f9c82b [ 4323.874387] RSP: 002b:00007ffda4b93ed8 EFLAGS: 00000246 ORIG_RAX: 0000000000000032 [ 4323.875496] RAX: ffffffffffffffda RBX: 00000000068d85b0 RCX: 00007fd156f9c82b [ 4323.876583] RDX: 0000000000000010 RSI: 0000000000000001 RDI: 000000000000002c [ 4323.877679] RBP: 00007ffda4b93f20 R08: 0000000000000010 R09: 0000000000000000 [ 4323.878774] R10: 00007ffda4b93eb0 R11: 0000000000000246 R12: 0000000000411fa0 [ 4323.879866] R13: 00007ffda4b94370 R14: 0000000000000000 R15: 0000000000000000 [ 4323.880952] </TASK> [ 4334.246892] rcu: INFO: rcu_preempt detected expedited stalls on CPUs/tasks: { P47486 } 61561 jiffies s: 3197 root: 0x0/T [ 4334.248460] rcu: blocking rcu_node structures (internal RCU debug): [ 4424.359789] INFO: task new_name:47486 blocked for more than 122 seconds. [ 4424.361458] Tainted: G W O 6.6.0-rc7+ #159 [ 4424.362418] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. [ 4424.363570] task:new_name state:D stack:0 pid:47486 ppid:47485 flags:0x00000000 [ 4424.364799] Call Trace: [ 4424.365430] <TASK> [ 4424.368934] __schedule+0x265/0x770 [ 4424.369663] ? __raw_spin_lock_irqsave+0x23/0x60 [ 4424.370499] schedule+0x6a/0xf0 [ 4424.371189] __lock_sock+0x7d/0xc0 [ 4424.371906] ? __pfx_autoremove_wake_function+0x10/0x10 [ 4424.372794] lock_sock_nested+0x54/0x60 [ 4424.373532] do_ip_setsockopt+0x10b4/0x11a0 [ 4424.374302] ? srso_alias_return_thunk+0x5/0x7f [ 4424.375127] ? srso_alias_return_thunk+0x5/0x7f [ 4424.375947] sol_ip_sockopt+0x3c/0x80 [ 4424.376653] __bpf_setsockopt+0x5f/0xa0 [ 4424.377381] bpf_sock_ops_setsockopt+0x19/0x30 [ 4424.378184] bpf_prog_e2095e3611d57c9f_bpf_test_sockopt_int+0xbc/0x127 [ 4424.379190] bpf_prog_493685a3bae00bbd_bpf_test_ip_sockopt+0x49/0x4f [ 4424.380188] bpf_prog_e80acd4ed535c5e2_skops_sockopt+0x433/0xe95 [ 4424.381150] ? migrate_disable+0x7b/0x90 [ 4424.381885] __cgroup_bpf_run_filter_sock_ops+0x91/0x140 [ 4424.382762] __inet_listen_sk+0x106/0x130 [ 4424.383515] ? lock_sock_nested+0x43/0x60 [ 4424.384290] inet_listen+0x4d/0x70 [ 4424.384992] __sys_listen+0x75/0xc0 [ 4424.385690] __x64_sys_listen+0x18/0x20 [ 4424.386425] do_syscall_64+0x3e/0x90 [ 4424.387150] entry_SYSCALL_64_after_hwframe+0x6e/0xd8 [ 4424.387991] RIP: 0033:0x7fd156f9c82b [ 4424.388702] RSP: 002b:00007ffda4b93ed8 EFLAGS: 00000246 ORIG_RAX: 0000000000000032 [ 4424.389804] RAX: ffffffffffffffda RBX: 00000000068d85b0 RCX: 00007fd156f9c82b [ 4424.390862] RDX: 0000000000000010 RSI: 0000000000000001 RDI: 000000000000002c [ 4424.391913] RBP: 00007ffda4b93f20 R08: 0000000000000010 R09: 0000000000000000 [ 4424.392987] R10: 00007ffda4b93eb0 R11: 0000000000000246 R12: 0000000000411fa0 [ 4424.394047] R13: 00007ffda4b94370 R14: 0000000000000000 R15: 0000000000000000 [ 4424.395117] </TASK> [ 4424.395701] INFO: task dtprobed:49514 blocked for more than 122 seconds. [ 4424.396751] Tainted: G W O 6.6.0-rc7+ #159 [ 4424.397696] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. [ 4424.398858] task:dtprobed state:D stack:0 pid:49514 ppid:1 flags:0x00000002 [ 4424.400036] Call Trace: [ 4424.400651] <TASK> [ 4424.401258] __schedule+0x265/0x770 [ 4424.401975] ? srso_alias_return_thunk+0x5/0x7f [ 4424.402795] schedule+0x6a/0xf0 [ 4424.403472] synchronize_rcu_expedited+0x185/0x200 [ 4424.404313] ? __pfx_wait_rcu_exp_gp+0x10/0x10 [ 4424.405101] ? __pfx_autoremove_wake_function+0x10/0x10 [ 4424.405966] ? srso_alias_return_thunk+0x5/0x7f [ 4424.406758] ? preempt_count_add+0x52/0xc0 [ 4424.407499] namespace_unlock+0xd6/0x1b0 [ 4424.408234] put_mnt_ns+0x74/0xa0 [ 4424.408903] free_nsproxy+0x1f/0x1b0 [ 4424.409609] exit_task_namespaces+0x6f/0x90 [ 4424.410388] do_exit+0x28b/0x520 [ 4424.411041] ? srso_alias_return_thunk+0x5/0x7f [ 4424.411837] ? srso_alias_return_thunk+0x5/0x7f [ 4424.412635] do_group_exit+0x39/0x90 [ 4424.413351] __x64_sys_exit_group+0x1c/0x20 [ 4424.414106] do_syscall_64+0x3e/0x90 [ 4424.414810] entry_SYSCALL_64_after_hwframe+0x6e/0xd8 [ 4424.415650] RIP: 0033:0x7ff70a952cf6 [ 4424.416373] RSP: 002b:00007ffeecec40b8 EFLAGS: 00000246 ORIG_RAX: 00000000000000e7 [ 4424.417494] RAX: ffffffffffffffda RBX: 00007ff70ac14820 RCX: 00007ff70a952cf6 [ 4424.418560] RDX: 0000000000000002 RSI: 000000000000003c RDI: 0000000000000002 [ 4424.419636] RBP: 0000000000000002 R08: 00000000000000e7 R09: ffffffffffffff68 [ 4424.420689] R10: 00007ffeecec3fbf R11: 0000000000000246 R12: 00007ff70ac14820 [ 4424.421768] R13: 0000000000000001 R14: 00007ff70ac1a310 R15: 0000000000000000 [ 4424.422837] </TASK> >> Thanks, >> Eduard.