> > Hi, this is your Linux kernel regression tracker. Top-posting for once, > to make this easily accessible to everyone. > > On 06.01.23 12:11, Linux kernel regression tracking (#adding) wrote: > > On 06.01.23 08:55, Jaroslav Pulchart wrote: > >> Hello, > >> > >> I would like to report a >= 6.1.2 some network regression (looks like > >> NIC us not UP) on our Dell R7525 servers with E810 NICs. The issue was > >> observed after I updated 6.1.0 to 6.1.2 or newer (tested up to newest > >> 6.1.4-rc1). The system is not accesible and all services are in D > >> state after each reboot. > > Can anyone please provide a status on this? It seems to take quite a > while to get this regression fixed, which is unfortunate. Or was > progress made somewhere and I just missed it? > > I noticed Tony tried to address this in mainline, but the last thing I'm > aware of is "Please ignore/drop this. Just saw that this change doesn't > solve the issue." here: > FYI: We are building 6.1.y with the provided patch to fix the regression in our environment. > https://lore.kernel.org/all/b944d1d4-7f90-dcef-231c-91bb031a4275@xxxxxxxxx/#t > > Should the backport to 6.1.y (425c9bd06b7a ) maybe be dropped to at > least resolve the issue there until this is fixed in mainline? Or would > that cause a regression as well? > > Ciao, Thorsten (wearing his 'the Linux kernel's regression tracker' hat) > -- > Everything you wanna know about Linux kernel regression tracking: > https://linux-regtracking.leemhuis.info/about/#tldr > If I did something stupid, please tell me, as explained on that page. > > #regzbot poke > > >> [ 257.625207] Tainted: G E 6.1.4-0.gdc.el9.x86_64 #1 > >> [ 257.631911] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > >> disables this message. > >> [ 257.639740] task:kworker/u192:1 state:D stack:0 pid:11 > >> ppid:2 flags:0x00004000 > >> [ 257.648095] Workqueue: netns cleanup_net > >> [ 257.652029] Call Trace: > >> [ 257.654481] <TASK> > >> [ 257.656589] __schedule+0x1eb/0x630 > >> [ 257.660087] schedule+0x5a/0xd0 > >> [ 257.663233] schedule_preempt_disabled+0x11/0x20 > >> [ 257.667851] __mutex_lock.constprop.0+0x372/0x6c0 > >> [ 257.672561] rdma_dev_change_netns+0x25/0x120 [ib_core] > >> [ 257.677821] rdma_dev_exit_net+0x139/0x1e0 [ib_core] > >> [ 257.682804] ops_exit_list+0x30/0x70 > >> [ 257.686382] cleanup_net+0x213/0x3b0 > >> [ 257.689964] process_one_work+0x1e2/0x3b0 > >> [ 257.693984] ? rescuer_thread+0x390/0x390 > >> [ 257.697995] worker_thread+0x50/0x3a0 > >> [ 257.701661] ? rescuer_thread+0x390/0x390 > >> [ 257.705674] kthread+0xd6/0x100 > >> [ 257.708819] ? kthread_complete_and_exit+0x20/0x20 > >> [ 257.713613] ret_from_fork+0x1f/0x30 > >> [ 257.717192] </TASK> > >> [ 257.719496] INFO: task kworker/87:0:470 blocked for more than 122 seconds. > >> [ 257.726423] Tainted: G E 6.1.4-0.gdc.el9.x86_64 #1 > >> [ 257.733123] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > >> disables this message. > >> [ 257.740949] task:kworker/87:0 state:D stack:0 pid:470 > >> ppid:2 flags:0x00004000 > >> [ 257.749307] Workqueue: events linkwatch_event > >> [ 257.753672] Call Trace: > >> [ 257.756124] <TASK> > >> [ 257.758228] __schedule+0x1eb/0x630 > >> [ 257.761723] schedule+0x5a/0xd0 > >> [ 257.764867] schedule_preempt_disabled+0x11/0x20 > >> [ 257.769487] __mutex_lock.constprop.0+0x372/0x6c0 > >> [ 257.774196] ? pick_next_task+0x57/0x9b0 > >> [ 257.778127] ? finish_task_switch.isra.0+0x8f/0x2a0 > >> [ 257.783007] linkwatch_event+0xa/0x30 > >> [ 257.786674] process_one_work+0x1e2/0x3b0 > >> [ 257.790687] worker_thread+0x50/0x3a0 > >> [ 257.794352] ? rescuer_thread+0x390/0x390 > >> [ 257.798365] kthread+0xd6/0x100 > >> [ 257.801513] ? kthread_complete_and_exit+0x20/0x20 > >> [ 257.806303] ret_from_fork+0x1f/0x30 > >> [ 257.809885] </TASK> > >> [ 257.812109] INFO: task kworker/39:1:614 blocked for more than 123 seconds. > >> [ 257.818984] Tainted: G E 6.1.4-0.gdc.el9.x86_64 #1 > >> [ 257.825686] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > >> disables this message. > >> [ 257.833519] task:kworker/39:1 state:D stack:0 pid:614 > >> ppid:2 flags:0x00004000 > >> [ 257.841869] Workqueue: infiniband ib_cache_event_task [ib_core] > >> [ 257.847802] Call Trace: > >> [ 257.850252] <TASK> > >> [ 257.852360] __schedule+0x1eb/0x630 > >> [ 257.855851] schedule+0x5a/0xd0 > >> [ 257.858998] schedule_preempt_disabled+0x11/0x20 > >> [ 257.863617] __mutex_lock.constprop.0+0x372/0x6c0 > >> [ 257.868325] ib_get_eth_speed+0x65/0x190 [ib_core] > >> [ 257.873127] ? ib_cache_update.part.0+0x4b/0x2b0 [ib_core] > >> [ 257.878619] ? __kmem_cache_alloc_node+0x18c/0x2b0 > >> [ 257.883417] irdma_query_port+0xb3/0x110 [irdma] > >> [ 257.888051] ib_query_port+0xaa/0x100 [ib_core] > >> [ 257.892601] ib_cache_update.part.0+0x65/0x2b0 [ib_core] > >> [ 257.897924] ? pick_next_task+0x57/0x9b0 > >> [ 257.901855] ? dequeue_task_fair+0xb6/0x3c0 > >> [ 257.906043] ? finish_task_switch.isra.0+0x8f/0x2a0 > >> [ 257.910920] ib_cache_event_task+0x58/0x80 [ib_core] > >> [ 257.915906] process_one_work+0x1e2/0x3b0 > >> [ 257.919918] ? rescuer_thread+0x390/0x390 > >> [ 257.923931] worker_thread+0x50/0x3a0 > >> [ 257.927595] ? rescuer_thread+0x390/0x390 > >> [ 257.931609] kthread+0xd6/0x100 > >> [ 257.934755] ? kthread_complete_and_exit+0x20/0x20 > >> [ 257.939549] ret_from_fork+0x1f/0x30 > >> [ 257.943128] </TASK> > >> [ 257.945438] INFO: task NetworkManager:3387 blocked for more than 123 seconds. > >> [ 257.952577] Tainted: G E 6.1.4-0.gdc.el9.x86_64 #1 > >> [ 257.959274] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > >> disables this message. > >> [ 257.967099] task:NetworkManager state:D stack:0 pid:3387 > >> ppid:1 flags:0x00004002 > >> [ 257.975446] Call Trace: > >> [ 257.977901] <TASK> > >> [ 257.980004] __schedule+0x1eb/0x630 > >> [ 257.983498] schedule+0x5a/0xd0 > >> [ 257.986641] schedule_timeout+0x11d/0x160 > >> [ 257.990654] __wait_for_common+0x90/0x1e0 > >> [ 257.994666] ? usleep_range_state+0x90/0x90 > >> [ 257.998854] __flush_workqueue+0x13a/0x3f0 > >> [ 258.002955] ? __kernfs_remove.part.0+0x11e/0x1e0 > >> [ 258.007661] ib_cache_cleanup_one+0x1c/0xe0 [ib_core] > >> [ 258.012721] __ib_unregister_device+0x62/0xa0 [ib_core] > >> [ 258.017959] ib_unregister_device+0x22/0x30 [ib_core] > >> [ 258.023024] irdma_remove+0x1a/0x60 [irdma] > >> [ 258.027223] auxiliary_bus_remove+0x18/0x30 > >> [ 258.031414] device_release_driver_internal+0x1aa/0x230 > >> [ 258.036643] bus_remove_device+0xd8/0x150 > >> [ 258.040654] device_del+0x18b/0x3f0 > >> [ 258.044149] ice_unplug_aux_dev+0x42/0x60 [ice] > >> [ 258.048707] ice_lag_changeupper_event+0x287/0x2a0 [ice] > >> [ 258.054038] ice_lag_event_handler+0x51/0x130 [ice] > >> [ 258.058930] raw_notifier_call_chain+0x41/0x60 > >> [ 258.063381] __netdev_upper_dev_link+0x1a0/0x370 > >> [ 258.068008] netdev_master_upper_dev_link+0x3d/0x60 > >> [ 258.072886] bond_enslave+0xd16/0x16f0 [bonding] > >> [ 258.077517] ? nla_put+0x28/0x40 > >> [ 258.080756] do_setlink+0x26c/0xc10 > >> [ 258.084249] ? avc_alloc_node+0x27/0x180 > >> [ 258.088173] ? __nla_validate_parse+0x141/0x190 > >> [ 258.092708] __rtnl_newlink+0x53a/0x620 > >> [ 258.096549] rtnl_newlink+0x44/0x70 > >> [ 258.100040] rtnetlink_rcv_msg+0x159/0x3d0 > >> [ 258.104140] ? rtnl_calcit.isra.0+0x140/0x140 > >> [ 258.108496] netlink_rcv_skb+0x4e/0x100 > >> [ 258.112338] netlink_unicast+0x23b/0x360 > >> [ 258.116264] netlink_sendmsg+0x24e/0x4b0 > >> [ 258.120191] sock_sendmsg+0x5f/0x70 > >> [ 258.123684] ____sys_sendmsg+0x241/0x2c0 > >> [ 258.127609] ? copy_msghdr_from_user+0x6d/0xa0 > >> [ 258.132054] ___sys_sendmsg+0x88/0xd0 > >> [ 258.135722] ? ___sys_recvmsg+0x88/0xd0 > >> [ 258.139559] ? wake_up_q+0x4a/0x90 > >> [ 258.142967] ? rseq_get_rseq_cs.isra.0+0x16/0x220 > >> [ 258.147673] ? __fget_light+0xa4/0x130 > >> [ 258.151434] __sys_sendmsg+0x59/0xa0 > >> [ 258.155012] do_syscall_64+0x38/0x90 > >> [ 258.158591] entry_SYSCALL_64_after_hwframe+0x63/0xcd > >> [ 258.163645] RIP: 0033:0x7ff23714fa7d > >> [ 258.167226] RSP: 002b:00007ffdddfc8c70 EFLAGS: 00000293 ORIG_RAX: > >> 000000000000002e > >> [ 258.174798] RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007ff23714fa7d > >> [ 258.181933] RDX: 0000000000000000 RSI: 00007ffdddfc8cb0 RDI: 000000000000000d > >> [ 258.189063] RBP: 00005572f5d77040 R08: 0000000000000000 R09: 0000000000000000 > >> [ 258.196197] R10: 0000000000000000 R11: 0000000000000293 R12: 00007ffdddfc8e1c > >> [ 258.203332] R13: 00007ffdddfc8e20 R14: 0000000000000000 R15: 00007ffdddfc8e28 > >> [ 258.210464] </TASK> > >> ... > >> > >> I bisected the issue to a commit > >> "425c9bd06b7a70796d880828d15c11321bdfb76d" (RDMA/irdma: Report the > >> correct link speed). Reverting this commit in my kernel build "fix" > >> the issue and the server has a working network again. > > > > Thanks for the report. To be sure the issue doesn't fall through the > > cracks unnoticed, I'm adding it to regzbot, the Linux kernel regression > > tracking bot: > > > > #regzbot ^introduced 425c9bd06b7a7079 > > #regzbot title RDMA/irdma: network stopped working > > #regzbot ignore-activity > > > > This isn't a regression? This issue or a fix for it are already > > discussed somewhere else? It was fixed already? You want to clarify when > > the regression started to happen? Or point out I got the title or > > something else totally wrong? Then just reply and tell me -- ideally > > while also telling regzbot about it, as explained by the page listed in > > the footer of this mail. > > > > Reminder for developers: When fixing the issue, add 'Link:' tags > > pointing to the report (see page linked in footer for details). > > > > Ciao, Thorsten (wearing his 'the Linux kernel's regression tracker' hat) > > -- > > Everything you wanna know about Linux kernel regression tracking: > > https://linux-regtracking.leemhuis.info/about/#tldr > > That page also explains what to do if mails like this annoy you. > > > >