[BUG REPORT] writeback: soft lockup encountered on a next-20240327 kernel

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

Executing fstest on XFS on a next-20240327 kernel resulted in two of my test
VMs getting into soft lockup state.

watchdog: BUG: soft lockup - CPU#0 stuck for 64383s! [kworker/u16:8:1648676]
Modules linked in: overlay dm_zero dm_thin_pool dm_persistent_data dm_bio_prison dm_snapshot dm_bufio dm_flakey loop nft_redir ipt_REJECT xt_comment xt_owner nft_compat nft_fib_inet nft_fib_ipv4 nft_fib_ipv6 nft_fib nft_reject_inet nf_reject_ipv4 nf_reject_ipv6 nft_reject nft_ct nft_chain_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 rfkill ip_set cuse vfat fat ext4 mbcache jbd2 intel_rapl_msr intel_rapl_common kvm_amd ccp bochs drm_vram_helper drm_kms_helper kvm drm_ttm_helper pcspkr pvpanic_mmio ttm pvpanic i2c_piix4 joydev sch_fq_codel drm fuse xfs nvme_tcp nvme_fabrics nvme_core sd_mod t10_pi crc64_rocksoft_generic crc64_rocksoft sg virtio_net net_failover failover virtio_scsi crct10dif_pclmul crc32_pclmul ata_generic pata_acpi ata_piix ghash_clmulni_intel sha512_ssse3 sha256_ssse3 sha1_ssse3 libata virtio_pci virtio_pci_legacy_dev virtio_pci_modern_dev serio_raw dm_multipath btrfs blake2b_generic xor zstd_compress raid6_pq sunrpc dm_mirror dm_region_hash dm_log dm_mod be2iscsi bnx2i cnic uio
 cxgb4i cxgb4 tls cxgb3i cxgb3 mdio libcxgbi libcxgb qla4xxx iscsi_boot_sysfs iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi qemu_fw_cfg aesni_intel crypto_simd cryptd [last unloaded: scsi_debug]
CPU: 0 PID: 1648676 Comm: kworker/u16:8 Kdump: loaded Tainted: G             L     6.9.0-rc1-next-20240327+ #1
Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.6.6 08/22/2023
Workqueue: writeback wb_update_bandwidth_workfn
RIP: 0010:__pv_queued_spin_lock_slowpath+0x4d5/0xc30
Code: eb c6 45 01 01 41 bc 00 80 00 00 48 c1 e9 03 83 e3 07 41 be 01 00 00 00 48 b8 00 00 00 00 00 fc ff df 4c 8d 2c 01 eb 0c f3 90 <41> 83 ec 01 0f 84 82 04 00 00 41 0f b6 45 00 38 d8 7f 08 84 c0 0f
RSP: 0018:ffffc90007167b18 EFLAGS: 00000206
RAX: 0000000000000003 RBX: 0000000000000000 RCX: 1ffff110211bda18
RDX: 0000000000000001 RSI: 0000000000000003 RDI: ffff888108ded0c0
RBP: ffff888108ded0c0 R08: 0000000000000001 R09: ffffed10211bda18
R10: ffff888108ded0c0 R11: 000000000000000a R12: 0000000000005218
R13: ffffed10211bda18 R14: 0000000000000001 R15: ffff8883ef246bc0
FS:  0000000000000000(0000) GS:ffff8883ef200000(0000) knlGS:0000000000000000
CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00007f9fa7534500 CR3: 0000000146556000 CR4: 0000000000350ef0
Call Trace:
 <IRQ>
 ? watchdog_timer_fn+0x2e2/0x3b0
 ? __pfx_watchdog_timer_fn+0x10/0x10
 ? __hrtimer_run_queues+0x300/0x6d0
 ? __pfx___hrtimer_run_queues+0x10/0x10
 ? __pfx___raw_spin_lock_irqsave+0x10/0x10
 ? srso_return_thunk+0x5/0x5f
 ? srso_return_thunk+0x5/0x5f
 ? ktime_get_update_offsets_now+0x73/0x280
 ? hrtimer_interrupt+0x2ce/0x770
 ? __sysvec_apic_timer_interrupt+0x90/0x2c0
 ? sysvec_apic_timer_interrupt+0x69/0x90
 </IRQ>
 <TASK>
 _raw_spin_lock+0xd0/0xe0
 __wb_update_bandwidth+0x72/0x600
 wb_update_bandwidth+0x97/0xd0
 process_one_work+0x60d/0x1020
 worker_thread+0x795/0x1290
 kthread+0x2a9/0x380
 ret_from_fork+0x34/0x70
 ret_from_fork_asm+0x1a/0x30
 </TASK>

I am unable to retrieve any other debug information since the machines are not
accessible. Hence, I am not sure about the exact test which caused the above
issue.

-- 
Chandan





[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux