Hello, syzbot found the following issue on: HEAD commit: bab0d47c riscv: kexec: Fix W=1 build warnings git tree: git://git.kernel.org/pub/scm/linux/kernel/git/riscv/linux.git fixes console output: https://syzkaller.appspot.com/x/log.txt?x=10f59535d00000 kernel config: https://syzkaller.appspot.com/x/.config?x=e610cd256ee3a8 dashboard link: https://syzkaller.appspot.com/bug?extid=06b228c6b9c37dcd3d79 userspace arch: riscv64 IMPORTANT: if you fix the issue, please add the following tag to the commit: Reported-by: syzbot+06b228c6b9c37dcd3d79@xxxxxxxxxxxxxxxxxxxxxxxxx ------------[ cut here ]------------ WARNING: CPU: 1 PID: 2996 at mm/vmalloc.c:448 vmap_pages_pte_range mm/vmalloc.c:448 [inline] WARNING: CPU: 1 PID: 2996 at mm/vmalloc.c:448 vmap_pages_pmd_range mm/vmalloc.c:471 [inline] WARNING: CPU: 1 PID: 2996 at mm/vmalloc.c:448 vmap_pages_pud_range mm/vmalloc.c:489 [inline] WARNING: CPU: 1 PID: 2996 at mm/vmalloc.c:448 vmap_pages_p4d_range mm/vmalloc.c:507 [inline] WARNING: CPU: 1 PID: 2996 at mm/vmalloc.c:448 vmap_small_pages_range_noflush+0x2fa/0x38e mm/vmalloc.c:529 Modules linked in: CPU: 1 PID: 2996 Comm: dhcpcd Not tainted 5.13.0-rc1-syzkaller-00629-gbab0d47c0ebb #0 Hardware name: riscv-virtio,qemu (DT) epc : vmap_pages_pte_range mm/vmalloc.c:448 [inline] epc : vmap_pages_pmd_range mm/vmalloc.c:471 [inline] epc : vmap_pages_pud_range mm/vmalloc.c:489 [inline] epc : vmap_pages_p4d_range mm/vmalloc.c:507 [inline] epc : vmap_small_pages_range_noflush+0x2fa/0x38e mm/vmalloc.c:529 ra : vmap_pages_pte_range mm/vmalloc.c:448 [inline] ra : vmap_pages_pmd_range mm/vmalloc.c:471 [inline] ra : vmap_pages_pud_range mm/vmalloc.c:489 [inline] ra : vmap_pages_p4d_range mm/vmalloc.c:507 [inline] ra : vmap_small_pages_range_noflush+0x2fa/0x38e mm/vmalloc.c:529 epc : ffffffff8036c3e4 ra : ffffffff8036c3e4 sp : ffffffe00984b900 gp : ffffffff845906e0 tp : ffffffe007d817c0 t0 : ffffffe008f1d000 t1 : ffffffc4011e39ff t2 : 0000004000010015 s0 : ffffffe00984b9b0 s1 : ffffffcf02244700 a0 : 0000000000000000 a1 : 00000000000f0000 a2 : 0000000000000002 a3 : ffffffff8036c3e4 a4 : ffffffe007d827c0 a5 : 0000000000000000 a6 : 0000000000f00000 a7 : f310d86cfc0c1a00 s2 : ffffffe005a00a28 s3 : ffffffff85b45000 s4 : ffffffff85b46000 s5 : 0000000000000000 s6 : 0000000000000200 s7 : ffffffe07fdfd168 s8 : ffffffe00e09f528 s9 : 0000003100000000 s10: ffffffe008eb8fa0 s11: 00000000000000c7 t3 : 0000000000000000 t4 : 0000000000000040 t5 : ffffffc4011e3a00 t6 : ffffffd00067e2d0 status: 0000000000000120 badaddr: 0000000000000000 cause: 0000000000000003 [<ffffffff8036c3e4>] vmap_pages_pte_range mm/vmalloc.c:448 [inline] [<ffffffff8036c3e4>] vmap_pages_pmd_range mm/vmalloc.c:471 [inline] [<ffffffff8036c3e4>] vmap_pages_pud_range mm/vmalloc.c:489 [inline] [<ffffffff8036c3e4>] vmap_pages_p4d_range mm/vmalloc.c:507 [inline] [<ffffffff8036c3e4>] vmap_small_pages_range_noflush+0x2fa/0x38e mm/vmalloc.c:529 [<ffffffff80375a7a>] vmap_pages_range_noflush mm/vmalloc.c:558 [inline] [<ffffffff80375a7a>] vmap_pages_range mm/vmalloc.c:592 [inline] [<ffffffff80375a7a>] __vmalloc_area_node mm/vmalloc.c:2829 [inline] [<ffffffff80375a7a>] __vmalloc_node_range+0x396/0x582 mm/vmalloc.c:2915 [<ffffffff80013c06>] bpf_jit_alloc_exec+0x46/0x52 arch/riscv/net/bpf_jit_core.c:171 [<ffffffff801f3d2a>] bpf_jit_binary_alloc+0xac/0x172 kernel/bpf/core.c:872 [<ffffffff80013a3a>] bpf_int_jit_compile+0x754/0x8da arch/riscv/net/bpf_jit_core.c:108 [<ffffffff801f59e0>] bpf_prog_select_runtime+0x258/0x2e4 kernel/bpf/core.c:1867 [<ffffffff821f9af2>] bpf_migrate_filter+0x1d6/0x23c net/core/filter.c:1294 [<ffffffff82201854>] bpf_prepare_filter net/core/filter.c:1342 [inline] [<ffffffff82201854>] __get_filter+0x1d6/0x2d0 net/core/filter.c:1511 [<ffffffff82202992>] sk_attach_filter+0x22/0x11a net/core/filter.c:1526 [<ffffffff82154d82>] sock_setsockopt+0x18c4/0x1c2c net/core/sock.c:1068 [<ffffffff82145172>] __sys_setsockopt+0x2de/0x33c net/socket.c:2113 [<ffffffff8214520a>] __do_sys_setsockopt net/socket.c:2128 [inline] [<ffffffff8214520a>] sys_setsockopt+0x3a/0x4c net/socket.c:2125 [<ffffffff8000562c>] ret_from_syscall+0x0/0x2 irq event stamp: 946 hardirqs last enabled at (945): [<ffffffff8037e5e4>] rmqueue_pcplist mm/page_alloc.c:3506 [inline] hardirqs last enabled at (945): [<ffffffff8037e5e4>] rmqueue mm/page_alloc.c:3529 [inline] hardirqs last enabled at (945): [<ffffffff8037e5e4>] get_page_from_freelist+0xc50/0xf66 mm/page_alloc.c:3991 hardirqs last disabled at (946): [<ffffffff80005570>] _save_context+0x80/0x90 softirqs last enabled at (942): [<ffffffff82b272a0>] softirq_handle_end kernel/softirq.c:402 [inline] softirqs last enabled at (942): [<ffffffff82b272a0>] __do_softirq+0x5e0/0x8c4 kernel/softirq.c:588 softirqs last disabled at (921): [<ffffffff80036760>] do_softirq_own_stack include/asm-generic/softirq_stack.h:10 [inline] softirqs last disabled at (921): [<ffffffff80036760>] invoke_softirq kernel/softirq.c:440 [inline] softirqs last disabled at (921): [<ffffffff80036760>] __irq_exit_rcu kernel/softirq.c:637 [inline] softirqs last disabled at (921): [<ffffffff80036760>] irq_exit+0x1a0/0x1b6 kernel/softirq.c:661 ---[ end trace f164002e4a3f575f ]--- ------------[ cut here ]------------ WARNING: CPU: 1 PID: 2996 at mm/vmalloc.c:305 vunmap_pte_range mm/vmalloc.c:305 [inline] WARNING: CPU: 1 PID: 2996 at mm/vmalloc.c:305 vunmap_pmd_range mm/vmalloc.c:329 [inline] WARNING: CPU: 1 PID: 2996 at mm/vmalloc.c:305 vunmap_pud_range mm/vmalloc.c:354 [inline] WARNING: CPU: 1 PID: 2996 at mm/vmalloc.c:305 vunmap_p4d_range mm/vmalloc.c:377 [inline] WARNING: CPU: 1 PID: 2996 at mm/vmalloc.c:305 vunmap_range_noflush+0x206/0x308 mm/vmalloc.c:408 Modules linked in: CPU: 1 PID: 2996 Comm: dhcpcd Tainted: G W 5.13.0-rc1-syzkaller-00629-gbab0d47c0ebb #0 Hardware name: riscv-virtio,qemu (DT) epc : vunmap_pte_range mm/vmalloc.c:305 [inline] epc : vunmap_pmd_range mm/vmalloc.c:329 [inline] epc : vunmap_pud_range mm/vmalloc.c:354 [inline] epc : vunmap_p4d_range mm/vmalloc.c:377 [inline] epc : vunmap_range_noflush+0x206/0x308 mm/vmalloc.c:408 ra : vunmap_pte_range mm/vmalloc.c:305 [inline] ra : vunmap_pmd_range mm/vmalloc.c:329 [inline] ra : vunmap_pud_range mm/vmalloc.c:354 [inline] ra : vunmap_p4d_range mm/vmalloc.c:377 [inline] ra : vunmap_range_noflush+0x206/0x308 mm/vmalloc.c:408 epc : ffffffff80372b18 ra : ffffffff80372b18 sp : ffffffe00984b820 gp : ffffffff845906e0 tp : ffffffe007d817c0 t0 : ffffffe008f1d000 t1 : 0000000000000001 t2 : 0000004000010015 s0 : ffffffe00984b8b0 s1 : ffffffe005a00a28 a0 : 0000000000000000 a1 : 00000000000f0000 a2 : 0000000000000002 a3 : ffffffff80372b18 a4 : ffffffe007d827c0 a5 : 0000000000000000 a6 : 0000000000f00000 a7 : ffffffff80374a74 s2 : ffffffff85b45000 s3 : ffffffff85b47000 s4 : 0000000000000000 s5 : ffffffe07fdfd168 s6 : 0000000000001000 s7 : ffffffff85b47000 s8 : ffffffff85b46fff s9 : 0000000000000000 s10: 0000000000000200 s11: ffffffff83858890 t3 : f310d86cfc0c1a00 t4 : 0000000000000040 t5 : ffffffc4011e3a00 t6 : ffffffd00067e2d0 status: 0000000000000120 badaddr: 0000000000000000 cause: 0000000000000003 [<ffffffff80372b18>] vunmap_pte_range mm/vmalloc.c:305 [inline] [<ffffffff80372b18>] vunmap_pmd_range mm/vmalloc.c:329 [inline] [<ffffffff80372b18>] vunmap_pud_range mm/vmalloc.c:354 [inline] [<ffffffff80372b18>] vunmap_p4d_range mm/vmalloc.c:377 [inline] [<ffffffff80372b18>] vunmap_range_noflush+0x206/0x308 mm/vmalloc.c:408 [<ffffffff80372c4a>] free_unmap_vmap_area+0x30/0x68 mm/vmalloc.c:1722 [<ffffffff80374b82>] remove_vm_area+0x150/0x152 mm/vmalloc.c:2462 [<ffffffff80374dce>] vm_remove_mappings mm/vmalloc.c:2491 [inline] [<ffffffff80374dce>] __vunmap+0x24a/0x616 mm/vmalloc.c:2556 [<ffffffff8037525a>] __vfree+0x70/0xf8 mm/vmalloc.c:2613 [<ffffffff80375b9c>] __vmalloc_area_node mm/vmalloc.c:2840 [inline] [<ffffffff80375b9c>] __vmalloc_node_range+0x4b8/0x582 mm/vmalloc.c:2915 [<ffffffff80013c06>] bpf_jit_alloc_exec+0x46/0x52 arch/riscv/net/bpf_jit_core.c:171 [<ffffffff801f3d2a>] bpf_jit_binary_alloc+0xac/0x172 kernel/bpf/core.c:872 [<ffffffff80013a3a>] bpf_int_jit_compile+0x754/0x8da arch/riscv/net/bpf_jit_core.c:108 [<ffffffff801f59e0>] bpf_prog_select_runtime+0x258/0x2e4 kernel/bpf/core.c:1867 [<ffffffff821f9af2>] bpf_migrate_filter+0x1d6/0x23c net/core/filter.c:1294 [<ffffffff82201854>] bpf_prepare_filter net/core/filter.c:1342 [inline] [<ffffffff82201854>] __get_filter+0x1d6/0x2d0 net/core/filter.c:1511 [<ffffffff82202992>] sk_attach_filter+0x22/0x11a net/core/filter.c:1526 [<ffffffff82154d82>] sock_setsockopt+0x18c4/0x1c2c net/core/sock.c:1068 [<ffffffff82145172>] __sys_setsockopt+0x2de/0x33c net/socket.c:2113 [<ffffffff8214520a>] __do_sys_setsockopt net/socket.c:2128 [inline] [<ffffffff8214520a>] sys_setsockopt+0x3a/0x4c net/socket.c:2125 [<ffffffff8000562c>] ret_from_syscall+0x0/0x2 irq event stamp: 964 hardirqs last enabled at (963): [<ffffffff82b264ac>] __raw_spin_unlock_irqrestore include/linux/spinlock_api_smp.h:160 [inline] hardirqs last enabled at (963): [<ffffffff82b264ac>] _raw_spin_unlock_irqrestore+0x68/0x98 kernel/locking/spinlock.c:191 hardirqs last disabled at (964): [<ffffffff80005570>] _save_context+0x80/0x90 softirqs last enabled at (956): [<ffffffff82b272a0>] softirq_handle_end kernel/softirq.c:402 [inline] softirqs last enabled at (956): [<ffffffff82b272a0>] __do_softirq+0x5e0/0x8c4 kernel/softirq.c:588 softirqs last disabled at (949): [<ffffffff80036760>] do_softirq_own_stack include/asm-generic/softirq_stack.h:10 [inline] softirqs last disabled at (949): [<ffffffff80036760>] invoke_softirq kernel/softirq.c:440 [inline] softirqs last disabled at (949): [<ffffffff80036760>] __irq_exit_rcu kernel/softirq.c:637 [inline] softirqs last disabled at (949): [<ffffffff80036760>] irq_exit+0x1a0/0x1b6 kernel/softirq.c:661 ---[ end trace f164002e4a3f5760 ]--- --- This report is generated by a bot. It may contain errors. See https://goo.gl/tpsmEJ for more information about syzbot. syzbot engineers can be reached at syzkaller@xxxxxxxxxxxxxxxx. syzbot will keep track of this issue. See: https://goo.gl/tpsmEJ#status for how to communicate with syzbot.