Hi, On 8/29/2023 7:46 PM, Björn Töpel wrote: > Björn Töpel <bjorn@xxxxxxxxxx> writes: > >> Hou Tao <houtao@xxxxxxxxxxxxxxx> writes: >> >>> Hi, >>> >>> On 8/27/2023 10:53 PM, Yonghong Song wrote: >>>> >>>> On 8/27/23 1:37 AM, Björn Töpel wrote: >>>>> Björn Töpel <bjorn@xxxxxxxxxx> writes: >>>>> >>>>>> Hou Tao <houtao@xxxxxxxxxxxxxxx> writes: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> On 8/26/2023 5:23 PM, Björn Töpel wrote: >>>>>>>> Hou Tao <houtao@xxxxxxxxxxxxxxx> writes: >>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> On 8/25/2023 11:28 PM, Yonghong Song wrote: >>>>>>>>>> On 8/25/23 3:32 AM, Björn Töpel wrote: >>>>>>>>>>> I'm chasing a workqueue hang on RISC-V/qemu (TCG), using the bpf >>>>>>>>>>> selftests on bpf-next 9e3b47abeb8f. >>>>>>>>>>> >>>>>>>>>>> I'm able to reproduce the hang by multiple runs of: >>>>>>>>>>> | ./test_progs -a link_api -a linked_list >>>>>>>>>>> I'm currently investigating that. >>>>>>>>>>> >>>>>>>>>>> But! Sometimes (every blue moon) I get a warn_on_once hit: >>>>>>>>>>> | ------------[ cut here ]------------ >>>>>>>>>>> | WARNING: CPU: 3 PID: 261 at kernel/bpf/memalloc.c:342 >>>>>>>>>>> bpf_mem_refill+0x1fc/0x206 >>>>>>>>>>> | Modules linked in: bpf_testmod(OE) >>>>>>>>>>> | CPU: 3 PID: 261 Comm: test_progs-cpuv Tainted: G OE >>>>>>>>>>> N 6.5.0-rc5-01743-gdcb152bb8328 #2 >>>>>>>>>>> | Hardware name: riscv-virtio,qemu (DT) >>>>>>>>>>> | epc : bpf_mem_refill+0x1fc/0x206 >>>>>>>>>>> | ra : irq_work_single+0x68/0x70 >>>>>>>>>>> | epc : ffffffff801b1bc4 ra : ffffffff8015fe84 sp : >>>>>>>>>>> ff2000000001be20 >>>>>>>>>>> | gp : ffffffff82d26138 tp : ff6000008477a800 t0 : >>>>>>>>>>> 0000000000046600 >>>>>>>>>>> | t1 : ffffffff812b6ddc t2 : 0000000000000000 s0 : >>>>>>>>>>> ff2000000001be70 >>>>>>>>>>> | s1 : ff5ffffffffe8998 a0 : ff5ffffffffe8998 a1 : >>>>>>>>>>> ff600003fef4b000 >>>>>>>>>>> | a2 : 000000000000003f a3 : ffffffff80008250 a4 : >>>>>>>>>>> 0000000000000060 >>>>>>>>>>> | a5 : 0000000000000080 a6 : 0000000000000000 a7 : >>>>>>>>>>> 0000000000735049 >>>>>>>>>>> | s2 : ff5ffffffffe8998 s3 : 0000000000000022 s4 : >>>>>>>>>>> 0000000000001000 >>>>>>>>>>> | s5 : 0000000000000007 s6 : ff5ffffffffe8570 s7 : >>>>>>>>>>> ffffffff82d6bd30 >>>>>>>>>>> | s8 : 000000000000003f s9 : ffffffff82d2c5e8 s10: >>>>>>>>>>> 000000000000ffff >>>>>>>>>>> | s11: ffffffff82d2c5d8 t3 : ffffffff81ea8f28 t4 : >>>>>>>>>>> 0000000000000000 >>>>>>>>>>> | t5 : ff6000008fd28278 t6 : 0000000000040000 >>>>>>>>>>> | status: 0000000200000100 badaddr: 0000000000000000 cause: >>>>>>>>>>> 0000000000000003 >>>>>>>>>>> | [<ffffffff801b1bc4>] bpf_mem_refill+0x1fc/0x206 >>>>>>>>>>> | [<ffffffff8015fe84>] irq_work_single+0x68/0x70 >>>>>>>>>>> | [<ffffffff8015feb4>] irq_work_run_list+0x28/0x36 >>>>>>>>>>> | [<ffffffff8015fefa>] irq_work_run+0x38/0x66 >>>>>>>>>>> | [<ffffffff8000828a>] handle_IPI+0x3a/0xb4 >>>>>>>>>>> | [<ffffffff800a5c3a>] handle_percpu_devid_irq+0xa4/0x1f8 >>>>>>>>>>> | [<ffffffff8009fafa>] generic_handle_domain_irq+0x28/0x36 >>>>>>>>>>> | [<ffffffff800ae570>] ipi_mux_process+0xac/0xfa >>>>>>>>>>> | [<ffffffff8000a8ea>] sbi_ipi_handle+0x2e/0x88 >>>>>>>>>>> | [<ffffffff8009fafa>] generic_handle_domain_irq+0x28/0x36 >>>>>>>>>>> | [<ffffffff807ee70e>] riscv_intc_irq+0x36/0x4e >>>>>>>>>>> | [<ffffffff812b5d3a>] handle_riscv_irq+0x54/0x86 >>>>>>>>>>> | [<ffffffff812b6904>] do_irq+0x66/0x98 >>>>>>>>>>> | ---[ end trace 0000000000000000 ]--- >>>>>>>>>>> >>>>>>>>>>> Code: >>>>>>>>>>> | static void free_bulk(struct bpf_mem_cache *c) >>>>>>>>>>> | { >>>>>>>>>>> | struct bpf_mem_cache *tgt = c->tgt; >>>>>>>>>>> | struct llist_node *llnode, *t; >>>>>>>>>>> | unsigned long flags; >>>>>>>>>>> | int cnt; >>>>>>>>>>> | >>>>>>>>>>> | WARN_ON_ONCE(tgt->unit_size != c->unit_size); >>>>>>>>>>> | ... >>>>>>>>>>> >>>>>>>>>>> I'm not well versed in the memory allocator; Before I dive into >>>>>>>>>>> it -- >>>>>>>>>>> has anyone else hit it? Ideas on why the warn_on_once is hit? >>>>>>>>>> Maybe take a look at the patch >>>>>>>>>> 822fb26bdb55 bpf: Add a hint to allocated objects. >>>>>>>>>> >>>>>>>>>> In the above patch, we have >>>>>>>>>> >>>>>>>>>> + /* >>>>>>>>>> + * Remember bpf_mem_cache that allocated this object. >>>>>>>>>> + * The hint is not accurate. >>>>>>>>>> + */ >>>>>>>>>> + c->tgt = *(struct bpf_mem_cache **)llnode; >>>>>>>>>> >>>>>>>>>> I suspect that the warning may be related to the above. >>>>>>>>>> I tried the above ./test_progs command line (running multiple >>>>>>>>>> at the same time) and didn't trigger the issue. >>>>>>>>> The extra 8-bytes before the freed pointer is used to save the >>>>>>>>> pointer >>>>>>>>> of the original bpf memory allocator where the freed pointer came >>>>>>>>> from, >>>>>>>>> so unit_free() could free the pointer back to the original >>>>>>>>> allocator to >>>>>>>>> prevent alloc-and-free unbalance. >>>>>>>>> >>>>>>>>> I suspect that a wrong pointer was passed to bpf_obj_drop, but do >>>>>>>>> not >>>>>>>>> find anything suspicious after checking linked_list. Another >>>>>>>>> possibility >>>>>>>>> is that there is write-after-free problem which corrupts the extra >>>>>>>>> 8-bytes before the freed pointer. Could you please apply the >>>>>>>>> following >>>>>>>>> debug patch to check whether or not the extra 8-bytes are >>>>>>>>> corrupted ? >>>>>>>> Thanks for getting back! >>>>>>>> >>>>>>>> I took your patch for a run, and there's a hit: >>>>>>>> | bad cache ff5ffffffffe8570: got size 96 work >>>>>>>> ffffffff801b19c8, cache ff5ffffffffe8980 exp size 128 work >>>>>>>> ffffffff801b19c8 >>>>>>> The extra 8-bytes are not corrupted. Both of these two >>>>>>> bpf_mem_cache are >>>>>>> valid and there are in the cache array defined in bpf_mem_caches. BPF >>>>>>> memory allocator allocated the pointer from 96-bytes sized-cache, >>>>>>> but it >>>>>>> tried to free the pointer through 128-bytes sized-cache. >>>>>>> >>>>>>> Now I suspect there is no 96-bytes slab in your system and ksize(ptr - >>>>>>> LLIST_NODE_SZ) returns 128, so bpf memory allocator selected the >>>>>>> 128-byte sized-cache instead of 96-bytes sized-cache. Could you please >>>>>>> check the value of KMALLOC_MIN_SIZE in your kernel .config and >>>>>>> using the >>>>>>> following command to check whether there is 96-bytes slab in your >>>>>>> system: >>>>>> KMALLOC_MIN_SIZE is 64. >>>>>> >>>>>>> $ cat /proc/slabinfo |grep kmalloc-96 >>>>>>> dma-kmalloc-96 0 0 96 42 1 : tunables 0 0 >>>>>>> 0 : slabdata 0 0 0 >>>>>>> kmalloc-96 1865 2268 96 42 1 : tunables 0 0 >>>>>>> 0 : slabdata 54 54 0 >>>>>>> >>>>>>> In my system, slab has 96-bytes cached, so grep outputs something, >>>>>>> but I >>>>>>> think there will no output in your system. >>>>>> You're right! No kmalloc-96. >>>>> To get rid of the warning, limit available sizes from >>>>> bpf_mem_alloc_init()? >>> It is not enough. We need to adjust size_index accordingly during >>> initialization. Could you please try the attached patch below ? It is >>> not a formal patch and I am considering to disable prefilling for these >>> redirected bpf_mem_caches. >> Sorry for the slow response; I'll take it for a spin today. Thanks for the test. > Hmm, on a related note, RISC-V has this change queued up [1], which will > introduce -96 and friends. Are there any other archs supporting the BPF > allocator where this is a concern? If not, I suggest simply leaving the > code as is. Thanks for the information. I am not a micro-architecture expert, but it seems arm32 sets KMALLOC_MIN_SIZE as 64 or 128. Beside the arch related setting, If switching from slub to the deprecated slab, the similar problem will be reported, because CONFIG_SLAB will set KMALLOC_MIN_SIZE as 32 (KMALLOC_SHIFT_LOW is 5). > > > Björn > > [1] https://lore.kernel.org/all/20230718152214.2907-1-jszhang@xxxxxxxxxx/