On Mon, Jul 29, 2019 at 04:21:39PM -0700, sathyanarayanan.kuppuswamy@xxxxxxxxxxxxxxx wrote: > From: Kuppuswamy Sathyanarayanan <sathyanarayanan.kuppuswamy@xxxxxxxxxxxxxxx> > > Recent changes to the vmalloc code by Commit 68ad4a330433 > ("mm/vmalloc.c: keep track of free blocks for vmap allocation") can > cause spurious percpu allocation failures. These, in turn, can result in > panic()s in the slub code. One such possible panic was reported by > Dave Hansen in following link https://lkml.org/lkml/2019/6/19/939. > Another related panic observed is, > > RIP: 0033:0x7f46f7441b9b > Call Trace: > dump_stack+0x61/0x80 > pcpu_alloc.cold.30+0x22/0x4f > mem_cgroup_css_alloc+0x110/0x650 > cgroup_apply_control_enable+0x133/0x330 > cgroup_mkdir+0x41b/0x500 > kernfs_iop_mkdir+0x5a/0x90 > vfs_mkdir+0x102/0x1b0 > do_mkdirat+0x7d/0xf0 > do_syscall_64+0x5b/0x180 > entry_SYSCALL_64_after_hwframe+0x44/0xa9 > > VMALLOC memory manager divides the entire VMALLOC space (VMALLOC_START > to VMALLOC_END) into multiple VM areas (struct vm_areas), and it mainly > uses two lists (vmap_area_list & free_vmap_area_list) to track the used > and free VM areas in VMALLOC space. And pcpu_get_vm_areas(offsets[], > sizes[], nr_vms, align) function is used for allocating congruent VM > areas for percpu memory allocator. In order to not conflict with VMALLOC > users, pcpu_get_vm_areas allocates VM areas near the end of the VMALLOC > space. So the search for free vm_area for the given requirement starts > near VMALLOC_END and moves upwards towards VMALLOC_START. > > Prior to commit 68ad4a330433, the search for free vm_area in > pcpu_get_vm_areas() involves following two main steps. > > Step 1: > Find a aligned "base" adress near VMALLOC_END. > va = free vm area near VMALLOC_END > Step 2: > Loop through number of requested vm_areas and check, > Step 2.1: > if (base < VMALLOC_START) > 1. fail with error > Step 2.2: > // end is offsets[area] + sizes[area] > if (base + end > va->vm_end) > 1. Move the base downwards and repeat Step 2 > Step 2.3: > if (base + start < va->vm_start) > 1. Move to previous free vm_area node, find aligned > base address and repeat Step 2 > > But Commit 68ad4a330433 removed Step 2.2 and modified Step 2.3 as below: > > Step 2.3: > if (base + start < va->vm_start || base + end > va->vm_end) > 1. Move to previous free vm_area node, find aligned > base address and repeat Step 2 > > Above change is the root cause of spurious percpu memory allocation > failures. For example, consider a case where a relatively large vm_area > (~ 30 TB) was ignored in free vm_area search because it did not pass the > base + end < vm->vm_end boundary check. Ignoring such large free > vm_area's would lead to not finding free vm_area within boundary of > VMALLOC_start to VMALLOC_END which in turn leads to allocation failures. > > So modify the search algorithm to include Step 2.2. > > Fixes: 68ad4a330433 ("mm/vmalloc.c: keep track of free blocks for vmap allocation") > Signed-off-by: Kuppuswamy Sathyanarayanan <sathyanarayanan.kuppuswamy@xxxxxxxxxxxxxxx> > --- > mm/vmalloc.c | 12 +++++++++++- > 1 file changed, 11 insertions(+), 1 deletion(-) > > diff --git a/mm/vmalloc.c b/mm/vmalloc.c > index 4fa8d84599b0..1faa45a38c08 100644 > --- a/mm/vmalloc.c > +++ b/mm/vmalloc.c > @@ -3269,10 +3269,20 @@ struct vm_struct **pcpu_get_vm_areas(const unsigned long *offsets, > if (va == NULL) > goto overflow; > > + /* > + * If required width exeeds current VA block, move > + * base downwards and then recheck. > + */ > + if (base + end > va->va_end) { > + base = pvm_determine_end_from_reverse(&va, align) - end; > + term_area = area; > + continue; > + } > + > /* > * If this VA does not fit, move base downwards and recheck. > */ > - if (base + start < va->va_start || base + end > va->va_end) { > + if (base + start < va->va_start) { > va = node_to_va(rb_prev(&va->rb_node)); > base = pvm_determine_end_from_reverse(&va, align) - end; > term_area = area; > -- > 2.21.0 > I guess it is NUMA related issue, i mean when we have several areas/sizes/offsets. Is that correct? Thank you! -- Vlad Rezki