On Thu, Jan 19, 2023 at 11:02:25AM +0100, Christoph Hellwig wrote: > vunmap only needs to find and free the vmap_area and vm_strut, so open > code that there and merge the rest of the code into vfree. > > Signed-off-by: Christoph Hellwig <hch@xxxxxx> > --- > mm/vmalloc.c | 85 ++++++++++++++++++++++++++-------------------------- > 1 file changed, 42 insertions(+), 43 deletions(-) > > diff --git a/mm/vmalloc.c b/mm/vmalloc.c > index 4cb189bdd51499..791d906d7e407c 100644 > --- a/mm/vmalloc.c > +++ b/mm/vmalloc.c > @@ -2666,45 +2666,6 @@ static void va_remove_mappings(struct vm_struct *area, int deallocate_pages) > set_area_direct_map(area, set_direct_map_default_noflush); > } > > -static void __vunmap(const void *addr, int deallocate_pages) > -{ > - struct vm_struct *area; > - > - if (!addr) > - return; > - > - area = remove_vm_area(addr); > - if (unlikely(!area)) { > - WARN(1, KERN_ERR "Trying to vfree() nonexistent vm area (%p)\n", > - addr); > - return; > - } > - > - va_remove_mappings(area, deallocate_pages); > - > - if (deallocate_pages) { > - int i; > - > - for (i = 0; i < area->nr_pages; i++) { > - struct page *page = area->pages[i]; > - > - BUG_ON(!page); > - mod_memcg_page_state(page, MEMCG_VMALLOC, -1); > - /* > - * High-order allocs for huge vmallocs are split, so > - * can be freed as an array of order-0 allocations > - */ > - __free_pages(page, 0); > - cond_resched(); > - } > - atomic_long_sub(area->nr_pages, &nr_vmalloc_pages); > - > - kvfree(area->pages); > - } > - > - kfree(area); > -} > - > static void delayed_vfree_work(struct work_struct *w) > { > struct vfree_deferred *p = container_of(w, struct vfree_deferred, wq); > @@ -2757,6 +2718,9 @@ void vfree_atomic(const void *addr) > */ > void vfree(const void *addr) > { > + struct vm_struct *vm; > + int i; > + > if (unlikely(in_interrupt())) { > vfree_atomic(addr); > return; > @@ -2766,8 +2730,32 @@ void vfree(const void *addr) > kmemleak_free(addr); > might_sleep(); > > - if (addr) > - __vunmap(addr, 1); > + if (!addr) > + return; > + > + vm = remove_vm_area(addr); > + if (unlikely(!vm)) { > + WARN(1, KERN_ERR "Trying to vfree() nonexistent vm area (%p)\n", > + addr); > + return; > + } > + > + va_remove_mappings(vm, true); > + for (i = 0; i < vm->nr_pages; i++) { > + struct page *page = vm->pages[i]; > + > + BUG_ON(!page); > + mod_memcg_page_state(page, MEMCG_VMALLOC, -1); > + /* > + * High-order allocs for huge vmallocs are split, so > + * can be freed as an array of order-0 allocations > + */ > + __free_pages(page, 0); > + cond_resched(); > + } > + atomic_long_sub(vm->nr_pages, &nr_vmalloc_pages); > + kvfree(vm->pages); > + kfree(vm); > } > EXPORT_SYMBOL(vfree); > > @@ -2782,10 +2770,21 @@ EXPORT_SYMBOL(vfree); > */ > void vunmap(const void *addr) > { > + struct vm_struct *vm; > + > BUG_ON(in_interrupt()); > might_sleep(); > - if (addr) > - __vunmap(addr, 0); > + > + if (!addr) > + return; > + vm = remove_vm_area(addr); > + if (unlikely(!vm)) { > + WARN(1, KERN_ERR "Trying to vunmap() nonexistent vm area (%p)\n", > + addr); > + return; > + } > + WARN_ON_ONCE(vm->flags & VM_FLUSH_RESET_PERMS); > + kfree(vm); > } > EXPORT_SYMBOL(vunmap); > > -- > 2.39.0 > After this patch same check in the end of the vunmap() becomes odd because we fail a vmap() call on its entry if VM_FLUSH_RESET_PERMS flag is set. See the [1] patch in this series. Is there any reason for such duplication? -- Uladzislau Rezki