On Mon, Jun 06, 2022 at 04:39:07PM +0800, Baoquan He wrote: > In __purge_vmap_area_lazy(), rename local_pure_list to local_purge_list. > > Signed-off-by: Baoquan He <bhe@xxxxxxxxxx> > --- > mm/vmalloc.c | 12 ++++++------ > 1 file changed, 6 insertions(+), 6 deletions(-) > > diff --git a/mm/vmalloc.c b/mm/vmalloc.c > index c7e1634ff2b9..11dfc897de40 100644 > --- a/mm/vmalloc.c > +++ b/mm/vmalloc.c > @@ -1673,32 +1673,32 @@ static void purge_fragmented_blocks_allcpus(void); > static bool __purge_vmap_area_lazy(unsigned long start, unsigned long end) > { > unsigned long resched_threshold; > - struct list_head local_pure_list; > + struct list_head local_purge_list; > struct vmap_area *va, *n_va; > > lockdep_assert_held(&vmap_purge_lock); > > spin_lock(&purge_vmap_area_lock); > purge_vmap_area_root = RB_ROOT; > - list_replace_init(&purge_vmap_area_list, &local_pure_list); > + list_replace_init(&purge_vmap_area_list, &local_purge_list); > spin_unlock(&purge_vmap_area_lock); > > - if (unlikely(list_empty(&local_pure_list))) > + if (unlikely(list_empty(&local_purge_list))) > return false; > > start = min(start, > - list_first_entry(&local_pure_list, > + list_first_entry(&local_purge_list, > struct vmap_area, list)->va_start); > > end = max(end, > - list_last_entry(&local_pure_list, > + list_last_entry(&local_purge_list, > struct vmap_area, list)->va_end); > > flush_tlb_kernel_range(start, end); > resched_threshold = lazy_max_pages() << 1; > > spin_lock(&free_vmap_area_lock); > - list_for_each_entry_safe(va, n_va, &local_pure_list, list) { > + list_for_each_entry_safe(va, n_va, &local_purge_list, list) { > unsigned long nr = (va->va_end - va->va_start) >> PAGE_SHIFT; > unsigned long orig_start = va->va_start; > unsigned long orig_end = va->va_end; > -- > 2.34.1 > Makes sense to me: Reviewed-by: Uladzislau Rezki (Sony) <urezki@xxxxxxxxx> -- Uladzislau Rezki