The quilt patch titled Subject: mm-dont-account-memmap-per-node-v5 has been removed from the -mm tree. Its filename was mm-dont-account-memmap-per-node-v5.patch This patch was dropped because it was folded into mm-dont-account-memmap-per-node.patch ------------------------------------------------------ From: Pasha Tatashin <pasha.tatashin@xxxxxxxxxx> Subject: mm-dont-account-memmap-per-node-v5 Date: Fri, 9 Aug 2024 19:10:20 +0000 address a few nits from David Hildenbrand Link: https://lkml.kernel.org/r/20240809191020.1142142-4-pasha.tatashin@xxxxxxxxxx Signed-off-by: Pasha Tatashin <pasha.tatashin@xxxxxxxxxx> Cc: Alison Schofield <alison.schofield@xxxxxxxxx> Cc: Dan Williams <dan.j.williams@xxxxxxxxx> Cc: David Hildenbrand <david@xxxxxxxxxx> Cc: David Rientjes <rientjes@xxxxxxxxxx> Cc: Domenico Cerasuolo <cerasuolodomenico@xxxxxxxxx> Cc: Fan Ni <fan.ni@xxxxxxxxxxx> Cc: Joel Granados <j.granados@xxxxxxxxxxx> Cc: Johannes Weiner <hannes@xxxxxxxxxxx> Cc: Li Zhijian <lizhijian@xxxxxxxxxxx> Cc: Matthew Wilcox (Oracle) <willy@xxxxxxxxxxxxx> Cc: Mike Rapoport <rppt@xxxxxxxxxx> Cc: Muchun Song <muchun.song@xxxxxxxxx> Cc: Nhat Pham <nphamcs@xxxxxxxxx> Cc: Sourav Panda <souravpanda@xxxxxxxxxx> Cc: Vlastimil Babka <vbabka@xxxxxxx> Cc: Yi Zhang <yi.zhang@xxxxxxxxxx> Cc: Yosry Ahmed <yosryahmed@xxxxxxxxxx> Signed-off-by: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> --- include/linux/vmstat.h | 8 ++++---- mm/hugetlb_vmemmap.c | 8 ++++---- mm/mm_init.c | 2 +- mm/page_ext.c | 6 +++--- mm/sparse-vmemmap.c | 4 ++-- mm/sparse.c | 4 ++-- mm/vmstat.c | 19 ++++++++++--------- 7 files changed, 26 insertions(+), 25 deletions(-) --- a/include/linux/vmstat.h~mm-dont-account-memmap-per-node-v5 +++ a/include/linux/vmstat.h @@ -38,8 +38,8 @@ struct reclaim_stat { enum vm_stat_item { NR_DIRTY_THRESHOLD, NR_DIRTY_BG_THRESHOLD, - NR_MEMMAP, /* page metadata allocated through buddy allocator */ - NR_MEMMAP_BOOT, /* page metadata allocated through boot allocator */ + NR_MEMMAP_PAGES, /* page metadata allocated through buddy allocator */ + NR_MEMMAP_BOOT_PAGES, /* page metadata allocated through boot allocator */ NR_VM_STAT_ITEMS, }; @@ -620,6 +620,6 @@ static inline void lruvec_stat_sub_folio lruvec_stat_mod_folio(folio, idx, -folio_nr_pages(folio)); } -void mod_memmap_boot(long delta); -void mod_memmap(long delta); +void memmap_boot_pages_add(long delta); +void memmap_pages_add(long delta); #endif /* _LINUX_VMSTAT_H */ --- a/mm/hugetlb_vmemmap.c~mm-dont-account-memmap-per-node-v5 +++ a/mm/hugetlb_vmemmap.c @@ -185,10 +185,10 @@ static int vmemmap_remap_range(unsigned static inline void free_vmemmap_page(struct page *page) { if (PageReserved(page)) { - mod_memmap_boot(-1); + memmap_boot_pages_add(-1); free_bootmem_page(page); } else { - mod_memmap(-1); + memmap_pages_add(-1); __free_page(page); } } @@ -341,7 +341,7 @@ static int vmemmap_remap_free(unsigned l copy_page(page_to_virt(walk.reuse_page), (void *)walk.reuse_addr); list_add(&walk.reuse_page->lru, vmemmap_pages); - mod_memmap(1); + memmap_pages_add(1); } /* @@ -396,7 +396,7 @@ static int alloc_vmemmap_page_list(unsig goto out; list_add(&page->lru, list); } - mod_memmap(nr_pages); + memmap_pages_add(nr_pages); return 0; out: --- a/mm/mm_init.c~mm-dont-account-memmap-per-node-v5 +++ a/mm/mm_init.c @@ -1623,7 +1623,7 @@ static void __init alloc_node_mem_map(st panic("Failed to allocate %ld bytes for node %d memory map\n", size, pgdat->node_id); pgdat->node_mem_map = map + offset; - mod_memmap_boot(DIV_ROUND_UP(size, PAGE_SIZE)); + memmap_boot_pages_add(DIV_ROUND_UP(size, PAGE_SIZE)); pr_debug("%s: node %d, pgdat %08lx, node_mem_map %08lx\n", __func__, pgdat->node_id, (unsigned long)pgdat, (unsigned long)pgdat->node_mem_map); --- a/mm/page_ext.c~mm-dont-account-memmap-per-node-v5 +++ a/mm/page_ext.c @@ -214,7 +214,7 @@ static int __init alloc_node_page_ext(in return -ENOMEM; NODE_DATA(nid)->node_page_ext = base; total_usage += table_size; - mod_memmap_boot(DIV_ROUND_UP(table_size, PAGE_SIZE)); + memmap_boot_pages_add(DIV_ROUND_UP(table_size, PAGE_SIZE)); return 0; } @@ -275,7 +275,7 @@ static void *__meminit alloc_page_ext(si addr = vzalloc_node(size, nid); if (addr) - mod_memmap(DIV_ROUND_UP(size, PAGE_SIZE)); + memmap_pages_add(DIV_ROUND_UP(size, PAGE_SIZE)); return addr; } @@ -322,7 +322,7 @@ static void free_page_ext(void *addr) struct page *page; table_size = page_ext_size * PAGES_PER_SECTION; - mod_memmap(-1L * (DIV_ROUND_UP(table_size, PAGE_SIZE))); + memmap_pages_add(-1L * (DIV_ROUND_UP(table_size, PAGE_SIZE))); if (is_vmalloc_addr(addr)) { vfree(addr); --- a/mm/sparse.c~mm-dont-account-memmap-per-node-v5 +++ a/mm/sparse.c @@ -463,7 +463,7 @@ static void __init sparse_buffer_init(un sparsemap_buf = memmap_alloc(size, section_map_size(), addr, nid, true); sparsemap_buf_end = sparsemap_buf + size; #ifndef CONFIG_SPARSEMEM_VMEMMAP - mod_memmap_boot(DIV_ROUND_UP(size, PAGE_SIZE)); + memmap_boot_pages_add(DIV_ROUND_UP(size, PAGE_SIZE)); #endif } @@ -643,7 +643,7 @@ static void depopulate_section_memmap(un unsigned long start = (unsigned long) pfn_to_page(pfn); unsigned long end = start + nr_pages * sizeof(struct page); - mod_memmap(-1L * (DIV_ROUND_UP(end - start, PAGE_SIZE))); + memmap_pages_add(-1L * (DIV_ROUND_UP(end - start, PAGE_SIZE))); vmemmap_free(start, end, altmap); } static void free_map_bootmem(struct page *memmap) --- a/mm/sparse-vmemmap.c~mm-dont-account-memmap-per-node-v5 +++ a/mm/sparse-vmemmap.c @@ -470,9 +470,9 @@ struct page * __meminit __populate_secti return NULL; if (system_state == SYSTEM_BOOTING) - mod_memmap_boot(DIV_ROUND_UP(end - start, PAGE_SIZE)); + memmap_boot_pages_add(DIV_ROUND_UP(end - start, PAGE_SIZE)); else - mod_memmap(DIV_ROUND_UP(end - start, PAGE_SIZE)); + memmap_pages_add(DIV_ROUND_UP(end - start, PAGE_SIZE)); return pfn_to_page(pfn); } --- a/mm/vmstat.c~mm-dont-account-memmap-per-node-v5 +++ a/mm/vmstat.c @@ -1035,19 +1035,20 @@ unsigned long node_page_state(struct pgl /* * Count number of pages "struct page" and "struct page_ext" consume. - * nr_memmap_boot: # of pages allocated by boot allocator & not part of MemTotal - * nr_memmap: # of pages that were allocated by buddy allocator + * nr_memmap_boot_pages: # of pages allocated by boot allocator + * nr_memmap_pages: # of pages that were allocated by buddy allocator */ -static atomic_long_t nr_memmap_boot, nr_memmap; +static atomic_long_t nr_memmap_boot_pages = ATOMIC_LONG_INIT(0); +static atomic_long_t nr_memmap_pages = ATOMIC_LONG_INIT(0); -void mod_memmap_boot(long delta) +void memmap_boot_pages_add(long delta) { - atomic_long_add(delta, &nr_memmap_boot); + atomic_long_add(delta, &nr_memmap_boot_pages); } -void mod_memmap(long delta) +void memmap_pages_add(long delta) { - atomic_long_add(delta, &nr_memmap); + atomic_long_add(delta, &nr_memmap_pages); } #ifdef CONFIG_COMPACTION @@ -1844,8 +1845,8 @@ static void *vmstat_start(struct seq_fil global_dirty_limits(v + NR_DIRTY_BG_THRESHOLD, v + NR_DIRTY_THRESHOLD); - v[NR_MEMMAP_BOOT] = atomic_long_read(&nr_memmap_boot); - v[NR_MEMMAP] = atomic_long_read(&nr_memmap); + v[NR_MEMMAP_PAGES] = atomic_long_read(&nr_memmap_pages); + v[NR_MEMMAP_BOOT_PAGES] = atomic_long_read(&nr_memmap_boot_pages); v += NR_VM_STAT_ITEMS; #ifdef CONFIG_VM_EVENT_COUNTERS _ Patches currently in -mm which might be from pasha.tatashin@xxxxxxxxxx are mm-dont-account-memmap-on-failure.patch mm-add-system-wide-stats-items-category.patch mm-dont-account-memmap-per-node.patch memcg-increase-the-valid-index-range-for-memcg-stats-v5.patch vmstat-kernel-stack-usage-histogram.patch task_stack-uninline-stack_not_used.patch