On Mon, Mar 25, 2024 at 10:56:46PM +0800, Baoquan He wrote: > Since the current calculation of calc_nr_kernel_pages() has taken into > consideration of kernel reserved memory, no need to have > arch_reserved_kernel_pages() any more. > > Signed-off-by: Baoquan He <bhe@xxxxxxxxxx> Reviewed-by: Mike Rapoport (IBM) <rppt@xxxxxxxxxx> > --- > arch/powerpc/include/asm/mmu.h | 4 ---- > arch/powerpc/kernel/fadump.c | 5 ----- > include/linux/mm.h | 3 --- > mm/mm_init.c | 12 ------------ > 4 files changed, 24 deletions(-) > > diff --git a/arch/powerpc/include/asm/mmu.h b/arch/powerpc/include/asm/mmu.h > index 3b72c7ed24cf..aa5c0fd5edb1 100644 > --- a/arch/powerpc/include/asm/mmu.h > +++ b/arch/powerpc/include/asm/mmu.h > @@ -406,9 +406,5 @@ extern void *abatron_pteptrs[2]; > #include <asm/nohash/mmu.h> > #endif > > -#if defined(CONFIG_FA_DUMP) || defined(CONFIG_PRESERVE_FA_DUMP) > -#define __HAVE_ARCH_RESERVED_KERNEL_PAGES > -#endif > - > #endif /* __KERNEL__ */ > #endif /* _ASM_POWERPC_MMU_H_ */ > diff --git a/arch/powerpc/kernel/fadump.c b/arch/powerpc/kernel/fadump.c > index d14eda1e8589..ae8c7619e597 100644 > --- a/arch/powerpc/kernel/fadump.c > +++ b/arch/powerpc/kernel/fadump.c > @@ -1735,8 +1735,3 @@ static void __init fadump_reserve_crash_area(u64 base) > memblock_reserve(mstart, msize); > } > } > - > -unsigned long __init arch_reserved_kernel_pages(void) > -{ > - return memblock_reserved_size() / PAGE_SIZE; > -} > diff --git a/include/linux/mm.h b/include/linux/mm.h > index ad19350e1538..ab1ba0a31429 100644 > --- a/include/linux/mm.h > +++ b/include/linux/mm.h > @@ -3221,9 +3221,6 @@ static inline void show_mem(void) > extern long si_mem_available(void); > extern void si_meminfo(struct sysinfo * val); > extern void si_meminfo_node(struct sysinfo *val, int nid); > -#ifdef __HAVE_ARCH_RESERVED_KERNEL_PAGES > -extern unsigned long arch_reserved_kernel_pages(void); > -#endif > > extern __printf(3, 4) > void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...); > diff --git a/mm/mm_init.c b/mm/mm_init.c > index e269a724f70e..089dc60159e9 100644 > --- a/mm/mm_init.c > +++ b/mm/mm_init.c > @@ -2373,17 +2373,6 @@ void __init page_alloc_init_late(void) > page_alloc_sysctl_init(); > } > > -#ifndef __HAVE_ARCH_RESERVED_KERNEL_PAGES > -/* > - * Returns the number of pages that arch has reserved but > - * is not known to alloc_large_system_hash(). > - */ > -static unsigned long __init arch_reserved_kernel_pages(void) > -{ > - return 0; > -} > -#endif > - > /* > * Adaptive scale is meant to reduce sizes of hash tables on large memory > * machines. As memory size is increased the scale is also increased but at > @@ -2426,7 +2415,6 @@ void *__init alloc_large_system_hash(const char *tablename, > if (!numentries) { > /* round applicable memory size up to nearest megabyte */ > numentries = nr_kernel_pages; > - numentries -= arch_reserved_kernel_pages(); > > /* It isn't necessary when PAGE_SIZE >= 1MB */ > if (PAGE_SIZE < SZ_1M) > -- > 2.41.0 > -- Sincerely yours, Mike.