From: zhong jiang <zhongjiang@xxxxxxxxxx> when required_kernelcore decrease to zero, we should exit the loop in time. because It will waste time to scan the remainder node. Signed-off-by: zhong jiang <zhongjiang@xxxxxxxxxx> --- mm/page_alloc.c | 10 +++++++--- 1 file changed, 7 insertions(+), 3 deletions(-) diff --git a/mm/page_alloc.c b/mm/page_alloc.c index ea759b9..be7df17 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -6093,7 +6093,7 @@ static unsigned long __init early_calculate_totalpages(void) unsigned long pages = end_pfn - start_pfn; totalpages += pages; - if (pages) + if (!node_isset(nid, node_states[N_MEMORY]) && pages) node_set_state(nid, N_MEMORY); } return totalpages; @@ -6115,6 +6115,7 @@ static void __init find_zone_movable_pfns_for_nodes(void) unsigned long totalpages = early_calculate_totalpages(); int usable_nodes = nodes_weight(node_states[N_MEMORY]); struct memblock_region *r; + bool avoid_loop = false; /* Need to find movable_zone earlier when movable_node is specified. */ find_usable_zone_for_movable(); @@ -6275,6 +6276,8 @@ restart: required_kernelcore -= min(required_kernelcore, size_pages); kernelcore_remaining -= size_pages; + if (!required_kernelcore && avoid_loop) + goto out2; if (!kernelcore_remaining) break; } @@ -6287,9 +6290,10 @@ restart: * satisfied */ usable_nodes--; - if (usable_nodes && required_kernelcore > usable_nodes) + if (usable_nodes && required_kernelcore > usable_nodes) { + avoid_loop = true; goto restart; - + } out2: /* Align start of ZONE_MOVABLE on all nids to MAX_ORDER_NR_PAGES */ for (nid = 0; nid < MAX_NUMNODES; nid++) -- 1.8.3.1 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>