Re: [PATCH] mm: Fix endless reclaim on machines with unaccepted memory.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Jul 17, 2024 at 02:06:46PM +0200, Michal Hocko wrote:
> Please try to investigate this further. The patch as is looks rather
> questionable to me TBH. Spilling unaccepted memory into the reclaim
> seems like something we should avoid if possible as this is something

Okay, I believe I have a better understanding of the situation:

- __alloc_pages_bulk() takes pages from the free list without accepting
  more memory. This can cause number of free pages to fall below the
  watermark.

  This issue can be resolved by accepting more memory in
  __alloc_pages_bulk() if the watermark check fails.

  The problem is not only related to unallocated memory. I think the
  deferred page initialization mechanism could result in premature OOM if
  __alloc_pages_bulk() allocates pages faster than deferred page
  initialization can add them to the free lists. However, this scenario is
  unlikely.

- There is nothing that compels the kernel to accept more memory after the
  watermarks have been calculated in __setup_per_zone_wmarks(). This can
  put us under the watermark.

  This issue can be resolved by accepting memory up to the watermark after
  the watermarks have been initialized.

- Once kswapd is started, it will begin spinning if we are below the
  watermark and there is no memory that can be reclaimed. Once the above
  problems are fixed, the issue will be resolved.

- The kernel needs to accept memory up to the PROMO watermark. This will
  prevent unaccepted memory from interfering with NUMA balancing.

The patch below addresses the issues I listed earlier. It is not yet ready
for application. Please see the issues listed below.

Andrew, please drop the current patch.

There are a few more things I am worried about:

- The current get_page_from_freelist() and patched __alloc_pages_bulk()
  only try to accept memory if the requested (alloc_flags & ALLOC_WMARK_MASK)
  watermark check fails. For example, if a requested allocation with
  ALLOC_WMARK_MIN is called, we will not try to accept more memory, which
  could potentially put us under the high/promo watermark and cause the
  following kswapd start to get us into an endless loop.

  Do we want to make memory acceptance in these paths independent of
  alloc_flags?

- __isolate_free_page() removes a page from the free list without
  accepting new memory. The function is called with the zone lock taken.
  It is bad idea to accept memory while holding the zone lock, but
  the alternative of pushing the accept to the caller is not much better.

  I have not observed any issues caused by __isolate_free_page() in
  practice, but there is no reason why it couldn't potentially cause
  problems.
 
- The function take_pages_off_buddy() also removes pages from the free
  list without accepting new memory. Unlike the function
  __isolate_free_page(), it is called without the zone lock being held, so
  we can accept memory there. I believe we should do so.

I understand why adding unaccepted memory handling into the reclaim path
is questionable. However, it may be the best way to handle cases like
__isolate_free_page() and possibly others in the future that directly take
memory from free lists.

Any thoughts?

I am still new to reclaim code and may be overlooking something
significant. Please correct any misconceptions you see.

diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h
index c11b7cde81ef..5e0bdfbe2f1f 100644
--- a/include/linux/mmzone.h
+++ b/include/linux/mmzone.h
@@ -667,6 +667,7 @@ enum zone_watermarks {
 #define min_wmark_pages(z) (z->_watermark[WMARK_MIN] + z->watermark_boost)
 #define low_wmark_pages(z) (z->_watermark[WMARK_LOW] + z->watermark_boost)
 #define high_wmark_pages(z) (z->_watermark[WMARK_HIGH] + z->watermark_boost)
+#define promo_wmark_pages(z) (z->_watermark[WMARK_PROMO] + z->watermark_boost)
 #define wmark_pages(z, i) (z->_watermark[i] + z->watermark_boost)
 
 /*
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index c62805dbd608..d537c633c6e9 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -1748,7 +1748,7 @@ static bool pgdat_free_space_enough(struct pglist_data *pgdat)
 			continue;
 
 		if (zone_watermark_ok(zone, 0,
-				      wmark_pages(zone, WMARK_PROMO) + enough_wmark,
+				      promo_wmark_pages(zone) + enough_wmark,
 				      ZONE_MOVABLE, 0))
 			return true;
 	}
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index 14d39f34d336..b744743d14a2 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -4462,6 +4462,22 @@ unsigned long __alloc_pages_bulk(gfp_t gfp, int preferred_nid,
 				alloc_flags, gfp)) {
 			break;
 		}
+
+		if (has_unaccepted_memory()) {
+			if (try_to_accept_memory(zone, 0))
+				break;
+		}
+
+#ifdef CONFIG_DEFERRED_STRUCT_PAGE_INIT
+		/*
+		 * Watermark failed for this zone, but see if we can
+		 * grow this zone if it contains deferred pages.
+		 */
+		if (deferred_pages_enabled()) {
+			if (_deferred_grow_zone(zone, 0))
+				break;
+		}
+#endif
 	}
 
 	/*
@@ -5899,6 +5915,9 @@ static void __setup_per_zone_wmarks(void)
 		zone->_watermark[WMARK_PROMO] = high_wmark_pages(zone) + tmp;
 
 		spin_unlock_irqrestore(&zone->lock, flags);
+
+		if (managed_zone(zone))
+			try_to_accept_memory(zone, 0);
 	}
 
 	/* update totalreserve_pages */
@@ -6866,8 +6885,8 @@ static bool try_to_accept_memory(struct zone *zone, unsigned int order)
 	long to_accept;
 	int ret = false;
 
-	/* How much to accept to get to high watermark? */
-	to_accept = high_wmark_pages(zone) -
+	/* How much to accept to get to promo watermark? */
+	to_accept = wmark_pages(zone, WMARK_PROMO) -
 		    (zone_page_state(zone, NR_FREE_PAGES) -
 		    __zone_watermark_unusable_free(zone, order, 0));
 
diff --git a/mm/vmscan.c b/mm/vmscan.c
index 3ef654addd44..d20242e36904 100644
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -6607,7 +6607,7 @@ static bool pgdat_balanced(pg_data_t *pgdat, int order, int highest_zoneidx)
 			continue;
 
 		if (sysctl_numa_balancing_mode & NUMA_BALANCING_MEMORY_TIERING)
-			mark = wmark_pages(zone, WMARK_PROMO);
+			mark = promo_wmark_pages(zone);
 		else
 			mark = high_wmark_pages(zone);
 		if (zone_watermark_ok_safe(zone, order, mark, highest_zoneidx))
-- 
  Kiryl Shutsemau / Kirill A. Shutemov




[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux