Re: [PATCH v2 08/11] mm/memory-failure: Convert hwpoison_user_mappings to take a folio

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 4/8/2024 12:42 PM, Matthew Wilcox (Oracle) wrote:

Pass the folio from the callers, and use it throughout instead of hpage.
Saves dozens of calls to compound_head().

Acked-by: Miaohe Lin <linmiaohe@xxxxxxxxxx>
Signed-off-by: Matthew Wilcox (Oracle) <willy@xxxxxxxxxxxxx>
---
  mm/memory-failure.c | 30 +++++++++++++++---------------
  1 file changed, 15 insertions(+), 15 deletions(-)

diff --git a/mm/memory-failure.c b/mm/memory-failure.c
index 1c7c73776604..fae0b42f0aaf 100644
--- a/mm/memory-failure.c
+++ b/mm/memory-failure.c
@@ -1543,24 +1543,24 @@ static int get_hwpoison_page(struct page *p, unsigned long flags)
   * Do all that is necessary to remove user space mappings. Unmap
   * the pages and send SIGBUS to the processes if the data was dirty.
   */
-static bool hwpoison_user_mappings(struct page *p, unsigned long pfn,
-				  int flags, struct page *hpage)
+static bool hwpoison_user_mappings(struct folio *folio, struct page *p,
+		unsigned long pfn, int flags)
  {
-	struct folio *folio = page_folio(hpage);
  	enum ttu_flags ttu = TTU_IGNORE_MLOCK | TTU_SYNC | TTU_HWPOISON;
  	struct address_space *mapping;
  	LIST_HEAD(tokill);
  	bool unmap_success;
  	int forcekill;
-	bool mlocked = PageMlocked(hpage);
+	bool mlocked = folio_test_mlocked(folio);
/*
  	 * Here we are interested only in user-mapped pages, so skip any
  	 * other types of pages.
  	 */
-	if (PageReserved(p) || PageSlab(p) || PageTable(p) || PageOffline(p))
+	if (folio_test_reserved(folio) || folio_test_slab(folio) ||
+	    folio_test_pgtable(folio) || folio_test_offline(folio))
  		return true;
-	if (!(PageLRU(hpage) || PageHuge(p)))
+	if (!(folio_test_lru(folio) || folio_test_hugetlb(folio)))
  		return true;
/*
@@ -1570,7 +1570,7 @@ static bool hwpoison_user_mappings(struct page *p, unsigned long pfn,
  	if (!page_mapped(p))
  		return true;
- if (PageSwapCache(p)) {
+	if (folio_test_swapcache(folio)) {
  		pr_err("%#lx: keeping poisoned page in swap cache\n", pfn);
  		ttu &= ~TTU_HWPOISON;
  	}
@@ -1581,11 +1581,11 @@ static bool hwpoison_user_mappings(struct page *p, unsigned long pfn,
  	 * XXX: the dirty test could be racy: set_page_dirty() may not always
  	 * be called inside page lock (it's recommended but not enforced).
  	 */
-	mapping = page_mapping(hpage);
-	if (!(flags & MF_MUST_KILL) && !PageDirty(hpage) && mapping &&
+	mapping = folio_mapping(folio);
+	if (!(flags & MF_MUST_KILL) && !folio_test_dirty(folio) && mapping &&
  	    mapping_can_writeback(mapping)) {
-		if (page_mkclean(hpage)) {
-			SetPageDirty(hpage);
+		if (folio_mkclean(folio)) {
+			folio_set_dirty(folio);
  		} else {
  			ttu &= ~TTU_HWPOISON;
  			pr_info("%#lx: corrupted page was clean: dropped without side effects\n",
@@ -1600,7 +1600,7 @@ static bool hwpoison_user_mappings(struct page *p, unsigned long pfn,
  	 */
  	collect_procs(folio, p, &tokill, flags & MF_ACTION_REQUIRED);
- if (PageHuge(hpage) && !PageAnon(hpage)) {
+	if (folio_test_hugetlb(folio) && !folio_test_anon(folio)) {
  		/*
  		 * For hugetlb pages in shared mappings, try_to_unmap
  		 * could potentially call huge_pmd_unshare.  Because of
@@ -1640,7 +1640,7 @@ static bool hwpoison_user_mappings(struct page *p, unsigned long pfn,
  	 * use a more force-full uncatchable kill to prevent
  	 * any accesses to the poisoned memory.
  	 */
-	forcekill = PageDirty(hpage) || (flags & MF_MUST_KILL) ||
+	forcekill = folio_test_dirty(folio) || (flags & MF_MUST_KILL) ||
  		    !unmap_success;
  	kill_procs(&tokill, forcekill, !unmap_success, pfn, flags);
@@ -2084,7 +2084,7 @@ static int try_memory_failure_hugetlb(unsigned long pfn, int flags, int *hugetlb page_flags = folio->flags; - if (!hwpoison_user_mappings(p, pfn, flags, &folio->page)) {
+	if (!hwpoison_user_mappings(folio, p, pfn, flags)) {
  		folio_unlock(folio);
  		return action_result(pfn, MF_MSG_UNMAP_FAILED, MF_IGNORED);
  	}
@@ -2351,7 +2351,7 @@ int memory_failure(unsigned long pfn, int flags)
  	 * Now take care of user space mappings.
  	 * Abort on fail: __filemap_remove_folio() assumes unmapped page.
  	 */
-	if (!hwpoison_user_mappings(p, pfn, flags, p)) {
+	if (!hwpoison_user_mappings(folio, p, pfn, flags)) {
  		res = action_result(pfn, MF_MSG_UNMAP_FAILED, MF_IGNORED);
  		goto unlock_page;
  	}

Looks good to me.

Reviewed-by: Jane Chu <jane.chu@xxxxxxxxxx>

-jane





[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux