The patch titled Subject: rmap: peplace two calls to compound_order with folio_order has been added to the -mm mm-unstable branch. Its filename is rmap-replace-two-calls-to-compound_order-with-folio_order.patch This patch will shortly appear at https://git.kernel.org/pub/scm/linux/kernel/git/akpm/25-new.git/tree/patches/rmap-replace-two-calls-to-compound_order-with-folio_order.patch This patch will later appear in the mm-unstable branch at git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm Before you just go and hit "reply", please: a) Consider who else should be cc'ed b) Prefer to cc a suitable mailing list as well c) Ideally: find the original patch on the mailing list and do a reply-to-all to that, adding suitable additional cc's *** Remember to use Documentation/process/submit-checklist.rst when testing your code *** The -mm tree is included into linux-next via the mm-everything branch at git://git.kernel.org/pub/scm/linux/kernel/git/akpm/mm and is updated there every 2-3 working days ------------------------------------------------------ From: "Matthew Wilcox (Oracle)" <willy@xxxxxxxxxxxxx> Subject: rmap: peplace two calls to compound_order with folio_order Date: Thu, 15 Feb 2024 20:53:05 +0000 Removes two unnecessary conversions from folio to page. Should be no difference in behaviour. Link: https://lkml.kernel.org/r/20240215205307.674707-1-willy@xxxxxxxxxxxxx Signed-off-by: Matthew Wilcox (Oracle) <willy@xxxxxxxxxxxxx> Signed-off-by: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> --- mm/rmap.c | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) --- a/mm/rmap.c~rmap-replace-two-calls-to-compound_order-with-folio_order +++ a/mm/rmap.c @@ -2169,7 +2169,7 @@ static bool try_to_migrate_one(struct fo swp_pte = pte_swp_mkuffd_wp(swp_pte); set_pte_at(mm, pvmw.address, pvmw.pte, swp_pte); trace_set_migration_pte(pvmw.address, pte_val(swp_pte), - compound_order(&folio->page)); + folio_order(folio)); /* * No need to invalidate here it will synchronize on * against the special swap migration pte. @@ -2261,7 +2261,7 @@ static bool try_to_migrate_one(struct fo else set_pte_at(mm, address, pvmw.pte, swp_pte); trace_set_migration_pte(address, pte_val(swp_pte), - compound_order(&folio->page)); + folio_order(folio)); /* * No need to invalidate here it will synchronize on * against the special swap migration pte. _ Patches currently in -mm which might be from willy@xxxxxxxxxxxxx are memcg-convert-mem_cgroup_move_charge_pte_range-to-use-a-folio.patch memcg-return-the-folio-in-union-mc_target.patch memcg-use-a-folio-in-get_mctgt_type.patch memcg-use-a-folio-in-get_mctgt_type_thp.patch mm-add-pfn_swap_entry_folio.patch proc-use-pfn_swap_entry_folio-where-obvious.patch mprotect-use-pfn_swap_entry_folio.patch highmem-add-kernel-doc-for-memcpy__folio.patch rmap-replace-two-calls-to-compound_order-with-folio_order.patch bounds-support-non-power-of-two-config_nr_cpus.patch