The patch titled Subject: mm/thp: narrow lru locking has been added to the -mm tree. Its filename is mm-thp-narrow-lru-locking.patch This patch should soon appear at http://ozlabs.org/~akpm/mmots/broken-out/mm-thp-narrow-lru-locking.patch and later at http://ozlabs.org/~akpm/mmotm/broken-out/mm-thp-narrow-lru-locking.patch Before you just go and hit "reply", please: a) Consider who else should be cc'ed b) Prefer to cc a suitable mailing list as well c) Ideally: find the original patch on the mailing list and do a reply-to-all to that, adding suitable additional cc's *** Remember to use Documentation/process/submit-checklist.rst when testing your code *** The -mm tree is included into linux-next and is updated there every 3-4 working days ------------------------------------------------------ From: Alex Shi <alex.shi@xxxxxxxxxxxxxxxxx> Subject: mm/thp: narrow lru locking Lru locking just guards the lru list and subpage's Mlocked. Including other things can't give help just delay the locking release. So narrow the locking for early lock release and better code meaning. Link: http://lkml.kernel.org/r/1583146830-169516-7-git-send-email-alex.shi@xxxxxxxxxxxxxxxxx Signed-off-by: Alex Shi <alex.shi@xxxxxxxxxxxxxxxxx> Cc: Kirill A. Shutemov <kirill@xxxxxxxxxxxxx> Cc: Andrea Arcangeli <aarcange@xxxxxxxxxx> Cc: Johannes Weiner <hannes@xxxxxxxxxxx> Cc: Daniel Jordan <daniel.m.jordan@xxxxxxxxxx> Cc: Hugh Dickins <hughd@xxxxxxxxxx> Cc: Konstantin Khlebnikov <khlebnikov@xxxxxxxxxxxxxx> Cc: Matthew Wilcox <willy@xxxxxxxxxxxxx> Cc: Michal Hocko <mhocko@xxxxxxxxxx> Cc: Mike Kravetz <kravetz@xxxxxxxxxx> Cc: Tejun Heo <tj@xxxxxxxxxx> Cc: Vladimir Davydov <vdavydov.dev@xxxxxxxxx> Cc: Yang Shi <yang.shi@xxxxxxxxxxxxxxxxx> Signed-off-by: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> --- mm/huge_memory.c | 17 +++++++---------- 1 file changed, 7 insertions(+), 10 deletions(-) --- a/mm/huge_memory.c~mm-thp-narrow-lru-locking +++ a/mm/huge_memory.c @@ -2559,13 +2559,14 @@ static void __split_huge_page_tail(struc } static void __split_huge_page(struct page *page, struct list_head *list, - pgoff_t end, unsigned long flags) + pgoff_t end) { struct page *head = compound_head(page); pg_data_t *pgdat = page_pgdat(head); struct lruvec *lruvec; struct address_space *swap_cache = NULL; unsigned long offset = 0; + unsigned long flags; int i; lruvec = mem_cgroup_page_lruvec(head, pgdat); @@ -2581,6 +2582,9 @@ static void __split_huge_page(struct pag xa_lock(&swap_cache->i_pages); } + /* Lru list would be changed, don't care head's LRU bit. */ + spin_lock_irqsave(&pgdat->lru_lock, flags); + for (i = HPAGE_PMD_NR - 1; i >= 1; i--) { __split_huge_page_tail(head, i, lruvec, list); /* Some pages can be beyond i_size: drop them from page cache */ @@ -2598,6 +2602,7 @@ static void __split_huge_page(struct pag head + i, 0); } } + spin_unlock_irqrestore(&pgdat->lru_lock, flags); ClearPageCompound(head); @@ -2618,8 +2623,6 @@ static void __split_huge_page(struct pag xa_unlock(&head->mapping->i_pages); } - spin_unlock_irqrestore(&pgdat->lru_lock, flags); - remap_page(head); for (i = 0; i < HPAGE_PMD_NR; i++) { @@ -2757,13 +2760,11 @@ bool can_split_huge_page(struct page *pa int split_huge_page_to_list(struct page *page, struct list_head *list) { struct page *head = compound_head(page); - struct pglist_data *pgdata = NODE_DATA(page_to_nid(head)); struct deferred_split *ds_queue = get_deferred_split_queue(head); struct anon_vma *anon_vma = NULL; struct address_space *mapping = NULL; int count, mapcount, extra_pins, ret; bool mlocked; - unsigned long flags; pgoff_t end; VM_BUG_ON_PAGE(is_huge_zero_page(head), head); @@ -2829,9 +2830,6 @@ int split_huge_page_to_list(struct page if (mlocked) lru_add_drain(); - /* prevent PageLRU to go away from under us, and freeze lru stats */ - spin_lock_irqsave(&pgdata->lru_lock, flags); - if (mapping) { XA_STATE(xas, &mapping->i_pages, page_index(head)); @@ -2861,7 +2859,7 @@ int split_huge_page_to_list(struct page __dec_node_page_state(head, NR_FILE_THPS); } - __split_huge_page(page, list, end, flags); + __split_huge_page(page, list, end); if (PageSwapCache(head)) { swp_entry_t entry = { .val = page_private(head) }; @@ -2880,7 +2878,6 @@ int split_huge_page_to_list(struct page spin_unlock(&ds_queue->split_queue_lock); fail: if (mapping) xa_unlock(&mapping->i_pages); - spin_unlock_irqrestore(&pgdata->lru_lock, flags); remap_page(head); ret = -EBUSY; } _ Patches currently in -mm which might be from alex.shi@xxxxxxxxxxxxxxxxx are ocfs2-remove-fs_ocfs2_nm.patch ocfs2-remove-unused-macros.patch ocfs2-use-ocfs2_sec_bits-in-macro.patch ocfs2-remove-dlm_lock_is_remote.patch ocfs2-remove-useless-err.patch mm-vmscan-remove-unnecessary-lruvec-adding.patch mm-memcg-fold-lock_page_lru-into-commit_charge.patch mm-page_idle-no-unlikely-double-check-for-idle-page-counting.patch mm-thp-move-lru_add_page_tail-func-to-huge_memoryc.patch mm-thp-clean-up-lru_add_page_tail.patch mm-thp-narrow-lru-locking.patch