This is in preparation for only handling kmem pages in __split_huge_page(). Signed-off-by: Matthew Wilcox (Oracle) <willy@xxxxxxxxxxxxx> --- mm/huge_memory.c | 11 +++++++++-- 1 file changed, 9 insertions(+), 2 deletions(-) diff --git a/mm/huge_memory.c b/mm/huge_memory.c index f92068864469..44d25a74b611 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -3234,6 +3234,10 @@ static void __split_huge_page_tail(struct folio *folio, int tail, folio_set_large_rmappable(new_folio); } +#ifdef CONFIG_MEMCG + new_folio->memcg_data = folio->memcg_data; +#endif + /* Finally unfreeze refcount. Additional reference from page cache. */ page_ref_unfreeze(page_tail, 1 + ((!folio_test_anon(folio) || folio_test_swapcache(folio)) ? @@ -3267,8 +3271,11 @@ static void __split_huge_page(struct page *page, struct list_head *list, int order = folio_order(folio); unsigned int nr = 1 << order; - /* complete memcg works before add pages to LRU */ - split_page_memcg(head, order, new_order); +#ifdef CONFIG_MEMCG + if (folio_memcg_charged(folio)) + css_get_many(&folio_memcg(folio)->css, + (1 << (order - new_order)) - 1); +#endif if (folio_test_anon(folio) && folio_test_swapcache(folio)) { offset = swap_cache_index(folio->swap); -- 2.43.0