On Mon, Oct 16, 2023 at 04:26:30PM -0400, Johannes Weiner wrote: > On Mon, Oct 16, 2023 at 03:49:49PM -0400, Zi Yan wrote: > > On 16 Oct 2023, at 14:51, Johannes Weiner wrote: > > > > > On Mon, Oct 16, 2023 at 11:00:33AM -0400, Zi Yan wrote: > > >> On 16 Oct 2023, at 10:37, Johannes Weiner wrote: > > >> > > >>> On Mon, Oct 16, 2023 at 09:35:34AM -0400, Zi Yan wrote: > > >>>>> The attached patch has all the suggested changes, let me know how it > > >>>>> looks to you. Thanks. > > >>>> > > >>>> The one I sent has free page accounting issues. The attached one fixes them. > > >>> > > >>> Do you still have the warnings? I wonder what went wrong. > > >> > > >> No warnings. But something with the code: > > >> > > >> 1. in your version, split_free_page() is called without changing any pageblock > > >> migratetypes, then split_free_page() is just a no-op, since the page is > > >> just deleted from the free list, then freed via different orders. Buddy allocator > > >> will merge them back. > > > > > > Hm not quite. > > > > > > If it's the tail block of a buddy, I update its type before > > > splitting. The splitting loop looks up the type of each block for > > > sorting it onto freelists. > > > > > > If it's the head block, yes I split it first according to its old > > > type. But then I let it fall through to scanning the block, which will > > > find that buddy, update its type and move it. > > > > That is the issue, since split_free_page() assumes the pageblocks of > > that free page have different types. It basically just free the page > > with different small orders summed up to the original free page order. > > If all pageblocks of the free page have the same migratetype, __free_one_page() > > will merge these small order pages back to the original order free page. > > duh, of course, you're right. Thanks for patiently explaining this. > > > >> 2. in my version, I set pageblock migratetype to new_mt before split_free_page(), > > >> but it causes free page accounting issues, since in the case of head, free pages > > >> are deleted from new_mt when they are in old_mt free list and the accounting > > >> decreases new_mt free page number instead of old_mt one. > > > > > > Right, that makes sense. > > > > > >> Basically, split_free_page() is awkward as it relies on preset migratetypes, > > >> which changes migratetypes without deleting the free pages from the list first. > > >> That is why I came up with the new split_free_page() below. > > > > > > Yeah, the in-between thing is bad. Either it fixes the migratetype > > > before deletion, or it doesn't do the deletion. I'm thinking it would > > > be simpler to move the deletion out instead. > > > > Yes and no. After deletion, a free page no longer has PageBuddy set and > > has buddy_order information cleared. Either we reset PageBuddy and order > > to the deleted free page, or split_free_page() needs to be changed to > > accept pages without the information (basically remove the PageBuddy > > and order check code). > > Good point, that requires extra care. > > It's correct in the code now, but it deserves a comment, especially > because of the "buddy" naming in the new split function. > > > >> Hmm, if CONFIG_ARCH_FORCE_MAX_ORDER can make a buddy have more than one > > >> pageblock and in turn makes an in-use page have more than one pageblock, > > >> we will have problems. Since in isolate_single_pageblock(), an in-use page > > >> can have part of its pageblock set to a different migratetype and be freed, > > >> causing the free page with unmatched migratetypes. We might need to > > >> free pages at pageblock_order if their orders are bigger than pageblock_order. > > > > > > Is this a practical issue? You mentioned that right now only gigantic > > > pages can be larger than a pageblock, and those are freed in order-0 > > > chunks. > > > > Only if the system allocates a page (non hugetlb pages) with >pageblock_order > > and frees it with the same order. I just do not know if such pages exist on > > other arch than x86. Maybe I just think too much. > > Hm, I removed LRU pages from the handling (and added the warning) but > I left in PageMovable(). The only users are z3fold, zsmalloc and > memory ballooning. AFAICS none of them can be bigger than a pageblock. > Let me remove that and add a warning for that case as well. > > This way, we only attempt to migrate hugetlb, where we know the free > path - and get warnings for anything else that's larger than expected. > > This seems like the safest option. On the off chance that there is a > regression, it won't jeopardize anybody's systems, while the warning > provides all the information we need to debug what's going on. This delta on top? diff --git a/mm/page_alloc.c b/mm/page_alloc.c index b5292ad9860c..0da7c61af37e 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -1628,7 +1628,7 @@ static int move_freepages_block(struct zone *zone, struct page *page, } #ifdef CONFIG_MEMORY_ISOLATION -/* Look for a multi-block buddy that straddles start_pfn */ +/* Look for a buddy that straddles start_pfn */ static unsigned long find_large_buddy(unsigned long start_pfn) { int order = 0; @@ -1652,7 +1652,7 @@ static unsigned long find_large_buddy(unsigned long start_pfn) return start_pfn; } -/* Split a multi-block buddy into its individual pageblocks */ +/* Split a multi-block free page into its individual pageblocks */ static void split_large_buddy(struct zone *zone, struct page *page, unsigned long pfn, int order) { @@ -1661,6 +1661,9 @@ static void split_large_buddy(struct zone *zone, struct page *page, VM_WARN_ON_ONCE(order < pageblock_order); VM_WARN_ON_ONCE(pfn & (pageblock_nr_pages - 1)); + /* Caller removed page from freelist, buddy info cleared! */ + VM_WARN_ON_ONCE(PageBuddy(page)); + while (pfn != end_pfn) { int mt = get_pfnblock_migratetype(page, pfn); diff --git a/mm/page_isolation.c b/mm/page_isolation.c index b4d53545496d..c8b3c0699683 100644 --- a/mm/page_isolation.c +++ b/mm/page_isolation.c @@ -399,14 +399,8 @@ static int isolate_single_pageblock(unsigned long boundary_pfn, int flags, continue; } - VM_WARN_ON_ONCE_PAGE(PageLRU(page), page); - #if defined CONFIG_COMPACTION || defined CONFIG_CMA - /* - * hugetlb, and movable compound pages can be - * migrated. Otherwise, fail the isolation. - */ - if (PageHuge(page) || __PageMovable(page)) { + if (PageHuge(page)) { struct compact_control cc = { .nr_migratepages = 0, .order = -1, @@ -426,9 +420,19 @@ static int isolate_single_pageblock(unsigned long boundary_pfn, int flags, pfn = head_pfn + nr_pages; continue; - } else + } + + /* + * These pages are movable too, but they're + * not expected to exceed pageblock_order. + * + * Let us know when they do, so we can add + * proper free and split handling for them. + */ + VM_WARN_ON_ONCE_PAGE(PageLRU(page), page); + VM_WARN_ON_ONCE_PAGE(__PageMovable(page), page); #endif - goto failed; + goto failed; } pfn++;