On 26/03/2024 17:23, Zi Yan wrote: > On 26 Mar 2024, at 12:10, Pankaj Raghav (Samsung) wrote: > >> On Mon, Mar 25, 2024 at 07:06:04PM +0000, Matthew Wilcox wrote: >>> On Wed, Mar 13, 2024 at 06:02:49PM +0100, Pankaj Raghav (Samsung) wrote: >>>> From: Pankaj Raghav <p.raghav@xxxxxxxxxxx> >>>> >>>> As we don't have a way to split a folio to a any given lower folio >>>> order yet, avoid splitting the folio in split_huge_page_to_list() if it >>>> has a minimum folio order requirement. >>> >>> FYI, Zi Yan's patch to do that is now in Andrew's tree. >>> c010d47f107f609b9f4d6a103b6dfc53889049e9 in current linux-next (dated >>> Feb 26) >> >> Yes, I started playing with the patches but I am getting a race condition >> resulting in a null-ptr-deref for which I don't have a good answer for >> yet. >> >> @zi yan Did you encounter any issue like this when you were testing? >> >> I did the following change (just a prototype) instead of this patch: >> >> diff --git a/mm/huge_memory.c b/mm/huge_memory.c >> index 9859aa4f7553..63ee7b6ed03d 100644 >> --- a/mm/huge_memory.c >> +++ b/mm/huge_memory.c >> @@ -3041,6 +3041,10 @@ int split_huge_page_to_list_to_order(struct page *page, struct list_head *list, >> { >> struct folio *folio = page_folio(page); >> struct deferred_split *ds_queue = get_deferred_split_queue(folio); >> + unsigned int mapping_min_order = mapping_min_folio_order(folio->mapping); > > I am not sure if this is right. Since folio can be anonymous and folio->mapping > does not point to struct address_space. > >> + >> + if (!folio_test_anon(folio)) Hmm, but I update the new_order only if it is not anonymous. Do you think it is still wrong? >> + new_order = max_t(unsigned int, mapping_min_order, new_order); >> /* reset xarray order to new order after split */ >> XA_STATE_ORDER(xas, &folio->mapping->i_pages, folio->index, new_order); >> struct anon_vma *anon_vma = NULL; >> @@ -3117,6 +3121,8 @@ int split_huge_page_to_list_to_order(struct page *page, struct list_head *list, >> goto out; >> } >> >> + // XXX: Remove it later >> + VM_WARN_ON_FOLIO((new_order < mapping_min_order), folio); >> gfp = current_gfp_context(mapping_gfp_mask(mapping) & >> GFP_RECLAIM_MASK); >> >> I am getting a random null-ptr deref when I run generic/176 multiple >> times with different blksizes. I still don't have a minimal reproducer >> for this issue. Race condition during writeback: >> >> filemap_get_folios_tag+0x171/0x5c0: >> arch_atomic_read at arch/x86/include/asm/atomic.h:23 >> (inlined by) raw_atomic_read at include/linux/atomic/atomic-arch-fallback.h:457 >> (inlined by) raw_atomic_fetch_add_unless at include/linux/atomic/atomic-arch-fallback.h:2426 >> (inlined by) raw_atomic_add_unless at include/linux/atomic/atomic-arch-fallback.h:2456 >> (inlined by) atomic_add_unless at include/linux/atomic/atomic-instrumented.h:1518 >> (inlined by) page_ref_add_unless at include/linux/page_ref.h:238 >> (inlined by) folio_ref_add_unless at include/linux/page_ref.h:247 >> (inlined by) folio_ref_try_add_rcu at include/linux/page_ref.h:280 >> (inlined by) folio_try_get_rcu at include/linux/page_ref.h:313 >> (inlined by) find_get_entry at mm/filemap.c:1984 >> (inlined by) filemap_get_folios_tag at mm/filemap.c:2222 >> >> >> >> [ 537.863105] ================================================================== >> [ 537.863968] BUG: KASAN: null-ptr-deref in filemap_get_folios_tag+0x171/0x5c0 >> [ 537.864581] Write of size 4 at addr 0000000000000036 by task kworker/u32:5/366 >> [ 537.865123] >> [ 537.865293] CPU: 6 PID: 366 Comm: kworker/u32:5 Not tainted 6.8.0-11739-g7d0c6e7b5a7d-dirty #795 >> [ 537.867201] Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS rel-1.16.3-0-ga6ed6b701f0a-prebuilt.qemu.org 04/01/2014 >> [ 537.868444] Workqueue: writeback wb_workfn (flush-254:32) >> [ 537.869055] Call Trace: >> [ 537.869341] <TASK> >> [ 537.869569] dump_stack_lvl+0x4f/0x70 >> [ 537.869938] kasan_report+0xbd/0xf0 >> [ 537.870293] ? filemap_get_folios_tag+0x171/0x5c0 >> [ 537.870767] ? filemap_get_folios_tag+0x171/0x5c0 >> [ 537.871578] kasan_check_range+0x101/0x1b0 >> [ 537.871893] filemap_get_folios_tag+0x171/0x5c0 >> [ 537.872269] ? __pfx_filemap_get_folios_tag+0x10/0x10 >> [ 537.872857] ? __pfx___submit_bio+0x10/0x10 >> [ 537.873326] ? mlock_drain_local+0x234/0x3f0 >> [ 537.873938] writeback_iter+0x59a/0xe00 >> [ 537.874477] ? __pfx_iomap_do_writepage+0x10/0x10 >> [ 537.874969] write_cache_pages+0x7f/0x100 >> [ 537.875396] ? __pfx_write_cache_pages+0x10/0x10 >> [ 537.875892] ? do_raw_spin_lock+0x12d/0x270 >> [ 537.876345] ? __pfx_do_raw_spin_lock+0x10/0x10 >> [ 537.876804] iomap_writepages+0x88/0xf0 >> [ 537.877186] xfs_vm_writepages+0x120/0x190 >> [ 537.877705] ? __pfx_xfs_vm_writepages+0x10/0x10 >> [ 537.878161] ? lock_release+0x36f/0x670 >> [ 537.878521] ? __wb_calc_thresh+0xe5/0x3b0 >> [ 537.878892] ? __pfx_lock_release+0x10/0x10 >> [ 537.879308] do_writepages+0x170/0x7a0 >> [ 537.879676] ? __pfx_do_writepages+0x10/0x10 >> [ 537.880182] ? writeback_sb_inodes+0x312/0xe40 >> [ 537.880689] ? reacquire_held_locks+0x1f1/0x4a0 >> [ 537.881193] ? writeback_sb_inodes+0x312/0xe40 >> [ 537.881665] ? find_held_lock+0x2d/0x110 >> [ 537.882104] ? lock_release+0x36f/0x670 >> [ 537.883344] ? wbc_attach_and_unlock_inode+0x3b8/0x710 >> [ 537.883853] ? __pfx_lock_release+0x10/0x10 >> [ 537.884229] ? __pfx_lock_release+0x10/0x10 >> [ 537.884604] ? lock_acquire+0x138/0x2f0 >> [ 537.884952] __writeback_single_inode+0xd4/0xa60 >> [ 537.885369] writeback_sb_inodes+0x4cf/0xe40 >> [ 537.885760] ? __pfx_writeback_sb_inodes+0x10/0x10 >> [ 537.886208] ? __pfx_move_expired_inodes+0x10/0x10 >> [ 537.886640] __writeback_inodes_wb+0xb4/0x200 >> [ 537.887037] wb_writeback+0x55b/0x7c0 >> [ 537.887372] ? __pfx_wb_writeback+0x10/0x10 >> [ 537.887750] ? lock_acquire+0x138/0x2f0 >> [ 537.888094] ? __pfx_register_lock_class+0x10/0x10 >> [ 537.888521] wb_workfn+0x648/0xbb0 >> [ 537.888835] ? __pfx_wb_workfn+0x10/0x10 >> [ 537.889192] ? lock_acquire+0x128/0x2f0 >> [ 537.889539] ? lock_acquire+0x138/0x2f0 >> [ 537.889890] process_one_work+0x7ff/0x1710 >> [ 537.890272] ? __pfx_process_one_work+0x10/0x10 >> [ 537.890685] ? assign_work+0x16c/0x240 >> [ 537.891026] worker_thread+0x6e8/0x12b0 >> [ 537.891381] ? __pfx_worker_thread+0x10/0x10 >> [ 537.891768] kthread+0x2ad/0x380 >> [ 537.892064] ? __pfx_kthread+0x10/0x10 >> [ 537.892403] ret_from_fork+0x2d/0x70 >> [ 537.892728] ? __pfx_kthread+0x10/0x10 >> [ 537.893068] ret_from_fork_asm+0x1a/0x30 >> [ 537.893434] </TASK> > > > -- > Best Regards, > Yan, Zi