On 9/6/22 17:11, Vlastimil Babka wrote: > On 9/6/22 16:56, Hyeonggon Yoo wrote: >> On Tue, Sep 06, 2022 at 03:51:01PM +0800, kernel test robot wrote: >>> Greeting, >>> >>> FYI, we noticed the following commit (built with gcc-11): >>> >>> commit: 3c4cafa313d978b31a1d5dc17c323074b19a1d63 ("mm/sl[au]b: rearrange >>> struct slab fields to allow larger rcu_head") >>> git://git.kernel.org/cgit/linux/kernel/git/vbabka/slab.git >>> for-6.1/fit_rcu_head >>> >>> in testcase: fio-basic >>> version: fio-x86_64-3.15-1_20220903 >>> with following parameters: >>> >>> disk: 2pmem >>> fs: xfs >>> runtime: 200s >>> nr_task: 50% >>> time_based: tb >>> rw: randrw >>> bs: 2M >>> ioengine: mmap >>> test_size: 200G >>> cpufreq_governor: performance >>> >>> test-description: Fio is a tool that will spawn a number of threads or >>> processes doing a particular type of I/O action as specified by the user. >>> test-url:https://github.com/axboe/fio >>> >>> >>> on test machine: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ >>> 2.10GHz (Cascade Lake) with 512G memory >>> >>> caused below changes (please refer to attached dmesg/kmsg for entire >>> log/backtrace): >>> >>> >>> [ 304.700893][ C40] perf: interrupt took too long (12747 > 12477), >>> lowering kernel.perf_event_max_sample_rate to 15000 >>> [ 305.015834][ C40] perf: interrupt took too long (15947 > 15933), >>> lowering kernel.perf_event_max_sample_rate to 12000 >>> [ 305.954702][ C40] perf: interrupt took too long (19968 > 19933), >>> lowering kernel.perf_event_max_sample_rate to 10000 >>> [ 309.554949][ C31] perf: interrupt took too long (25118 > 24960), >>> lowering kernel.perf_event_max_sample_rate to 7000 >>> [ 315.068744][ C95] sched: RT throttling activated >>> [ 317.121806][ T590] general protection fault, probably for >>> non-canonical address 0xdead000000000120: 0000 [#1] SMP NOPTI >>> [ 317.133291][ T590] CPU: 61 PID: 590 Comm: kcompactd0 Tainted: G >>> S 6.0.0-rc2-00002-g3c4cafa313d9 #1 >>> [ 317.144084][ T590] Hardware name: Intel Corporation >>> S2600WFT/S2600WFT, BIOS SE5C620.86B.02.01.0008.031920191559 03/19/2019 >>> [ 317.155668][ T590] RIP: 0010:isolate_movable_page (mm/migrate.c:103) >>> [ 317.162016][ T590] Code: ba 28 00 0f 82 88 00 00 00 48 89 ef e8 e2 3a >>> f8 ff 84 c0 74 74 48 8b 45 00 a9 00 00 04 00 75 69 48 8b 45 18 44 89 e6 >>> 48 89 ef <48> 8b 40 fe ff d0 0f 1f 00 84 c0 74 52 48 8b 45 00 a9 00 00 04 00 >>> All code >>> ======== >>> 0: ba 28 00 0f 82 mov $0x820f0028,%edx >>> 5: 88 00 mov %al,(%rax) >>> 7: 00 00 add %al,(%rax) >>> 9: 48 89 ef mov %rbp,%rdi >>> c: e8 e2 3a f8 ff callq 0xfffffffffff83af3 >>> 11: 84 c0 test %al,%al >>> 13: 74 74 je 0x89 >>> 15: 48 8b 45 00 mov 0x0(%rbp),%rax >>> 19: a9 00 00 04 00 test $0x40000,%eax >>> 1e: 75 69 jne 0x89 >>> 20: 48 8b 45 18 mov 0x18(%rbp),%rax >>> 24: 44 89 e6 mov %r12d,%esi >>> 27: 48 89 ef mov %rbp,%rdi >>> 2a:* 48 8b 40 fe mov -0x2(%rax),%rax <-- >>> trapping instruction >>> 2e: ff d0 callq *%rax >>> 30: 0f 1f 00 nopl (%rax) >>> 33: 84 c0 test %al,%al >>> 35: 74 52 je 0x89 >>> 37: 48 8b 45 00 mov 0x0(%rbp),%rax >>> 3b: a9 00 00 04 00 test $0x40000,%eax >>> >>> Code starting with the faulting instruction >>> =========================================== >>> 0: 48 8b 40 fe mov -0x2(%rax),%rax >>> 4: ff d0 callq *%rax >>> 6: 0f 1f 00 nopl (%rax) >>> 9: 84 c0 test %al,%al >>> b: 74 52 je 0x5f >>> d: 48 8b 45 00 mov 0x0(%rbp),%rax >>> 11: a9 00 00 04 00 test $0x40000,%eax >>> [ 317.182354][ T590] RSP: 0018:ffffc9000e1d3c78 EFLAGS: 00010246 >>> [ 317.188668][ T590] RAX: dead000000000122 RBX: ffffea0004031034 RCX: >>> 000000000000000c >>> [ 317.196890][ T590] RDX: dead000000000101 RSI: 000000000000000c RDI: >>> ffffea0004031000 >>> [ 317.205273][ T590] RBP: ffffea0004031000 R08: 0000000004031000 R09: >>> 0000000000000004 >>> [ 317.213752][ T590] R10: 00000000000066b6 R11: 0000000000000004 R12: >>> 000000000000000c >>> [ 317.222384][ T590] R13: ffffea0004031000 R14: 0000000000100c40 R15: >>> ffffc9000e1d3df0 >>> [ 317.230679][ T590] FS: 0000000000000000(0000) >>> GS:ffff88c04ff40000(0000) knlGS:0000000000000000 >>> [ 317.239896][ T590] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 >>> [ 317.247098][ T590] CR2: 0000000000451c00 CR3: 0000008064ca4002 CR4: >>> 00000000007706e0 >>> [ 317.255788][ T590] DR0: 0000000000000000 DR1: 0000000000000000 DR2: >>> 0000000000000000 >>> [ 317.264256][ T590] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: >>> 0000000000000400 >>> [ 317.272772][ T590] PKRU: 55555554 >>> [ 317.276783][ T590] Call Trace: >>> [ 317.280932][ T590] <TASK> >>> [ 317.284315][ T590] isolate_migratepages_block (mm/compaction.c:982) >>> [ 317.290702][ T590] isolate_migratepages (mm/compaction.c:1960) >>> [ 317.296278][ T590] compact_zone (mm/compaction.c:2393) >>> [ 317.301202][ T590] proactive_compact_node (mm/compaction.c:2661 >>> (discriminator 2)) >> Hmm... Let's debug. >> >> FYI, simply echo 1 > /proc/sys/vm/compact_memory invokes same bug on my test >> environment. >> >> the 'mops' is invalid address in mm/migrate.c:103. >> >> Hmm, why is this slab page confused as movable page? >> -> Because page->'mapping' and slab->slabs field has same offset. >> >> I think this is invoked because lowest two bits of slab->slabs is not 0. >> >> Vlastimil, any thoughts? > > Yeah, slabs->slabs could do that, and the remedy would be to exchange it > with the slab->next field. > However the report points to the value dead000000000122 which is > LIST_POISON2, which unfortunately contains the lower bit after 4c6080cd6f8b > ("lib/list: tweak LIST_POISON2 for better code generation on x86_64") > > Probably the simplest fix would be to check for PageSlab() before > __PageMovable(). So I've done with the patch below, that I added to the for-6.1/fit_rcu_head branch in slab.git. It's not very nice though with all the new membarriers. I hope it's at least correct... > But heads up for Joel - if your rcu_head debugging info series (didn't > check) has something like a counter in the 3rd 64bit word, where bit 1 can > thus be set, it can cause the same issue fooling the __PageMovable() check. ----8<---- >From d6f9fbb33b908eb8162cc1f6ce7f7c970d0f285f Mon Sep 17 00:00:00 2001 From: Vlastimil Babka <vbabka@xxxxxxx> Date: Fri, 9 Sep 2022 12:03:10 +0200 Subject: [PATCH 2/3] mm/migrate: make isolate_movable_page() skip slab pages In the next commit we want to rearrange struct slab fields to allow a larger rcu_head. Afterwards, the page->mapping field will overlap with SLUB's "struct list_head slab_list", where the value of prev pointer can become LIST_POISON2, which is 0x122 + POISON_POINTER_DELTA. Unfortunately the bit 1 being set can confuse PageMovable() to be a false positive and cause a GPF as reported by lkp [1]. To fix this, make isolate_movable_page() skip pages with the PageSlab flag set. This is a bit tricky as we need to add memory barriers to SLAB and SLUB's page allocation and freeing, and their counterparts to isolate_movable_page(). [1] https://lore.kernel.org/all/208c1757-5edd-fd42-67d4-1940cc43b50f@xxxxxxxxx/ Reported-by: kernel test robot <yujie.liu@xxxxxxxxx> Signed-off-by: Vlastimil Babka <vbabka@xxxxxxx> --- mm/compaction.c | 2 +- mm/migrate.c | 12 +++++++++++- mm/slab.c | 6 +++++- mm/slub.c | 6 +++++- 4 files changed, 22 insertions(+), 4 deletions(-) diff --git a/mm/compaction.c b/mm/compaction.c index 640fa76228dd..b697c207beec 100644 --- a/mm/compaction.c +++ b/mm/compaction.c @@ -972,7 +972,7 @@ isolate_migratepages_block(struct compact_control *cc, unsigned long low_pfn, * __PageMovable can return false positive so we need * to verify it under page_lock. */ - if (unlikely(__PageMovable(page)) && + if (unlikely(!PageSlab(page) && __PageMovable(page)) && !PageIsolated(page)) { if (locked) { unlock_page_lruvec_irqrestore(locked, flags); diff --git a/mm/migrate.c b/mm/migrate.c index 6a1597c92261..7f661b45d431 100644 --- a/mm/migrate.c +++ b/mm/migrate.c @@ -78,7 +78,7 @@ int isolate_movable_page(struct page *page, isolate_mode_t mode) * assumes anybody doesn't touch PG_lock of newly allocated page * so unconditionally grabbing the lock ruins page's owner side. */ - if (unlikely(!__PageMovable(page))) + if (unlikely(!__PageMovable(page) || PageSlab(page))) goto out_putpage; /* * As movable pages are not isolated from LRU lists, concurrent @@ -94,9 +94,19 @@ int isolate_movable_page(struct page *page, isolate_mode_t mode) if (unlikely(!trylock_page(page))) goto out_putpage; + if (unlikely(PageSlab(page))) + goto out_no_isolated; + /* Pairs with smp_wmb() in slab freeing, e.g. SLUB's __free_slab() */ + smp_rmb(); + if (!PageMovable(page) || PageIsolated(page)) goto out_no_isolated; + /* Pairs with smp_wmb() in slab allocation, e.g. SLUB's alloc_slab_page() */ + smp_rmb(); + if (unlikely(PageSlab(page))) + goto out_no_isolated; + mops = page_movable_ops(page); VM_BUG_ON_PAGE(!mops, page); diff --git a/mm/slab.c b/mm/slab.c index 10e96137b44f..25e9a6ef4f74 100644 --- a/mm/slab.c +++ b/mm/slab.c @@ -1370,6 +1370,8 @@ static struct slab *kmem_getpages(struct kmem_cache *cachep, gfp_t flags, account_slab(slab, cachep->gfporder, cachep, flags); __folio_set_slab(folio); + /* Make the flag visible before any changes to folio->mapping */ + smp_wmb(); /* Record if ALLOC_NO_WATERMARKS was set when allocating the slab */ if (sk_memalloc_socks() && page_is_pfmemalloc(folio_page(folio, 0))) slab_set_pfmemalloc(slab); @@ -1387,9 +1389,11 @@ static void kmem_freepages(struct kmem_cache *cachep, struct slab *slab) BUG_ON(!folio_test_slab(folio)); __slab_clear_pfmemalloc(slab); - __folio_clear_slab(folio); page_mapcount_reset(folio_page(folio, 0)); folio->mapping = NULL; + /* Make the mapping reset visible before clearing the flag */ + smp_wmb(); + __folio_clear_slab(folio); if (current->reclaim_state) current->reclaim_state->reclaimed_slab += 1 << order; diff --git a/mm/slub.c b/mm/slub.c index d86be1b0d09f..2f9cb6e67de3 100644 --- a/mm/slub.c +++ b/mm/slub.c @@ -1830,6 +1830,8 @@ static inline struct slab *alloc_slab_page(gfp_t flags, int node, slab = folio_slab(folio); __folio_set_slab(folio); + /* Make the flag visible before any changes to folio->mapping */ + smp_wmb(); if (page_is_pfmemalloc(folio_page(folio, 0))) slab_set_pfmemalloc(slab); @@ -2037,8 +2039,10 @@ static void __free_slab(struct kmem_cache *s, struct slab *slab) int pages = 1 << order; __slab_clear_pfmemalloc(slab); - __folio_clear_slab(folio); folio->mapping = NULL; + /* Make the mapping reset visible before clearing the flag */ + smp_wmb(); + __folio_clear_slab(folio); if (current->reclaim_state) current->reclaim_state->reclaimed_slab += pages; unaccount_slab(slab, order, s); -- 2.37.3