On Thu, 4 Mar 2021 15:57:20 -0800 "Luck, Tony" <tony.luck@xxxxxxxxx> wrote: > On Thu, Mar 04, 2021 at 02:45:24PM +0800, Aili Yao wrote: > > > > if your methods works, should it be like this? > > > > > > > > 1582 pteval = swp_entry_to_pte(make_hwpoison_entry(subpage)); > > > > 1583 if (PageHuge(page)) { > > > > 1584 hugetlb_count_sub(compound_nr(page), mm); > > > > 1585 set_huge_swap_pte_at(mm, address, > > > > 1586 pvmw.pte, pteval, > > > > 1587 vma_mmu_pagesize(vma)); > > > > 1588 } else { > > > > 1589 dec_mm_counter(mm, mm_counter(page)); > > > > 1590 set_pte_at(mm, address, pvmw.pte, pteval); > > > > 1591 } > > > > > > > > the page fault check if it's a poison page using is_hwpoison_entry(), > > > > > > > > > > And if it works, does we need some locking mechanism before we call walk_page_range(); > > > if we lock, does we need to process the blocking interrupted error as other places will do? > > > > > > > And another thing: > > Do we need a call to flush_tlb_page(vma, address) to make the pte changes into effect? > > Thanks for all the pointers. I added them to the patch (see below). > [The pmd/pud cases may need some tweaking ... but just trying to get > the 4K page case working first] > > I tried testing by skipping the call to memory_failure() and just > using this new code to search the page tables for current page and > marking it hwpoison (to simulate the case where 2nd process gets the > early return from memory_failure(). Something is still missing because I get: > > [ 481.911298] mce: pte_entry: matched pfn - mark poison & zap pte > [ 481.917935] MCE: Killing einj_mem_uc:5555 due to hardware memory corruption fault at 7fe64b33b400 > [ 482.933775] BUG: Bad page cache in process einj_mem_uc pfn:408b6d6 > [ 482.940777] page:0000000013ea6e96 refcount:3 mapcount:1 mapping:00000000e3a069d9 index:0x0 pfn:0x408b6d6 > [ 482.951355] memcg:ffff94a809834000 > [ 482.955153] aops:shmem_aops ino:3c04 > [ 482.959142] flags: 0x97ffffc0880015(locked|uptodate|lru|swapbacked|hwpoison) > [ 482.967018] raw: 0097ffffc0880015 ffff94c80e93ec00 ffff94c80e93ec00 ffff94c80a9b25a8 > [ 482.975666] raw: 0000000000000000 0000000000000000 0000000300000000 ffff94a809834000 > [ 482.984310] page dumped because: still mapped when deleted >From the walk, it seems we have got the virtual address, can we just send a SIGBUS with it? > commit e5de44560b33e2d407704243566253a70f858a59 > Author: Tony Luck <tony.luck@xxxxxxxxx> > Date: Tue Mar 2 15:06:33 2021 -0800 > > x86/mce: Handle races between machine checks > > When multiple CPUs hit the same poison memory there is a race. The > first CPU into memory_failure() atomically marks the page as poison > and continues processing to hunt down all the tasks that map this page > so that the virtual addresses can be marked not-present and SIGBUS > sent to the task that did the access. > > Later CPUs get an early return from memory_failure() and may return > to user mode and access the poison again. > > Add a new argument to memory_failure() so that it can indicate when > the race has been lost. Fix kill_me_maybe() to scan page tables in > this case to unmap pages. > > diff --git a/arch/x86/kernel/cpu/mce/core.c b/arch/x86/kernel/cpu/mce/core.c > index 7962355436da..a52c6a772de2 100644 > --- a/arch/x86/kernel/cpu/mce/core.c > +++ b/arch/x86/kernel/cpu/mce/core.c > @@ -28,8 +28,12 @@ > #include <linux/sysfs.h> > #include <linux/types.h> > #include <linux/slab.h> > +#include <linux/hugetlb.h> > +#include <linux/swap.h> > +#include <linux/swapops.h> > #include <linux/init.h> > #include <linux/kmod.h> > +#include <linux/pagewalk.h> > #include <linux/poll.h> > #include <linux/nmi.h> > #include <linux/cpu.h> > @@ -637,6 +641,7 @@ static int uc_decode_notifier(struct notifier_block *nb, unsigned long val, > { > struct mce *mce = (struct mce *)data; > unsigned long pfn; > + int already = 0; > > if (!mce || !mce_usable_address(mce)) > return NOTIFY_DONE; > @@ -646,8 +651,9 @@ static int uc_decode_notifier(struct notifier_block *nb, unsigned long val, > return NOTIFY_DONE; > > pfn = mce->addr >> PAGE_SHIFT; > - if (!memory_failure(pfn, 0)) { > - set_mce_nospec(pfn, whole_page(mce)); > + if (!memory_failure(pfn, 0, &already)) { > + if (!already) > + set_mce_nospec(pfn, whole_page(mce)); > mce->kflags |= MCE_HANDLED_UC; > } > > @@ -1248,6 +1254,79 @@ static void __mc_scan_banks(struct mce *m, struct pt_regs *regs, struct mce *fin > *m = *final; > } > > +static int pte_entry(pte_t *pte, unsigned long addr, unsigned long next, struct mm_walk *walk) > +{ > + u64 pfn = (u64)walk->private; > + struct page *page; > + pte_t pteval; > + > + if (pte_pfn(*pte) == pfn) { > +pr_info("pte_entry: matched pfn - mark poison & zap pte\n"); > + page = pfn_to_page(pfn); > + lock_page(page); > +SetPageHWPoison(page); > + pteval = swp_entry_to_pte(make_hwpoison_entry(page)); > + dec_mm_counter(walk->mm, mm_counter(page)); > + set_pte_at(current->mm, addr, pte, pteval); > + unlock_page(page); > + flush_tlb_page(walk->vma, addr); > + } > + > + return 0; > +} > + > +static int pmd_entry(pmd_t *pmd, unsigned long addr, unsigned long next, struct mm_walk *walk) > +{ > + int shift = PMD_SHIFT - PAGE_SHIFT; > + u64 pfn = (u64)walk->private; > + struct page *page; > + pte_t pteval; > + > + if (!pmd_large(*pmd)) > + return 0; > + > + if (pmd_pfn(*pmd) >> shift == pfn >> shift) { > + page = pfn_to_page(pfn); > + lock_page(page); > + pteval = swp_entry_to_pte(make_hwpoison_entry(page)); > + hugetlb_count_sub(compound_nr(page), walk->mm); > + set_huge_swap_pte_at(walk->mm, addr, (pte_t *)pmd, pteval, vma_mmu_pagesize(walk->vma)); > + unlock_page(page); > + flush_tlb_page(walk->vma, addr); > + } > + > + return 0; > +} > + > +static int pud_entry(pud_t *pud, unsigned long addr, unsigned long next, struct mm_walk *walk) > +{ > + int shift = PUD_SHIFT - PAGE_SHIFT; > + u64 pfn = (u64)walk->private; > + struct page *page; > + pte_t pteval; > + > + if (!pud_large(*pud)) > + return 0; > + > + if (pud_pfn(*pud) >> shift == pfn >> shift) { > + page = pfn_to_page(pfn); > + lock_page(page); > + pteval = swp_entry_to_pte(make_hwpoison_entry(page)); > + hugetlb_count_sub(compound_nr(page), walk->mm); > + set_huge_swap_pte_at(walk->mm, addr, (pte_t *)pud, pteval, vma_mmu_pagesize(walk->vma)); > + unlock_page(page); > + flush_tlb_page(walk->vma, addr); > + } > + > + return 0; > +} > + > +static struct mm_walk_ops walk = { > + .pte_entry = pte_entry, > + .pmd_entry = pmd_entry, > + .pud_entry = pud_entry > +}; > + > static void kill_me_now(struct callback_head *ch) > { > force_sig(SIGBUS); > @@ -1257,15 +1336,22 @@ static void kill_me_maybe(struct callback_head *cb) > { > struct task_struct *p = container_of(cb, struct task_struct, mce_kill_me); > int flags = MF_ACTION_REQUIRED; > + int already = 0; > > pr_err("Uncorrected hardware memory error in user-access at %llx", p->mce_addr); > > if (!p->mce_ripv) > flags |= MF_MUST_KILL; > > - if (!memory_failure(p->mce_addr >> PAGE_SHIFT, flags) && > + if (!memory_failure(p->mce_addr >> PAGE_SHIFT, flags, &already) && > !(p->mce_kflags & MCE_IN_KERNEL_COPYIN)) { > - set_mce_nospec(p->mce_addr >> PAGE_SHIFT, p->mce_whole_page); > + if (already) { > + mmap_read_lock(current->mm); > + walk_page_range(current->mm, 0, TASK_SIZE_MAX, &walk, (void *)(p->mce_addr >> PAGE_SHIFT)); > + mmap_read_unlock(current->mm); > + } else { > + set_mce_nospec(p->mce_addr >> PAGE_SHIFT, p->mce_whole_page); > + } > sync_core(); > return; > } > @@ -1452,7 +1538,7 @@ noinstr void do_machine_check(struct pt_regs *regs) > EXPORT_SYMBOL_GPL(do_machine_check); > > #ifndef CONFIG_MEMORY_FAILURE > -int memory_failure(unsigned long pfn, int flags) > +int memory_failure(unsigned long pfn, int flags, int *already) > { > /* mce_severity() should not hand us an ACTION_REQUIRED error */ > BUG_ON(flags & MF_ACTION_REQUIRED); > diff --git a/drivers/base/memory.c b/drivers/base/memory.c > index f35298425575..144500983656 100644 > --- a/drivers/base/memory.c > +++ b/drivers/base/memory.c > @@ -480,7 +480,7 @@ static ssize_t hard_offline_page_store(struct device *dev, > if (kstrtoull(buf, 0, &pfn) < 0) > return -EINVAL; > pfn >>= PAGE_SHIFT; > - ret = memory_failure(pfn, 0); > + ret = memory_failure(pfn, 0, NULL); > return ret ? ret : count; > } > > diff --git a/include/linux/mm.h b/include/linux/mm.h > index 77e64e3eac80..beaa6e871cbe 100644 > --- a/include/linux/mm.h > +++ b/include/linux/mm.h > @@ -3003,7 +3003,7 @@ enum mf_flags { > MF_MUST_KILL = 1 << 2, > MF_SOFT_OFFLINE = 1 << 3, > }; > -extern int memory_failure(unsigned long pfn, int flags); > +extern int memory_failure(unsigned long pfn, int flags, int *already); > extern void memory_failure_queue(unsigned long pfn, int flags); > extern void memory_failure_queue_kick(int cpu); > extern int unpoison_memory(unsigned long pfn); > diff --git a/mm/hwpoison-inject.c b/mm/hwpoison-inject.c > index 1ae1ebc2b9b1..bfd5151dcd3f 100644 > --- a/mm/hwpoison-inject.c > +++ b/mm/hwpoison-inject.c > @@ -48,7 +48,7 @@ static int hwpoison_inject(void *data, u64 val) > > inject: > pr_info("Injecting memory failure at pfn %#lx\n", pfn); > - return memory_failure(pfn, 0); > + return memory_failure(pfn, 0, NULL); > } > > static int hwpoison_unpoison(void *data, u64 val) > diff --git a/mm/madvise.c b/mm/madvise.c > index df692d2e35d4..09f569fed68d 100644 > --- a/mm/madvise.c > +++ b/mm/madvise.c > @@ -908,7 +908,7 @@ static int madvise_inject_error(int behavior, > } else { > pr_info("Injecting memory failure for pfn %#lx at process virtual address %#lx\n", > pfn, start); > - ret = memory_failure(pfn, MF_COUNT_INCREASED); > + ret = memory_failure(pfn, MF_COUNT_INCREASED, NULL); > } > > if (ret) > diff --git a/mm/memory-failure.c b/mm/memory-failure.c > index 24210c9bd843..9a8911aa5fc9 100644 > --- a/mm/memory-failure.c > +++ b/mm/memory-failure.c > @@ -1398,7 +1398,7 @@ static int memory_failure_dev_pagemap(unsigned long pfn, int flags, > * Must run in process context (e.g. a work queue) with interrupts > * enabled and no spinlocks hold. > */ > -int memory_failure(unsigned long pfn, int flags) > +int memory_failure(unsigned long pfn, int flags, int *already) > { > struct page *p; > struct page *hpage; > @@ -1428,6 +1428,8 @@ int memory_failure(unsigned long pfn, int flags) > if (PageHuge(p)) > return memory_failure_hugetlb(pfn, flags); > if (TestSetPageHWPoison(p)) { > + if (already) > + *already = 1; > pr_err("Memory failure: %#lx: already hardware poisoned\n", > pfn); > return 0; > @@ -1634,7 +1636,7 @@ static void memory_failure_work_func(struct work_struct *work) > if (entry.flags & MF_SOFT_OFFLINE) > soft_offline_page(entry.pfn, entry.flags); > else > - memory_failure(entry.pfn, entry.flags); > + memory_failure(entry.pfn, entry.flags, NULL); > } > } > -- Thanks! Aili Yao