On Fri, 5 Mar 2021 09:30:16 +0800 Aili Yao <yaoaili@xxxxxxxxxxxx> wrote: > On Thu, 4 Mar 2021 15:57:20 -0800 > "Luck, Tony" <tony.luck@xxxxxxxxx> wrote: > > > On Thu, Mar 04, 2021 at 02:45:24PM +0800, Aili Yao wrote: > > > > > if your methods works, should it be like this? > > > > > > > > > > 1582 pteval = swp_entry_to_pte(make_hwpoison_entry(subpage)); > > > > > 1583 if (PageHuge(page)) { > > > > > 1584 hugetlb_count_sub(compound_nr(page), mm); > > > > > 1585 set_huge_swap_pte_at(mm, address, > > > > > 1586 pvmw.pte, pteval, > > > > > 1587 vma_mmu_pagesize(vma)); > > > > > 1588 } else { > > > > > 1589 dec_mm_counter(mm, mm_counter(page)); > > > > > 1590 set_pte_at(mm, address, pvmw.pte, pteval); > > > > > 1591 } > > > > > > > > > > the page fault check if it's a poison page using is_hwpoison_entry(), > > > > > > > > > > > > > And if it works, does we need some locking mechanism before we call walk_page_range(); > > > > if we lock, does we need to process the blocking interrupted error as other places will do? > > > > > > > > > > And another thing: > > > Do we need a call to flush_tlb_page(vma, address) to make the pte changes into effect? > > > > Thanks for all the pointers. I added them to the patch (see below). > > [The pmd/pud cases may need some tweaking ... but just trying to get > > the 4K page case working first] > > > > I tried testing by skipping the call to memory_failure() and just > > using this new code to search the page tables for current page and > > marking it hwpoison (to simulate the case where 2nd process gets the > > early return from memory_failure(). Something is still missing because I get: > > > > [ 481.911298] mce: pte_entry: matched pfn - mark poison & zap pte > > [ 481.917935] MCE: Killing einj_mem_uc:5555 due to hardware memory corruption fault at 7fe64b33b400 > > [ 482.933775] BUG: Bad page cache in process einj_mem_uc pfn:408b6d6 > > [ 482.940777] page:0000000013ea6e96 refcount:3 mapcount:1 mapping:00000000e3a069d9 index:0x0 pfn:0x408b6d6 > > [ 482.951355] memcg:ffff94a809834000 > > [ 482.955153] aops:shmem_aops ino:3c04 > > [ 482.959142] flags: 0x97ffffc0880015(locked|uptodate|lru|swapbacked|hwpoison) > > [ 482.967018] raw: 0097ffffc0880015 ffff94c80e93ec00 ffff94c80e93ec00 ffff94c80a9b25a8 > > [ 482.975666] raw: 0000000000000000 0000000000000000 0000000300000000 ffff94a809834000 > > [ 482.984310] page dumped because: still mapped when deleted > > From the walk, it seems we have got the virtual address, can we just send a SIGBUS with it? Does this walk proper for other memory-failure error cases, can it be applied to if (p->mce_vaddr != (void __user *)-1l) branch in kill_me_maybe()? > > commit e5de44560b33e2d407704243566253a70f858a59 > > Author: Tony Luck <tony.luck@xxxxxxxxx> > > Date: Tue Mar 2 15:06:33 2021 -0800 > > > > x86/mce: Handle races between machine checks > > > > When multiple CPUs hit the same poison memory there is a race. The > > first CPU into memory_failure() atomically marks the page as poison > > and continues processing to hunt down all the tasks that map this page > > so that the virtual addresses can be marked not-present and SIGBUS > > sent to the task that did the access. > > > > Later CPUs get an early return from memory_failure() and may return > > to user mode and access the poison again. > > > > Add a new argument to memory_failure() so that it can indicate when > > the race has been lost. Fix kill_me_maybe() to scan page tables in > > this case to unmap pages. > > > > diff --git a/arch/x86/kernel/cpu/mce/core.c b/arch/x86/kernel/cpu/mce/core.c > > index 7962355436da..a52c6a772de2 100644 > > --- a/arch/x86/kernel/cpu/mce/core.c > > +++ b/arch/x86/kernel/cpu/mce/core.c > > @@ -28,8 +28,12 @@ > > #include <linux/sysfs.h> > > #include <linux/types.h> > > #include <linux/slab.h> > > +#include <linux/hugetlb.h> > > +#include <linux/swap.h> > > +#include <linux/swapops.h> > > #include <linux/init.h> > > #include <linux/kmod.h> > > +#include <linux/pagewalk.h> > > #include <linux/poll.h> > > #include <linux/nmi.h> > > #include <linux/cpu.h> > > @@ -637,6 +641,7 @@ static int uc_decode_notifier(struct notifier_block *nb, unsigned long val, > > { > > struct mce *mce = (struct mce *)data; > > unsigned long pfn; > > + int already = 0; > > > > if (!mce || !mce_usable_address(mce)) > > return NOTIFY_DONE; > > @@ -646,8 +651,9 @@ static int uc_decode_notifier(struct notifier_block *nb, unsigned long val, > > return NOTIFY_DONE; > > > > pfn = mce->addr >> PAGE_SHIFT; > > - if (!memory_failure(pfn, 0)) { > > - set_mce_nospec(pfn, whole_page(mce)); > > + if (!memory_failure(pfn, 0, &already)) { > > + if (!already) > > + set_mce_nospec(pfn, whole_page(mce)); > > mce->kflags |= MCE_HANDLED_UC; > > } > > > > @@ -1248,6 +1254,79 @@ static void __mc_scan_banks(struct mce *m, struct pt_regs *regs, struct mce *fin > > *m = *final; > > } > > > > +static int pte_entry(pte_t *pte, unsigned long addr, unsigned long next, struct mm_walk *walk) > > +{ > > + u64 pfn = (u64)walk->private; > > + struct page *page; > > + pte_t pteval; > > + > > + if (pte_pfn(*pte) == pfn) { > > +pr_info("pte_entry: matched pfn - mark poison & zap pte\n"); > > + page = pfn_to_page(pfn); > > + lock_page(page); > > +SetPageHWPoison(page); > > + pteval = swp_entry_to_pte(make_hwpoison_entry(page)); > > + dec_mm_counter(walk->mm, mm_counter(page)); > > + set_pte_at(current->mm, addr, pte, pteval); > > + unlock_page(page); > > + flush_tlb_page(walk->vma, addr); > > + } > > + > > + return 0; > > +} > > + > > +static int pmd_entry(pmd_t *pmd, unsigned long addr, unsigned long next, struct mm_walk *walk) > > +{ > > + int shift = PMD_SHIFT - PAGE_SHIFT; > > + u64 pfn = (u64)walk->private; > > + struct page *page; > > + pte_t pteval; > > + > > + if (!pmd_large(*pmd)) > > + return 0; > > + > > + if (pmd_pfn(*pmd) >> shift == pfn >> shift) { > > + page = pfn_to_page(pfn); > > + lock_page(page); > > + pteval = swp_entry_to_pte(make_hwpoison_entry(page)); > > + hugetlb_count_sub(compound_nr(page), walk->mm); > > + set_huge_swap_pte_at(walk->mm, addr, (pte_t *)pmd, pteval, vma_mmu_pagesize(walk->vma)); > > + unlock_page(page); > > + flush_tlb_page(walk->vma, addr); > > + } > > + > > + return 0; > > +} > > + > > +static int pud_entry(pud_t *pud, unsigned long addr, unsigned long next, struct mm_walk *walk) > > +{ > > + int shift = PUD_SHIFT - PAGE_SHIFT; > > + u64 pfn = (u64)walk->private; > > + struct page *page; > > + pte_t pteval; > > + > > + if (!pud_large(*pud)) > > + return 0; > > + > > + if (pud_pfn(*pud) >> shift == pfn >> shift) { > > + page = pfn_to_page(pfn); > > + lock_page(page); > > + pteval = swp_entry_to_pte(make_hwpoison_entry(page)); > > + hugetlb_count_sub(compound_nr(page), walk->mm); > > + set_huge_swap_pte_at(walk->mm, addr, (pte_t *)pud, pteval, vma_mmu_pagesize(walk->vma)); > > + unlock_page(page); > > + flush_tlb_page(walk->vma, addr); > > + } > > + > > + return 0; > > +} > > + > > +static struct mm_walk_ops walk = { > > + .pte_entry = pte_entry, > > + .pmd_entry = pmd_entry, > > + .pud_entry = pud_entry > > +}; > > + > > static void kill_me_now(struct callback_head *ch) > > { > > force_sig(SIGBUS); > > @@ -1257,15 +1336,22 @@ static void kill_me_maybe(struct callback_head *cb) > > { > > struct task_struct *p = container_of(cb, struct task_struct, mce_kill_me); > > int flags = MF_ACTION_REQUIRED; > > + int already = 0; > > > > pr_err("Uncorrected hardware memory error in user-access at %llx", p->mce_addr); > > > > if (!p->mce_ripv) > > flags |= MF_MUST_KILL; > > > > - if (!memory_failure(p->mce_addr >> PAGE_SHIFT, flags) && > > + if (!memory_failure(p->mce_addr >> PAGE_SHIFT, flags, &already) && > > !(p->mce_kflags & MCE_IN_KERNEL_COPYIN)) { > > - set_mce_nospec(p->mce_addr >> PAGE_SHIFT, p->mce_whole_page); > > + if (already) { > > + mmap_read_lock(current->mm); > > + walk_page_range(current->mm, 0, TASK_SIZE_MAX, &walk, (void *)(p->mce_addr >> PAGE_SHIFT)); > > + mmap_read_unlock(current->mm); > > + } else { > > + set_mce_nospec(p->mce_addr >> PAGE_SHIFT, p->mce_whole_page); > > + } > > sync_core(); > > return; > > } > > @@ -1452,7 +1538,7 @@ noinstr void do_machine_check(struct pt_regs *regs) > > EXPORT_SYMBOL_GPL(do_machine_check); > > > > #ifndef CONFIG_MEMORY_FAILURE > > -int memory_failure(unsigned long pfn, int flags) > > +int memory_failure(unsigned long pfn, int flags, int *already) > > { > > /* mce_severity() should not hand us an ACTION_REQUIRED error */ > > BUG_ON(flags & MF_ACTION_REQUIRED); > > diff --git a/drivers/base/memory.c b/drivers/base/memory.c > > index f35298425575..144500983656 100644 > > --- a/drivers/base/memory.c > > +++ b/drivers/base/memory.c > > @@ -480,7 +480,7 @@ static ssize_t hard_offline_page_store(struct device *dev, > > if (kstrtoull(buf, 0, &pfn) < 0) > > return -EINVAL; > > pfn >>= PAGE_SHIFT; > > - ret = memory_failure(pfn, 0); > > + ret = memory_failure(pfn, 0, NULL); > > return ret ? ret : count; > > } > > > > diff --git a/include/linux/mm.h b/include/linux/mm.h > > index 77e64e3eac80..beaa6e871cbe 100644 > > --- a/include/linux/mm.h > > +++ b/include/linux/mm.h > > @@ -3003,7 +3003,7 @@ enum mf_flags { > > MF_MUST_KILL = 1 << 2, > > MF_SOFT_OFFLINE = 1 << 3, > > }; > > -extern int memory_failure(unsigned long pfn, int flags); > > +extern int memory_failure(unsigned long pfn, int flags, int *already); > > extern void memory_failure_queue(unsigned long pfn, int flags); > > extern void memory_failure_queue_kick(int cpu); > > extern int unpoison_memory(unsigned long pfn); > > diff --git a/mm/hwpoison-inject.c b/mm/hwpoison-inject.c > > index 1ae1ebc2b9b1..bfd5151dcd3f 100644 > > --- a/mm/hwpoison-inject.c > > +++ b/mm/hwpoison-inject.c > > @@ -48,7 +48,7 @@ static int hwpoison_inject(void *data, u64 val) > > > > inject: > > pr_info("Injecting memory failure at pfn %#lx\n", pfn); > > - return memory_failure(pfn, 0); > > + return memory_failure(pfn, 0, NULL); > > } > > > > static int hwpoison_unpoison(void *data, u64 val) > > diff --git a/mm/madvise.c b/mm/madvise.c > > index df692d2e35d4..09f569fed68d 100644 > > --- a/mm/madvise.c > > +++ b/mm/madvise.c > > @@ -908,7 +908,7 @@ static int madvise_inject_error(int behavior, > > } else { > > pr_info("Injecting memory failure for pfn %#lx at process virtual address %#lx\n", > > pfn, start); > > - ret = memory_failure(pfn, MF_COUNT_INCREASED); > > + ret = memory_failure(pfn, MF_COUNT_INCREASED, NULL); > > } > > > > if (ret) > > diff --git a/mm/memory-failure.c b/mm/memory-failure.c > > index 24210c9bd843..9a8911aa5fc9 100644 > > --- a/mm/memory-failure.c > > +++ b/mm/memory-failure.c > > @@ -1398,7 +1398,7 @@ static int memory_failure_dev_pagemap(unsigned long pfn, int flags, > > * Must run in process context (e.g. a work queue) with interrupts > > * enabled and no spinlocks hold. > > */ > > -int memory_failure(unsigned long pfn, int flags) > > +int memory_failure(unsigned long pfn, int flags, int *already) > > { > > struct page *p; > > struct page *hpage; > > @@ -1428,6 +1428,8 @@ int memory_failure(unsigned long pfn, int flags) > > if (PageHuge(p)) > > return memory_failure_hugetlb(pfn, flags); > > if (TestSetPageHWPoison(p)) { > > + if (already) > > + *already = 1; > > pr_err("Memory failure: %#lx: already hardware poisoned\n", > > pfn); > > return 0; > > @@ -1634,7 +1636,7 @@ static void memory_failure_work_func(struct work_struct *work) > > if (entry.flags & MF_SOFT_OFFLINE) > > soft_offline_page(entry.pfn, entry.flags); > > else > > - memory_failure(entry.pfn, entry.flags); > > + memory_failure(entry.pfn, entry.flags, NULL); > > } > > } > > > -- Thanks! Aili Yao