Hi Peter, On Fri, Aug 11, 2017 at 03:30:20PM +0200, Peter Zijlstra wrote: > On Tue, Aug 01, 2017 at 05:08:17PM -0700, Nadav Amit wrote: > > void tlb_finish_mmu(struct mmu_gather *tlb, > > unsigned long start, unsigned long end) > > { > > - arch_tlb_finish_mmu(tlb, start, end); > > + /* > > + * If there are parallel threads are doing PTE changes on same range > > + * under non-exclusive lock(e.g., mmap_sem read-side) but defer TLB > > + * flush by batching, a thread has stable TLB entry can fail to flush > > + * the TLB by observing pte_none|!pte_dirty, for example so flush TLB > > + * forcefully if we detect parallel PTE batching threads. > > + */ > > + bool force = mm_tlb_flush_nested(tlb->mm); > > + > > + arch_tlb_finish_mmu(tlb, start, end, force); > > } > > I don't understand the comment nor the ordering. What guarantees we see > the increment if we need to? How about this about commenting part? >From 05f06fd6aba14447a9ca2df8b810fbcf9a58e14b Mon Sep 17 00:00:00 2001 From: Minchan Kim <minchan@xxxxxxxxxx> Date: Mon, 14 Aug 2017 10:16:56 +0900 Subject: [PATCH] mm: add describable comment for TLB batch race [1] is a rather subtle/complicated bug so that it's hard to understand it with limited code comment. This patch adds a sequence diagaram to explain the problem more easily, I hope. [1] 99baac21e458, mm: fix MADV_[FREE|DONTNEED] TLB flush miss problem Cc: Peter Zijlstra <peterz@xxxxxxxxxxxxx> Cc: Nadav Amit <namit@xxxxxxxxxx> Cc: Mel Gorman <mgorman@xxxxxxxxxxxxxxxxxxx> Signed-off-by: Minchan Kim <minchan@xxxxxxxxxx> --- mm/memory.c | 25 +++++++++++++++++++++++++ 1 file changed, 25 insertions(+) diff --git a/mm/memory.c b/mm/memory.c index bcbe56f52163..f571b0eb9816 100644 --- a/mm/memory.c +++ b/mm/memory.c @@ -413,12 +413,37 @@ void tlb_gather_mmu(struct mmu_gather *tlb, struct mm_struct *mm, void tlb_finish_mmu(struct mmu_gather *tlb, unsigned long start, unsigned long end) { + + /* * If there are parallel threads are doing PTE changes on same range * under non-exclusive lock(e.g., mmap_sem read-side) but defer TLB * flush by batching, a thread has stable TLB entry can fail to flush * the TLB by observing pte_none|!pte_dirty, for example so flush TLB * forcefully if we detect parallel PTE batching threads. + * + * Example: MADV_DONTNEED stale TLB problem on same range + * + * CPU 0 CPU 1 + * *a = 1; + * MADV_DONTNEED + * MADV_DONTNEED tlb_gather_mmu + * tlb_gather_mmu + * down_read(mmap_sem) down_read(mmap_sem) + * pte_lock + * pte_get_and_clear + * tlb_remove_tlb_entry + * pte_unlock + * pte_lock + * found out the pte is none + * pte_unlock + * tlb_finish_mmu doesn't flush + * + * Access the address with stale TLB + * *a = 2;ie, success without segfault + * tlb_finish_mmu flush on range + * but it is too late. + * */ bool force = mm_tlb_flush_nested(tlb->mm); -- 2.7.4 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>