On Mon, Dec 30, 2024 at 6:53 PM Rik van Riel <riel@xxxxxxxxxxx> wrote: > With AMD TCE (translation cache extensions) only the intermediate mappings > that cover the address range zapped by INVLPG / INVLPGB get invalidated, > rather than all intermediate mappings getting zapped at every TLB invalidation. > > This can help reduce the TLB miss rate, by keeping more intermediate > mappings in the cache. > > >From the AMD manual: > > Translation Cache Extension (TCE) Bit. Bit 15, read/write. Setting this bit > to 1 changes how the INVLPG, INVLPGB, and INVPCID instructions operate on > TLB entries. When this bit is 0, these instructions remove the target PTE > from the TLB as well as all upper-level table entries that are cached > in the TLB, whether or not they are associated with the target PTE. > When this bit is set, these instructions will remove the target PTE and > only those upper-level entries that lead to the target PTE in > the page table hierarchy, leaving unrelated upper-level entries intact. How does this patch interact with KVM SVM guests? In particular, will this patch cause TLB flushes performed by guest kernels to behave differently? [...] > diff --git a/arch/x86/mm/tlb.c b/arch/x86/mm/tlb.c > index 454a370494d3..585d0731ca9f 100644 > --- a/arch/x86/mm/tlb.c > +++ b/arch/x86/mm/tlb.c > @@ -477,7 +477,7 @@ static void broadcast_tlb_flush(struct flush_tlb_info *info) > if (info->stride_shift > PMD_SHIFT) > maxnr = 1; > > - if (info->end == TLB_FLUSH_ALL) { > + if (info->end == TLB_FLUSH_ALL || info->freed_tables) { > invlpgb_flush_single_pcid(kern_pcid(asid)); > /* Do any CPUs supporting INVLPGB need PTI? */ > if (static_cpu_has(X86_FEATURE_PTI)) > @@ -1110,7 +1110,7 @@ static void flush_tlb_func(void *info) > * > * The only question is whether to do a full or partial flush. > * > - * We do a partial flush if requested and two extra conditions > + * We do a partial flush if requested and three extra conditions > * are met: > * > * 1. f->new_tlb_gen == local_tlb_gen + 1. We have an invariant that > @@ -1137,10 +1137,14 @@ static void flush_tlb_func(void *info) > * date. By doing a full flush instead, we can increase > * local_tlb_gen all the way to mm_tlb_gen and we can probably > * avoid another flush in the very near future. > + * > + * 3. No page tables were freed. If page tables were freed, a full > + * flush ensures intermediate translations in the TLB get flushed. > */ Why is this necessary - do we ever issue TLB flushes that are intended to zap upper-level entries which are not covered by the specified address range? When, for example, free_pmd_range() gets rid of a page table, it calls pmd_free_tlb(), which sets tlb->freed_tables and does tlb_flush_pud_range(tlb, address, PAGE_SIZE).