On Thu, Aug 08, 2024 at 02:31:19PM -0700, Sean Christopherson wrote: > On Thu, Aug 08, 2024, Peter Xu wrote: > > Hi, Sean, > > > > On Thu, Aug 08, 2024 at 08:33:59AM -0700, Sean Christopherson wrote: > > > On Wed, Aug 07, 2024, Peter Xu wrote: > > > > mprotect() does mmu notifiers in PMD levels. It's there since 2014 of > > > > commit a5338093bfb4 ("mm: move mmu notifier call from change_protection to > > > > change_pmd_range"). > > > > > > > > At that time, the issue was that NUMA balancing can be applied on a huge > > > > range of VM memory, even if nothing was populated. The notification can be > > > > avoided in this case if no valid pmd detected, which includes either THP or > > > > a PTE pgtable page. > > > > > > > > Now to pave way for PUD handling, this isn't enough. We need to generate > > > > mmu notifications even on PUD entries properly. mprotect() is currently > > > > broken on PUD (e.g., one can easily trigger kernel error with dax 1G > > > > mappings already), this is the start to fix it. > > > > > > > > To fix that, this patch proposes to push such notifications to the PUD > > > > layers. > > > > > > > > There is risk on regressing the problem Rik wanted to resolve before, but I > > > > think it shouldn't really happen, and I still chose this solution because > > > > of a few reasons: > > > > > > > > 1) Consider a large VM that should definitely contain more than GBs of > > > > memory, it's highly likely that PUDs are also none. In this case there > > > > > > I don't follow this. Did you mean to say it's highly likely that PUDs are *NOT* > > > none? > > > > I did mean the original wordings. > > > > Note that in the previous case Rik worked on, it's about a mostly empty VM > > got NUMA hint applied. So I did mean "PUDs are also none" here, with the > > hope that when the numa hint applies on any part of the unpopulated guest > > memory, it'll find nothing in PUDs. Here it's mostly not about a huge PUD > > mapping as long as the guest memory is not backed by DAX (since only DAX > > supports 1G huge pud so far, while hugetlb has its own path here in > > mprotect, so it must be things like anon or shmem), but a PUD entry that > > contains pmd pgtables. For that part, I was trying to justify "no pmd > > pgtable installed" with the fact that "a large VM that should definitely > > contain more than GBs of memory", it means the PUD range should hopefully > > never been accessed, so even the pmd pgtable entry should be missing. > > Ah, now I get what you were saying. > > Problem is, walking the rmaps for the shadow MMU doesn't benefit (much) from > empty PUDs, because KVM needs to blindly walk the rmaps for every gfn covered by > the PUD to see if there are any SPTEs in any shadow MMUs mapping that gfn. And > that walk is done without ever yielding, which I suspect is the source of the > soft lockups of yore. > > And there's no way around that conundrum (walking rmaps), at least not without a > major rewrite in KVM. In a nested TDP scenario, KVM's stage-2 page tables (for > L2) key off of L2 gfns, not L1 gfns, and so the only way to find mappings is > through the rmaps. I think the hope here is when the whole PUDs being hinted are empty without pgtable installed, there'll be no mmu notifier to be kicked off at all. To be explicit, I meant after this patch applied, the pud loop for numa hints look like this: FOR_EACH_PUD() { ... if (pud_none(pud)) continue; if (!range.start) { mmu_notifier_range_init(&range, MMU_NOTIFY_PROTECTION_VMA, 0, vma->vm_mm, addr, end); mmu_notifier_invalidate_range_start(&range); } ... } So the hope is that pud_none() is always true for the hinted area (just like it used to be when pmd_none() can be hopefully true always), then we skip the mmu notifier as a whole (including KVM's)! Thanks, -- Peter Xu