On Tue, Jul 18, 2023 at 04:44:44PM -0700, Sean Christopherson wrote: May I know why KVM now needs to register to callback .change_pte()? As also commented in kvm_mmu_notifier_change_pte(), .change_pte() must be surrounded by .invalidate_range_{start,end}(). While kvm_mmu_notifier_invalidate_range_start() has called kvm_unmap_gfn_range() to zap all leaf SPTEs, and page fault path will not install new SPTEs successfully before kvm_mmu_notifier_invalidate_range_end(), kvm_set_spte_gfn() should not be able to find any shadow present leaf entries to update PFN. Or could we just delete completely "kvm_handle_hva_range(mn, address, address + 1, pte, kvm_set_spte_gfn);" from kvm_mmu_notifier_change_pte() ? > diff --git a/arch/arm64/kvm/mmu.c b/arch/arm64/kvm/mmu.c > index 6db9ef288ec3..55f03a68f1cd 100644 > --- a/arch/arm64/kvm/mmu.c > +++ b/arch/arm64/kvm/mmu.c > @@ -1721,7 +1721,7 @@ bool kvm_unmap_gfn_range(struct kvm *kvm, struct kvm_gfn_range *range) > > bool kvm_set_spte_gfn(struct kvm *kvm, struct kvm_gfn_range *range) > { > - kvm_pfn_t pfn = pte_pfn(range->pte); > + kvm_pfn_t pfn = pte_pfn(range->arg.pte); > > if (!kvm->arch.mmu.pgt) > return false; > diff --git a/arch/mips/kvm/mmu.c b/arch/mips/kvm/mmu.c > index e8c08988ed37..7b2ac1319d70 100644 > --- a/arch/mips/kvm/mmu.c > +++ b/arch/mips/kvm/mmu.c > @@ -447,7 +447,7 @@ bool kvm_unmap_gfn_range(struct kvm *kvm, struct kvm_gfn_range *range) > bool kvm_set_spte_gfn(struct kvm *kvm, struct kvm_gfn_range *range) > { > gpa_t gpa = range->start << PAGE_SHIFT; > - pte_t hva_pte = range->pte; > + pte_t hva_pte = range->arg.pte; > pte_t *gpa_pte = kvm_mips_pte_for_gpa(kvm, NULL, gpa); > pte_t old_pte; > > diff --git a/arch/riscv/kvm/mmu.c b/arch/riscv/kvm/mmu.c > index f2eb47925806..857f4312b0f8 100644 > --- a/arch/riscv/kvm/mmu.c > +++ b/arch/riscv/kvm/mmu.c > @@ -559,7 +559,7 @@ bool kvm_unmap_gfn_range(struct kvm *kvm, struct kvm_gfn_range *range) > bool kvm_set_spte_gfn(struct kvm *kvm, struct kvm_gfn_range *range) > { > int ret; > - kvm_pfn_t pfn = pte_pfn(range->pte); > + kvm_pfn_t pfn = pte_pfn(range->arg.pte); > > if (!kvm->arch.pgd) > return false; > diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c > index ec169f5c7dce..d72f2b20f430 100644 > --- a/arch/x86/kvm/mmu/mmu.c > +++ b/arch/x86/kvm/mmu/mmu.c > @@ -1588,7 +1588,7 @@ static __always_inline bool kvm_handle_gfn_range(struct kvm *kvm, > for_each_slot_rmap_range(range->slot, PG_LEVEL_4K, KVM_MAX_HUGEPAGE_LEVEL, > range->start, range->end - 1, &iterator) > ret |= handler(kvm, iterator.rmap, range->slot, iterator.gfn, > - iterator.level, range->pte); > + iterator.level, range->arg.pte); > > return ret; > } > diff --git a/arch/x86/kvm/mmu/tdp_mmu.c b/arch/x86/kvm/mmu/tdp_mmu.c > index 512163d52194..6250bd3d20c1 100644 > --- a/arch/x86/kvm/mmu/tdp_mmu.c > +++ b/arch/x86/kvm/mmu/tdp_mmu.c > @@ -1241,7 +1241,7 @@ static bool set_spte_gfn(struct kvm *kvm, struct tdp_iter *iter, > u64 new_spte; > > /* Huge pages aren't expected to be modified without first being zapped. */ > - WARN_ON(pte_huge(range->pte) || range->start + 1 != range->end); > + WARN_ON(pte_huge(range->arg.pte) || range->start + 1 != range->end); > > if (iter->level != PG_LEVEL_4K || > !is_shadow_present_pte(iter->old_spte)) > @@ -1255,9 +1255,9 @@ static bool set_spte_gfn(struct kvm *kvm, struct tdp_iter *iter, > */ > tdp_mmu_iter_set_spte(kvm, iter, 0); > > - if (!pte_write(range->pte)) { > + if (!pte_write(range->arg.pte)) { > new_spte = kvm_mmu_changed_pte_notifier_make_spte(iter->old_spte, > - pte_pfn(range->pte)); > + pte_pfn(range->arg.pte)); > > tdp_mmu_iter_set_spte(kvm, iter, new_spte); > }