On Tue, Aug 18, 2020 at 8:20 AM Sean Christopherson <sean.j.christopherson@xxxxxxxxx> wrote: > I'd prefer to handle this on the switch from L2->L1. It avoids adding a > kvm_x86_ops and yet another sequence of four VMWRITEs, e.g. I think this > will do the trick. > > diff --git a/arch/x86/kvm/vmx/nested.c b/arch/x86/kvm/vmx/nested.c > index 9c74a732b08d..67465f0ca1b9 100644 > --- a/arch/x86/kvm/vmx/nested.c > +++ b/arch/x86/kvm/vmx/nested.c > @@ -4356,6 +4356,9 @@ void nested_vmx_vmexit(struct kvm_vcpu *vcpu, u32 vm_exit_reason, > if (kvm_check_request(KVM_REQ_TLB_FLUSH_CURRENT, vcpu)) > kvm_vcpu_flush_tlb_current(vcpu); > > + if (enable_ept && is_pae_paging(vcpu)) > + ept_load_pdptrs(vcpu); > + Are the mmu->pdptrs[] guaranteed to be valid at this point? If L2 has PAE paging enabled, and it has modified CR3 without a VM-exit, where are the current PDPTE values read from the vmcs02 into mmu->pdptrs[]?