On Mon, 2021-06-28 at 12:44 +0200, Vitaly Kuznetsov wrote: > In case nested state was saved/resored while in SMM, > nested_load_control_from_vmcb12() which sets svm->nested.ctl was never > called and the first nested_vmcb_check_controls() (either from > nested_svm_vmrun() or from svm_set_nested_state() if save/restore > cycle is repeated) is doomed to fail. I don't like the commit description. I propose something like that: If the VM was migrated while in SMM, no nested state was saved/restored, and therefore svm_leave_smm has to load both save and control area of the vmcb12. Save area is already loaded from HSAVE area, so now load the control area as well from the vmcb12. However if you like to, feel free to leave the commit message as is. My point is that while in SMM, SVM is fully disabled, so not only svm->nested.ctl is not set but no nested state is loaded/stored at all. Also this makes the svm_leave_smm even more dangerous versus errors, as I said in previos patch. Since its return value is ignored, and we are loading here the guest vmcb01 which can change under our feet, lots of fun can happen (enter_svm_guest_mode result isn't really checked). Best regards, Maxim Levitsky > > Signed-off-by: Vitaly Kuznetsov <vkuznets@xxxxxxxxxx> > --- > arch/x86/kvm/svm/nested.c | 4 ++-- > arch/x86/kvm/svm/svm.c | 7 ++++++- > arch/x86/kvm/svm/svm.h | 2 ++ > 3 files changed, 10 insertions(+), 3 deletions(-) > > diff --git a/arch/x86/kvm/svm/nested.c b/arch/x86/kvm/svm/nested.c > index a1dec2c40181..6549e40155fa 100644 > --- a/arch/x86/kvm/svm/nested.c > +++ b/arch/x86/kvm/svm/nested.c > @@ -304,8 +304,8 @@ static bool nested_vmcb_valid_sregs(struct kvm_vcpu *vcpu, > return true; > } > > -static void nested_load_control_from_vmcb12(struct vcpu_svm *svm, > - struct vmcb_control_area *control) > +void nested_load_control_from_vmcb12(struct vcpu_svm *svm, > + struct vmcb_control_area *control) > { > copy_vmcb_control_area(&svm->nested.ctl, control); > > diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c > index fbf1b352a9bb..525b07873927 100644 > --- a/arch/x86/kvm/svm/svm.c > +++ b/arch/x86/kvm/svm/svm.c > @@ -4344,6 +4344,7 @@ static int svm_leave_smm(struct kvm_vcpu *vcpu, const char *smstate) > u64 saved_efer = GET_SMSTATE(u64, smstate, 0x7ed0); > u64 guest = GET_SMSTATE(u64, smstate, 0x7ed8); > u64 vmcb12_gpa = GET_SMSTATE(u64, smstate, 0x7ee0); > + struct vmcb *vmcb12; > > if (guest) { > if (!guest_cpuid_has(vcpu, X86_FEATURE_SVM)) > @@ -4359,7 +4360,11 @@ static int svm_leave_smm(struct kvm_vcpu *vcpu, const char *smstate) > if (svm_allocate_nested(svm)) > return 1; > > - ret = enter_svm_guest_mode(vcpu, vmcb12_gpa, map.hva); > + vmcb12 = map.hva; > + > + nested_load_control_from_vmcb12(svm, &vmcb12->control); > + > + ret = enter_svm_guest_mode(vcpu, vmcb12_gpa, vmcb12); > kvm_vcpu_unmap(vcpu, &map, true); > > /* > diff --git a/arch/x86/kvm/svm/svm.h b/arch/x86/kvm/svm/svm.h > index ff2dac2b23b6..13f2d465ca36 100644 > --- a/arch/x86/kvm/svm/svm.h > +++ b/arch/x86/kvm/svm/svm.h > @@ -481,6 +481,8 @@ int nested_svm_check_permissions(struct kvm_vcpu *vcpu); > int nested_svm_check_exception(struct vcpu_svm *svm, unsigned nr, > bool has_error_code, u32 error_code); > int nested_svm_exit_special(struct vcpu_svm *svm); > +void nested_load_control_from_vmcb12(struct vcpu_svm *svm, > + struct vmcb_control_area *control); > void nested_sync_control_from_vmcb02(struct vcpu_svm *svm); > void nested_vmcb02_compute_g_pat(struct vcpu_svm *svm); > void svm_switch_vmcb(struct vcpu_svm *svm, struct kvm_vmcb_info *target_vmcb);