Maxim Levitsky <mlevitsk@xxxxxxxxxx> writes: > On Mon, 2021-05-03 at 17:08 +0200, Vitaly Kuznetsov wrote: >> It now looks like a bad idea to not restore eVMCS mapping directly from >> vmx_set_nested_state(). The restoration path now depends on whether KVM >> will continue executing L2 (vmx_get_nested_state_pages()) or will have to >> exit to L1 (nested_vmx_vmexit()), this complicates error propagation and >> diverges too much from the 'native' path when 'nested.current_vmptr' is >> set directly from vmx_get_nested_state_pages(). >> >> The existing solution postponing eVMCS mapping also seems to be fragile. >> In multiple places the code checks whether 'vmx->nested.hv_evmcs' is not >> NULL to distinguish between eVMCS and non-eVMCS cases. All these checks >> are 'incomplete' as we have a weird 'eVMCS is in use but not yet mapped' >> state. >> >> Also, in case vmx_get_nested_state() is called right after >> vmx_set_nested_state() without executing the guest first, the resulting >> state is going to be incorrect as 'KVM_STATE_NESTED_EVMCS' flag will be >> missing. >> >> Fix all these issues by making eVMCS restoration path closer to its >> 'native' sibling by putting eVMCS GPA to 'struct kvm_vmx_nested_state_hdr'. >> To avoid ABI incompatibility, do not introduce a new flag and keep the >> original eVMCS mapping path through KVM_REQ_GET_NESTED_STATE_PAGES in >> place. To distinguish between 'new' and 'old' formats consider eVMCS >> GPA == 0 as an unset GPA (thus forcing KVM_REQ_GET_NESTED_STATE_PAGES >> path). While technically possible, it seems to be an extremely unlikely >> case. >> >> Signed-off-by: Vitaly Kuznetsov <vkuznets@xxxxxxxxxx> >> --- >> arch/x86/include/uapi/asm/kvm.h | 2 ++ >> arch/x86/kvm/vmx/nested.c | 27 +++++++++++++++++++++------ >> 2 files changed, 23 insertions(+), 6 deletions(-) >> >> diff --git a/arch/x86/include/uapi/asm/kvm.h b/arch/x86/include/uapi/asm/kvm.h >> index 0662f644aad9..3845977b739e 100644 >> --- a/arch/x86/include/uapi/asm/kvm.h >> +++ b/arch/x86/include/uapi/asm/kvm.h >> @@ -441,6 +441,8 @@ struct kvm_vmx_nested_state_hdr { >> >> __u32 flags; >> __u64 preemption_timer_deadline; >> + >> + __u64 evmcs_pa; >> }; >> >> struct kvm_svm_nested_state_data { >> diff --git a/arch/x86/kvm/vmx/nested.c b/arch/x86/kvm/vmx/nested.c >> index 37fdc34f7afc..4261cf4755c8 100644 >> --- a/arch/x86/kvm/vmx/nested.c >> +++ b/arch/x86/kvm/vmx/nested.c >> @@ -6019,6 +6019,7 @@ static int vmx_get_nested_state(struct kvm_vcpu *vcpu, >> .hdr.vmx.vmxon_pa = -1ull, >> .hdr.vmx.vmcs12_pa = -1ull, >> .hdr.vmx.preemption_timer_deadline = 0, >> + .hdr.vmx.evmcs_pa = -1ull, >> }; >> struct kvm_vmx_nested_state_data __user *user_vmx_nested_state = >> &user_kvm_nested_state->data.vmx[0]; >> @@ -6037,8 +6038,10 @@ static int vmx_get_nested_state(struct kvm_vcpu *vcpu, >> if (vmx_has_valid_vmcs12(vcpu)) { >> kvm_state.size += sizeof(user_vmx_nested_state->vmcs12); >> >> - if (vmx->nested.hv_evmcs) >> + if (vmx->nested.hv_evmcs) { >> kvm_state.flags |= KVM_STATE_NESTED_EVMCS; >> + kvm_state.hdr.vmx.evmcs_pa = vmx->nested.hv_evmcs_vmptr; >> + } >> >> if (is_guest_mode(vcpu) && >> nested_cpu_has_shadow_vmcs(vmcs12) && >> @@ -6230,13 +6233,25 @@ static int vmx_set_nested_state(struct kvm_vcpu *vcpu, >> >> set_current_vmptr(vmx, kvm_state->hdr.vmx.vmcs12_pa); >> } else if (kvm_state->flags & KVM_STATE_NESTED_EVMCS) { >> + u64 evmcs_gpa = kvm_state->hdr.vmx.evmcs_pa; >> + >> /* >> - * nested_vmx_handle_enlightened_vmptrld() cannot be called >> - * directly from here as HV_X64_MSR_VP_ASSIST_PAGE may not be >> - * restored yet. EVMCS will be mapped from >> - * nested_get_vmcs12_pages(). >> + * EVMCS GPA == 0 most likely indicates that the migration data is >> + * coming from an older KVM which doesn't support 'evmcs_pa' in >> + * 'struct kvm_vmx_nested_state_hdr'. >> */ >> - kvm_make_request(KVM_REQ_GET_NESTED_STATE_PAGES, vcpu); >> + if (evmcs_gpa && (evmcs_gpa != -1ull) && >> + (__nested_vmx_handle_enlightened_vmptrld(vcpu, evmcs_gpa, false) != >> + EVMPTRLD_SUCCEEDED)) { >> + return -EINVAL; >> + } else if (!evmcs_gpa) { >> + /* >> + * EVMCS GPA can't be acquired from VP assist page here because >> + * HV_X64_MSR_VP_ASSIST_PAGE may not be restored yet. >> + * EVMCS will be mapped from nested_get_evmcs_page(). >> + */ >> + kvm_make_request(KVM_REQ_GET_NESTED_STATE_PAGES, vcpu); >> + } >> } else { >> return -EINVAL; >> } > > Hi everyone! > > Let me expalin my concern about this patch and also ask if I understand this correctly. > > In a nutshell if I understand this correctly, we are not allowed to access any guest > memory while setting the nested state. > > Now, if I understand correctly as well, the reason for the above, > is that the userspace is allowed to set the nested state first, then fiddle with > the KVM memslots, maybe even update the guest memory and only later do the KVM_RUN ioctl, Currently, userspace is free to restore the guest in any order indeed. I've probably missed post-copy but even the fact that guest MSRs can be restored after restoring nested state doesn't make our life easier. > > And so this is the major reason why the KVM_REQ_GET_NESTED_STATE_PAGES > request exists in the first place. > > If that is correct I assume that we either have to keep loading the EVMCS page on > KVM_REQ_GET_NESTED_STATE_PAGES request, or we want to include the EVMCS itself > in the migration state in addition to its physical address, similar to how we treat > the VMCS12 and the VMCB12. Keeping eVMCS load from KVM_REQ_GET_NESTED_STATE_PAGES is OK I believe (or at least I still don't see a reason for us to carry a copy in the migration data). What I still don't like is the transient state after vmx_set_nested_state(): - vmx->nested.current_vmptr is -1ull because no 'real' vmptrld was done (we skip set_current_vmptr() when KVM_STATE_NESTED_EVMCS) - vmx->nested.hv_evmcs/vmx->nested.hv_evmcs_vmptr are also NULL because we haven't performed nested_vmx_handle_enlightened_vmptrld() yet. I know of at least one real problem with this state: in case vmx_get_nested_state() happens before KVM_RUN the resulting state won't have KVM_STATE_NESTED_EVMCS flag and this is incorrect. Take a look at the check in nested_vmx_fail() for example: if (vmx->nested.current_vmptr == -1ull && !vmx->nested.hv_evmcs) return nested_vmx_failInvalid(vcpu); this also seems off (I'm not sure it matters in any context but still). > > I personally tinkered with qemu to try and reproduce this situation > and in my tests I wasn't able to make it update the memory > map after the load of the nested state but prior to KVM_RUN > but neither I wasn't able to prove that this can't happen. Userspace has multiple ways to mess with the state of course, in KVM we only need to make sure we don't crash :-) On migration, well behaving userspace is supposed to restore exactly what it got though. The restoration sequence may vary. > > In addition to that I don't know how qemu behaves when it does > guest ram post-copy because so far I haven't tried to tinker with it. > > Finally other userspace hypervisors exist, and they might rely on assumption > as well. > > Looking forward for any comments, > Best regards, > Maxim Levitsky > > > -- Vitaly