On Mon, 2022-07-18 at 16:07 +0300, Maxim Levitsky wrote: > On Tue, 2022-06-14 at 20:47 +0000, Sean Christopherson wrote: > > Move the definition of "struct kvm_queued_exception" out of kvm_vcpu_arch > > in anticipation of adding a second instance in kvm_vcpu_arch to handle > > exceptions that occur when vectoring an injected exception and are > > morphed to VM-Exit instead of leading to #DF. > > > > Opportunistically take advantage of the churn to rename "nr" to "vector". > > > > No functional change intended. > > > > Signed-off-by: Sean Christopherson <seanjc@xxxxxxxxxx> > > --- > > > ... > > > Is this change below intentional? My memory on nested_apf_token is quite rusty, but at least > if possible, I would prefer this to be done in separate patch. Sorry, I replied to the wrong mail, but the newer version also has the same issue. (It should be v3 btw.) Best regards, Maxim Levitsky > > > Best regards, > Maxim Levitsky > > > - else if (svm->vcpu.arch.exception.has_payload) > > - vmcb->control.exit_info_2 = svm->vcpu.arch.exception.payload; > > + if (ex->vector == PF_VECTOR) { > > + if (ex->has_payload) > > + vmcb->control.exit_info_2 = ex->payload; > > else > > - vmcb->control.exit_info_2 = svm->vcpu.arch.cr2; > > - } else if (nr == DB_VECTOR) { > > + vmcb->control.exit_info_2 = vcpu->arch.cr2; > > + } else if (ex->vector == DB_VECTOR) { > > /* See inject_pending_event. */ > > - kvm_deliver_exception_payload(&svm->vcpu); > > - if (svm->vcpu.arch.dr7 & DR7_GD) { > > - svm->vcpu.arch.dr7 &= ~DR7_GD; > > - kvm_update_dr7(&svm->vcpu); > > + kvm_deliver_exception_payload(vcpu, ex); > > + > > + if (vcpu->arch.dr7 & DR7_GD) { > > + vcpu->arch.dr7 &= ~DR7_GD; > > + kvm_update_dr7(vcpu); > > } > > - } else > > - WARN_ON(svm->vcpu.arch.exception.has_payload); > > + } else { > > + WARN_ON(ex->has_payload); > > + } > > > > nested_svm_vmexit(svm); > > } > > @@ -1372,7 +1373,7 @@ static int svm_check_nested_events(struct kvm_vcpu *vcpu) > > return -EBUSY; > > if (!nested_exit_on_exception(svm)) > > return 0; > > - nested_svm_inject_exception_vmexit(svm); > > + nested_svm_inject_exception_vmexit(vcpu); > > return 0; > > } > > > > diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c > > index ca39f76ca44b..6b80046a014f 100644 > > --- a/arch/x86/kvm/svm/svm.c > > +++ b/arch/x86/kvm/svm/svm.c > > @@ -432,22 +432,20 @@ static int svm_update_soft_interrupt_rip(struct kvm_vcpu *vcpu) > > > > static void svm_inject_exception(struct kvm_vcpu *vcpu) > > { > > + struct kvm_queued_exception *ex = &vcpu->arch.exception; > > struct vcpu_svm *svm = to_svm(vcpu); > > - unsigned nr = vcpu->arch.exception.nr; > > - bool has_error_code = vcpu->arch.exception.has_error_code; > > - u32 error_code = vcpu->arch.exception.error_code; > > > > - kvm_deliver_exception_payload(vcpu); > > + kvm_deliver_exception_payload(vcpu, ex); > > > > - if (kvm_exception_is_soft(nr) && > > + if (kvm_exception_is_soft(ex->vector) && > > svm_update_soft_interrupt_rip(vcpu)) > > return; > > > > - svm->vmcb->control.event_inj = nr > > + svm->vmcb->control.event_inj = ex->vector > > | SVM_EVTINJ_VALID > > - | (has_error_code ? SVM_EVTINJ_VALID_ERR : 0) > > + | (ex->has_error_code ? SVM_EVTINJ_VALID_ERR : 0) > > | SVM_EVTINJ_TYPE_EXEPT; > > - svm->vmcb->control.event_inj_err = error_code; > > + svm->vmcb->control.event_inj_err = ex->error_code; > > } > > > > static void svm_init_erratum_383(void) > > diff --git a/arch/x86/kvm/vmx/nested.c b/arch/x86/kvm/vmx/nested.c > > index 7b644513c82b..fafdcbfeca1f 100644 > > --- a/arch/x86/kvm/vmx/nested.c > > +++ b/arch/x86/kvm/vmx/nested.c > > @@ -445,29 +445,27 @@ static bool nested_vmx_is_page_fault_vmexit(struct vmcs12 *vmcs12, > > */ > > static int nested_vmx_check_exception(struct kvm_vcpu *vcpu, unsigned long *exit_qual) > > { > > + struct kvm_queued_exception *ex = &vcpu->arch.exception; > > struct vmcs12 *vmcs12 = get_vmcs12(vcpu); > > - unsigned int nr = vcpu->arch.exception.nr; > > - bool has_payload = vcpu->arch.exception.has_payload; > > - unsigned long payload = vcpu->arch.exception.payload; > > > > - if (nr == PF_VECTOR) { > > - if (vcpu->arch.exception.nested_apf) { > > + if (ex->vector == PF_VECTOR) { > > + if (ex->nested_apf) { > > *exit_qual = vcpu->arch.apf.nested_apf_token; > > return 1; > > } > > - if (nested_vmx_is_page_fault_vmexit(vmcs12, > > - vcpu->arch.exception.error_code)) { > > - *exit_qual = has_payload ? payload : vcpu->arch.cr2; > > + if (nested_vmx_is_page_fault_vmexit(vmcs12, ex->error_code)) { > > + *exit_qual = ex->has_payload ? ex->payload : vcpu->arch.cr2; > > return 1; > > } > > - } else if (vmcs12->exception_bitmap & (1u << nr)) { > > - if (nr == DB_VECTOR) { > > - if (!has_payload) { > > - payload = vcpu->arch.dr6; > > - payload &= ~DR6_BT; > > - payload ^= DR6_ACTIVE_LOW; > > + } else if (vmcs12->exception_bitmap & (1u << ex->vector)) { > > + if (ex->vector == DB_VECTOR) { > > + if (ex->has_payload) { > > + *exit_qual = ex->payload; > > + } else { > > + *exit_qual = vcpu->arch.dr6; > > + *exit_qual &= ~DR6_BT; > > + *exit_qual ^= DR6_ACTIVE_LOW; > > } > > - *exit_qual = payload; > > } else > > *exit_qual = 0; > > return 1; > > @@ -3724,7 +3722,7 @@ static void vmcs12_save_pending_event(struct kvm_vcpu *vcpu, > > is_double_fault(exit_intr_info))) { > > vmcs12->idt_vectoring_info_field = 0; > > } else if (vcpu->arch.exception.injected) { > > - nr = vcpu->arch.exception.nr; > > + nr = vcpu->arch.exception.vector; > > idt_vectoring = nr | VECTORING_INFO_VALID_MASK; > > > > if (kvm_exception_is_soft(nr)) { > > @@ -3828,11 +3826,11 @@ static int vmx_complete_nested_posted_interrupt(struct kvm_vcpu *vcpu) > > static void nested_vmx_inject_exception_vmexit(struct kvm_vcpu *vcpu, > > unsigned long exit_qual) > > { > > + struct kvm_queued_exception *ex = &vcpu->arch.exception; > > + u32 intr_info = ex->vector | INTR_INFO_VALID_MASK; > > struct vmcs12 *vmcs12 = get_vmcs12(vcpu); > > - unsigned int nr = vcpu->arch.exception.nr; > > - u32 intr_info = nr | INTR_INFO_VALID_MASK; > > > > - if (vcpu->arch.exception.has_error_code) { > > + if (ex->has_error_code) { > > /* > > * Intel CPUs will never generate an error code with bits 31:16 > > * set, and more importantly VMX disallows setting bits 31:16 > > @@ -3840,11 +3838,11 @@ static void nested_vmx_inject_exception_vmexit(struct kvm_vcpu *vcpu, > > * mimic hardware and avoid inducing failure on nested VM-Entry > > * if L1 chooses to inject the exception back to L2. > > */ > > - vmcs12->vm_exit_intr_error_code = (u16)vcpu->arch.exception.error_code; > > + vmcs12->vm_exit_intr_error_code = (u16)ex->error_code; > > intr_info |= INTR_INFO_DELIVER_CODE_MASK; > > } > > > > - if (kvm_exception_is_soft(nr)) > > + if (kvm_exception_is_soft(ex->vector)) > > intr_info |= INTR_TYPE_SOFT_EXCEPTION; > > else > > intr_info |= INTR_TYPE_HARD_EXCEPTION; > > @@ -3875,7 +3873,7 @@ static void nested_vmx_inject_exception_vmexit(struct kvm_vcpu *vcpu, > > static inline unsigned long vmx_get_pending_dbg_trap(struct kvm_vcpu *vcpu) > > { > > if (!vcpu->arch.exception.pending || > > - vcpu->arch.exception.nr != DB_VECTOR) > > + vcpu->arch.exception.vector != DB_VECTOR) > > return 0; > > > > /* General Detect #DBs are always fault-like. */ > > diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c > > index 26b863c78a9f..7ef5659a1bbd 100644 > > --- a/arch/x86/kvm/vmx/vmx.c > > +++ b/arch/x86/kvm/vmx/vmx.c > > @@ -1585,7 +1585,7 @@ static void vmx_update_emulated_instruction(struct kvm_vcpu *vcpu) > > */ > > if (nested_cpu_has_mtf(vmcs12) && > > (!vcpu->arch.exception.pending || > > - vcpu->arch.exception.nr == DB_VECTOR)) > > + vcpu->arch.exception.vector == DB_VECTOR)) > > vmx->nested.mtf_pending = true; > > else > > vmx->nested.mtf_pending = false; > > @@ -1612,15 +1612,13 @@ static void vmx_clear_hlt(struct kvm_vcpu *vcpu) > > > > static void vmx_inject_exception(struct kvm_vcpu *vcpu) > > { > > + struct kvm_queued_exception *ex = &vcpu->arch.exception; > > + u32 intr_info = ex->vector | INTR_INFO_VALID_MASK; > > struct vcpu_vmx *vmx = to_vmx(vcpu); > > - unsigned nr = vcpu->arch.exception.nr; > > - bool has_error_code = vcpu->arch.exception.has_error_code; > > - u32 error_code = vcpu->arch.exception.error_code; > > - u32 intr_info = nr | INTR_INFO_VALID_MASK; > > > > - kvm_deliver_exception_payload(vcpu); > > + kvm_deliver_exception_payload(vcpu, ex); > > > > - if (has_error_code) { > > + if (ex->has_error_code) { > > /* > > * Despite the error code being architecturally defined as 32 > > * bits, and the VMCS field being 32 bits, Intel CPUs and thus > > @@ -1630,21 +1628,21 @@ static void vmx_inject_exception(struct kvm_vcpu *vcpu) > > * the upper bits to avoid VM-Fail, losing information that > > * does't really exist is preferable to killing the VM. > > */ > > - vmcs_write32(VM_ENTRY_EXCEPTION_ERROR_CODE, (u16)error_code); > > + vmcs_write32(VM_ENTRY_EXCEPTION_ERROR_CODE, (u16)ex->error_code); > > intr_info |= INTR_INFO_DELIVER_CODE_MASK; > > } > > > > if (vmx->rmode.vm86_active) { > > int inc_eip = 0; > > - if (kvm_exception_is_soft(nr)) > > + if (kvm_exception_is_soft(ex->vector)) > > inc_eip = vcpu->arch.event_exit_inst_len; > > - kvm_inject_realmode_interrupt(vcpu, nr, inc_eip); > > + kvm_inject_realmode_interrupt(vcpu, ex->vector, inc_eip); > > return; > > } > > > > WARN_ON_ONCE(vmx->emulation_required); > > > > - if (kvm_exception_is_soft(nr)) { > > + if (kvm_exception_is_soft(ex->vector)) { > > vmcs_write32(VM_ENTRY_INSTRUCTION_LEN, > > vmx->vcpu.arch.event_exit_inst_len); > > intr_info |= INTR_TYPE_SOFT_EXCEPTION; > > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > > index b63421d511c5..511c0c8af80e 100644 > > --- a/arch/x86/kvm/x86.c > > +++ b/arch/x86/kvm/x86.c > > @@ -557,16 +557,13 @@ static int exception_type(int vector) > > return EXCPT_FAULT; > > } > > > > -void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu) > > +void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu, > > + struct kvm_queued_exception *ex) > > { > > - unsigned nr = vcpu->arch.exception.nr; > > - bool has_payload = vcpu->arch.exception.has_payload; > > - unsigned long payload = vcpu->arch.exception.payload; > > - > > - if (!has_payload) > > + if (!ex->has_payload) > > return; > > > > - switch (nr) { > > + switch (ex->vector) { > > case DB_VECTOR: > > /* > > * "Certain debug exceptions may clear bit 0-3. The > > @@ -591,8 +588,8 @@ void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu) > > * So they need to be flipped for DR6. > > */ > > vcpu->arch.dr6 |= DR6_ACTIVE_LOW; > > - vcpu->arch.dr6 |= payload; > > - vcpu->arch.dr6 ^= payload & DR6_ACTIVE_LOW; > > + vcpu->arch.dr6 |= ex->payload; > > + vcpu->arch.dr6 ^= ex->payload & DR6_ACTIVE_LOW; > > > > /* > > * The #DB payload is defined as compatible with the 'pending > > @@ -603,12 +600,12 @@ void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu) > > vcpu->arch.dr6 &= ~BIT(12); > > break; > > case PF_VECTOR: > > - vcpu->arch.cr2 = payload; > > + vcpu->arch.cr2 = ex->payload; > > break; > > } > > > > - vcpu->arch.exception.has_payload = false; > > - vcpu->arch.exception.payload = 0; > > + ex->has_payload = false; > > + ex->payload = 0; > > } > > EXPORT_SYMBOL_GPL(kvm_deliver_exception_payload); > > > > @@ -647,17 +644,18 @@ static void kvm_multiple_exception(struct kvm_vcpu *vcpu, > > vcpu->arch.exception.injected = false; > > } > > vcpu->arch.exception.has_error_code = has_error; > > - vcpu->arch.exception.nr = nr; > > + vcpu->arch.exception.vector = nr; > > vcpu->arch.exception.error_code = error_code; > > vcpu->arch.exception.has_payload = has_payload; > > vcpu->arch.exception.payload = payload; > > if (!is_guest_mode(vcpu)) > > - kvm_deliver_exception_payload(vcpu); > > + kvm_deliver_exception_payload(vcpu, > > + &vcpu->arch.exception); > > return; > > } > > > > /* to check exception */ > > - prev_nr = vcpu->arch.exception.nr; > > + prev_nr = vcpu->arch.exception.vector; > > if (prev_nr == DF_VECTOR) { > > /* triple fault -> shutdown */ > > kvm_make_request(KVM_REQ_TRIPLE_FAULT, vcpu); > > @@ -675,7 +673,7 @@ static void kvm_multiple_exception(struct kvm_vcpu *vcpu, > > vcpu->arch.exception.pending = true; > > vcpu->arch.exception.injected = false; > > vcpu->arch.exception.has_error_code = true; > > - vcpu->arch.exception.nr = DF_VECTOR; > > + vcpu->arch.exception.vector = DF_VECTOR; > > vcpu->arch.exception.error_code = 0; > > vcpu->arch.exception.has_payload = false; > > vcpu->arch.exception.payload = 0; > > @@ -4886,25 +4884,24 @@ static int kvm_vcpu_ioctl_x86_set_mce(struct kvm_vcpu *vcpu, > > static void kvm_vcpu_ioctl_x86_get_vcpu_events(struct kvm_vcpu *vcpu, > > struct kvm_vcpu_events *events) > > { > > + struct kvm_queued_exception *ex = &vcpu->arch.exception; > > + > > process_nmi(vcpu); > > > > if (kvm_check_request(KVM_REQ_SMI, vcpu)) > > process_smi(vcpu); > > > > /* > > - * In guest mode, payload delivery should be deferred, > > - * so that the L1 hypervisor can intercept #PF before > > - * CR2 is modified (or intercept #DB before DR6 is > > - * modified under nVMX). Unless the per-VM capability, > > - * KVM_CAP_EXCEPTION_PAYLOAD, is set, we may not defer the delivery of > > - * an exception payload and handle after a KVM_GET_VCPU_EVENTS. Since we > > - * opportunistically defer the exception payload, deliver it if the > > - * capability hasn't been requested before processing a > > - * KVM_GET_VCPU_EVENTS. > > + * In guest mode, payload delivery should be deferred if the exception > > + * will be intercepted by L1, e.g. KVM should not modifying CR2 if L1 > > + * intercepts #PF, ditto for DR6 and #DBs. If the per-VM capability, > > + * KVM_CAP_EXCEPTION_PAYLOAD, is not set, userspace may or may not > > + * propagate the payload and so it cannot be safely deferred. Deliver > > + * the payload if the capability hasn't been requested. > > */ > > if (!vcpu->kvm->arch.exception_payload_enabled && > > - vcpu->arch.exception.pending && vcpu->arch.exception.has_payload) > > - kvm_deliver_exception_payload(vcpu); > > + ex->pending && ex->has_payload) > > + kvm_deliver_exception_payload(vcpu, ex); > > > > /* > > * The API doesn't provide the instruction length for software > > @@ -4912,26 +4909,25 @@ static void kvm_vcpu_ioctl_x86_get_vcpu_events(struct kvm_vcpu *vcpu, > > * isn't advanced, we should expect to encounter the exception > > * again. > > */ > > - if (kvm_exception_is_soft(vcpu->arch.exception.nr)) { > > + if (kvm_exception_is_soft(ex->vector)) { > > events->exception.injected = 0; > > events->exception.pending = 0; > > } else { > > - events->exception.injected = vcpu->arch.exception.injected; > > - events->exception.pending = vcpu->arch.exception.pending; > > + events->exception.injected = ex->injected; > > + events->exception.pending = ex->pending; > > /* > > * For ABI compatibility, deliberately conflate > > * pending and injected exceptions when > > * KVM_CAP_EXCEPTION_PAYLOAD isn't enabled. > > */ > > if (!vcpu->kvm->arch.exception_payload_enabled) > > - events->exception.injected |= > > - vcpu->arch.exception.pending; > > + events->exception.injected |= ex->pending; > > } > > - events->exception.nr = vcpu->arch.exception.nr; > > - events->exception.has_error_code = vcpu->arch.exception.has_error_code; > > - events->exception.error_code = vcpu->arch.exception.error_code; > > - events->exception_has_payload = vcpu->arch.exception.has_payload; > > - events->exception_payload = vcpu->arch.exception.payload; > > + events->exception.nr = ex->vector; > > + events->exception.has_error_code = ex->has_error_code; > > + events->exception.error_code = ex->error_code; > > + events->exception_has_payload = ex->has_payload; > > + events->exception_payload = ex->payload; > > > > events->interrupt.injected = > > vcpu->arch.interrupt.injected && !vcpu->arch.interrupt.soft; > > @@ -5003,7 +4999,7 @@ static int kvm_vcpu_ioctl_x86_set_vcpu_events(struct kvm_vcpu *vcpu, > > process_nmi(vcpu); > > vcpu->arch.exception.injected = events->exception.injected; > > vcpu->arch.exception.pending = events->exception.pending; > > - vcpu->arch.exception.nr = events->exception.nr; > > + vcpu->arch.exception.vector = events->exception.nr; > > vcpu->arch.exception.has_error_code = events->exception.has_error_code; > > vcpu->arch.exception.error_code = events->exception.error_code; > > vcpu->arch.exception.has_payload = events->exception_has_payload; > > @@ -9497,7 +9493,7 @@ int kvm_check_nested_events(struct kvm_vcpu *vcpu) > > > > static void kvm_inject_exception(struct kvm_vcpu *vcpu) > > { > > - trace_kvm_inj_exception(vcpu->arch.exception.nr, > > + trace_kvm_inj_exception(vcpu->arch.exception.vector, > > vcpu->arch.exception.has_error_code, > > vcpu->arch.exception.error_code, > > vcpu->arch.exception.injected); > > @@ -9569,12 +9565,12 @@ static int inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit) > > * describe the behavior of General Detect #DBs, which are > > * fault-like. They do _not_ set RF, a la code breakpoints. > > */ > > - if (exception_type(vcpu->arch.exception.nr) == EXCPT_FAULT) > > + if (exception_type(vcpu->arch.exception.vector) == EXCPT_FAULT) > > __kvm_set_rflags(vcpu, kvm_get_rflags(vcpu) | > > X86_EFLAGS_RF); > > > > - if (vcpu->arch.exception.nr == DB_VECTOR) { > > - kvm_deliver_exception_payload(vcpu); > > + if (vcpu->arch.exception.vector == DB_VECTOR) { > > + kvm_deliver_exception_payload(vcpu, &vcpu->arch.exception); > > if (vcpu->arch.dr7 & DR7_GD) { > > vcpu->arch.dr7 &= ~DR7_GD; > > kvm_update_dr7(vcpu); > > diff --git a/arch/x86/kvm/x86.h b/arch/x86/kvm/x86.h > > index 501b884b8cc4..dc2af0146220 100644 > > --- a/arch/x86/kvm/x86.h > > +++ b/arch/x86/kvm/x86.h > > @@ -286,7 +286,8 @@ int kvm_write_guest_virt_system(struct kvm_vcpu *vcpu, > > > > int handle_ud(struct kvm_vcpu *vcpu); > > > > -void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu); > > +void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu, > > + struct kvm_queued_exception *ex); > > > > void kvm_vcpu_mtrr_init(struct kvm_vcpu *vcpu); > > u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn); >