Re: [PATCH v4 12/24] KVM: x86: Make kvm_queued_exception a properly named, visible struct

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Sat, 2022-07-23 at 00:51 +0000, Sean Christopherson wrote:
> Move the definition of "struct kvm_queued_exception" out of kvm_vcpu_arch
> in anticipation of adding a second instance in kvm_vcpu_arch to handle
> exceptions that occur when vectoring an injected exception and are
> morphed to VM-Exit instead of leading to #DF.
> 
> Opportunistically take advantage of the churn to rename "nr" to "vector".
> 
> No functional change intended.
> 
> Signed-off-by: Sean Christopherson <seanjc@xxxxxxxxxx>
> ---
>  arch/x86/include/asm/kvm_host.h | 23 +++++-----
>  arch/x86/kvm/svm/nested.c       | 47 ++++++++++---------
>  arch/x86/kvm/svm/svm.c          | 14 +++---
>  arch/x86/kvm/vmx/nested.c       | 42 +++++++++--------
>  arch/x86/kvm/vmx/vmx.c          | 20 ++++-----
>  arch/x86/kvm/x86.c              | 80 ++++++++++++++++-----------------
>  arch/x86/kvm/x86.h              |  3 +-
>  7 files changed, 113 insertions(+), 116 deletions(-)
> 
> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
> index dbb9eab979d4..0a6a05e25f24 100644
> --- a/arch/x86/include/asm/kvm_host.h
> +++ b/arch/x86/include/asm/kvm_host.h
> @@ -639,6 +639,17 @@ struct kvm_vcpu_xen {
>  	struct timer_list poll_timer;
>  };
>  
> +struct kvm_queued_exception {
> +	bool pending;
> +	bool injected;
> +	bool has_error_code;
> +	u8 vector;
> +	u32 error_code;
> +	unsigned long payload;
> +	bool has_payload;
> +	u8 nested_apf;
> +};
> +
>  struct kvm_vcpu_arch {
>  	/*
>  	 * rip and regs accesses must go through
> @@ -737,16 +748,8 @@ struct kvm_vcpu_arch {
>  
>  	u8 event_exit_inst_len;
>  
> -	struct kvm_queued_exception {
> -		bool pending;
> -		bool injected;
> -		bool has_error_code;
> -		u8 nr;
> -		u32 error_code;
> -		unsigned long payload;
> -		bool has_payload;
> -		u8 nested_apf;
> -	} exception;
> +	/* Exceptions to be injected to the guest. */
> +	struct kvm_queued_exception exception;
>  
>  	struct kvm_queued_interrupt {
>  		bool injected;
> diff --git a/arch/x86/kvm/svm/nested.c b/arch/x86/kvm/svm/nested.c
> index 76dcc8a3e849..8f991592d277 100644
> --- a/arch/x86/kvm/svm/nested.c
> +++ b/arch/x86/kvm/svm/nested.c
> @@ -468,7 +468,7 @@ static void nested_save_pending_event_to_vmcb12(struct vcpu_svm *svm,
>  	unsigned int nr;
>  
>  	if (vcpu->arch.exception.injected) {
> -		nr = vcpu->arch.exception.nr;
> +		nr = vcpu->arch.exception.vector;
>  		exit_int_info = nr | SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_EXEPT;
>  
>  		if (vcpu->arch.exception.has_error_code) {
> @@ -1306,42 +1306,45 @@ int nested_svm_check_permissions(struct kvm_vcpu *vcpu)
>  
>  static bool nested_exit_on_exception(struct vcpu_svm *svm)
>  {
> -	unsigned int nr = svm->vcpu.arch.exception.nr;
> +	unsigned int vector = svm->vcpu.arch.exception.vector;
>  
> -	return (svm->nested.ctl.intercepts[INTERCEPT_EXCEPTION] & BIT(nr));
> +	return (svm->nested.ctl.intercepts[INTERCEPT_EXCEPTION] & BIT(vector));
>  }
>  
> -static void nested_svm_inject_exception_vmexit(struct vcpu_svm *svm)
> +static void nested_svm_inject_exception_vmexit(struct kvm_vcpu *vcpu)
>  {
> -	unsigned int nr = svm->vcpu.arch.exception.nr;
> +	struct kvm_queued_exception *ex = &vcpu->arch.exception;
> +	struct vcpu_svm *svm = to_svm(vcpu);
>  	struct vmcb *vmcb = svm->vmcb;
>  
> -	vmcb->control.exit_code = SVM_EXIT_EXCP_BASE + nr;
> +	vmcb->control.exit_code = SVM_EXIT_EXCP_BASE + ex->vector;
>  	vmcb->control.exit_code_hi = 0;
>  
> -	if (svm->vcpu.arch.exception.has_error_code)
> -		vmcb->control.exit_info_1 = svm->vcpu.arch.exception.error_code;
> +	if (ex->has_error_code)
> +		vmcb->control.exit_info_1 = ex->error_code;
>  
>  	/*
>  	 * EXITINFO2 is undefined for all exception intercepts other
>  	 * than #PF.
>  	 */
> -	if (nr == PF_VECTOR) {
> -		if (svm->vcpu.arch.exception.nested_apf)
> -			vmcb->control.exit_info_2 = svm->vcpu.arch.apf.nested_apf_token;
> -		else if (svm->vcpu.arch.exception.has_payload)
> -			vmcb->control.exit_info_2 = svm->vcpu.arch.exception.payload;
> +	if (ex->vector == PF_VECTOR) {
> +		if (ex->nested_apf)
> +			vmcb->control.exit_info_2 = vcpu->arch.apf.nested_apf_token;
> +		else if (ex->has_payload)
> +			vmcb->control.exit_info_2 = ex->payload;
>  		else
> -			vmcb->control.exit_info_2 = svm->vcpu.arch.cr2;
> -	} else if (nr == DB_VECTOR) {
> +			vmcb->control.exit_info_2 = vcpu->arch.cr2;
> +	} else if (ex->vector == DB_VECTOR) {
>  		/* See inject_pending_event.  */
> -		kvm_deliver_exception_payload(&svm->vcpu);
> -		if (svm->vcpu.arch.dr7 & DR7_GD) {
> -			svm->vcpu.arch.dr7 &= ~DR7_GD;
> -			kvm_update_dr7(&svm->vcpu);
> +		kvm_deliver_exception_payload(vcpu, ex);
> +
> +		if (vcpu->arch.dr7 & DR7_GD) {
> +			vcpu->arch.dr7 &= ~DR7_GD;
> +			kvm_update_dr7(vcpu);
>  		}
> -	} else
> -		WARN_ON(svm->vcpu.arch.exception.has_payload);
> +	} else {
> +		WARN_ON(ex->has_payload);
> +	}
>  
>  	nested_svm_vmexit(svm);
>  }
> @@ -1379,7 +1382,7 @@ static int svm_check_nested_events(struct kvm_vcpu *vcpu)
>                          return -EBUSY;
>  		if (!nested_exit_on_exception(svm))
>  			return 0;
> -		nested_svm_inject_exception_vmexit(svm);
> +		nested_svm_inject_exception_vmexit(vcpu);
>  		return 0;
>  	}
>  
> diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c
> index e73d79ae0e45..74cbe177e0d1 100644
> --- a/arch/x86/kvm/svm/svm.c
> +++ b/arch/x86/kvm/svm/svm.c
> @@ -456,22 +456,20 @@ static int svm_update_soft_interrupt_rip(struct kvm_vcpu *vcpu)
>  
>  static void svm_inject_exception(struct kvm_vcpu *vcpu)
>  {
> +	struct kvm_queued_exception *ex = &vcpu->arch.exception;
>  	struct vcpu_svm *svm = to_svm(vcpu);
> -	unsigned nr = vcpu->arch.exception.nr;
> -	bool has_error_code = vcpu->arch.exception.has_error_code;
> -	u32 error_code = vcpu->arch.exception.error_code;
>  
> -	kvm_deliver_exception_payload(vcpu);
> +	kvm_deliver_exception_payload(vcpu, ex);
>  
> -	if (kvm_exception_is_soft(nr) &&
> +	if (kvm_exception_is_soft(ex->vector) &&
>  	    svm_update_soft_interrupt_rip(vcpu))
>  		return;
>  
> -	svm->vmcb->control.event_inj = nr
> +	svm->vmcb->control.event_inj = ex->vector
>  		| SVM_EVTINJ_VALID
> -		| (has_error_code ? SVM_EVTINJ_VALID_ERR : 0)
> +		| (ex->has_error_code ? SVM_EVTINJ_VALID_ERR : 0)
>  		| SVM_EVTINJ_TYPE_EXEPT;
> -	svm->vmcb->control.event_inj_err = error_code;
> +	svm->vmcb->control.event_inj_err = ex->error_code;
>  }
>  
>  static void svm_init_erratum_383(void)
> diff --git a/arch/x86/kvm/vmx/nested.c b/arch/x86/kvm/vmx/nested.c
> index 17df0c31f0b5..0f5a7aec82a2 100644
> --- a/arch/x86/kvm/vmx/nested.c
> +++ b/arch/x86/kvm/vmx/nested.c
> @@ -446,29 +446,27 @@ static bool nested_vmx_is_page_fault_vmexit(struct vmcs12 *vmcs12,
>   */
>  static int nested_vmx_check_exception(struct kvm_vcpu *vcpu, unsigned long *exit_qual)
>  {
> +	struct kvm_queued_exception *ex = &vcpu->arch.exception;
>  	struct vmcs12 *vmcs12 = get_vmcs12(vcpu);
> -	unsigned int nr = vcpu->arch.exception.nr;
> -	bool has_payload = vcpu->arch.exception.has_payload;
> -	unsigned long payload = vcpu->arch.exception.payload;
>  
> -	if (nr == PF_VECTOR) {
> -		if (vcpu->arch.exception.nested_apf) {
> +	if (ex->vector == PF_VECTOR) {
> +		if (ex->nested_apf) {
>  			*exit_qual = vcpu->arch.apf.nested_apf_token;
>  			return 1;
>  		}
> -		if (nested_vmx_is_page_fault_vmexit(vmcs12,
> -						    vcpu->arch.exception.error_code)) {
> -			*exit_qual = has_payload ? payload : vcpu->arch.cr2;
> +		if (nested_vmx_is_page_fault_vmexit(vmcs12, ex->error_code)) {
> +			*exit_qual = ex->has_payload ? ex->payload : vcpu->arch.cr2;
>  			return 1;
>  		}
> -	} else if (vmcs12->exception_bitmap & (1u << nr)) {
> -		if (nr == DB_VECTOR) {
> -			if (!has_payload) {
> -				payload = vcpu->arch.dr6;
> -				payload &= ~DR6_BT;
> -				payload ^= DR6_ACTIVE_LOW;
> +	} else if (vmcs12->exception_bitmap & (1u << ex->vector)) {
> +		if (ex->vector == DB_VECTOR) {
> +			if (ex->has_payload) {
> +				*exit_qual = ex->payload;
> +			} else {
> +				*exit_qual = vcpu->arch.dr6;
> +				*exit_qual &= ~DR6_BT;
> +				*exit_qual ^= DR6_ACTIVE_LOW;
>  			}
> -			*exit_qual = payload;
>  		} else
>  			*exit_qual = 0;
>  		return 1;
> @@ -3718,7 +3716,7 @@ static void vmcs12_save_pending_event(struct kvm_vcpu *vcpu,
>  	     is_double_fault(exit_intr_info))) {
>  		vmcs12->idt_vectoring_info_field = 0;
>  	} else if (vcpu->arch.exception.injected) {
> -		nr = vcpu->arch.exception.nr;
> +		nr = vcpu->arch.exception.vector;
>  		idt_vectoring = nr | VECTORING_INFO_VALID_MASK;
>  
>  		if (kvm_exception_is_soft(nr)) {
> @@ -3822,11 +3820,11 @@ static int vmx_complete_nested_posted_interrupt(struct kvm_vcpu *vcpu)
>  static void nested_vmx_inject_exception_vmexit(struct kvm_vcpu *vcpu,
>  					       unsigned long exit_qual)
>  {
> +	struct kvm_queued_exception *ex = &vcpu->arch.exception;
> +	u32 intr_info = ex->vector | INTR_INFO_VALID_MASK;
>  	struct vmcs12 *vmcs12 = get_vmcs12(vcpu);
> -	unsigned int nr = vcpu->arch.exception.nr;
> -	u32 intr_info = nr | INTR_INFO_VALID_MASK;
>  
> -	if (vcpu->arch.exception.has_error_code) {
> +	if (ex->has_error_code) {
>  		/*
>  		 * Intel CPUs do not generate error codes with bits 31:16 set,
>  		 * and more importantly VMX disallows setting bits 31:16 in the
> @@ -3836,11 +3834,11 @@ static void nested_vmx_inject_exception_vmexit(struct kvm_vcpu *vcpu,
>  		 * generate "full" 32-bit error codes, so KVM allows userspace
>  		 * to inject exception error codes with bits 31:16 set.
>  		 */
> -		vmcs12->vm_exit_intr_error_code = (u16)vcpu->arch.exception.error_code;
> +		vmcs12->vm_exit_intr_error_code = (u16)ex->error_code;
>  		intr_info |= INTR_INFO_DELIVER_CODE_MASK;
>  	}
>  
> -	if (kvm_exception_is_soft(nr))
> +	if (kvm_exception_is_soft(ex->vector))
>  		intr_info |= INTR_TYPE_SOFT_EXCEPTION;
>  	else
>  		intr_info |= INTR_TYPE_HARD_EXCEPTION;
> @@ -3871,7 +3869,7 @@ static void nested_vmx_inject_exception_vmexit(struct kvm_vcpu *vcpu,
>  static inline unsigned long vmx_get_pending_dbg_trap(struct kvm_vcpu *vcpu)
>  {
>  	if (!vcpu->arch.exception.pending ||
> -	    vcpu->arch.exception.nr != DB_VECTOR)
> +	    vcpu->arch.exception.vector != DB_VECTOR)
>  		return 0;
>  
>  	/* General Detect #DBs are always fault-like. */
> diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c
> index 14f75e4003d3..4cfe6646476b 100644
> --- a/arch/x86/kvm/vmx/vmx.c
> +++ b/arch/x86/kvm/vmx/vmx.c
> @@ -1585,7 +1585,7 @@ static void vmx_update_emulated_instruction(struct kvm_vcpu *vcpu)
>  	 */
>  	if (nested_cpu_has_mtf(vmcs12) &&
>  	    (!vcpu->arch.exception.pending ||
> -	     vcpu->arch.exception.nr == DB_VECTOR))
> +	     vcpu->arch.exception.vector == DB_VECTOR))
>  		vmx->nested.mtf_pending = true;
>  	else
>  		vmx->nested.mtf_pending = false;
> @@ -1612,15 +1612,13 @@ static void vmx_clear_hlt(struct kvm_vcpu *vcpu)
>  
>  static void vmx_inject_exception(struct kvm_vcpu *vcpu)
>  {
> +	struct kvm_queued_exception *ex = &vcpu->arch.exception;
> +	u32 intr_info = ex->vector | INTR_INFO_VALID_MASK;
>  	struct vcpu_vmx *vmx = to_vmx(vcpu);
> -	unsigned nr = vcpu->arch.exception.nr;
> -	bool has_error_code = vcpu->arch.exception.has_error_code;
> -	u32 error_code = vcpu->arch.exception.error_code;
> -	u32 intr_info = nr | INTR_INFO_VALID_MASK;
>  
> -	kvm_deliver_exception_payload(vcpu);
> +	kvm_deliver_exception_payload(vcpu, ex);
>  
> -	if (has_error_code) {
> +	if (ex->has_error_code) {
>  		/*
>  		 * Despite the error code being architecturally defined as 32
>  		 * bits, and the VMCS field being 32 bits, Intel CPUs and thus
> @@ -1631,21 +1629,21 @@ static void vmx_inject_exception(struct kvm_vcpu *vcpu)
>  		 * the upper bits to avoid VM-Fail, losing information that
>  		 * does't really exist is preferable to killing the VM.
>  		 */
> -		vmcs_write32(VM_ENTRY_EXCEPTION_ERROR_CODE, (u16)error_code);
> +		vmcs_write32(VM_ENTRY_EXCEPTION_ERROR_CODE, (u16)ex->error_code);
>  		intr_info |= INTR_INFO_DELIVER_CODE_MASK;
>  	}
>  
>  	if (vmx->rmode.vm86_active) {
>  		int inc_eip = 0;
> -		if (kvm_exception_is_soft(nr))
> +		if (kvm_exception_is_soft(ex->vector))
>  			inc_eip = vcpu->arch.event_exit_inst_len;
> -		kvm_inject_realmode_interrupt(vcpu, nr, inc_eip);
> +		kvm_inject_realmode_interrupt(vcpu, ex->vector, inc_eip);
>  		return;
>  	}
>  
>  	WARN_ON_ONCE(vmx->emulation_required);
>  
> -	if (kvm_exception_is_soft(nr)) {
> +	if (kvm_exception_is_soft(ex->vector)) {
>  		vmcs_write32(VM_ENTRY_INSTRUCTION_LEN,
>  			     vmx->vcpu.arch.event_exit_inst_len);
>  		intr_info |= INTR_TYPE_SOFT_EXCEPTION;
> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> index a61b6cbd7194..027fc518ba75 100644
> --- a/arch/x86/kvm/x86.c
> +++ b/arch/x86/kvm/x86.c
> @@ -557,16 +557,13 @@ static int exception_type(int vector)
>  	return EXCPT_FAULT;
>  }
>  
> -void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu)
> +void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu,
> +				   struct kvm_queued_exception *ex)
>  {
> -	unsigned nr = vcpu->arch.exception.nr;
> -	bool has_payload = vcpu->arch.exception.has_payload;
> -	unsigned long payload = vcpu->arch.exception.payload;
> -
> -	if (!has_payload)
> +	if (!ex->has_payload)
>  		return;
>  
> -	switch (nr) {
> +	switch (ex->vector) {
>  	case DB_VECTOR:
>  		/*
>  		 * "Certain debug exceptions may clear bit 0-3.  The
> @@ -591,8 +588,8 @@ void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu)
>  		 * So they need to be flipped for DR6.
>  		 */
>  		vcpu->arch.dr6 |= DR6_ACTIVE_LOW;
> -		vcpu->arch.dr6 |= payload;
> -		vcpu->arch.dr6 ^= payload & DR6_ACTIVE_LOW;
> +		vcpu->arch.dr6 |= ex->payload;
> +		vcpu->arch.dr6 ^= ex->payload & DR6_ACTIVE_LOW;
>  
>  		/*
>  		 * The #DB payload is defined as compatible with the 'pending
> @@ -603,12 +600,12 @@ void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu)
>  		vcpu->arch.dr6 &= ~BIT(12);
>  		break;
>  	case PF_VECTOR:
> -		vcpu->arch.cr2 = payload;
> +		vcpu->arch.cr2 = ex->payload;
>  		break;
>  	}
>  
> -	vcpu->arch.exception.has_payload = false;
> -	vcpu->arch.exception.payload = 0;
> +	ex->has_payload = false;
> +	ex->payload = 0;
>  }
>  EXPORT_SYMBOL_GPL(kvm_deliver_exception_payload);
>  
> @@ -647,17 +644,18 @@ static void kvm_multiple_exception(struct kvm_vcpu *vcpu,
>  			vcpu->arch.exception.injected = false;
>  		}
>  		vcpu->arch.exception.has_error_code = has_error;
> -		vcpu->arch.exception.nr = nr;
> +		vcpu->arch.exception.vector = nr;
>  		vcpu->arch.exception.error_code = error_code;
>  		vcpu->arch.exception.has_payload = has_payload;
>  		vcpu->arch.exception.payload = payload;
>  		if (!is_guest_mode(vcpu))
> -			kvm_deliver_exception_payload(vcpu);
> +			kvm_deliver_exception_payload(vcpu,
> +						      &vcpu->arch.exception);
>  		return;
>  	}
>  
>  	/* to check exception */
> -	prev_nr = vcpu->arch.exception.nr;
> +	prev_nr = vcpu->arch.exception.vector;
>  	if (prev_nr == DF_VECTOR) {
>  		/* triple fault -> shutdown */
>  		kvm_make_request(KVM_REQ_TRIPLE_FAULT, vcpu);
> @@ -675,7 +673,7 @@ static void kvm_multiple_exception(struct kvm_vcpu *vcpu,
>  		vcpu->arch.exception.pending = true;
>  		vcpu->arch.exception.injected = false;
>  		vcpu->arch.exception.has_error_code = true;
> -		vcpu->arch.exception.nr = DF_VECTOR;
> +		vcpu->arch.exception.vector = DF_VECTOR;
>  		vcpu->arch.exception.error_code = 0;
>  		vcpu->arch.exception.has_payload = false;
>  		vcpu->arch.exception.payload = 0;
> @@ -5006,25 +5004,24 @@ static int kvm_vcpu_ioctl_x86_set_mce(struct kvm_vcpu *vcpu,
>  static void kvm_vcpu_ioctl_x86_get_vcpu_events(struct kvm_vcpu *vcpu,
>  					       struct kvm_vcpu_events *events)
>  {
> +	struct kvm_queued_exception *ex = &vcpu->arch.exception;
> +
>  	process_nmi(vcpu);
>  
>  	if (kvm_check_request(KVM_REQ_SMI, vcpu))
>  		process_smi(vcpu);
>  
>  	/*
> -	 * In guest mode, payload delivery should be deferred,
> -	 * so that the L1 hypervisor can intercept #PF before
> -	 * CR2 is modified (or intercept #DB before DR6 is
> -	 * modified under nVMX). Unless the per-VM capability,
> -	 * KVM_CAP_EXCEPTION_PAYLOAD, is set, we may not defer the delivery of
> -	 * an exception payload and handle after a KVM_GET_VCPU_EVENTS. Since we
> -	 * opportunistically defer the exception payload, deliver it if the
> -	 * capability hasn't been requested before processing a
> -	 * KVM_GET_VCPU_EVENTS.
> +	 * In guest mode, payload delivery should be deferred if the exception
> +	 * will be intercepted by L1, e.g. KVM should not modifying CR2 if L1
> +	 * intercepts #PF, ditto for DR6 and #DBs.  If the per-VM capability,
> +	 * KVM_CAP_EXCEPTION_PAYLOAD, is not set, userspace may or may not
> +	 * propagate the payload and so it cannot be safely deferred.  Deliver
> +	 * the payload if the capability hasn't been requested.
>  	 */
>  	if (!vcpu->kvm->arch.exception_payload_enabled &&
> -	    vcpu->arch.exception.pending && vcpu->arch.exception.has_payload)
> -		kvm_deliver_exception_payload(vcpu);
> +	    ex->pending && ex->has_payload)
> +		kvm_deliver_exception_payload(vcpu, ex);
>  
>  	/*
>  	 * The API doesn't provide the instruction length for software
> @@ -5032,26 +5029,25 @@ static void kvm_vcpu_ioctl_x86_get_vcpu_events(struct kvm_vcpu *vcpu,
>  	 * isn't advanced, we should expect to encounter the exception
>  	 * again.
>  	 */
> -	if (kvm_exception_is_soft(vcpu->arch.exception.nr)) {
> +	if (kvm_exception_is_soft(ex->vector)) {
>  		events->exception.injected = 0;
>  		events->exception.pending = 0;
>  	} else {
> -		events->exception.injected = vcpu->arch.exception.injected;
> -		events->exception.pending = vcpu->arch.exception.pending;
> +		events->exception.injected = ex->injected;
> +		events->exception.pending = ex->pending;
>  		/*
>  		 * For ABI compatibility, deliberately conflate
>  		 * pending and injected exceptions when
>  		 * KVM_CAP_EXCEPTION_PAYLOAD isn't enabled.
>  		 */
>  		if (!vcpu->kvm->arch.exception_payload_enabled)
> -			events->exception.injected |=
> -				vcpu->arch.exception.pending;
> +			events->exception.injected |= ex->pending;
>  	}
> -	events->exception.nr = vcpu->arch.exception.nr;
> -	events->exception.has_error_code = vcpu->arch.exception.has_error_code;
> -	events->exception.error_code = vcpu->arch.exception.error_code;
> -	events->exception_has_payload = vcpu->arch.exception.has_payload;
> -	events->exception_payload = vcpu->arch.exception.payload;
> +	events->exception.nr = ex->vector;
> +	events->exception.has_error_code = ex->has_error_code;
> +	events->exception.error_code = ex->error_code;
> +	events->exception_has_payload = ex->has_payload;
> +	events->exception_payload = ex->payload;
>  
>  	events->interrupt.injected =
>  		vcpu->arch.interrupt.injected && !vcpu->arch.interrupt.soft;
> @@ -5123,7 +5119,7 @@ static int kvm_vcpu_ioctl_x86_set_vcpu_events(struct kvm_vcpu *vcpu,
>  	process_nmi(vcpu);
>  	vcpu->arch.exception.injected = events->exception.injected;
>  	vcpu->arch.exception.pending = events->exception.pending;
> -	vcpu->arch.exception.nr = events->exception.nr;
> +	vcpu->arch.exception.vector = events->exception.nr;
>  	vcpu->arch.exception.has_error_code = events->exception.has_error_code;
>  	vcpu->arch.exception.error_code = events->exception.error_code;
>  	vcpu->arch.exception.has_payload = events->exception_has_payload;
> @@ -9665,7 +9661,7 @@ int kvm_check_nested_events(struct kvm_vcpu *vcpu)
>  
>  static void kvm_inject_exception(struct kvm_vcpu *vcpu)
>  {
> -	trace_kvm_inj_exception(vcpu->arch.exception.nr,
> +	trace_kvm_inj_exception(vcpu->arch.exception.vector,
>  				vcpu->arch.exception.has_error_code,
>  				vcpu->arch.exception.error_code,
>  				vcpu->arch.exception.injected);
> @@ -9737,12 +9733,12 @@ static int inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit)
>  		 * describe the behavior of General Detect #DBs, which are
>  		 * fault-like.  They do _not_ set RF, a la code breakpoints.
>  		 */
> -		if (exception_type(vcpu->arch.exception.nr) == EXCPT_FAULT)
> +		if (exception_type(vcpu->arch.exception.vector) == EXCPT_FAULT)
>  			__kvm_set_rflags(vcpu, kvm_get_rflags(vcpu) |
>  					     X86_EFLAGS_RF);
>  
> -		if (vcpu->arch.exception.nr == DB_VECTOR) {
> -			kvm_deliver_exception_payload(vcpu);
> +		if (vcpu->arch.exception.vector == DB_VECTOR) {
> +			kvm_deliver_exception_payload(vcpu, &vcpu->arch.exception);
>  			if (vcpu->arch.dr7 & DR7_GD) {
>  				vcpu->arch.dr7 &= ~DR7_GD;
>  				kvm_update_dr7(vcpu);
> diff --git a/arch/x86/kvm/x86.h b/arch/x86/kvm/x86.h
> index 1926d2cb8e79..4147d27f9fbc 100644
> --- a/arch/x86/kvm/x86.h
> +++ b/arch/x86/kvm/x86.h
> @@ -286,7 +286,8 @@ int kvm_write_guest_virt_system(struct kvm_vcpu *vcpu,
>  
>  int handle_ud(struct kvm_vcpu *vcpu);
>  
> -void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu);
> +void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu,
> +				   struct kvm_queued_exception *ex);
>  
>  void kvm_vcpu_mtrr_init(struct kvm_vcpu *vcpu);
>  u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn);


To be honest, it is still a bit hard to review this patch because it does too much things
at once.

Anyway, while I can't be 100% sure I didn't miss anything, I haven't found any issues either, so:

Reviewed-by: Maxim Levitsky <mlevitsk@xxxxxxxxxx>

Best regards,
	Maxim Levitsky







[Index of Archives]     [KVM ARM]     [KVM ia64]     [KVM ppc]     [Virtualization Tools]     [Spice Development]     [Libvirt]     [Libvirt Users]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite Questions]     [Linux Kernel]     [Linux SCSI]     [XFree86]

  Powered by Linux