Re: [PATCH v2 2/2] KVM: x86: hyperv: optimize sparse VP set processing

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 10/10/2018 17:14, Vitaly Kuznetsov wrote:
> Rewrite kvm_hv_flush_tlb()/send_ipi_vcpus_mask() making them cleaner and
> somewhat more optimal.
> 
> hv_vcpu_in_sparse_set() is converted to sparse_set_to_vcpu_mask()
> which copies sparse banks u64-at-a-time and then, depending on the
> num_mismatched_vp_indexes value, returns immediately or does
> vp index to vcpu index conversion by walking all vCPUs.
> 
> To support the change and make kvm_hv_send_ipi() look similar to
> kvm_hv_flush_tlb() send_ipi_vcpus_mask() is introduced.
> 
> Suggested-by: Roman Kagan <rkagan@xxxxxxxxxxxxx>
> Signed-off-by: Vitaly Kuznetsov <vkuznets@xxxxxxxxxx>
> Reviewed-by: Roman Kagan <rkagan@xxxxxxxxxxxxx>
> ---
>  arch/x86/kvm/hyperv.c | 163 +++++++++++++++++-------------------------
>  1 file changed, 65 insertions(+), 98 deletions(-)
> 
> diff --git a/arch/x86/kvm/hyperv.c b/arch/x86/kvm/hyperv.c
> index f1ae1777519c..5e74ba6a1e7a 100644
> --- a/arch/x86/kvm/hyperv.c
> +++ b/arch/x86/kvm/hyperv.c
> @@ -36,6 +36,8 @@
>  
>  #include "trace.h"
>  
> +#define KVM_HV_MAX_SPARSE_VCPU_SET_BITS DIV_ROUND_UP(KVM_MAX_VCPUS, 64)
> +
>  static inline u64 synic_read_sint(struct kvm_vcpu_hv_synic *synic, int sint)
>  {
>  	return atomic64_read(&synic->sint[sint]);
> @@ -1277,37 +1279,47 @@ int kvm_hv_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host)
>  		return kvm_hv_get_msr(vcpu, msr, pdata, host);
>  }
>  
> -static __always_inline bool hv_vcpu_in_sparse_set(struct kvm_vcpu_hv *hv_vcpu,
> -						  u64 sparse_banks[],
> -						  u64 valid_bank_mask)
> +static __always_inline unsigned long *sparse_set_to_vcpu_mask(
> +	struct kvm *kvm, u64 *sparse_banks, u64 valid_bank_mask,
> +	u64 *vp_bitmap, unsigned long *vcpu_bitmap)
>  {
> -	int bank = hv_vcpu->vp_index / 64, sbank;
> -
> -	if (bank >= 64)
> -		return false;
> +	struct kvm_hv *hv = &kvm->arch.hyperv;
> +	struct kvm_vcpu *vcpu;
> +	int i, bank, sbank = 0;
>  
> -	if (!(valid_bank_mask & BIT_ULL(bank)))
> -		return false;
> +	memset(vp_bitmap, 0,
> +	       KVM_HV_MAX_SPARSE_VCPU_SET_BITS * sizeof(*vp_bitmap));
> +	for_each_set_bit(bank, (unsigned long *)&valid_bank_mask,
> +			 KVM_HV_MAX_SPARSE_VCPU_SET_BITS)
> +		vp_bitmap[bank] = sparse_banks[sbank++];
>  
> -	/* Sparse bank number equals to the number of set bits before it */
> -	sbank = bitmap_weight((unsigned long *)&valid_bank_mask, bank);
> +	if (likely(!atomic_read(&hv->num_mismatched_vp_indexes))) {
> +		/* for all vcpus vp_index == vcpu_idx */
> +		return (unsigned long *)vp_bitmap;
> +	}
>  
> -	return !!(sparse_banks[sbank] & BIT_ULL(hv_vcpu->vp_index % 64));
> +	bitmap_zero(vcpu_bitmap, KVM_MAX_VCPUS);
> +	kvm_for_each_vcpu(i, vcpu, kvm) {
> +		if (test_bit(vcpu_to_hv_vcpu(vcpu)->vp_index,
> +			     (unsigned long *)vp_bitmap))
> +			__set_bit(i, vcpu_bitmap);
> +	}
> +	return vcpu_bitmap;
>  }
>  
>  static u64 kvm_hv_flush_tlb(struct kvm_vcpu *current_vcpu, u64 ingpa,
>  			    u16 rep_cnt, bool ex)
>  {
>  	struct kvm *kvm = current_vcpu->kvm;
> -	struct kvm_hv *hv = &kvm->arch.hyperv;
>  	struct kvm_vcpu_hv *hv_vcpu = &current_vcpu->arch.hyperv;
>  	struct hv_tlb_flush_ex flush_ex;
>  	struct hv_tlb_flush flush;
> -	struct kvm_vcpu *vcpu;
> -	unsigned long vcpu_bitmap[BITS_TO_LONGS(KVM_MAX_VCPUS)] = {0};
> +	u64 vp_bitmap[KVM_HV_MAX_SPARSE_VCPU_SET_BITS];
> +	DECLARE_BITMAP(vcpu_bitmap, KVM_MAX_VCPUS);
> +	unsigned long *vcpu_mask;
>  	u64 valid_bank_mask;
>  	u64 sparse_banks[64];
> -	int sparse_banks_len, i, bank, sbank;
> +	int sparse_banks_len;
>  	bool all_cpus;
>  
>  	if (!ex) {
> @@ -1350,54 +1362,19 @@ static u64 kvm_hv_flush_tlb(struct kvm_vcpu *current_vcpu, u64 ingpa,
>  			return HV_STATUS_INVALID_HYPERCALL_INPUT;
>  	}
>  
> -	/*
> -	 * vcpu->arch.cr3 may not be up-to-date for running vCPUs so we can't
> -	 * analyze it here, flush TLB regardless of the specified address space.
> -	 */
>  	cpumask_clear(&hv_vcpu->tlb_flush);
>  
> -	if (all_cpus) {
> -		kvm_make_vcpus_request_mask(kvm,
> -				    KVM_REQ_TLB_FLUSH | KVM_REQUEST_NO_WAKEUP,
> -				    NULL, &hv_vcpu->tlb_flush);
> -		goto ret_success;
> -	}
> -
> -	if (atomic_read(&hv->num_mismatched_vp_indexes)) {
> -		kvm_for_each_vcpu(i, vcpu, kvm) {
> -			if (hv_vcpu_in_sparse_set(&vcpu->arch.hyperv,
> -						  sparse_banks,
> -						  valid_bank_mask))
> -				__set_bit(i, vcpu_bitmap);
> -		}
> -		goto flush_request;
> -	}
> +	vcpu_mask = all_cpus ? NULL :
> +		sparse_set_to_vcpu_mask(kvm, sparse_banks, valid_bank_mask,
> +					vp_bitmap, vcpu_bitmap);
>  
>  	/*
> -	 * num_mismatched_vp_indexes is zero so every vcpu has
> -	 * vp_index == vcpu_idx.
> +	 * vcpu->arch.cr3 may not be up-to-date for running vCPUs so we can't
> +	 * analyze it here, flush TLB regardless of the specified address space.
>  	 */
> -	sbank = 0;
> -	for_each_set_bit(bank, (unsigned long *)&valid_bank_mask,
> -			 BITS_PER_LONG) {
> -		for_each_set_bit(i,
> -				 (unsigned long *)&sparse_banks[sbank],
> -				 BITS_PER_LONG) {
> -			u32 vp_index = bank * 64 + i;
> -
> -			/* A non-existent vCPU was specified */
> -			if (vp_index >= KVM_MAX_VCPUS)
> -				return HV_STATUS_INVALID_HYPERCALL_INPUT;
> -
> -			__set_bit(vp_index, vcpu_bitmap);
> -		}
> -		sbank++;
> -	}
> -
> -flush_request:
>  	kvm_make_vcpus_request_mask(kvm,
>  				    KVM_REQ_TLB_FLUSH | KVM_REQUEST_NO_WAKEUP,
> -				    vcpu_bitmap, &hv_vcpu->tlb_flush);
> +				    vcpu_mask, &hv_vcpu->tlb_flush);
>  
>  ret_success:
>  	/* We always do full TLB flush, set rep_done = rep_cnt. */
> @@ -1405,18 +1382,36 @@ static u64 kvm_hv_flush_tlb(struct kvm_vcpu *current_vcpu, u64 ingpa,
>  		((u64)rep_cnt << HV_HYPERCALL_REP_COMP_OFFSET);
>  }
>  
> +static void send_ipi_vcpus_mask(struct kvm *kvm, u32 vector,
> +				unsigned long *vcpu_bitmap)
> +{
> +	struct kvm_lapic_irq irq = {.delivery_mode = APIC_DM_FIXED,
> +				    .vector = vector};
> +	struct kvm_vcpu *vcpu;
> +	int i;
> +
> +	kvm_for_each_vcpu(i, vcpu, kvm) {
> +		if (vcpu_bitmap && !test_bit(i, vcpu_bitmap))
> +			continue;
> +
> +		/* We fail only when APIC is disabled */
> +		kvm_apic_set_irq(vcpu, &irq, NULL);
> +	}
> +}
> +
>  static u64 kvm_hv_send_ipi(struct kvm_vcpu *current_vcpu, u64 ingpa, u64 outgpa,
>  			   bool ex, bool fast)
>  {
>  	struct kvm *kvm = current_vcpu->kvm;
> -	struct kvm_hv *hv = &kvm->arch.hyperv;
>  	struct hv_send_ipi_ex send_ipi_ex;
>  	struct hv_send_ipi send_ipi;
> -	struct kvm_vcpu *vcpu;
> +	u64 vp_bitmap[KVM_HV_MAX_SPARSE_VCPU_SET_BITS];
> +	DECLARE_BITMAP(vcpu_bitmap, KVM_MAX_VCPUS);
> +	unsigned long *vcpu_mask;
>  	unsigned long valid_bank_mask;
>  	u64 sparse_banks[64];
> -	int sparse_banks_len, bank, i, sbank;
> -	struct kvm_lapic_irq irq = {.delivery_mode = APIC_DM_FIXED};
> +	int sparse_banks_len;
> +	u32 vector;
>  	bool all_cpus;
>  
>  	if (!ex) {
> @@ -1425,18 +1420,18 @@ static u64 kvm_hv_send_ipi(struct kvm_vcpu *current_vcpu, u64 ingpa, u64 outgpa,
>  						    sizeof(send_ipi))))
>  				return HV_STATUS_INVALID_HYPERCALL_INPUT;
>  			sparse_banks[0] = send_ipi.cpu_mask;
> -			irq.vector = send_ipi.vector;
> +			vector = send_ipi.vector;
>  		} else {
>  			/* 'reserved' part of hv_send_ipi should be 0 */
>  			if (unlikely(ingpa >> 32 != 0))
>  				return HV_STATUS_INVALID_HYPERCALL_INPUT;
>  			sparse_banks[0] = outgpa;
> -			irq.vector = (u32)ingpa;
> +			vector = (u32)ingpa;
>  		}
>  		all_cpus = false;
>  		valid_bank_mask = BIT_ULL(0);
>  
> -		trace_kvm_hv_send_ipi(irq.vector, sparse_banks[0]);
> +		trace_kvm_hv_send_ipi(vector, sparse_banks[0]);
>  	} else {
>  		if (unlikely(kvm_read_guest(kvm, ingpa, &send_ipi_ex,
>  					    sizeof(send_ipi_ex))))
> @@ -1446,7 +1441,7 @@ static u64 kvm_hv_send_ipi(struct kvm_vcpu *current_vcpu, u64 ingpa, u64 outgpa,
>  					 send_ipi_ex.vp_set.format,
>  					 send_ipi_ex.vp_set.valid_bank_mask);
>  
> -		irq.vector = send_ipi_ex.vector;
> +		vector = send_ipi_ex.vector;
>  		valid_bank_mask = send_ipi_ex.vp_set.valid_bank_mask;
>  		sparse_banks_len = bitmap_weight(&valid_bank_mask, 64) *
>  			sizeof(sparse_banks[0]);
> @@ -1465,42 +1460,14 @@ static u64 kvm_hv_send_ipi(struct kvm_vcpu *current_vcpu, u64 ingpa, u64 outgpa,
>  			return HV_STATUS_INVALID_HYPERCALL_INPUT;
>  	}
>  
> -	if ((irq.vector < HV_IPI_LOW_VECTOR) ||
> -	    (irq.vector > HV_IPI_HIGH_VECTOR))
> +	if ((vector < HV_IPI_LOW_VECTOR) || (vector > HV_IPI_HIGH_VECTOR))
>  		return HV_STATUS_INVALID_HYPERCALL_INPUT;
>  
> -	if (all_cpus || atomic_read(&hv->num_mismatched_vp_indexes)) {
> -		kvm_for_each_vcpu(i, vcpu, kvm) {
> -			if (all_cpus || hv_vcpu_in_sparse_set(
> -				    &vcpu->arch.hyperv, sparse_banks,
> -				    valid_bank_mask)) {
> -				/* We fail only when APIC is disabled */
> -				kvm_apic_set_irq(vcpu, &irq, NULL);
> -			}
> -		}
> -		goto ret_success;
> -	}
> +	vcpu_mask = all_cpus ? NULL :
> +		sparse_set_to_vcpu_mask(kvm, sparse_banks, valid_bank_mask,
> +					vp_bitmap, vcpu_bitmap);
>  
> -	/*
> -	 * num_mismatched_vp_indexes is zero so every vcpu has
> -	 * vp_index == vcpu_idx.
> -	 */
> -	sbank = 0;
> -	for_each_set_bit(bank, (unsigned long *)&valid_bank_mask, 64) {
> -		for_each_set_bit(i, (unsigned long *)&sparse_banks[sbank], 64) {
> -			u32 vp_index = bank * 64 + i;
> -			struct kvm_vcpu *vcpu =
> -				get_vcpu_by_vpidx(kvm, vp_index);
> -
> -			/* Unknown vCPU specified */
> -			if (!vcpu)
> -				continue;
> -
> -			/* We fail only when APIC is disabled */
> -			kvm_apic_set_irq(vcpu, &irq, NULL);
> -		}
> -		sbank++;
> -	}
> +	send_ipi_vcpus_mask(kvm, vector, vcpu_mask);
>  
>  ret_success:
>  	return HV_STATUS_SUCCESS;
> 

Queued, thanks.  I renamed send_ipi_vcpus_mask to kvm_send_ipi_to_many,
just because. :)

Paolo



[Index of Archives]     [KVM ARM]     [KVM ia64]     [KVM ppc]     [Virtualization Tools]     [Spice Development]     [Libvirt]     [Libvirt Users]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite Questions]     [Linux Kernel]     [Linux SCSI]     [XFree86]

  Powered by Linux