From: Dapeng Mi <dapeng1.mi@xxxxxxxxxxxxxxx> Event selectors for GP counters and fixed counters control MSR are intercepted for the purpose of security, i.e., preventing guest from using unallowed events to steal information or take advantages of any CPU errata. Other than event selectors, disable PMU counter MSR interception specified in guest CPUID, counter MSR index outside of exported range will still be intercepted. Global registers like global_ctrl will passthrough only if pmu version is greater than 1. Signed-off-by: Dapeng Mi <dapeng1.mi@xxxxxxxxxxxxxxx> Signed-off-by: Xiong Zhang <xiong.y.zhang@xxxxxxxxxxxxxxx> --- arch/x86/kvm/cpuid.c | 3 +-- arch/x86/kvm/vmx/pmu_intel.c | 47 ++++++++++++++++++++++++++++++++++++ 2 files changed, 48 insertions(+), 2 deletions(-) diff --git a/arch/x86/kvm/cpuid.c b/arch/x86/kvm/cpuid.c index b577ba649feb..99e6cca67beb 100644 --- a/arch/x86/kvm/cpuid.c +++ b/arch/x86/kvm/cpuid.c @@ -382,8 +382,7 @@ static void kvm_vcpu_after_set_cpuid(struct kvm_vcpu *vcpu) kvm_pmu_refresh(vcpu); - if (is_passthrough_pmu_enabled(vcpu)) - kvm_pmu_passthrough_pmu_msrs(vcpu); + kvm_pmu_passthrough_pmu_msrs(vcpu); vcpu->arch.cr4_guest_rsvd_bits = __cr4_reserved_bits(guest_cpuid_has, vcpu); diff --git a/arch/x86/kvm/vmx/pmu_intel.c b/arch/x86/kvm/vmx/pmu_intel.c index ed79cbba1edc..8e8d1f2aa5e5 100644 --- a/arch/x86/kvm/vmx/pmu_intel.c +++ b/arch/x86/kvm/vmx/pmu_intel.c @@ -740,6 +740,52 @@ static bool intel_is_rdpmc_passthru_allowed(struct kvm_vcpu *vcpu) return true; } +/* + * Setup PMU MSR interception for both mediated passthrough vPMU and legacy + * emulated vPMU. Note that this function is called after each time userspace + * set CPUID. + */ +static void intel_passthrough_pmu_msrs(struct kvm_vcpu *vcpu) +{ + bool msr_intercept = !is_passthrough_pmu_enabled(vcpu); + struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); + int i; + + /* + * Unexposed PMU MSRs are intercepted by default. However, + * KVM_SET_CPUID{,2} may be invoked multiple times. To ensure MSR + * interception is correct after each call of setting CPUID, explicitly + * touch msr bitmap for each PMU MSR. + */ + for (i = 0; i < kvm_pmu_cap.num_counters_gp; i++) { + if (i >= pmu->nr_arch_gp_counters) + msr_intercept = true; + vmx_set_intercept_for_msr(vcpu, MSR_IA32_PERFCTR0 + i, MSR_TYPE_RW, msr_intercept); + if (fw_writes_is_enabled(vcpu)) + vmx_set_intercept_for_msr(vcpu, MSR_IA32_PMC0 + i, MSR_TYPE_RW, msr_intercept); + else + vmx_set_intercept_for_msr(vcpu, MSR_IA32_PMC0 + i, MSR_TYPE_RW, true); + } + + msr_intercept = !is_passthrough_pmu_enabled(vcpu); + for (i = 0; i < kvm_pmu_cap.num_counters_fixed; i++) { + if (i >= pmu->nr_arch_fixed_counters) + msr_intercept = true; + vmx_set_intercept_for_msr(vcpu, MSR_CORE_PERF_FIXED_CTR0 + i, MSR_TYPE_RW, msr_intercept); + } + + if (pmu->version > 1 && is_passthrough_pmu_enabled(vcpu) && + pmu->nr_arch_gp_counters == kvm_pmu_cap.num_counters_gp && + pmu->nr_arch_fixed_counters == kvm_pmu_cap.num_counters_fixed) + msr_intercept = false; + else + msr_intercept = true; + + vmx_set_intercept_for_msr(vcpu, MSR_CORE_PERF_GLOBAL_STATUS, MSR_TYPE_RW, msr_intercept); + vmx_set_intercept_for_msr(vcpu, MSR_CORE_PERF_GLOBAL_CTRL, MSR_TYPE_RW, msr_intercept); + vmx_set_intercept_for_msr(vcpu, MSR_CORE_PERF_GLOBAL_OVF_CTRL, MSR_TYPE_RW, msr_intercept); +} + struct kvm_pmu_ops intel_pmu_ops __initdata = { .rdpmc_ecx_to_pmc = intel_rdpmc_ecx_to_pmc, .msr_idx_to_pmc = intel_msr_idx_to_pmc, @@ -752,6 +798,7 @@ struct kvm_pmu_ops intel_pmu_ops __initdata = { .deliver_pmi = intel_pmu_deliver_pmi, .cleanup = intel_pmu_cleanup, .is_rdpmc_passthru_allowed = intel_is_rdpmc_passthru_allowed, + .passthrough_pmu_msrs = intel_passthrough_pmu_msrs, .EVENTSEL_EVENT = ARCH_PERFMON_EVENTSEL_EVENT, .MAX_NR_GP_COUNTERS = KVM_INTEL_PMC_MAX_GENERIC, .MIN_NR_GP_COUNTERS = 1, -- 2.45.0.rc1.225.g2a3ae87e7f-goog