Sean Christopherson <sean.j.christopherson@xxxxxxxxx> writes: > Move VMX's hardware_setup() below its vmx_x86_ops definition so that a > future patch can refactor hardware_setup() to modify vmx_x86_ops > directly instead of indirectly modifying the ops via the global > kvm_x86_ops. > > No functional change intended. > > Signed-off-by: Sean Christopherson <sean.j.christopherson@xxxxxxxxx> > --- > arch/x86/kvm/vmx/vmx.c | 346 ++++++++++++++++++++--------------------- > 1 file changed, 173 insertions(+), 173 deletions(-) > > diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c > index ffcdcc86f5b7..82dab775d520 100644 > --- a/arch/x86/kvm/vmx/vmx.c > +++ b/arch/x86/kvm/vmx/vmx.c > @@ -7652,179 +7652,6 @@ static bool vmx_apic_init_signal_blocked(struct kvm_vcpu *vcpu) > return to_vmx(vcpu)->nested.vmxon; > } > > -static __init int hardware_setup(void) > -{ > - unsigned long host_bndcfgs; > - struct desc_ptr dt; > - int r, i, ept_lpage_level; > - > - store_idt(&dt); > - host_idt_base = dt.address; > - > - for (i = 0; i < ARRAY_SIZE(vmx_msr_index); ++i) > - kvm_define_shared_msr(i, vmx_msr_index[i]); > - > - if (setup_vmcs_config(&vmcs_config, &vmx_capability) < 0) > - return -EIO; > - > - if (boot_cpu_has(X86_FEATURE_NX)) > - kvm_enable_efer_bits(EFER_NX); > - > - if (boot_cpu_has(X86_FEATURE_MPX)) { > - rdmsrl(MSR_IA32_BNDCFGS, host_bndcfgs); > - WARN_ONCE(host_bndcfgs, "KVM: BNDCFGS in host will be lost"); > - } > - > - if (!cpu_has_vmx_mpx()) > - supported_xcr0 &= ~(XFEATURE_MASK_BNDREGS | > - XFEATURE_MASK_BNDCSR); > - > - if (!cpu_has_vmx_vpid() || !cpu_has_vmx_invvpid() || > - !(cpu_has_vmx_invvpid_single() || cpu_has_vmx_invvpid_global())) > - enable_vpid = 0; > - > - if (!cpu_has_vmx_ept() || > - !cpu_has_vmx_ept_4levels() || > - !cpu_has_vmx_ept_mt_wb() || > - !cpu_has_vmx_invept_global()) > - enable_ept = 0; > - > - if (!cpu_has_vmx_ept_ad_bits() || !enable_ept) > - enable_ept_ad_bits = 0; > - > - if (!cpu_has_vmx_unrestricted_guest() || !enable_ept) > - enable_unrestricted_guest = 0; > - > - if (!cpu_has_vmx_flexpriority()) > - flexpriority_enabled = 0; > - > - if (!cpu_has_virtual_nmis()) > - enable_vnmi = 0; > - > - /* > - * set_apic_access_page_addr() is used to reload apic access > - * page upon invalidation. No need to do anything if not > - * using the APIC_ACCESS_ADDR VMCS field. > - */ > - if (!flexpriority_enabled) > - kvm_x86_ops->set_apic_access_page_addr = NULL; > - > - if (!cpu_has_vmx_tpr_shadow()) > - kvm_x86_ops->update_cr8_intercept = NULL; > - > -#if IS_ENABLED(CONFIG_HYPERV) > - if (ms_hyperv.nested_features & HV_X64_NESTED_GUEST_MAPPING_FLUSH > - && enable_ept) { > - kvm_x86_ops->tlb_remote_flush = hv_remote_flush_tlb; > - kvm_x86_ops->tlb_remote_flush_with_range = > - hv_remote_flush_tlb_with_range; > - } > -#endif > - > - if (!cpu_has_vmx_ple()) { > - ple_gap = 0; > - ple_window = 0; > - ple_window_grow = 0; > - ple_window_max = 0; > - ple_window_shrink = 0; > - } > - > - if (!cpu_has_vmx_apicv()) { > - enable_apicv = 0; > - kvm_x86_ops->sync_pir_to_irr = NULL; > - } > - > - if (cpu_has_vmx_tsc_scaling()) { > - kvm_has_tsc_control = true; > - kvm_max_tsc_scaling_ratio = KVM_VMX_TSC_MULTIPLIER_MAX; > - kvm_tsc_scaling_ratio_frac_bits = 48; > - } > - > - set_bit(0, vmx_vpid_bitmap); /* 0 is reserved for host */ > - > - if (enable_ept) > - vmx_enable_tdp(); > - > - if (!enable_ept) > - ept_lpage_level = 0; > - else if (cpu_has_vmx_ept_1g_page()) > - ept_lpage_level = PT_PDPE_LEVEL; > - else if (cpu_has_vmx_ept_2m_page()) > - ept_lpage_level = PT_DIRECTORY_LEVEL; > - else > - ept_lpage_level = PT_PAGE_TABLE_LEVEL; > - kvm_configure_mmu(enable_ept, ept_lpage_level); > - > - /* > - * Only enable PML when hardware supports PML feature, and both EPT > - * and EPT A/D bit features are enabled -- PML depends on them to work. > - */ > - if (!enable_ept || !enable_ept_ad_bits || !cpu_has_vmx_pml()) > - enable_pml = 0; > - > - if (!enable_pml) { > - kvm_x86_ops->slot_enable_log_dirty = NULL; > - kvm_x86_ops->slot_disable_log_dirty = NULL; > - kvm_x86_ops->flush_log_dirty = NULL; > - kvm_x86_ops->enable_log_dirty_pt_masked = NULL; > - } > - > - if (!cpu_has_vmx_preemption_timer()) > - enable_preemption_timer = false; > - > - if (enable_preemption_timer) { > - u64 use_timer_freq = 5000ULL * 1000 * 1000; > - u64 vmx_msr; > - > - rdmsrl(MSR_IA32_VMX_MISC, vmx_msr); > - cpu_preemption_timer_multi = > - vmx_msr & VMX_MISC_PREEMPTION_TIMER_RATE_MASK; > - > - if (tsc_khz) > - use_timer_freq = (u64)tsc_khz * 1000; > - use_timer_freq >>= cpu_preemption_timer_multi; > - > - /* > - * KVM "disables" the preemption timer by setting it to its max > - * value. Don't use the timer if it might cause spurious exits > - * at a rate faster than 0.1 Hz (of uninterrupted guest time). > - */ > - if (use_timer_freq > 0xffffffffu / 10) > - enable_preemption_timer = false; > - } > - > - if (!enable_preemption_timer) { > - kvm_x86_ops->set_hv_timer = NULL; > - kvm_x86_ops->cancel_hv_timer = NULL; > - kvm_x86_ops->request_immediate_exit = __kvm_request_immediate_exit; > - } > - > - kvm_set_posted_intr_wakeup_handler(wakeup_handler); > - > - kvm_mce_cap_supported |= MCG_LMCE_P; > - > - if (pt_mode != PT_MODE_SYSTEM && pt_mode != PT_MODE_HOST_GUEST) > - return -EINVAL; > - if (!enable_ept || !cpu_has_vmx_intel_pt()) > - pt_mode = PT_MODE_SYSTEM; > - > - if (nested) { > - nested_vmx_setup_ctls_msrs(&vmcs_config.nested, > - vmx_capability.ept); > - > - r = nested_vmx_hardware_setup(kvm_vmx_exit_handlers); > - if (r) > - return r; > - } > - > - vmx_set_cpu_caps(); > - > - r = alloc_kvm_area(); > - if (r) > - nested_vmx_hardware_unsetup(); > - return r; > -} > - > static __exit void hardware_unsetup(void) > { > if (nested) > @@ -7978,6 +7805,179 @@ static struct kvm_x86_ops vmx_x86_ops __ro_after_init = { > .apic_init_signal_blocked = vmx_apic_init_signal_blocked, > }; > > +static __init int hardware_setup(void) > +{ > + unsigned long host_bndcfgs; > + struct desc_ptr dt; > + int r, i, ept_lpage_level; > + > + store_idt(&dt); > + host_idt_base = dt.address; > + > + for (i = 0; i < ARRAY_SIZE(vmx_msr_index); ++i) > + kvm_define_shared_msr(i, vmx_msr_index[i]); > + > + if (setup_vmcs_config(&vmcs_config, &vmx_capability) < 0) > + return -EIO; > + > + if (boot_cpu_has(X86_FEATURE_NX)) > + kvm_enable_efer_bits(EFER_NX); > + > + if (boot_cpu_has(X86_FEATURE_MPX)) { > + rdmsrl(MSR_IA32_BNDCFGS, host_bndcfgs); > + WARN_ONCE(host_bndcfgs, "KVM: BNDCFGS in host will be lost"); > + } > + > + if (!cpu_has_vmx_mpx()) > + supported_xcr0 &= ~(XFEATURE_MASK_BNDREGS | > + XFEATURE_MASK_BNDCSR); > + > + if (!cpu_has_vmx_vpid() || !cpu_has_vmx_invvpid() || > + !(cpu_has_vmx_invvpid_single() || cpu_has_vmx_invvpid_global())) > + enable_vpid = 0; > + > + if (!cpu_has_vmx_ept() || > + !cpu_has_vmx_ept_4levels() || > + !cpu_has_vmx_ept_mt_wb() || > + !cpu_has_vmx_invept_global()) > + enable_ept = 0; > + > + if (!cpu_has_vmx_ept_ad_bits() || !enable_ept) > + enable_ept_ad_bits = 0; > + > + if (!cpu_has_vmx_unrestricted_guest() || !enable_ept) > + enable_unrestricted_guest = 0; > + > + if (!cpu_has_vmx_flexpriority()) > + flexpriority_enabled = 0; > + > + if (!cpu_has_virtual_nmis()) > + enable_vnmi = 0; > + > + /* > + * set_apic_access_page_addr() is used to reload apic access > + * page upon invalidation. No need to do anything if not > + * using the APIC_ACCESS_ADDR VMCS field. > + */ > + if (!flexpriority_enabled) > + kvm_x86_ops->set_apic_access_page_addr = NULL; > + > + if (!cpu_has_vmx_tpr_shadow()) > + kvm_x86_ops->update_cr8_intercept = NULL; > + > +#if IS_ENABLED(CONFIG_HYPERV) > + if (ms_hyperv.nested_features & HV_X64_NESTED_GUEST_MAPPING_FLUSH > + && enable_ept) { > + kvm_x86_ops->tlb_remote_flush = hv_remote_flush_tlb; > + kvm_x86_ops->tlb_remote_flush_with_range = > + hv_remote_flush_tlb_with_range; > + } > +#endif > + > + if (!cpu_has_vmx_ple()) { > + ple_gap = 0; > + ple_window = 0; > + ple_window_grow = 0; > + ple_window_max = 0; > + ple_window_shrink = 0; > + } > + > + if (!cpu_has_vmx_apicv()) { > + enable_apicv = 0; > + kvm_x86_ops->sync_pir_to_irr = NULL; > + } > + > + if (cpu_has_vmx_tsc_scaling()) { > + kvm_has_tsc_control = true; > + kvm_max_tsc_scaling_ratio = KVM_VMX_TSC_MULTIPLIER_MAX; > + kvm_tsc_scaling_ratio_frac_bits = 48; > + } > + > + set_bit(0, vmx_vpid_bitmap); /* 0 is reserved for host */ > + > + if (enable_ept) > + vmx_enable_tdp(); > + > + if (!enable_ept) > + ept_lpage_level = 0; > + else if (cpu_has_vmx_ept_1g_page()) > + ept_lpage_level = PT_PDPE_LEVEL; > + else if (cpu_has_vmx_ept_2m_page()) > + ept_lpage_level = PT_DIRECTORY_LEVEL; > + else > + ept_lpage_level = PT_PAGE_TABLE_LEVEL; > + kvm_configure_mmu(enable_ept, ept_lpage_level); > + > + /* > + * Only enable PML when hardware supports PML feature, and both EPT > + * and EPT A/D bit features are enabled -- PML depends on them to work. > + */ > + if (!enable_ept || !enable_ept_ad_bits || !cpu_has_vmx_pml()) > + enable_pml = 0; > + > + if (!enable_pml) { > + kvm_x86_ops->slot_enable_log_dirty = NULL; > + kvm_x86_ops->slot_disable_log_dirty = NULL; > + kvm_x86_ops->flush_log_dirty = NULL; > + kvm_x86_ops->enable_log_dirty_pt_masked = NULL; > + } > + > + if (!cpu_has_vmx_preemption_timer()) > + enable_preemption_timer = false; > + > + if (enable_preemption_timer) { > + u64 use_timer_freq = 5000ULL * 1000 * 1000; > + u64 vmx_msr; > + > + rdmsrl(MSR_IA32_VMX_MISC, vmx_msr); > + cpu_preemption_timer_multi = > + vmx_msr & VMX_MISC_PREEMPTION_TIMER_RATE_MASK; > + > + if (tsc_khz) > + use_timer_freq = (u64)tsc_khz * 1000; > + use_timer_freq >>= cpu_preemption_timer_multi; > + > + /* > + * KVM "disables" the preemption timer by setting it to its max > + * value. Don't use the timer if it might cause spurious exits > + * at a rate faster than 0.1 Hz (of uninterrupted guest time). > + */ > + if (use_timer_freq > 0xffffffffu / 10) > + enable_preemption_timer = false; > + } > + > + if (!enable_preemption_timer) { > + kvm_x86_ops->set_hv_timer = NULL; > + kvm_x86_ops->cancel_hv_timer = NULL; > + kvm_x86_ops->request_immediate_exit = __kvm_request_immediate_exit; > + } > + > + kvm_set_posted_intr_wakeup_handler(wakeup_handler); > + > + kvm_mce_cap_supported |= MCG_LMCE_P; > + > + if (pt_mode != PT_MODE_SYSTEM && pt_mode != PT_MODE_HOST_GUEST) > + return -EINVAL; > + if (!enable_ept || !cpu_has_vmx_intel_pt()) > + pt_mode = PT_MODE_SYSTEM; > + > + if (nested) { > + nested_vmx_setup_ctls_msrs(&vmcs_config.nested, > + vmx_capability.ept); > + > + r = nested_vmx_hardware_setup(kvm_vmx_exit_handlers); > + if (r) > + return r; > + } > + > + vmx_set_cpu_caps(); > + > + r = alloc_kvm_area(); > + if (r) > + nested_vmx_hardware_unsetup(); > + return r; > +} > + > static struct kvm_x86_init_ops vmx_init_ops __initdata = { > .cpu_has_kvm_support = cpu_has_kvm_support, > .disabled_by_bios = vmx_disabled_by_bios, Reviewed-by: Vitaly Kuznetsov <vkuznets@xxxxxxxxxx> -- Vitaly