On Fri, Apr 19, 2024 at 1:30 PM Wei Wang <wei.w.wang@xxxxxxxxx> wrote: > +#define KVM_X86_CALL(func, ...) static_call(kvm_x86_##func)(__VA_ARGS__) Just #define KVM_X86_CALL(func) static_call(kvm_x86_##func) please, because having the parentheses around the arguments is a lot more readable (and works better with most editors, for example in vim you can use the "%" movement comment to find the matching parenthesis). Paolo > + > #define KVM_X86_OP(func) \ > DECLARE_STATIC_CALL(kvm_x86_##func, *(((struct kvm_x86_ops *)0)->func)); > #define KVM_X86_OP_OPTIONAL KVM_X86_OP > @@ -1875,7 +1877,7 @@ void kvm_arch_free_vm(struct kvm *kvm); > static inline int kvm_arch_flush_remote_tlbs(struct kvm *kvm) > { > if (kvm_x86_ops.flush_remote_tlbs && > - !static_call(kvm_x86_flush_remote_tlbs)(kvm)) > + !KVM_X86_CALL(flush_remote_tlbs, kvm)) > return 0; > else > return -ENOTSUPP; > @@ -1888,7 +1890,7 @@ static inline int kvm_arch_flush_remote_tlbs_range(struct kvm *kvm, gfn_t gfn, > if (!kvm_x86_ops.flush_remote_tlbs_range) > return -EOPNOTSUPP; > > - return static_call(kvm_x86_flush_remote_tlbs_range)(kvm, gfn, nr_pages); > + return KVM_X86_CALL(flush_remote_tlbs_range, kvm, gfn, nr_pages); > } > #endif /* CONFIG_HYPERV */ > > @@ -2280,12 +2282,12 @@ static inline bool kvm_irq_is_postable(struct kvm_lapic_irq *irq) > > static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) > { > - static_call(kvm_x86_vcpu_blocking)(vcpu); > + KVM_X86_CALL(vcpu_blocking, vcpu); > } > > static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) > { > - static_call(kvm_x86_vcpu_unblocking)(vcpu); > + KVM_X86_CALL(vcpu_unblocking, vcpu); > } > > static inline int kvm_cpu_get_apicid(int mps_cpu) > diff --git a/arch/x86/kvm/cpuid.c b/arch/x86/kvm/cpuid.c > index 77352a4abd87..ba15fb416aaf 100644 > --- a/arch/x86/kvm/cpuid.c > +++ b/arch/x86/kvm/cpuid.c > @@ -388,7 +388,7 @@ static void kvm_vcpu_after_set_cpuid(struct kvm_vcpu *vcpu) > vcpu->arch.cpuid_nent)); > > /* Invoke the vendor callback only after the above state is updated. */ > - static_call(kvm_x86_vcpu_after_set_cpuid)(vcpu); > + KVM_X86_CALL(vcpu_after_set_cpuid, vcpu); > > /* > * Except for the MMU, which needs to do its thing any vendor specific > diff --git a/arch/x86/kvm/hyperv.c b/arch/x86/kvm/hyperv.c > index 8a47f8541eab..5833e82221ea 100644 > --- a/arch/x86/kvm/hyperv.c > +++ b/arch/x86/kvm/hyperv.c > @@ -1417,7 +1417,7 @@ static int kvm_hv_set_msr_pw(struct kvm_vcpu *vcpu, u32 msr, u64 data, > } > > /* vmcall/vmmcall */ > - static_call(kvm_x86_patch_hypercall)(vcpu, instructions + i); > + KVM_X86_CALL(patch_hypercall, vcpu, instructions + i); > i += 3; > > /* ret */ > @@ -1985,7 +1985,7 @@ int kvm_hv_vcpu_flush_tlb(struct kvm_vcpu *vcpu) > */ > gva = entries[i] & PAGE_MASK; > for (j = 0; j < (entries[i] & ~PAGE_MASK) + 1; j++) > - static_call(kvm_x86_flush_tlb_gva)(vcpu, gva + j * PAGE_SIZE); > + KVM_X86_CALL(flush_tlb_gva, vcpu, gva + j * PAGE_SIZE); > > ++vcpu->stat.tlb_flush; > } > @@ -2526,7 +2526,7 @@ int kvm_hv_hypercall(struct kvm_vcpu *vcpu) > * hypercall generates UD from non zero cpl and real mode > * per HYPER-V spec > */ > - if (static_call(kvm_x86_get_cpl)(vcpu) != 0 || !is_protmode(vcpu)) { > + if (KVM_X86_CALL(get_cpl, vcpu) != 0 || !is_protmode(vcpu)) { > kvm_queue_exception(vcpu, UD_VECTOR); > return 1; > } > diff --git a/arch/x86/kvm/irq.c b/arch/x86/kvm/irq.c > index 7cf93d427484..2feaeec271cf 100644 > --- a/arch/x86/kvm/irq.c > +++ b/arch/x86/kvm/irq.c > @@ -157,7 +157,7 @@ void __kvm_migrate_timers(struct kvm_vcpu *vcpu) > { > __kvm_migrate_apic_timer(vcpu); > __kvm_migrate_pit_timer(vcpu); > - static_call(kvm_x86_migrate_timers)(vcpu); > + KVM_X86_CALL(migrate_timers, vcpu); > } > > bool kvm_arch_irqfd_allowed(struct kvm *kvm, struct kvm_irqfd *args) > diff --git a/arch/x86/kvm/kvm_cache_regs.h b/arch/x86/kvm/kvm_cache_regs.h > index 75eae9c4998a..2053ed0066ac 100644 > --- a/arch/x86/kvm/kvm_cache_regs.h > +++ b/arch/x86/kvm/kvm_cache_regs.h > @@ -98,7 +98,7 @@ static inline unsigned long kvm_register_read_raw(struct kvm_vcpu *vcpu, int reg > return 0; > > if (!kvm_register_is_available(vcpu, reg)) > - static_call(kvm_x86_cache_reg)(vcpu, reg); > + KVM_X86_CALL(cache_reg, vcpu, reg); > > return vcpu->arch.regs[reg]; > } > @@ -138,7 +138,7 @@ static inline u64 kvm_pdptr_read(struct kvm_vcpu *vcpu, int index) > might_sleep(); /* on svm */ > > if (!kvm_register_is_available(vcpu, VCPU_EXREG_PDPTR)) > - static_call(kvm_x86_cache_reg)(vcpu, VCPU_EXREG_PDPTR); > + KVM_X86_CALL(cache_reg, vcpu, VCPU_EXREG_PDPTR); > > return vcpu->arch.walk_mmu->pdptrs[index]; > } > @@ -153,7 +153,7 @@ static inline ulong kvm_read_cr0_bits(struct kvm_vcpu *vcpu, ulong mask) > ulong tmask = mask & KVM_POSSIBLE_CR0_GUEST_BITS; > if ((tmask & vcpu->arch.cr0_guest_owned_bits) && > !kvm_register_is_available(vcpu, VCPU_EXREG_CR0)) > - static_call(kvm_x86_cache_reg)(vcpu, VCPU_EXREG_CR0); > + KVM_X86_CALL(cache_reg, vcpu, VCPU_EXREG_CR0); > return vcpu->arch.cr0 & mask; > } > > @@ -175,7 +175,7 @@ static inline ulong kvm_read_cr4_bits(struct kvm_vcpu *vcpu, ulong mask) > ulong tmask = mask & KVM_POSSIBLE_CR4_GUEST_BITS; > if ((tmask & vcpu->arch.cr4_guest_owned_bits) && > !kvm_register_is_available(vcpu, VCPU_EXREG_CR4)) > - static_call(kvm_x86_cache_reg)(vcpu, VCPU_EXREG_CR4); > + KVM_X86_CALL(cache_reg, vcpu, VCPU_EXREG_CR4); > return vcpu->arch.cr4 & mask; > } > > @@ -190,7 +190,7 @@ static __always_inline bool kvm_is_cr4_bit_set(struct kvm_vcpu *vcpu, > static inline ulong kvm_read_cr3(struct kvm_vcpu *vcpu) > { > if (!kvm_register_is_available(vcpu, VCPU_EXREG_CR3)) > - static_call(kvm_x86_cache_reg)(vcpu, VCPU_EXREG_CR3); > + KVM_X86_CALL(cache_reg, vcpu, VCPU_EXREG_CR3); > return vcpu->arch.cr3; > } > > diff --git a/arch/x86/kvm/lapic.c b/arch/x86/kvm/lapic.c > index eaf840699d27..0c52cf912712 100644 > --- a/arch/x86/kvm/lapic.c > +++ b/arch/x86/kvm/lapic.c > @@ -728,8 +728,8 @@ static inline void apic_clear_irr(int vec, struct kvm_lapic *apic) > if (unlikely(apic->apicv_active)) { > /* need to update RVI */ > kvm_lapic_clear_vector(vec, apic->regs + APIC_IRR); > - static_call(kvm_x86_hwapic_irr_update)(apic->vcpu, > - apic_find_highest_irr(apic)); > + KVM_X86_CALL(hwapic_irr_update, > + apic->vcpu, apic_find_highest_irr(apic)); > } else { > apic->irr_pending = false; > kvm_lapic_clear_vector(vec, apic->regs + APIC_IRR); > @@ -755,7 +755,7 @@ static inline void apic_set_isr(int vec, struct kvm_lapic *apic) > * just set SVI. > */ > if (unlikely(apic->apicv_active)) > - static_call(kvm_x86_hwapic_isr_update)(vec); > + KVM_X86_CALL(hwapic_isr_update, vec); > else { > ++apic->isr_count; > BUG_ON(apic->isr_count > MAX_APIC_VECTOR); > @@ -800,7 +800,7 @@ static inline void apic_clear_isr(int vec, struct kvm_lapic *apic) > * and must be left alone. > */ > if (unlikely(apic->apicv_active)) > - static_call(kvm_x86_hwapic_isr_update)(apic_find_highest_isr(apic)); > + KVM_X86_CALL(hwapic_isr_update, apic_find_highest_isr(apic)); > else { > --apic->isr_count; > BUG_ON(apic->isr_count < 0); > @@ -936,7 +936,7 @@ static int apic_has_interrupt_for_ppr(struct kvm_lapic *apic, u32 ppr) > { > int highest_irr; > if (kvm_x86_ops.sync_pir_to_irr) > - highest_irr = static_call(kvm_x86_sync_pir_to_irr)(apic->vcpu); > + highest_irr = KVM_X86_CALL(sync_pir_to_irr, apic->vcpu); > else > highest_irr = apic_find_highest_irr(apic); > if (highest_irr == -1 || (highest_irr & 0xF0) <= ppr) > @@ -1328,8 +1328,8 @@ static int __apic_accept_irq(struct kvm_lapic *apic, int delivery_mode, > apic->regs + APIC_TMR); > } > > - static_call(kvm_x86_deliver_interrupt)(apic, delivery_mode, > - trig_mode, vector); > + KVM_X86_CALL(deliver_interrupt, > + apic, delivery_mode, trig_mode, vector); > break; > > case APIC_DM_REMRD: > @@ -2095,7 +2095,7 @@ static void cancel_hv_timer(struct kvm_lapic *apic) > { > WARN_ON(preemptible()); > WARN_ON(!apic->lapic_timer.hv_timer_in_use); > - static_call(kvm_x86_cancel_hv_timer)(apic->vcpu); > + KVM_X86_CALL(cancel_hv_timer, apic->vcpu); > apic->lapic_timer.hv_timer_in_use = false; > } > > @@ -2112,7 +2112,7 @@ static bool start_hv_timer(struct kvm_lapic *apic) > if (!ktimer->tscdeadline) > return false; > > - if (static_call(kvm_x86_set_hv_timer)(vcpu, ktimer->tscdeadline, &expired)) > + if (KVM_X86_CALL(set_hv_timer, vcpu, ktimer->tscdeadline, &expired)) > return false; > > ktimer->hv_timer_in_use = true; > @@ -2567,7 +2567,7 @@ void kvm_lapic_set_base(struct kvm_vcpu *vcpu, u64 value) > > if ((old_value ^ value) & (MSR_IA32_APICBASE_ENABLE | X2APIC_ENABLE)) { > kvm_make_request(KVM_REQ_APICV_UPDATE, vcpu); > - static_call(kvm_x86_set_virtual_apic_mode)(vcpu); > + KVM_X86_CALL(set_virtual_apic_mode, vcpu); > } > > apic->base_address = apic->vcpu->arch.apic_base & > @@ -2677,7 +2677,7 @@ void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event) > u64 msr_val; > int i; > > - static_call(kvm_x86_apicv_pre_state_restore)(vcpu); > + KVM_X86_CALL(apicv_pre_state_restore, vcpu); > > if (!init_event) { > msr_val = APIC_DEFAULT_PHYS_BASE | MSR_IA32_APICBASE_ENABLE; > @@ -2732,9 +2732,9 @@ void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event) > vcpu->arch.pv_eoi.msr_val = 0; > apic_update_ppr(apic); > if (apic->apicv_active) { > - static_call(kvm_x86_apicv_post_state_restore)(vcpu); > - static_call(kvm_x86_hwapic_irr_update)(vcpu, -1); > - static_call(kvm_x86_hwapic_isr_update)(-1); > + KVM_X86_CALL(apicv_post_state_restore, vcpu); > + KVM_X86_CALL(hwapic_irr_update, vcpu, -1); > + KVM_X86_CALL(hwapic_isr_update, -1); > } > > vcpu->arch.apic_arb_prio = 0; > @@ -2830,7 +2830,7 @@ int kvm_create_lapic(struct kvm_vcpu *vcpu, int timer_advance_ns) > vcpu->arch.apic = apic; > > if (kvm_x86_ops.alloc_apic_backing_page) > - apic->regs = static_call(kvm_x86_alloc_apic_backing_page)(vcpu); > + apic->regs = KVM_X86_CALL(alloc_apic_backing_page, vcpu); > else > apic->regs = (void *)get_zeroed_page(GFP_KERNEL_ACCOUNT); > if (!apic->regs) { > @@ -3014,7 +3014,7 @@ int kvm_apic_set_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s) > struct kvm_lapic *apic = vcpu->arch.apic; > int r; > > - static_call(kvm_x86_apicv_pre_state_restore)(vcpu); > + KVM_X86_CALL(apicv_pre_state_restore, vcpu); > > kvm_lapic_set_base(vcpu, vcpu->arch.apic_base); > /* set SPIV separately to get count of SW disabled APICs right */ > @@ -3041,9 +3041,10 @@ int kvm_apic_set_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s) > kvm_lapic_set_reg(apic, APIC_TMCCT, 0); > kvm_apic_update_apicv(vcpu); > if (apic->apicv_active) { > - static_call(kvm_x86_apicv_post_state_restore)(vcpu); > - static_call(kvm_x86_hwapic_irr_update)(vcpu, apic_find_highest_irr(apic)); > - static_call(kvm_x86_hwapic_isr_update)(apic_find_highest_isr(apic)); > + KVM_X86_CALL(apicv_post_state_restore, vcpu); > + KVM_X86_CALL(hwapic_irr_update, > + vcpu, apic_find_highest_irr(apic)); > + KVM_X86_CALL(hwapic_isr_update, apic_find_highest_isr(apic)); > } > kvm_make_request(KVM_REQ_EVENT, vcpu); > if (ioapic_in_kernel(vcpu->kvm)) > @@ -3331,7 +3332,8 @@ int kvm_apic_accept_events(struct kvm_vcpu *vcpu) > /* evaluate pending_events before reading the vector */ > smp_rmb(); > sipi_vector = apic->sipi_vector; > - static_call(kvm_x86_vcpu_deliver_sipi_vector)(vcpu, sipi_vector); > + KVM_X86_CALL(vcpu_deliver_sipi_vector, > + vcpu, sipi_vector); > vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE; > } > } > diff --git a/arch/x86/kvm/lapic.h b/arch/x86/kvm/lapic.h > index 0a0ea4b5dd8c..b56208a6864c 100644 > --- a/arch/x86/kvm/lapic.h > +++ b/arch/x86/kvm/lapic.h > @@ -236,7 +236,7 @@ static inline bool kvm_apic_has_pending_init_or_sipi(struct kvm_vcpu *vcpu) > static inline bool kvm_apic_init_sipi_allowed(struct kvm_vcpu *vcpu) > { > return !is_smm(vcpu) && > - !static_call(kvm_x86_apic_init_signal_blocked)(vcpu); > + !KVM_X86_CALL(apic_init_signal_blocked, vcpu); > } > > static inline bool kvm_lowest_prio_delivery(struct kvm_lapic_irq *irq) > diff --git a/arch/x86/kvm/mmu.h b/arch/x86/kvm/mmu.h > index 60f21bb4c27b..e13af03f5cc1 100644 > --- a/arch/x86/kvm/mmu.h > +++ b/arch/x86/kvm/mmu.h > @@ -161,8 +161,8 @@ static inline void kvm_mmu_load_pgd(struct kvm_vcpu *vcpu) > if (!VALID_PAGE(root_hpa)) > return; > > - static_call(kvm_x86_load_mmu_pgd)(vcpu, root_hpa, > - vcpu->arch.mmu->root_role.level); > + KVM_X86_CALL(load_mmu_pgd, vcpu, root_hpa, > + vcpu->arch.mmu->root_role.level); > } > > static inline void kvm_mmu_refresh_passthrough_bits(struct kvm_vcpu *vcpu, > @@ -197,7 +197,7 @@ static inline u8 permission_fault(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, > { > /* strip nested paging fault error codes */ > unsigned int pfec = access; > - unsigned long rflags = static_call(kvm_x86_get_rflags)(vcpu); > + unsigned long rflags = KVM_X86_CALL(get_rflags, vcpu); > > /* > * For explicit supervisor accesses, SMAP is disabled if EFLAGS.AC = 1. > diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c > index db007a4dffa2..112643a1fa05 100644 > --- a/arch/x86/kvm/mmu/mmu.c > +++ b/arch/x86/kvm/mmu/mmu.c > @@ -5626,7 +5626,7 @@ int kvm_mmu_load(struct kvm_vcpu *vcpu) > * stale entries. Flushing on alloc also allows KVM to skip the TLB > * flush when freeing a root (see kvm_tdp_mmu_put_root()). > */ > - static_call(kvm_x86_flush_tlb_current)(vcpu); > + KVM_X86_CALL(flush_tlb_current, vcpu); > out: > return r; > } > @@ -5962,7 +5962,7 @@ void kvm_mmu_invalidate_addr(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, > if (is_noncanonical_address(addr, vcpu)) > return; > > - static_call(kvm_x86_flush_tlb_gva)(vcpu, addr); > + KVM_X86_CALL(flush_tlb_gva, vcpu, addr); > } > > if (!mmu->sync_spte) > diff --git a/arch/x86/kvm/mmu/spte.c b/arch/x86/kvm/mmu/spte.c > index 4a599130e9c9..40a6368e4913 100644 > --- a/arch/x86/kvm/mmu/spte.c > +++ b/arch/x86/kvm/mmu/spte.c > @@ -190,8 +190,8 @@ bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, > spte |= PT_PAGE_SIZE_MASK; > > if (shadow_memtype_mask) > - spte |= static_call(kvm_x86_get_mt_mask)(vcpu, gfn, > - kvm_is_mmio_pfn(pfn)); > + spte |= KVM_X86_CALL(get_mt_mask, > + vcpu, gfn, kvm_is_mmio_pfn(pfn)); > if (host_writable) > spte |= shadow_host_writable_mask; > else > diff --git a/arch/x86/kvm/smm.c b/arch/x86/kvm/smm.c > index d06d43d8d2aa..745b3d0954c7 100644 > --- a/arch/x86/kvm/smm.c > +++ b/arch/x86/kvm/smm.c > @@ -200,11 +200,11 @@ static void enter_smm_save_state_32(struct kvm_vcpu *vcpu, > enter_smm_save_seg_32(vcpu, &smram->tr, &smram->tr_sel, VCPU_SREG_TR); > enter_smm_save_seg_32(vcpu, &smram->ldtr, &smram->ldtr_sel, VCPU_SREG_LDTR); > > - static_call(kvm_x86_get_gdt)(vcpu, &dt); > + KVM_X86_CALL(get_gdt, vcpu, &dt); > smram->gdtr.base = dt.address; > smram->gdtr.limit = dt.size; > > - static_call(kvm_x86_get_idt)(vcpu, &dt); > + KVM_X86_CALL(get_idt, vcpu, &dt); > smram->idtr.base = dt.address; > smram->idtr.limit = dt.size; > > @@ -220,7 +220,7 @@ static void enter_smm_save_state_32(struct kvm_vcpu *vcpu, > smram->smm_revision = 0x00020000; > smram->smbase = vcpu->arch.smbase; > > - smram->int_shadow = static_call(kvm_x86_get_interrupt_shadow)(vcpu); > + smram->int_shadow = KVM_X86_CALL(get_interrupt_shadow, vcpu); > } > > #ifdef CONFIG_X86_64 > @@ -250,13 +250,13 @@ static void enter_smm_save_state_64(struct kvm_vcpu *vcpu, > > enter_smm_save_seg_64(vcpu, &smram->tr, VCPU_SREG_TR); > > - static_call(kvm_x86_get_idt)(vcpu, &dt); > + KVM_X86_CALL(get_idt, vcpu, &dt); > smram->idtr.limit = dt.size; > smram->idtr.base = dt.address; > > enter_smm_save_seg_64(vcpu, &smram->ldtr, VCPU_SREG_LDTR); > > - static_call(kvm_x86_get_gdt)(vcpu, &dt); > + KVM_X86_CALL(get_gdt, vcpu, &dt); > smram->gdtr.limit = dt.size; > smram->gdtr.base = dt.address; > > @@ -267,7 +267,7 @@ static void enter_smm_save_state_64(struct kvm_vcpu *vcpu, > enter_smm_save_seg_64(vcpu, &smram->fs, VCPU_SREG_FS); > enter_smm_save_seg_64(vcpu, &smram->gs, VCPU_SREG_GS); > > - smram->int_shadow = static_call(kvm_x86_get_interrupt_shadow)(vcpu); > + smram->int_shadow = KVM_X86_CALL(get_interrupt_shadow, vcpu); > } > #endif > > @@ -297,7 +297,7 @@ void enter_smm(struct kvm_vcpu *vcpu) > * Kill the VM in the unlikely case of failure, because the VM > * can be in undefined state in this case. > */ > - if (static_call(kvm_x86_enter_smm)(vcpu, &smram)) > + if (KVM_X86_CALL(enter_smm, vcpu, &smram)) > goto error; > > kvm_smm_changed(vcpu, true); > @@ -305,24 +305,24 @@ void enter_smm(struct kvm_vcpu *vcpu) > if (kvm_vcpu_write_guest(vcpu, vcpu->arch.smbase + 0xfe00, &smram, sizeof(smram))) > goto error; > > - if (static_call(kvm_x86_get_nmi_mask)(vcpu)) > + if (KVM_X86_CALL(get_nmi_mask, vcpu)) > vcpu->arch.hflags |= HF_SMM_INSIDE_NMI_MASK; > else > - static_call(kvm_x86_set_nmi_mask)(vcpu, true); > + KVM_X86_CALL(set_nmi_mask, vcpu, true); > > kvm_set_rflags(vcpu, X86_EFLAGS_FIXED); > kvm_rip_write(vcpu, 0x8000); > > - static_call(kvm_x86_set_interrupt_shadow)(vcpu, 0); > + KVM_X86_CALL(set_interrupt_shadow, vcpu, 0); > > cr0 = vcpu->arch.cr0 & ~(X86_CR0_PE | X86_CR0_EM | X86_CR0_TS | X86_CR0_PG); > - static_call(kvm_x86_set_cr0)(vcpu, cr0); > + KVM_X86_CALL(set_cr0, vcpu, cr0); > > - static_call(kvm_x86_set_cr4)(vcpu, 0); > + KVM_X86_CALL(set_cr4, vcpu, 0); > > /* Undocumented: IDT limit is set to zero on entry to SMM. */ > dt.address = dt.size = 0; > - static_call(kvm_x86_set_idt)(vcpu, &dt); > + KVM_X86_CALL(set_idt, vcpu, &dt); > > if (WARN_ON_ONCE(kvm_set_dr(vcpu, 7, DR7_FIXED_1))) > goto error; > @@ -354,7 +354,7 @@ void enter_smm(struct kvm_vcpu *vcpu) > > #ifdef CONFIG_X86_64 > if (guest_cpuid_has(vcpu, X86_FEATURE_LM)) > - if (static_call(kvm_x86_set_efer)(vcpu, 0)) > + if (KVM_X86_CALL(set_efer, vcpu, 0)) > goto error; > #endif > > @@ -479,11 +479,11 @@ static int rsm_load_state_32(struct x86_emulate_ctxt *ctxt, > > dt.address = smstate->gdtr.base; > dt.size = smstate->gdtr.limit; > - static_call(kvm_x86_set_gdt)(vcpu, &dt); > + KVM_X86_CALL(set_gdt, vcpu, &dt); > > dt.address = smstate->idtr.base; > dt.size = smstate->idtr.limit; > - static_call(kvm_x86_set_idt)(vcpu, &dt); > + KVM_X86_CALL(set_idt, vcpu, &dt); > > rsm_load_seg_32(vcpu, &smstate->es, smstate->es_sel, VCPU_SREG_ES); > rsm_load_seg_32(vcpu, &smstate->cs, smstate->cs_sel, VCPU_SREG_CS); > @@ -501,7 +501,7 @@ static int rsm_load_state_32(struct x86_emulate_ctxt *ctxt, > if (r != X86EMUL_CONTINUE) > return r; > > - static_call(kvm_x86_set_interrupt_shadow)(vcpu, 0); > + KVM_X86_CALL(set_interrupt_shadow, vcpu, 0); > ctxt->interruptibility = (u8)smstate->int_shadow; > > return r; > @@ -535,13 +535,13 @@ static int rsm_load_state_64(struct x86_emulate_ctxt *ctxt, > > dt.size = smstate->idtr.limit; > dt.address = smstate->idtr.base; > - static_call(kvm_x86_set_idt)(vcpu, &dt); > + KVM_X86_CALL(set_idt, vcpu, &dt); > > rsm_load_seg_64(vcpu, &smstate->ldtr, VCPU_SREG_LDTR); > > dt.size = smstate->gdtr.limit; > dt.address = smstate->gdtr.base; > - static_call(kvm_x86_set_gdt)(vcpu, &dt); > + KVM_X86_CALL(set_gdt, vcpu, &dt); > > r = rsm_enter_protected_mode(vcpu, smstate->cr0, smstate->cr3, smstate->cr4); > if (r != X86EMUL_CONTINUE) > @@ -554,7 +554,7 @@ static int rsm_load_state_64(struct x86_emulate_ctxt *ctxt, > rsm_load_seg_64(vcpu, &smstate->fs, VCPU_SREG_FS); > rsm_load_seg_64(vcpu, &smstate->gs, VCPU_SREG_GS); > > - static_call(kvm_x86_set_interrupt_shadow)(vcpu, 0); > + KVM_X86_CALL(set_interrupt_shadow, vcpu, 0); > ctxt->interruptibility = (u8)smstate->int_shadow; > > return X86EMUL_CONTINUE; > @@ -576,7 +576,7 @@ int emulator_leave_smm(struct x86_emulate_ctxt *ctxt) > return X86EMUL_UNHANDLEABLE; > > if ((vcpu->arch.hflags & HF_SMM_INSIDE_NMI_MASK) == 0) > - static_call(kvm_x86_set_nmi_mask)(vcpu, false); > + KVM_X86_CALL(set_nmi_mask, vcpu, false); > > kvm_smm_changed(vcpu, false); > > @@ -628,7 +628,7 @@ int emulator_leave_smm(struct x86_emulate_ctxt *ctxt) > * state (e.g. enter guest mode) before loading state from the SMM > * state-save area. > */ > - if (static_call(kvm_x86_leave_smm)(vcpu, &smram)) > + if (KVM_X86_CALL(leave_smm, vcpu, &smram)) > return X86EMUL_UNHANDLEABLE; > > #ifdef CONFIG_X86_64 > diff --git a/arch/x86/kvm/trace.h b/arch/x86/kvm/trace.h > index c6b4b1728006..27096f001e54 100644 > --- a/arch/x86/kvm/trace.h > +++ b/arch/x86/kvm/trace.h > @@ -314,12 +314,9 @@ TRACE_EVENT(name, \ > __entry->guest_rip = kvm_rip_read(vcpu); \ > __entry->isa = isa; \ > __entry->vcpu_id = vcpu->vcpu_id; \ > - static_call(kvm_x86_get_exit_info)(vcpu, \ > - &__entry->exit_reason, \ > - &__entry->info1, \ > - &__entry->info2, \ > - &__entry->intr_info, \ > - &__entry->error_code); \ > + KVM_X86_CALL(get_exit_info, vcpu, &__entry->exit_reason, \ > + &__entry->info1, &__entry->info2, \ > + &__entry->intr_info, &__entry->error_code); \ > ), \ > \ > TP_printk("vcpu %u reason %s%s%s rip 0x%lx info1 0x%016llx " \ > @@ -828,7 +825,8 @@ TRACE_EVENT(kvm_emulate_insn, > ), > > TP_fast_assign( > - __entry->csbase = static_call(kvm_x86_get_segment_base)(vcpu, VCPU_SREG_CS); > + __entry->csbase = KVM_X86_CALL(get_segment_base, > + vcpu, VCPU_SREG_CS); > __entry->len = vcpu->arch.emulate_ctxt->fetch.ptr > - vcpu->arch.emulate_ctxt->fetch.data; > __entry->rip = vcpu->arch.emulate_ctxt->_eip - __entry->len; > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > index 4e6cbbab1e18..2467e053cb35 100644 > --- a/arch/x86/kvm/x86.c > +++ b/arch/x86/kvm/x86.c > @@ -839,7 +839,7 @@ EXPORT_SYMBOL_GPL(kvm_requeue_exception_e); > */ > bool kvm_require_cpl(struct kvm_vcpu *vcpu, int required_cpl) > { > - if (static_call(kvm_x86_get_cpl)(vcpu) <= required_cpl) > + if (KVM_X86_CALL(get_cpl, vcpu) <= required_cpl) > return true; > kvm_queue_exception_e(vcpu, GP_VECTOR, 0); > return false; > @@ -923,7 +923,7 @@ static bool kvm_is_valid_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) > if ((cr0 & X86_CR0_PG) && !(cr0 & X86_CR0_PE)) > return false; > > - return static_call(kvm_x86_is_valid_cr0)(vcpu, cr0); > + return KVM_X86_CALL(is_valid_cr0, vcpu, cr0); > } > > void kvm_post_set_cr0(struct kvm_vcpu *vcpu, unsigned long old_cr0, unsigned long cr0) > @@ -987,7 +987,7 @@ int kvm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) > > if (!is_pae(vcpu)) > return 1; > - static_call(kvm_x86_get_cs_db_l_bits)(vcpu, &cs_db, &cs_l); > + KVM_X86_CALL(get_cs_db_l_bits, vcpu, &cs_db, &cs_l); > if (cs_l) > return 1; > } > @@ -1001,7 +1001,7 @@ int kvm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0) > (is_64_bit_mode(vcpu) || kvm_is_cr4_bit_set(vcpu, X86_CR4_PCIDE))) > return 1; > > - static_call(kvm_x86_set_cr0)(vcpu, cr0); > + KVM_X86_CALL(set_cr0, vcpu, cr0); > > kvm_post_set_cr0(vcpu, old_cr0, cr0); > > @@ -1119,7 +1119,7 @@ static int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr) > int kvm_emulate_xsetbv(struct kvm_vcpu *vcpu) > { > /* Note, #UD due to CR4.OSXSAVE=0 has priority over the intercept. */ > - if (static_call(kvm_x86_get_cpl)(vcpu) != 0 || > + if (KVM_X86_CALL(get_cpl, vcpu) != 0 || > __kvm_set_xcr(vcpu, kvm_rcx_read(vcpu), kvm_read_edx_eax(vcpu))) { > kvm_inject_gp(vcpu, 0); > return 1; > @@ -1144,7 +1144,7 @@ EXPORT_SYMBOL_GPL(__kvm_is_valid_cr4); > static bool kvm_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) > { > return __kvm_is_valid_cr4(vcpu, cr4) && > - static_call(kvm_x86_is_valid_cr4)(vcpu, cr4); > + KVM_X86_CALL(is_valid_cr4, vcpu, cr4); > } > > void kvm_post_set_cr4(struct kvm_vcpu *vcpu, unsigned long old_cr4, unsigned long cr4) > @@ -1212,7 +1212,7 @@ int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4) > return 1; > } > > - static_call(kvm_x86_set_cr4)(vcpu, cr4); > + KVM_X86_CALL(set_cr4, vcpu, cr4); > > kvm_post_set_cr4(vcpu, old_cr4, cr4); > > @@ -1351,7 +1351,7 @@ void kvm_update_dr7(struct kvm_vcpu *vcpu) > dr7 = vcpu->arch.guest_debug_dr7; > else > dr7 = vcpu->arch.dr7; > - static_call(kvm_x86_set_dr7)(vcpu, dr7); > + KVM_X86_CALL(set_dr7, vcpu, dr7); > vcpu->arch.switch_db_regs &= ~KVM_DEBUGREG_BP_ENABLED; > if (dr7 & DR7_BP_EN_MASK) > vcpu->arch.switch_db_regs |= KVM_DEBUGREG_BP_ENABLED; > @@ -1694,7 +1694,7 @@ static int kvm_get_msr_feature(struct kvm_msr_entry *msr) > rdmsrl_safe(msr->index, &msr->data); > break; > default: > - return static_call(kvm_x86_get_msr_feature)(msr); > + return KVM_X86_CALL(get_msr_feature, msr); > } > return 0; > } > @@ -1768,7 +1768,7 @@ static int set_efer(struct kvm_vcpu *vcpu, struct msr_data *msr_info) > efer &= ~EFER_LMA; > efer |= vcpu->arch.efer & EFER_LMA; > > - r = static_call(kvm_x86_set_efer)(vcpu, efer); > + r = KVM_X86_CALL(set_efer, vcpu, efer); > if (r) { > WARN_ON(r > 0); > return r; > @@ -1898,7 +1898,7 @@ static int __kvm_set_msr(struct kvm_vcpu *vcpu, u32 index, u64 data, > msr.index = index; > msr.host_initiated = host_initiated; > > - return static_call(kvm_x86_set_msr)(vcpu, &msr); > + return KVM_X86_CALL(set_msr, vcpu, &msr); > } > > static int kvm_set_msr_ignored_check(struct kvm_vcpu *vcpu, > @@ -1940,7 +1940,7 @@ int __kvm_get_msr(struct kvm_vcpu *vcpu, u32 index, u64 *data, > msr.index = index; > msr.host_initiated = host_initiated; > > - ret = static_call(kvm_x86_get_msr)(vcpu, &msr); > + ret = KVM_X86_CALL(get_msr, vcpu, &msr); > if (!ret) > *data = msr.data; > return ret; > @@ -2008,7 +2008,7 @@ static int complete_emulated_rdmsr(struct kvm_vcpu *vcpu) > > static int complete_fast_msr_access(struct kvm_vcpu *vcpu) > { > - return static_call(kvm_x86_complete_emulated_msr)(vcpu, vcpu->run->msr.error); > + return KVM_X86_CALL(complete_emulated_msr, vcpu, vcpu->run->msr.error); > } > > static int complete_fast_rdmsr(struct kvm_vcpu *vcpu) > @@ -2072,7 +2072,7 @@ int kvm_emulate_rdmsr(struct kvm_vcpu *vcpu) > trace_kvm_msr_read_ex(ecx); > } > > - return static_call(kvm_x86_complete_emulated_msr)(vcpu, r); > + return KVM_X86_CALL(complete_emulated_msr, vcpu, r); > } > EXPORT_SYMBOL_GPL(kvm_emulate_rdmsr); > > @@ -2097,7 +2097,7 @@ int kvm_emulate_wrmsr(struct kvm_vcpu *vcpu) > trace_kvm_msr_write_ex(ecx, data); > } > > - return static_call(kvm_x86_complete_emulated_msr)(vcpu, r); > + return KVM_X86_CALL(complete_emulated_msr, vcpu, r); > } > EXPORT_SYMBOL_GPL(kvm_emulate_wrmsr); > > @@ -2625,12 +2625,12 @@ static void kvm_vcpu_write_tsc_offset(struct kvm_vcpu *vcpu, u64 l1_offset) > if (is_guest_mode(vcpu)) > vcpu->arch.tsc_offset = kvm_calc_nested_tsc_offset( > l1_offset, > - static_call(kvm_x86_get_l2_tsc_offset)(vcpu), > - static_call(kvm_x86_get_l2_tsc_multiplier)(vcpu)); > + KVM_X86_CALL(get_l2_tsc_offset, vcpu), > + KVM_X86_CALL(get_l2_tsc_multiplier, vcpu)); > else > vcpu->arch.tsc_offset = l1_offset; > > - static_call(kvm_x86_write_tsc_offset)(vcpu); > + KVM_X86_CALL(write_tsc_offset, vcpu); > } > > static void kvm_vcpu_write_tsc_multiplier(struct kvm_vcpu *vcpu, u64 l1_multiplier) > @@ -2641,12 +2641,12 @@ static void kvm_vcpu_write_tsc_multiplier(struct kvm_vcpu *vcpu, u64 l1_multipli > if (is_guest_mode(vcpu)) > vcpu->arch.tsc_scaling_ratio = kvm_calc_nested_tsc_multiplier( > l1_multiplier, > - static_call(kvm_x86_get_l2_tsc_multiplier)(vcpu)); > + KVM_X86_CALL(get_l2_tsc_multiplier, vcpu)); > else > vcpu->arch.tsc_scaling_ratio = l1_multiplier; > > if (kvm_caps.has_tsc_control) > - static_call(kvm_x86_write_tsc_multiplier)(vcpu); > + KVM_X86_CALL(write_tsc_multiplier, vcpu); > } > > static inline bool kvm_check_tsc_unstable(void) > @@ -3619,7 +3619,7 @@ static void kvmclock_reset(struct kvm_vcpu *vcpu) > static void kvm_vcpu_flush_tlb_all(struct kvm_vcpu *vcpu) > { > ++vcpu->stat.tlb_flush; > - static_call(kvm_x86_flush_tlb_all)(vcpu); > + KVM_X86_CALL(flush_tlb_all, vcpu); > > /* Flushing all ASIDs flushes the current ASID... */ > kvm_clear_request(KVM_REQ_TLB_FLUSH_CURRENT, vcpu); > @@ -3640,7 +3640,7 @@ static void kvm_vcpu_flush_tlb_guest(struct kvm_vcpu *vcpu) > kvm_mmu_sync_prev_roots(vcpu); > } > > - static_call(kvm_x86_flush_tlb_guest)(vcpu); > + KVM_X86_CALL(flush_tlb_guest, vcpu); > > /* > * Flushing all "guest" TLB is always a superset of Hyper-V's fine > @@ -3653,7 +3653,7 @@ static void kvm_vcpu_flush_tlb_guest(struct kvm_vcpu *vcpu) > static inline void kvm_vcpu_flush_tlb_current(struct kvm_vcpu *vcpu) > { > ++vcpu->stat.tlb_flush; > - static_call(kvm_x86_flush_tlb_current)(vcpu); > + KVM_X86_CALL(flush_tlb_current, vcpu); > } > > /* > @@ -4764,7 +4764,7 @@ int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext) > * fringe case that is not enabled except via specific settings > * of the module parameters. > */ > - r = static_call(kvm_x86_has_emulated_msr)(kvm, MSR_IA32_SMBASE); > + r = KVM_X86_CALL(has_emulated_msr, kvm, MSR_IA32_SMBASE); > break; > case KVM_CAP_NR_VCPUS: > r = min_t(unsigned int, num_online_cpus(), KVM_MAX_VCPUS); > @@ -5012,14 +5012,14 @@ void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu) > { > /* Address WBINVD may be executed by guest */ > if (need_emulate_wbinvd(vcpu)) { > - if (static_call(kvm_x86_has_wbinvd_exit)()) > + if (KVM_X86_CALL(has_wbinvd_exit)) > cpumask_set_cpu(cpu, vcpu->arch.wbinvd_dirty_mask); > else if (vcpu->cpu != -1 && vcpu->cpu != cpu) > smp_call_function_single(vcpu->cpu, > wbinvd_ipi, NULL, 1); > } > > - static_call(kvm_x86_vcpu_load)(vcpu, cpu); > + KVM_X86_CALL(vcpu_load, vcpu, cpu); > > /* Save host pkru register if supported */ > vcpu->arch.host_pkru = read_pkru(); > @@ -5127,14 +5127,14 @@ void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu) > srcu_read_unlock(&vcpu->kvm->srcu, idx); > } > > - static_call(kvm_x86_vcpu_put)(vcpu); > + KVM_X86_CALL(vcpu_put, vcpu); > vcpu->arch.last_host_tsc = rdtsc(); > } > > static int kvm_vcpu_ioctl_get_lapic(struct kvm_vcpu *vcpu, > struct kvm_lapic_state *s) > { > - static_call(kvm_x86_sync_pir_to_irr)(vcpu); > + KVM_X86_CALL(sync_pir_to_irr, vcpu); > > return kvm_apic_get_state(vcpu, s); > } > @@ -5251,7 +5251,7 @@ static int kvm_vcpu_ioctl_x86_setup_mce(struct kvm_vcpu *vcpu, > > kvm_apic_after_set_mcg_cap(vcpu); > > - static_call(kvm_x86_setup_mce)(vcpu); > + KVM_X86_CALL(setup_mce, vcpu); > out: > return r; > } > @@ -5411,11 +5411,11 @@ static void kvm_vcpu_ioctl_x86_get_vcpu_events(struct kvm_vcpu *vcpu, > events->interrupt.injected = > vcpu->arch.interrupt.injected && !vcpu->arch.interrupt.soft; > events->interrupt.nr = vcpu->arch.interrupt.nr; > - events->interrupt.shadow = static_call(kvm_x86_get_interrupt_shadow)(vcpu); > + events->interrupt.shadow = KVM_X86_CALL(get_interrupt_shadow, vcpu); > > events->nmi.injected = vcpu->arch.nmi_injected; > events->nmi.pending = kvm_get_nr_pending_nmis(vcpu); > - events->nmi.masked = static_call(kvm_x86_get_nmi_mask)(vcpu); > + events->nmi.masked = KVM_X86_CALL(get_nmi_mask, vcpu); > > /* events->sipi_vector is never valid when reporting to user space */ > > @@ -5497,8 +5497,8 @@ static int kvm_vcpu_ioctl_x86_set_vcpu_events(struct kvm_vcpu *vcpu, > vcpu->arch.interrupt.nr = events->interrupt.nr; > vcpu->arch.interrupt.soft = events->interrupt.soft; > if (events->flags & KVM_VCPUEVENT_VALID_SHADOW) > - static_call(kvm_x86_set_interrupt_shadow)(vcpu, > - events->interrupt.shadow); > + KVM_X86_CALL(set_interrupt_shadow, > + vcpu, events->interrupt.shadow); > > vcpu->arch.nmi_injected = events->nmi.injected; > if (events->flags & KVM_VCPUEVENT_VALID_NMI_PENDING) { > @@ -5507,7 +5507,7 @@ static int kvm_vcpu_ioctl_x86_set_vcpu_events(struct kvm_vcpu *vcpu, > if (events->nmi.pending) > kvm_make_request(KVM_REQ_NMI, vcpu); > } > - static_call(kvm_x86_set_nmi_mask)(vcpu, events->nmi.masked); > + KVM_X86_CALL(set_nmi_mask, vcpu, events->nmi.masked); > > if (events->flags & KVM_VCPUEVENT_VALID_SIPI_VECTOR && > lapic_in_kernel(vcpu)) > @@ -5842,7 +5842,7 @@ static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu, > if (!kvm_x86_ops.enable_l2_tlb_flush) > return -ENOTTY; > > - return static_call(kvm_x86_enable_l2_tlb_flush)(vcpu); > + return KVM_X86_CALL(enable_l2_tlb_flush, vcpu); > > case KVM_CAP_HYPERV_ENFORCE_CPUID: > return kvm_hv_set_enforce_cpuid(vcpu, cap->args[0]); > @@ -6314,14 +6314,14 @@ static int kvm_vm_ioctl_set_tss_addr(struct kvm *kvm, unsigned long addr) > > if (addr > (unsigned int)(-3 * PAGE_SIZE)) > return -EINVAL; > - ret = static_call(kvm_x86_set_tss_addr)(kvm, addr); > + ret = KVM_X86_CALL(set_tss_addr, kvm, addr); > return ret; > } > > static int kvm_vm_ioctl_set_identity_map_addr(struct kvm *kvm, > u64 ident_addr) > { > - return static_call(kvm_x86_set_identity_map_addr)(kvm, ident_addr); > + return KVM_X86_CALL(set_identity_map_addr, kvm, ident_addr); > } > > static int kvm_vm_ioctl_set_nr_mmu_pages(struct kvm *kvm, > @@ -6634,14 +6634,14 @@ int kvm_vm_ioctl_enable_cap(struct kvm *kvm, > if (!kvm_x86_ops.vm_copy_enc_context_from) > break; > > - r = static_call(kvm_x86_vm_copy_enc_context_from)(kvm, cap->args[0]); > + r = KVM_X86_CALL(vm_copy_enc_context_from, kvm, cap->args[0]); > break; > case KVM_CAP_VM_MOVE_ENC_CONTEXT_FROM: > r = -EINVAL; > if (!kvm_x86_ops.vm_move_enc_context_from) > break; > > - r = static_call(kvm_x86_vm_move_enc_context_from)(kvm, cap->args[0]); > + r = KVM_X86_CALL(vm_move_enc_context_from, kvm, cap->args[0]); > break; > case KVM_CAP_EXIT_HYPERCALL: > if (cap->args[0] & ~KVM_EXIT_HYPERCALL_VALID_MASK) { > @@ -7273,7 +7273,7 @@ int kvm_arch_vm_ioctl(struct file *filp, unsigned int ioctl, unsigned long arg) > if (!kvm_x86_ops.mem_enc_ioctl) > goto out; > > - r = static_call(kvm_x86_mem_enc_ioctl)(kvm, argp); > + r = KVM_X86_CALL(mem_enc_ioctl, kvm, argp); > break; > } > case KVM_MEMORY_ENCRYPT_REG_REGION: { > @@ -7287,7 +7287,7 @@ int kvm_arch_vm_ioctl(struct file *filp, unsigned int ioctl, unsigned long arg) > if (!kvm_x86_ops.mem_enc_register_region) > goto out; > > - r = static_call(kvm_x86_mem_enc_register_region)(kvm, ®ion); > + r = KVM_X86_CALL(mem_enc_register_region, kvm, ®ion); > break; > } > case KVM_MEMORY_ENCRYPT_UNREG_REGION: { > @@ -7301,7 +7301,8 @@ int kvm_arch_vm_ioctl(struct file *filp, unsigned int ioctl, unsigned long arg) > if (!kvm_x86_ops.mem_enc_unregister_region) > goto out; > > - r = static_call(kvm_x86_mem_enc_unregister_region)(kvm, ®ion); > + r = KVM_X86_CALL(mem_enc_unregister_region, > + kvm, ®ion); > break; > } > #ifdef CONFIG_KVM_HYPERV > @@ -7452,7 +7453,8 @@ static void kvm_init_msr_lists(void) > } > > for (i = 0; i < ARRAY_SIZE(emulated_msrs_all); i++) { > - if (!static_call(kvm_x86_has_emulated_msr)(NULL, emulated_msrs_all[i])) > + if (!KVM_X86_CALL(has_emulated_msr, > + NULL, emulated_msrs_all[i])) > continue; > > emulated_msrs[num_emulated_msrs++] = emulated_msrs_all[i]; > @@ -7511,13 +7513,13 @@ static int vcpu_mmio_read(struct kvm_vcpu *vcpu, gpa_t addr, int len, void *v) > void kvm_set_segment(struct kvm_vcpu *vcpu, > struct kvm_segment *var, int seg) > { > - static_call(kvm_x86_set_segment)(vcpu, var, seg); > + KVM_X86_CALL(set_segment, vcpu, var, seg); > } > > void kvm_get_segment(struct kvm_vcpu *vcpu, > struct kvm_segment *var, int seg) > { > - static_call(kvm_x86_get_segment)(vcpu, var, seg); > + KVM_X86_CALL(get_segment, vcpu, var, seg); > } > > gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u64 access, > @@ -7540,7 +7542,7 @@ gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva, > { > struct kvm_mmu *mmu = vcpu->arch.walk_mmu; > > - u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; > + u64 access = (KVM_X86_CALL(get_cpl, vcpu) == 3) ? PFERR_USER_MASK : 0; > return mmu->gva_to_gpa(vcpu, mmu, gva, access, exception); > } > EXPORT_SYMBOL_GPL(kvm_mmu_gva_to_gpa_read); > @@ -7550,7 +7552,7 @@ gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva, > { > struct kvm_mmu *mmu = vcpu->arch.walk_mmu; > > - u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; > + u64 access = (KVM_X86_CALL(get_cpl, vcpu) == 3) ? PFERR_USER_MASK : 0; > access |= PFERR_WRITE_MASK; > return mmu->gva_to_gpa(vcpu, mmu, gva, access, exception); > } > @@ -7603,7 +7605,7 @@ static int kvm_fetch_guest_virt(struct x86_emulate_ctxt *ctxt, > { > struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt); > struct kvm_mmu *mmu = vcpu->arch.walk_mmu; > - u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; > + u64 access = (KVM_X86_CALL(get_cpl, vcpu) == 3) ? PFERR_USER_MASK : 0; > unsigned offset; > int ret; > > @@ -7628,7 +7630,7 @@ int kvm_read_guest_virt(struct kvm_vcpu *vcpu, > gva_t addr, void *val, unsigned int bytes, > struct x86_exception *exception) > { > - u64 access = (static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0; > + u64 access = (KVM_X86_CALL(get_cpl, vcpu) == 3) ? PFERR_USER_MASK : 0; > > /* > * FIXME: this should call handle_emulation_failure if X86EMUL_IO_NEEDED > @@ -7651,7 +7653,7 @@ static int emulator_read_std(struct x86_emulate_ctxt *ctxt, > > if (system) > access |= PFERR_IMPLICIT_ACCESS; > - else if (static_call(kvm_x86_get_cpl)(vcpu) == 3) > + else if (KVM_X86_CALL(get_cpl, vcpu) == 3) > access |= PFERR_USER_MASK; > > return kvm_read_guest_virt_helper(addr, val, bytes, vcpu, access, exception); > @@ -7696,7 +7698,7 @@ static int emulator_write_std(struct x86_emulate_ctxt *ctxt, gva_t addr, void *v > > if (system) > access |= PFERR_IMPLICIT_ACCESS; > - else if (static_call(kvm_x86_get_cpl)(vcpu) == 3) > + else if (KVM_X86_CALL(get_cpl, vcpu) == 3) > access |= PFERR_USER_MASK; > > return kvm_write_guest_virt_helper(addr, val, bytes, vcpu, > @@ -7717,8 +7719,8 @@ EXPORT_SYMBOL_GPL(kvm_write_guest_virt_system); > static int kvm_check_emulate_insn(struct kvm_vcpu *vcpu, int emul_type, > void *insn, int insn_len) > { > - return static_call(kvm_x86_check_emulate_instruction)(vcpu, emul_type, > - insn, insn_len); > + return KVM_X86_CALL(check_emulate_instruction, > + vcpu, emul_type, insn, insn_len); > } > > int handle_ud(struct kvm_vcpu *vcpu) > @@ -7768,8 +7770,9 @@ static int vcpu_mmio_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva, > bool write) > { > struct kvm_mmu *mmu = vcpu->arch.walk_mmu; > - u64 access = ((static_call(kvm_x86_get_cpl)(vcpu) == 3) ? PFERR_USER_MASK : 0) > - | (write ? PFERR_WRITE_MASK : 0); > + u64 access = (KVM_X86_CALL(get_cpl, vcpu) == 3) ? PFERR_USER_MASK : 0; > + > + access |= (write ? PFERR_WRITE_MASK : 0); > > /* > * currently PKRU is only applied to ept enabled guest so > @@ -8195,7 +8198,7 @@ static int emulator_pio_out_emulated(struct x86_emulate_ctxt *ctxt, > > static unsigned long get_segment_base(struct kvm_vcpu *vcpu, int seg) > { > - return static_call(kvm_x86_get_segment_base)(vcpu, seg); > + return KVM_X86_CALL(get_segment_base, vcpu, seg); > } > > static void emulator_invlpg(struct x86_emulate_ctxt *ctxt, ulong address) > @@ -8208,7 +8211,7 @@ static int kvm_emulate_wbinvd_noskip(struct kvm_vcpu *vcpu) > if (!need_emulate_wbinvd(vcpu)) > return X86EMUL_CONTINUE; > > - if (static_call(kvm_x86_has_wbinvd_exit)()) { > + if (KVM_X86_CALL(has_wbinvd_exit)) { > int cpu = get_cpu(); > > cpumask_set_cpu(cpu, vcpu->arch.wbinvd_dirty_mask); > @@ -8312,27 +8315,27 @@ static int emulator_set_cr(struct x86_emulate_ctxt *ctxt, int cr, ulong val) > > static int emulator_get_cpl(struct x86_emulate_ctxt *ctxt) > { > - return static_call(kvm_x86_get_cpl)(emul_to_vcpu(ctxt)); > + return KVM_X86_CALL(get_cpl, emul_to_vcpu(ctxt)); > } > > static void emulator_get_gdt(struct x86_emulate_ctxt *ctxt, struct desc_ptr *dt) > { > - static_call(kvm_x86_get_gdt)(emul_to_vcpu(ctxt), dt); > + KVM_X86_CALL(get_gdt, emul_to_vcpu(ctxt), dt); > } > > static void emulator_get_idt(struct x86_emulate_ctxt *ctxt, struct desc_ptr *dt) > { > - static_call(kvm_x86_get_idt)(emul_to_vcpu(ctxt), dt); > + KVM_X86_CALL(get_idt, emul_to_vcpu(ctxt), dt); > } > > static void emulator_set_gdt(struct x86_emulate_ctxt *ctxt, struct desc_ptr *dt) > { > - static_call(kvm_x86_set_gdt)(emul_to_vcpu(ctxt), dt); > + KVM_X86_CALL(set_gdt, emul_to_vcpu(ctxt), dt); > } > > static void emulator_set_idt(struct x86_emulate_ctxt *ctxt, struct desc_ptr *dt) > { > - static_call(kvm_x86_set_idt)(emul_to_vcpu(ctxt), dt); > + KVM_X86_CALL(set_idt, emul_to_vcpu(ctxt), dt); > } > > static unsigned long emulator_get_cached_segment_base( > @@ -8479,8 +8482,8 @@ static int emulator_intercept(struct x86_emulate_ctxt *ctxt, > struct x86_instruction_info *info, > enum x86_intercept_stage stage) > { > - return static_call(kvm_x86_check_intercept)(emul_to_vcpu(ctxt), info, stage, > - &ctxt->exception); > + return KVM_X86_CALL(check_intercept, > + emul_to_vcpu(ctxt), info, stage, &ctxt->exception); > } > > static bool emulator_get_cpuid(struct x86_emulate_ctxt *ctxt, > @@ -8517,7 +8520,7 @@ static void emulator_write_gpr(struct x86_emulate_ctxt *ctxt, unsigned reg, ulon > > static void emulator_set_nmi_mask(struct x86_emulate_ctxt *ctxt, bool masked) > { > - static_call(kvm_x86_set_nmi_mask)(emul_to_vcpu(ctxt), masked); > + KVM_X86_CALL(set_nmi_mask, emul_to_vcpu(ctxt), masked); > } > > static bool emulator_is_smm(struct x86_emulate_ctxt *ctxt) > @@ -8562,7 +8565,8 @@ static gva_t emulator_get_untagged_addr(struct x86_emulate_ctxt *ctxt, > if (!kvm_x86_ops.get_untagged_addr) > return addr; > > - return static_call(kvm_x86_get_untagged_addr)(emul_to_vcpu(ctxt), addr, flags); > + return KVM_X86_CALL(get_untagged_addr, > + emul_to_vcpu(ctxt), addr, flags); > } > > static const struct x86_emulate_ops emulate_ops = { > @@ -8614,7 +8618,7 @@ static const struct x86_emulate_ops emulate_ops = { > > static void toggle_interruptibility(struct kvm_vcpu *vcpu, u32 mask) > { > - u32 int_shadow = static_call(kvm_x86_get_interrupt_shadow)(vcpu); > + u32 int_shadow = KVM_X86_CALL(get_interrupt_shadow, vcpu); > /* > * an sti; sti; sequence only disable interrupts for the first > * instruction. So, if the last instruction, be it emulated or > @@ -8625,7 +8629,7 @@ static void toggle_interruptibility(struct kvm_vcpu *vcpu, u32 mask) > if (int_shadow & mask) > mask = 0; > if (unlikely(int_shadow || mask)) { > - static_call(kvm_x86_set_interrupt_shadow)(vcpu, mask); > + KVM_X86_CALL(set_interrupt_shadow, vcpu, mask); > if (!mask) > kvm_make_request(KVM_REQ_EVENT, vcpu); > } > @@ -8666,7 +8670,7 @@ static void init_emulate_ctxt(struct kvm_vcpu *vcpu) > struct x86_emulate_ctxt *ctxt = vcpu->arch.emulate_ctxt; > int cs_db, cs_l; > > - static_call(kvm_x86_get_cs_db_l_bits)(vcpu, &cs_db, &cs_l); > + KVM_X86_CALL(get_cs_db_l_bits, vcpu, &cs_db, &cs_l); > > ctxt->gpa_available = false; > ctxt->eflags = kvm_get_rflags(vcpu); > @@ -8722,9 +8726,8 @@ static void prepare_emulation_failure_exit(struct kvm_vcpu *vcpu, u64 *data, > */ > memset(&info, 0, sizeof(info)); > > - static_call(kvm_x86_get_exit_info)(vcpu, (u32 *)&info[0], &info[1], > - &info[2], (u32 *)&info[3], > - (u32 *)&info[4]); > + KVM_X86_CALL(get_exit_info, vcpu, (u32 *)&info[0], &info[1], > + &info[2], (u32 *)&info[3], (u32 *)&info[4]); > > run->exit_reason = KVM_EXIT_INTERNAL_ERROR; > run->emulation_failure.suberror = KVM_INTERNAL_ERROR_EMULATION; > @@ -8801,7 +8804,7 @@ static int handle_emulation_failure(struct kvm_vcpu *vcpu, int emulation_type) > > kvm_queue_exception(vcpu, UD_VECTOR); > > - if (!is_guest_mode(vcpu) && static_call(kvm_x86_get_cpl)(vcpu) == 0) { > + if (!is_guest_mode(vcpu) && KVM_X86_CALL(get_cpl, vcpu) == 0) { > prepare_emulation_ctxt_failure_exit(vcpu); > return 0; > } > @@ -8959,10 +8962,10 @@ static int kvm_vcpu_do_singlestep(struct kvm_vcpu *vcpu) > > int kvm_skip_emulated_instruction(struct kvm_vcpu *vcpu) > { > - unsigned long rflags = static_call(kvm_x86_get_rflags)(vcpu); > + unsigned long rflags = KVM_X86_CALL(get_rflags, vcpu); > int r; > > - r = static_call(kvm_x86_skip_emulated_instruction)(vcpu); > + r = KVM_X86_CALL(skip_emulated_instruction, vcpu); > if (unlikely(!r)) > return 0; > > @@ -8994,7 +8997,7 @@ static bool kvm_is_code_breakpoint_inhibited(struct kvm_vcpu *vcpu) > * but AMD CPUs do not. MOV/POP SS blocking is rare, check that first > * to avoid the relatively expensive CPUID lookup. > */ > - shadow = static_call(kvm_x86_get_interrupt_shadow)(vcpu); > + shadow = KVM_X86_CALL(get_interrupt_shadow, vcpu); > return (shadow & KVM_X86_SHADOW_INT_MOV_SS) && > guest_cpuid_is_intel(vcpu); > } > @@ -9268,7 +9271,7 @@ int x86_emulate_instruction(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, > > writeback: > if (writeback) { > - unsigned long rflags = static_call(kvm_x86_get_rflags)(vcpu); > + unsigned long rflags = KVM_X86_CALL(get_rflags, vcpu); > toggle_interruptibility(vcpu, ctxt->interruptibility); > vcpu->arch.emulate_regs_need_sync_to_vcpu = false; > > @@ -9285,7 +9288,7 @@ int x86_emulate_instruction(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, > kvm_rip_write(vcpu, ctxt->eip); > if (r && (ctxt->tf || (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP))) > r = kvm_vcpu_do_singlestep(vcpu); > - static_call(kvm_x86_update_emulated_instruction)(vcpu); > + KVM_X86_CALL(update_emulated_instruction, vcpu); > __kvm_set_rflags(vcpu, ctxt->eflags); > } > > @@ -9684,7 +9687,7 @@ static int kvm_x86_check_processor_compatibility(void) > __cr4_reserved_bits(cpu_has, &boot_cpu_data)) > return -EIO; > > - return static_call(kvm_x86_check_processor_compatibility)(); > + return KVM_X86_CALL(check_processor_compatibility); > } > > static void kvm_x86_check_cpu_compat(void *ret) > @@ -9819,7 +9822,7 @@ int kvm_x86_vendor_init(struct kvm_x86_init_ops *ops) > > out_unwind_ops: > kvm_x86_ops.hardware_enable = NULL; > - static_call(kvm_x86_hardware_unsetup)(); > + KVM_X86_CALL(hardware_unsetup); > out_mmu_exit: > kvm_mmu_vendor_module_exit(); > out_free_percpu: > @@ -9850,7 +9853,7 @@ void kvm_x86_vendor_exit(void) > irq_work_sync(&pvclock_irq_work); > cancel_work_sync(&pvclock_gtod_work); > #endif > - static_call(kvm_x86_hardware_unsetup)(); > + KVM_X86_CALL(hardware_unsetup); > kvm_mmu_vendor_module_exit(); > free_percpu(user_return_msrs); > kmem_cache_destroy(x86_emulator_cache); > @@ -9976,7 +9979,8 @@ EXPORT_SYMBOL_GPL(kvm_apicv_activated); > bool kvm_vcpu_apicv_activated(struct kvm_vcpu *vcpu) > { > ulong vm_reasons = READ_ONCE(vcpu->kvm->arch.apicv_inhibit_reasons); > - ulong vcpu_reasons = static_call(kvm_x86_vcpu_get_apicv_inhibit_reasons)(vcpu); > + ulong vcpu_reasons = > + KVM_X86_CALL(vcpu_get_apicv_inhibit_reasons, vcpu); > > return (vm_reasons | vcpu_reasons) == 0; > } > @@ -10079,7 +10083,7 @@ int kvm_emulate_hypercall(struct kvm_vcpu *vcpu) > a3 &= 0xFFFFFFFF; > } > > - if (static_call(kvm_x86_get_cpl)(vcpu) != 0) { > + if (KVM_X86_CALL(get_cpl, vcpu) != 0) { > ret = -KVM_EPERM; > goto out; > } > @@ -10173,7 +10177,7 @@ static int emulator_fix_hypercall(struct x86_emulate_ctxt *ctxt) > return X86EMUL_PROPAGATE_FAULT; > } > > - static_call(kvm_x86_patch_hypercall)(vcpu, instruction); > + KVM_X86_CALL(patch_hypercall, vcpu, instruction); > > return emulator_write_emulated(ctxt, rip, instruction, 3, > &ctxt->exception); > @@ -10190,7 +10194,7 @@ static void post_kvm_run_save(struct kvm_vcpu *vcpu) > { > struct kvm_run *kvm_run = vcpu->run; > > - kvm_run->if_flag = static_call(kvm_x86_get_if_flag)(vcpu); > + kvm_run->if_flag = KVM_X86_CALL(get_if_flag, vcpu); > kvm_run->cr8 = kvm_get_cr8(vcpu); > kvm_run->apic_base = kvm_get_apic_base(vcpu); > > @@ -10225,7 +10229,7 @@ static void update_cr8_intercept(struct kvm_vcpu *vcpu) > > tpr = kvm_lapic_get_cr8(vcpu); > > - static_call(kvm_x86_update_cr8_intercept)(vcpu, tpr, max_irr); > + KVM_X86_CALL(update_cr8_intercept, vcpu, tpr, max_irr); > } > > > @@ -10255,7 +10259,7 @@ static void kvm_inject_exception(struct kvm_vcpu *vcpu) > vcpu->arch.exception.error_code, > vcpu->arch.exception.injected); > > - static_call(kvm_x86_inject_exception)(vcpu); > + KVM_X86_CALL(inject_exception, vcpu); > } > > /* > @@ -10341,9 +10345,9 @@ static int kvm_check_and_inject_events(struct kvm_vcpu *vcpu, > else if (kvm_is_exception_pending(vcpu)) > ; /* see above */ > else if (vcpu->arch.nmi_injected) > - static_call(kvm_x86_inject_nmi)(vcpu); > + KVM_X86_CALL(inject_nmi, vcpu); > else if (vcpu->arch.interrupt.injected) > - static_call(kvm_x86_inject_irq)(vcpu, true); > + KVM_X86_CALL(inject_irq, vcpu, true); > > /* > * Exceptions that morph to VM-Exits are handled above, and pending > @@ -10428,7 +10432,8 @@ static int kvm_check_and_inject_events(struct kvm_vcpu *vcpu, > */ > #ifdef CONFIG_KVM_SMM > if (vcpu->arch.smi_pending) { > - r = can_inject ? static_call(kvm_x86_smi_allowed)(vcpu, true) : -EBUSY; > + r = can_inject ? KVM_X86_CALL(smi_allowed, vcpu, true) : > + -EBUSY; > if (r < 0) > goto out; > if (r) { > @@ -10437,27 +10442,29 @@ static int kvm_check_and_inject_events(struct kvm_vcpu *vcpu, > enter_smm(vcpu); > can_inject = false; > } else > - static_call(kvm_x86_enable_smi_window)(vcpu); > + KVM_X86_CALL(enable_smi_window, vcpu); > } > #endif > > if (vcpu->arch.nmi_pending) { > - r = can_inject ? static_call(kvm_x86_nmi_allowed)(vcpu, true) : -EBUSY; > + r = can_inject ? KVM_X86_CALL(nmi_allowed, vcpu, true) : > + -EBUSY; > if (r < 0) > goto out; > if (r) { > --vcpu->arch.nmi_pending; > vcpu->arch.nmi_injected = true; > - static_call(kvm_x86_inject_nmi)(vcpu); > + KVM_X86_CALL(inject_nmi, vcpu); > can_inject = false; > - WARN_ON(static_call(kvm_x86_nmi_allowed)(vcpu, true) < 0); > + WARN_ON(KVM_X86_CALL(nmi_allowed, vcpu, true) < 0); > } > if (vcpu->arch.nmi_pending) > - static_call(kvm_x86_enable_nmi_window)(vcpu); > + KVM_X86_CALL(enable_nmi_window, vcpu); > } > > if (kvm_cpu_has_injectable_intr(vcpu)) { > - r = can_inject ? static_call(kvm_x86_interrupt_allowed)(vcpu, true) : -EBUSY; > + r = can_inject ? KVM_X86_CALL(interrupt_allowed, vcpu, true) : > + -EBUSY; > if (r < 0) > goto out; > if (r) { > @@ -10465,12 +10472,13 @@ static int kvm_check_and_inject_events(struct kvm_vcpu *vcpu, > > if (!WARN_ON_ONCE(irq == -1)) { > kvm_queue_interrupt(vcpu, irq, false); > - static_call(kvm_x86_inject_irq)(vcpu, false); > - WARN_ON(static_call(kvm_x86_interrupt_allowed)(vcpu, true) < 0); > + KVM_X86_CALL(inject_irq, vcpu, false); > + WARN_ON(KVM_X86_CALL(interrupt_allowed, > + vcpu, true) < 0); > } > } > if (kvm_cpu_has_injectable_intr(vcpu)) > - static_call(kvm_x86_enable_irq_window)(vcpu); > + KVM_X86_CALL(enable_irq_window, vcpu); > } > > if (is_guest_mode(vcpu) && > @@ -10516,7 +10524,7 @@ static void process_nmi(struct kvm_vcpu *vcpu) > * blocks NMIs). KVM will immediately inject one of the two NMIs, and > * will request an NMI window to handle the second NMI. > */ > - if (static_call(kvm_x86_get_nmi_mask)(vcpu) || vcpu->arch.nmi_injected) > + if (KVM_X86_CALL(get_nmi_mask, vcpu) || vcpu->arch.nmi_injected) > limit = 1; > else > limit = 2; > @@ -10525,14 +10533,14 @@ static void process_nmi(struct kvm_vcpu *vcpu) > * Adjust the limit to account for pending virtual NMIs, which aren't > * tracked in vcpu->arch.nmi_pending. > */ > - if (static_call(kvm_x86_is_vnmi_pending)(vcpu)) > + if (KVM_X86_CALL(is_vnmi_pending, vcpu)) > limit--; > > vcpu->arch.nmi_pending += atomic_xchg(&vcpu->arch.nmi_queued, 0); > vcpu->arch.nmi_pending = min(vcpu->arch.nmi_pending, limit); > > if (vcpu->arch.nmi_pending && > - (static_call(kvm_x86_set_vnmi_pending)(vcpu))) > + (KVM_X86_CALL(set_vnmi_pending, vcpu))) > vcpu->arch.nmi_pending--; > > if (vcpu->arch.nmi_pending) > @@ -10542,8 +10550,7 @@ static void process_nmi(struct kvm_vcpu *vcpu) > /* Return total number of NMIs pending injection to the VM */ > int kvm_get_nr_pending_nmis(struct kvm_vcpu *vcpu) > { > - return vcpu->arch.nmi_pending + > - static_call(kvm_x86_is_vnmi_pending)(vcpu); > + return vcpu->arch.nmi_pending + KVM_X86_CALL(is_vnmi_pending, vcpu); > } > > void kvm_make_scan_ioapic_request_mask(struct kvm *kvm, > @@ -10577,7 +10584,7 @@ void __kvm_vcpu_update_apicv(struct kvm_vcpu *vcpu) > > apic->apicv_active = activate; > kvm_apic_update_apicv(vcpu); > - static_call(kvm_x86_refresh_apicv_exec_ctrl)(vcpu); > + KVM_X86_CALL(refresh_apicv_exec_ctrl, vcpu); > > /* > * When APICv gets disabled, we may still have injected interrupts > @@ -10680,7 +10687,7 @@ static void vcpu_scan_ioapic(struct kvm_vcpu *vcpu) > if (irqchip_split(vcpu->kvm)) > kvm_scan_ioapic_routes(vcpu, vcpu->arch.ioapic_handled_vectors); > else { > - static_call(kvm_x86_sync_pir_to_irr)(vcpu); > + KVM_X86_CALL(sync_pir_to_irr, vcpu); > if (ioapic_in_kernel(vcpu->kvm)) > kvm_ioapic_scan_entry(vcpu, vcpu->arch.ioapic_handled_vectors); > } > @@ -10703,17 +10710,17 @@ static void vcpu_load_eoi_exitmap(struct kvm_vcpu *vcpu) > bitmap_or((ulong *)eoi_exit_bitmap, > vcpu->arch.ioapic_handled_vectors, > to_hv_synic(vcpu)->vec_bitmap, 256); > - static_call(kvm_x86_load_eoi_exitmap)(vcpu, eoi_exit_bitmap); > + KVM_X86_CALL(load_eoi_exitmap, vcpu, eoi_exit_bitmap); > return; > } > #endif > - static_call(kvm_x86_load_eoi_exitmap)( > - vcpu, (u64 *)vcpu->arch.ioapic_handled_vectors); > + KVM_X86_CALL(load_eoi_exitmap, vcpu, > + (u64 *)vcpu->arch.ioapic_handled_vectors); > } > > void kvm_arch_guest_memory_reclaimed(struct kvm *kvm) > { > - static_call(kvm_x86_guest_memory_reclaimed)(kvm); > + KVM_X86_CALL(guest_memory_reclaimed, kvm); > } > > static void kvm_vcpu_reload_apic_access_page(struct kvm_vcpu *vcpu) > @@ -10721,7 +10728,7 @@ static void kvm_vcpu_reload_apic_access_page(struct kvm_vcpu *vcpu) > if (!lapic_in_kernel(vcpu)) > return; > > - static_call(kvm_x86_set_apic_access_page_addr)(vcpu); > + KVM_X86_CALL(set_apic_access_page_addr, vcpu); > } > > /* > @@ -10885,10 +10892,10 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu) > if (kvm_check_request(KVM_REQ_APF_READY, vcpu)) > kvm_check_async_pf_completion(vcpu); > if (kvm_check_request(KVM_REQ_MSR_FILTER_CHANGED, vcpu)) > - static_call(kvm_x86_msr_filter_changed)(vcpu); > + KVM_X86_CALL(msr_filter_changed, vcpu); > > if (kvm_check_request(KVM_REQ_UPDATE_CPU_DIRTY_LOGGING, vcpu)) > - static_call(kvm_x86_update_cpu_dirty_logging)(vcpu); > + KVM_X86_CALL(update_cpu_dirty_logging, vcpu); > } > > if (kvm_check_request(KVM_REQ_EVENT, vcpu) || req_int_win || > @@ -10910,7 +10917,7 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu) > goto out; > } > if (req_int_win) > - static_call(kvm_x86_enable_irq_window)(vcpu); > + KVM_X86_CALL(enable_irq_window, vcpu); > > if (kvm_lapic_enabled(vcpu)) { > update_cr8_intercept(vcpu); > @@ -10925,7 +10932,7 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu) > > preempt_disable(); > > - static_call(kvm_x86_prepare_switch_to_guest)(vcpu); > + KVM_X86_CALL(prepare_switch_to_guest, vcpu); > > /* > * Disable IRQs before setting IN_GUEST_MODE. Posted interrupt > @@ -10961,7 +10968,7 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu) > * i.e. they can post interrupts even if APICv is temporarily disabled. > */ > if (kvm_lapic_enabled(vcpu)) > - static_call(kvm_x86_sync_pir_to_irr)(vcpu); > + KVM_X86_CALL(sync_pir_to_irr, vcpu); > > if (kvm_vcpu_exit_request(vcpu)) { > vcpu->mode = OUTSIDE_GUEST_MODE; > @@ -11005,12 +11012,13 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu) > WARN_ON_ONCE((kvm_vcpu_apicv_activated(vcpu) != kvm_vcpu_apicv_active(vcpu)) && > (kvm_get_apic_mode(vcpu) != LAPIC_MODE_DISABLED)); > > - exit_fastpath = static_call(kvm_x86_vcpu_run)(vcpu, req_immediate_exit); > + exit_fastpath = KVM_X86_CALL(vcpu_run, > + vcpu, req_immediate_exit); > if (likely(exit_fastpath != EXIT_FASTPATH_REENTER_GUEST)) > break; > > if (kvm_lapic_enabled(vcpu)) > - static_call(kvm_x86_sync_pir_to_irr)(vcpu); > + KVM_X86_CALL(sync_pir_to_irr, vcpu); > > if (unlikely(kvm_vcpu_exit_request(vcpu))) { > exit_fastpath = EXIT_FASTPATH_EXIT_HANDLED; > @@ -11029,7 +11037,7 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu) > */ > if (unlikely(vcpu->arch.switch_db_regs & KVM_DEBUGREG_WONT_EXIT)) { > WARN_ON(vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP); > - static_call(kvm_x86_sync_dirty_debug_regs)(vcpu); > + KVM_X86_CALL(sync_dirty_debug_regs, vcpu); > kvm_update_dr0123(vcpu); > kvm_update_dr7(vcpu); > } > @@ -11058,7 +11066,7 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu) > if (vcpu->arch.xfd_no_write_intercept) > fpu_sync_guest_vmexit_xfd_state(); > > - static_call(kvm_x86_handle_exit_irqoff)(vcpu); > + KVM_X86_CALL(handle_exit_irqoff, vcpu); > > if (vcpu->arch.guest_fpu.xfd_err) > wrmsrl(MSR_IA32_XFD_ERR, 0); > @@ -11104,13 +11112,13 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu) > if (vcpu->arch.apic_attention) > kvm_lapic_sync_from_vapic(vcpu); > > - r = static_call(kvm_x86_handle_exit)(vcpu, exit_fastpath); > + r = KVM_X86_CALL(handle_exit, vcpu, exit_fastpath); > return r; > > cancel_injection: > if (req_immediate_exit) > kvm_make_request(KVM_REQ_EVENT, vcpu); > - static_call(kvm_x86_cancel_injection)(vcpu); > + KVM_X86_CALL(cancel_injection, vcpu); > if (unlikely(vcpu->arch.apic_attention)) > kvm_lapic_sync_from_vapic(vcpu); > out: > @@ -11430,7 +11438,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu) > goto out; > } > > - r = static_call(kvm_x86_vcpu_pre_run)(vcpu); > + r = KVM_X86_CALL(vcpu_pre_run, vcpu); > if (r <= 0) > goto out; > > @@ -11550,10 +11558,10 @@ static void __get_sregs_common(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs) > kvm_get_segment(vcpu, &sregs->tr, VCPU_SREG_TR); > kvm_get_segment(vcpu, &sregs->ldt, VCPU_SREG_LDTR); > > - static_call(kvm_x86_get_idt)(vcpu, &dt); > + KVM_X86_CALL(get_idt, vcpu, &dt); > sregs->idt.limit = dt.size; > sregs->idt.base = dt.address; > - static_call(kvm_x86_get_gdt)(vcpu, &dt); > + KVM_X86_CALL(get_gdt, vcpu, &dt); > sregs->gdt.limit = dt.size; > sregs->gdt.base = dt.address; > > @@ -11749,27 +11757,27 @@ static int __set_sregs_common(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs, > > dt.size = sregs->idt.limit; > dt.address = sregs->idt.base; > - static_call(kvm_x86_set_idt)(vcpu, &dt); > + KVM_X86_CALL(set_idt, vcpu, &dt); > dt.size = sregs->gdt.limit; > dt.address = sregs->gdt.base; > - static_call(kvm_x86_set_gdt)(vcpu, &dt); > + KVM_X86_CALL(set_gdt, vcpu, &dt); > > vcpu->arch.cr2 = sregs->cr2; > *mmu_reset_needed |= kvm_read_cr3(vcpu) != sregs->cr3; > vcpu->arch.cr3 = sregs->cr3; > kvm_register_mark_dirty(vcpu, VCPU_EXREG_CR3); > - static_call(kvm_x86_post_set_cr3)(vcpu, sregs->cr3); > + KVM_X86_CALL(post_set_cr3, vcpu, sregs->cr3); > > kvm_set_cr8(vcpu, sregs->cr8); > > *mmu_reset_needed |= vcpu->arch.efer != sregs->efer; > - static_call(kvm_x86_set_efer)(vcpu, sregs->efer); > + KVM_X86_CALL(set_efer, vcpu, sregs->efer); > > *mmu_reset_needed |= kvm_read_cr0(vcpu) != sregs->cr0; > - static_call(kvm_x86_set_cr0)(vcpu, sregs->cr0); > + KVM_X86_CALL(set_cr0, vcpu, sregs->cr0); > > *mmu_reset_needed |= kvm_read_cr4(vcpu) != sregs->cr4; > - static_call(kvm_x86_set_cr4)(vcpu, sregs->cr4); > + KVM_X86_CALL(set_cr4, vcpu, sregs->cr4); > > if (update_pdptrs) { > idx = srcu_read_lock(&vcpu->kvm->srcu); > @@ -11943,7 +11951,7 @@ int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu, > */ > kvm_set_rflags(vcpu, rflags); > > - static_call(kvm_x86_update_exception_bitmap)(vcpu); > + KVM_X86_CALL(update_exception_bitmap, vcpu); > > kvm_arch_vcpu_guestdbg_update_apicv_inhibit(vcpu->kvm); > > @@ -12080,7 +12088,7 @@ int kvm_arch_vcpu_precreate(struct kvm *kvm, unsigned int id) > if (id >= kvm->arch.max_vcpu_ids) > return -EINVAL; > > - return static_call(kvm_x86_vcpu_precreate)(kvm); > + return KVM_X86_CALL(vcpu_precreate, kvm); > } > > int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu) > @@ -12151,7 +12159,7 @@ int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu) > vcpu->arch.hv_root_tdp = INVALID_PAGE; > #endif > > - r = static_call(kvm_x86_vcpu_create)(vcpu); > + r = KVM_X86_CALL(vcpu_create, vcpu); > if (r) > goto free_guest_fpu; > > @@ -12209,7 +12217,7 @@ void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu) > > kvmclock_reset(vcpu); > > - static_call(kvm_x86_vcpu_free)(vcpu); > + KVM_X86_CALL(vcpu_free, vcpu); > > kmem_cache_free(x86_emulator_cache, vcpu->arch.emulate_ctxt); > free_cpumask_var(vcpu->arch.wbinvd_dirty_mask); > @@ -12327,7 +12335,7 @@ void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) > cpuid_0x1 = kvm_find_cpuid_entry(vcpu, 1); > kvm_rdx_write(vcpu, cpuid_0x1 ? cpuid_0x1->eax : 0x600); > > - static_call(kvm_x86_vcpu_reset)(vcpu, init_event); > + KVM_X86_CALL(vcpu_reset, vcpu, init_event); > > kvm_set_rflags(vcpu, X86_EFLAGS_FIXED); > kvm_rip_write(vcpu, 0xfff0); > @@ -12346,10 +12354,10 @@ void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) > else > new_cr0 |= X86_CR0_NW | X86_CR0_CD; > > - static_call(kvm_x86_set_cr0)(vcpu, new_cr0); > - static_call(kvm_x86_set_cr4)(vcpu, 0); > - static_call(kvm_x86_set_efer)(vcpu, 0); > - static_call(kvm_x86_update_exception_bitmap)(vcpu); > + KVM_X86_CALL(set_cr0, vcpu, new_cr0); > + KVM_X86_CALL(set_cr4, vcpu, 0); > + KVM_X86_CALL(set_efer, vcpu, 0); > + KVM_X86_CALL(update_exception_bitmap, vcpu); > > /* > * On the standard CR0/CR4/EFER modification paths, there are several > @@ -12406,7 +12414,7 @@ int kvm_arch_hardware_enable(void) > if (ret) > return ret; > > - ret = static_call(kvm_x86_hardware_enable)(); > + ret = KVM_X86_CALL(hardware_enable); > if (ret != 0) > return ret; > > @@ -12488,7 +12496,7 @@ int kvm_arch_hardware_enable(void) > > void kvm_arch_hardware_disable(void) > { > - static_call(kvm_x86_hardware_disable)(); > + KVM_X86_CALL(hardware_disable); > drop_user_return_notifiers(); > } > > @@ -12511,7 +12519,7 @@ void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) > pmu->need_cleanup = true; > kvm_make_request(KVM_REQ_PMU, vcpu); > } > - static_call(kvm_x86_sched_in)(vcpu, cpu); > + KVM_X86_CALL(sched_in, vcpu, cpu); > } > > void kvm_arch_free_vm(struct kvm *kvm) > @@ -12539,7 +12547,7 @@ int kvm_arch_init_vm(struct kvm *kvm, unsigned long type) > > kvm_mmu_init_vm(kvm); > > - ret = static_call(kvm_x86_vm_init)(kvm); > + ret = KVM_X86_CALL(vm_init, kvm); > if (ret) > goto out_uninit_mmu; > > @@ -12713,7 +12721,7 @@ void kvm_arch_destroy_vm(struct kvm *kvm) > mutex_unlock(&kvm->slots_lock); > } > kvm_unload_vcpu_mmus(kvm); > - static_call(kvm_x86_vm_destroy)(kvm); > + KVM_X86_CALL(vm_destroy, kvm); > kvm_free_msr_filter(srcu_dereference_check(kvm->arch.msr_filter, &kvm->srcu, 1)); > kvm_pic_destroy(kvm); > kvm_ioapic_destroy(kvm); > @@ -13045,7 +13053,7 @@ void kvm_arch_commit_memory_region(struct kvm *kvm, > static inline bool kvm_guest_apic_has_interrupt(struct kvm_vcpu *vcpu) > { > return (is_guest_mode(vcpu) && > - static_call(kvm_x86_guest_apic_has_interrupt)(vcpu)); > + KVM_X86_CALL(guest_apic_has_interrupt, vcpu)); > } > > static inline bool kvm_vcpu_has_events(struct kvm_vcpu *vcpu) > @@ -13064,14 +13072,12 @@ static inline bool kvm_vcpu_has_events(struct kvm_vcpu *vcpu) > return true; > > if (kvm_test_request(KVM_REQ_NMI, vcpu) || > - (vcpu->arch.nmi_pending && > - static_call(kvm_x86_nmi_allowed)(vcpu, false))) > + (vcpu->arch.nmi_pending && KVM_X86_CALL(nmi_allowed, vcpu, false))) > return true; > > #ifdef CONFIG_KVM_SMM > if (kvm_test_request(KVM_REQ_SMI, vcpu) || > - (vcpu->arch.smi_pending && > - static_call(kvm_x86_smi_allowed)(vcpu, false))) > + (vcpu->arch.smi_pending && KVM_X86_CALL(smi_allowed, vcpu, false))) > return true; > #endif > > @@ -13105,7 +13111,7 @@ int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu) > bool kvm_arch_dy_has_pending_interrupt(struct kvm_vcpu *vcpu) > { > return kvm_vcpu_apicv_active(vcpu) && > - static_call(kvm_x86_dy_apicv_has_pending_interrupt)(vcpu); > + KVM_X86_CALL(dy_apicv_has_pending_interrupt, vcpu); > } > > bool kvm_arch_vcpu_preempted_in_kernel(struct kvm_vcpu *vcpu) > @@ -13133,7 +13139,7 @@ bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu) > if (vcpu->arch.guest_state_protected) > return true; > > - return static_call(kvm_x86_get_cpl)(vcpu) == 0; > + return KVM_X86_CALL(get_cpl, vcpu) == 0; > } > > unsigned long kvm_arch_vcpu_get_ip(struct kvm_vcpu *vcpu) > @@ -13148,7 +13154,7 @@ int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu) > > int kvm_arch_interrupt_allowed(struct kvm_vcpu *vcpu) > { > - return static_call(kvm_x86_interrupt_allowed)(vcpu, false); > + return KVM_X86_CALL(interrupt_allowed, vcpu, false); > } > > unsigned long kvm_get_linear_rip(struct kvm_vcpu *vcpu) > @@ -13174,7 +13180,7 @@ unsigned long kvm_get_rflags(struct kvm_vcpu *vcpu) > { > unsigned long rflags; > > - rflags = static_call(kvm_x86_get_rflags)(vcpu); > + rflags = KVM_X86_CALL(get_rflags, vcpu); > if (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP) > rflags &= ~X86_EFLAGS_TF; > return rflags; > @@ -13186,7 +13192,7 @@ static void __kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags) > if (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP && > kvm_is_linear_rip(vcpu, vcpu->arch.singlestep_rip)) > rflags |= X86_EFLAGS_TF; > - static_call(kvm_x86_set_rflags)(vcpu, rflags); > + KVM_X86_CALL(set_rflags, vcpu, rflags); > } > > void kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags) > @@ -13297,8 +13303,7 @@ static bool kvm_can_deliver_async_pf(struct kvm_vcpu *vcpu) > if (!kvm_pv_async_pf_enabled(vcpu)) > return false; > > - if (vcpu->arch.apf.send_user_only && > - static_call(kvm_x86_get_cpl)(vcpu) == 0) > + if (vcpu->arch.apf.send_user_only && KVM_X86_CALL(get_cpl, vcpu) == 0) > return false; > > if (is_guest_mode(vcpu)) { > @@ -13409,7 +13414,7 @@ bool kvm_arch_can_dequeue_async_page_present(struct kvm_vcpu *vcpu) > void kvm_arch_start_assignment(struct kvm *kvm) > { > if (atomic_inc_return(&kvm->arch.assigned_device_count) == 1) > - static_call(kvm_x86_pi_start_assignment)(kvm); > + KVM_X86_CALL(pi_start_assignment, kvm); > } > EXPORT_SYMBOL_GPL(kvm_arch_start_assignment); > > @@ -13472,9 +13477,8 @@ int kvm_arch_irq_bypass_add_producer(struct irq_bypass_consumer *cons, > > irqfd->producer = prod; > kvm_arch_start_assignment(irqfd->kvm); > - ret = static_call(kvm_x86_pi_update_irte)(irqfd->kvm, > - prod->irq, irqfd->gsi, 1); > - > + ret = KVM_X86_CALL(pi_update_irte, > + irqfd->kvm, prod->irq, irqfd->gsi, 1); > if (ret) > kvm_arch_end_assignment(irqfd->kvm); > > @@ -13497,7 +13501,8 @@ void kvm_arch_irq_bypass_del_producer(struct irq_bypass_consumer *cons, > * when the irq is masked/disabled or the consumer side (KVM > * int this case doesn't want to receive the interrupts. > */ > - ret = static_call(kvm_x86_pi_update_irte)(irqfd->kvm, prod->irq, irqfd->gsi, 0); > + ret = KVM_X86_CALL(pi_update_irte, > + irqfd->kvm, prod->irq, irqfd->gsi, 0); > if (ret) > printk(KERN_INFO "irq bypass consumer (token %p) unregistration" > " fails: %d\n", irqfd->consumer.token, ret); > @@ -13508,7 +13513,7 @@ void kvm_arch_irq_bypass_del_producer(struct irq_bypass_consumer *cons, > int kvm_arch_update_irqfd_routing(struct kvm *kvm, unsigned int host_irq, > uint32_t guest_irq, bool set) > { > - return static_call(kvm_x86_pi_update_irte)(kvm, host_irq, guest_irq, set); > + return KVM_X86_CALL(pi_update_irte, kvm, host_irq, guest_irq, set); > } > > bool kvm_arch_irqfd_route_changed(struct kvm_kernel_irq_routing_entry *old, > diff --git a/arch/x86/kvm/x86.h b/arch/x86/kvm/x86.h > index a8b71803777b..8260da45540b 100644 > --- a/arch/x86/kvm/x86.h > +++ b/arch/x86/kvm/x86.h > @@ -157,7 +157,7 @@ static inline bool is_64_bit_mode(struct kvm_vcpu *vcpu) > > if (!is_long_mode(vcpu)) > return false; > - static_call(kvm_x86_get_cs_db_l_bits)(vcpu, &cs_db, &cs_l); > + KVM_X86_CALL(get_cs_db_l_bits, vcpu, &cs_db, &cs_l); > return cs_l; > } > > diff --git a/arch/x86/kvm/xen.c b/arch/x86/kvm/xen.c > index f65b35a05d91..afa804d32166 100644 > --- a/arch/x86/kvm/xen.c > +++ b/arch/x86/kvm/xen.c > @@ -1270,7 +1270,7 @@ int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data) > instructions[0] = 0xb8; > > /* vmcall / vmmcall */ > - static_call(kvm_x86_patch_hypercall)(vcpu, instructions + 5); > + KVM_X86_CALL(patch_hypercall, vcpu, instructions + 5); > > /* ret */ > instructions[8] = 0xc3; > @@ -1650,7 +1650,7 @@ int kvm_xen_hypercall(struct kvm_vcpu *vcpu) > params[5] = (u64)kvm_r9_read(vcpu); > } > #endif > - cpl = static_call(kvm_x86_get_cpl)(vcpu); > + cpl = KVM_X86_CALL(get_cpl, vcpu); > trace_kvm_xen_hypercall(cpl, input, params[0], params[1], params[2], > params[3], params[4], params[5]); > > -- > 2.27.0 >