On Sun, Dec 06, 2015 at 05:07:14PM -0800, Mario Smarduch wrote: > This patch tracks armv7 and armv8 fp/simd hardware state with cptr_el2 register. > On vcpu_load for 32 bit guests enable FP access, and enable fp/simd > trapping for 32 and 64 bit guests. On first fp/simd access trap to handler > to save host and restore guest context, and clear trapping bits to enable vcpu > lazy mode. On vcpu_put if trap bits are clear save guest and restore host > context and also save 32 bit guest fpexc register. > > Signed-off-by: Mario Smarduch <m.smarduch@xxxxxxxxxxx> > --- > arch/arm/include/asm/kvm_emulate.h | 5 ++ > arch/arm/include/asm/kvm_host.h | 2 + > arch/arm/kvm/arm.c | 20 +++++-- > arch/arm64/include/asm/kvm_asm.h | 2 + > arch/arm64/include/asm/kvm_emulate.h | 15 +++-- > arch/arm64/include/asm/kvm_host.h | 16 +++++- > arch/arm64/kernel/asm-offsets.c | 1 + > arch/arm64/kvm/Makefile | 3 +- > arch/arm64/kvm/fpsimd_switch.S | 38 ++++++++++++ > arch/arm64/kvm/hyp.S | 108 +++++++++++++---------------------- > arch/arm64/kvm/hyp_head.S | 48 ++++++++++++++++ > 11 files changed, 181 insertions(+), 77 deletions(-) > create mode 100644 arch/arm64/kvm/fpsimd_switch.S > create mode 100644 arch/arm64/kvm/hyp_head.S > > diff --git a/arch/arm/include/asm/kvm_emulate.h b/arch/arm/include/asm/kvm_emulate.h > index 3de11a2..13feed5 100644 > --- a/arch/arm/include/asm/kvm_emulate.h > +++ b/arch/arm/include/asm/kvm_emulate.h > @@ -243,6 +243,11 @@ static inline unsigned long vcpu_data_host_to_guest(struct kvm_vcpu *vcpu, > } > } > > +static inline bool kvm_guest_vcpu_is_32bit(struct kvm_vcpu *vcpu) > +{ > + return true; > +} > + > #ifdef CONFIG_VFPv3 > /* Called from vcpu_load - save fpexc and enable guest access to fp/simd unit */ > static inline void kvm_enable_vcpu_fpexc(struct kvm_vcpu *vcpu) > diff --git a/arch/arm/include/asm/kvm_host.h b/arch/arm/include/asm/kvm_host.h > index ecc883a..720ae51 100644 > --- a/arch/arm/include/asm/kvm_host.h > +++ b/arch/arm/include/asm/kvm_host.h > @@ -227,6 +227,8 @@ int kvm_perf_teardown(void); > void kvm_mmu_wp_memory_region(struct kvm *kvm, int slot); > > struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr); > + > +static inline void kvm_save_guest_vcpu_fpexc(struct kvm_vcpu *vcpu) {} > void kvm_restore_host_vfp_state(struct kvm_vcpu *); > > static inline void kvm_arch_hardware_disable(void) {} > diff --git a/arch/arm/kvm/arm.c b/arch/arm/kvm/arm.c > index 1de07ab..dd59f8a 100644 > --- a/arch/arm/kvm/arm.c > +++ b/arch/arm/kvm/arm.c > @@ -292,8 +292,12 @@ void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu) > > kvm_arm_set_running_vcpu(vcpu); > > - /* Save and enable FPEXC before we load guest context */ > - kvm_enable_vcpu_fpexc(vcpu); > + /* > + * For 32bit guest executing on arm64, enable fp/simd access in > + * EL2. On arm32 save host fpexc and then enable fp/simd access. > + */ > + if (kvm_guest_vcpu_is_32bit(vcpu)) > + kvm_enable_vcpu_fpexc(vcpu); > > /* reset hyp cptr register to trap on tracing and vfp/simd access*/ > vcpu_reset_cptr(vcpu); > @@ -302,10 +306,18 @@ void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu) > void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu) > { > /* If the fp/simd registers are dirty save guest, restore host. */ > - if (kvm_vcpu_vfp_isdirty(vcpu)) > + if (kvm_vcpu_vfp_isdirty(vcpu)) { > kvm_restore_host_vfp_state(vcpu); > > - /* Restore host FPEXC trashed in vcpu_load */ > + /* > + * For 32bit guest on arm64 save the guest fpexc register > + * in EL2 mode. > + */ > + if (kvm_guest_vcpu_is_32bit(vcpu)) > + kvm_save_guest_vcpu_fpexc(vcpu); > + } > + > + /* For arm32 restore host FPEXC trashed in vcpu_load. */ > kvm_restore_host_fpexc(vcpu); > > /* > diff --git a/arch/arm64/include/asm/kvm_asm.h b/arch/arm64/include/asm/kvm_asm.h > index 5e37710..d53d069 100644 > --- a/arch/arm64/include/asm/kvm_asm.h > +++ b/arch/arm64/include/asm/kvm_asm.h > @@ -117,6 +117,8 @@ extern char __kvm_hyp_vector[]; > extern void __kvm_flush_vm_context(void); > extern void __kvm_tlb_flush_vmid_ipa(struct kvm *kvm, phys_addr_t ipa); > extern void __kvm_tlb_flush_vmid(struct kvm *kvm); > +extern void __kvm_vcpu_enable_fpexc32(void); > +extern void __kvm_vcpu_save_fpexc32(struct kvm_vcpu *vcpu); > > extern int __kvm_vcpu_run(struct kvm_vcpu *vcpu); > > diff --git a/arch/arm64/include/asm/kvm_emulate.h b/arch/arm64/include/asm/kvm_emulate.h > index 8dccbd7..bbbee9d 100644 > --- a/arch/arm64/include/asm/kvm_emulate.h > +++ b/arch/arm64/include/asm/kvm_emulate.h > @@ -290,13 +290,20 @@ static inline unsigned long vcpu_data_host_to_guest(struct kvm_vcpu *vcpu, > return data; /* Leave LE untouched */ > } > > -static inline void kvm_enable_vcpu_fpexc(struct kvm_vcpu *vcpu) {} > -static inline void kvm_restore_host_fpexc(struct kvm_vcpu *vcpu) {} > -static inline void vcpu_reset_cptr(struct kvm_vcpu *vcpu) {} > +static inline bool kvm_guest_vcpu_is_32bit(struct kvm_vcpu *vcpu) > +{ > + return !(vcpu->arch.hcr_el2 & HCR_RW); > +} > + > +static inline void vcpu_reset_cptr(struct kvm_vcpu *vcpu) > +{ > + vcpu->arch.cptr_el2 = CPTR_EL2_TTA | CPTR_EL2_TFP; > +} > + > > static inline bool kvm_vcpu_vfp_isdirty(struct kvm_vcpu *vcpu) > { > - return false; > + return !!(~vcpu->arch.cptr_el2 & CPTR_EL2_TFP); > } > > #endif /* __ARM64_KVM_EMULATE_H__ */ > diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h > index e16fd39..0c65393 100644 > --- a/arch/arm64/include/asm/kvm_host.h > +++ b/arch/arm64/include/asm/kvm_host.h > @@ -100,6 +100,7 @@ struct kvm_vcpu_arch { > /* HYP configuration */ > u64 hcr_el2; > u32 mdcr_el2; > + u32 cptr_el2; > > /* Exception Information */ > struct kvm_vcpu_fault_info fault; > @@ -248,7 +249,20 @@ static inline void kvm_arch_hardware_unsetup(void) {} > static inline void kvm_arch_sync_events(struct kvm *kvm) {} > static inline void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu) {} > static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} > -static inline void kvm_restore_host_vfp_state(struct kvm_vcpu *vcpu) {} > + > +static inline void kvm_enable_vcpu_fpexc(struct kvm_vcpu *vcpu) > +{ > + /* Enable FP/SIMD access from EL2 mode*/ > + kvm_call_hyp(__kvm_vcpu_enable_fpexc32); > +} > + > +static inline void kvm_save_guest_vcpu_fpexc(struct kvm_vcpu *vcpu) > +{ > + /* Save FPEXEC32_EL2 in EL2 mode */ > + kvm_call_hyp(__kvm_vcpu_save_fpexc32, vcpu); > +} > +static inline void kvm_restore_host_fpexc(struct kvm_vcpu *vcpu) {} > +void kvm_restore_host_vfp_state(struct kvm_vcpu *vcpu); > > void kvm_arm_init_debug(void); > void kvm_arm_setup_debug(struct kvm_vcpu *vcpu); > diff --git a/arch/arm64/kernel/asm-offsets.c b/arch/arm64/kernel/asm-offsets.c > index 8d89cf8..3c8d836 100644 > --- a/arch/arm64/kernel/asm-offsets.c > +++ b/arch/arm64/kernel/asm-offsets.c > @@ -123,6 +123,7 @@ int main(void) > DEFINE(DEBUG_WVR, offsetof(struct kvm_guest_debug_arch, dbg_wvr)); > DEFINE(VCPU_HCR_EL2, offsetof(struct kvm_vcpu, arch.hcr_el2)); > DEFINE(VCPU_MDCR_EL2, offsetof(struct kvm_vcpu, arch.mdcr_el2)); > + DEFINE(VCPU_CPTR_EL2, offsetof(struct kvm_vcpu, arch.cptr_el2)); > DEFINE(VCPU_IRQ_LINES, offsetof(struct kvm_vcpu, arch.irq_lines)); > DEFINE(VCPU_HOST_CONTEXT, offsetof(struct kvm_vcpu, arch.host_cpu_context)); > DEFINE(VCPU_HOST_DEBUG_STATE, offsetof(struct kvm_vcpu, arch.host_debug_state)); > diff --git a/arch/arm64/kvm/Makefile b/arch/arm64/kvm/Makefile > index 1949fe5..262b9a5 100644 > --- a/arch/arm64/kvm/Makefile > +++ b/arch/arm64/kvm/Makefile > @@ -17,7 +17,8 @@ kvm-$(CONFIG_KVM_ARM_HOST) += $(ARM)/psci.o $(ARM)/perf.o > > kvm-$(CONFIG_KVM_ARM_HOST) += emulate.o inject_fault.o regmap.o > kvm-$(CONFIG_KVM_ARM_HOST) += hyp.o hyp-init.o handle_exit.o > -kvm-$(CONFIG_KVM_ARM_HOST) += guest.o debug.o reset.o sys_regs.o sys_regs_generic_v8.o > +kvm-$(CONFIG_KVM_ARM_HOST) += guest.o debug.o reset.o sys_regs.o > +kvm-$(CONFIG_KVM_ARM_HOST) += sys_regs_generic_v8.o fpsimd_switch.o > > kvm-$(CONFIG_KVM_ARM_HOST) += $(KVM)/arm/vgic.o > kvm-$(CONFIG_KVM_ARM_HOST) += $(KVM)/arm/vgic-v2.o > diff --git a/arch/arm64/kvm/fpsimd_switch.S b/arch/arm64/kvm/fpsimd_switch.S > new file mode 100644 > index 0000000..5295512 > --- /dev/null > +++ b/arch/arm64/kvm/fpsimd_switch.S > @@ -0,0 +1,38 @@ > +/* > + * Copyright (C) 2012,2013 - ARM Ltd > + * Author: Marc Zyngier <marc.zyngier@xxxxxxx> > + * Is this copied code or new code? > + * This program is free software; you can redistribute it and/or modify > + * it under the terms of the GNU General Public License version 2 as > + * published by the Free Software Foundation. > + * > + * This program is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + * > + * You should have received a copy of the GNU General Public License > + * along with this program. If not, see <http://www.gnu.org/licenses/>. > + */ > + > +#include <linux/linkage.h> > + > +#include "hyp_head.S" > + > + .text > +/** > + * void kvm_restore_host_vfp_state(struct vcpu *vcpu) - > + * This function saves the guest, restores host, called from host. > + */ > +ENTRY(kvm_restore_host_vfp_state) > + push xzr, lr > + > + add x2, x0, #VCPU_CONTEXT > + bl __save_fpsimd > + > + ldr x2, [x0, #VCPU_HOST_CONTEXT] > + bl __restore_fpsimd > + > + pop xzr, lr > + ret > +ENDPROC(kvm_restore_host_vfp_state) > diff --git a/arch/arm64/kvm/hyp.S b/arch/arm64/kvm/hyp.S > index e583613..b8b1afb 100644 > --- a/arch/arm64/kvm/hyp.S > +++ b/arch/arm64/kvm/hyp.S > @@ -17,23 +17,7 @@ > > #include <linux/linkage.h> > > -#include <asm/alternative.h> > -#include <asm/asm-offsets.h> > -#include <asm/assembler.h> > -#include <asm/cpufeature.h> > -#include <asm/debug-monitors.h> > -#include <asm/esr.h> > -#include <asm/fpsimdmacros.h> > -#include <asm/kvm.h> > -#include <asm/kvm_arm.h> > -#include <asm/kvm_asm.h> > -#include <asm/kvm_mmu.h> > -#include <asm/memory.h> > - > -#define CPU_GP_REG_OFFSET(x) (CPU_GP_REGS + x) > -#define CPU_XREG_OFFSET(x) CPU_GP_REG_OFFSET(CPU_USER_PT_REGS + 8*x) > -#define CPU_SPSR_OFFSET(x) CPU_GP_REG_OFFSET(CPU_SPSR + 8*x) > -#define CPU_SYSREG_OFFSET(x) (CPU_SYSREGS + 8*x) > +#include "hyp_head.S" > > .text > .pushsection .hyp.text, "ax" > @@ -104,20 +88,6 @@ > restore_common_regs > .endm > > -.macro save_fpsimd > - // x2: cpu context address > - // x3, x4: tmp regs > - add x3, x2, #CPU_GP_REG_OFFSET(CPU_FP_REGS) > - fpsimd_save x3, 4 > -.endm > - > -.macro restore_fpsimd > - // x2: cpu context address > - // x3, x4: tmp regs > - add x3, x2, #CPU_GP_REG_OFFSET(CPU_FP_REGS) > - fpsimd_restore x3, 4 > -.endm > - > .macro save_guest_regs > // x0 is the vcpu address > // x1 is the return code, do not corrupt! > @@ -385,14 +355,6 @@ > tbz \tmp, #KVM_ARM64_DEBUG_DIRTY_SHIFT, \target > .endm > > -/* > - * Branch to target if CPTR_EL2.TFP bit is set (VFP/SIMD trapping enabled) > - */ > -.macro skip_fpsimd_state tmp, target > - mrs \tmp, cptr_el2 > - tbnz \tmp, #CPTR_EL2_TFP_SHIFT, \target > -.endm > - > .macro compute_debug_state target > // Compute debug state: If any of KDE, MDE or KVM_ARM64_DEBUG_DIRTY > // is set, we do a full save/restore cycle and disable trapping. > @@ -433,10 +395,6 @@ > mrs x5, ifsr32_el2 > stp x4, x5, [x3] > > - skip_fpsimd_state x8, 2f > - mrs x6, fpexc32_el2 > - str x6, [x3, #16] > -2: > skip_debug_state x8, 1f > mrs x7, dbgvcr32_el2 > str x7, [x3, #24] > @@ -467,22 +425,9 @@ > > .macro activate_traps > ldr x2, [x0, #VCPU_HCR_EL2] > - > - /* > - * We are about to set CPTR_EL2.TFP to trap all floating point > - * register accesses to EL2, however, the ARM ARM clearly states that > - * traps are only taken to EL2 if the operation would not otherwise > - * trap to EL1. Therefore, always make sure that for 32-bit guests, > - * we set FPEXC.EN to prevent traps to EL1, when setting the TFP bit. > - */ > - tbnz x2, #HCR_RW_SHIFT, 99f // open code skip_32bit_state > - mov x3, #(1 << 30) > - msr fpexc32_el2, x3 > - isb > -99: > msr hcr_el2, x2 > - mov x2, #CPTR_EL2_TTA > - orr x2, x2, #CPTR_EL2_TFP > + > + ldr w2, [x0, VCPU_CPTR_EL2] > msr cptr_el2, x2 > > mov x2, #(1 << 15) // Trap CP15 Cr=15 > @@ -668,15 +613,15 @@ __restore_debug: > > ret > > -__save_fpsimd: > - skip_fpsimd_state x3, 1f > +ENTRY(__save_fpsimd) > save_fpsimd > -1: ret > + ret > +ENDPROC(__save_fpsimd) > > -__restore_fpsimd: > - skip_fpsimd_state x3, 1f > +ENTRY(__restore_fpsimd) > restore_fpsimd > -1: ret > + ret > +ENDPROC(__restore_fpsimd) > > switch_to_guest_fpsimd: > push x4, lr > @@ -763,7 +708,6 @@ __kvm_vcpu_return: > add x2, x0, #VCPU_CONTEXT > > save_guest_regs > - bl __save_fpsimd > bl __save_sysregs > > skip_debug_state x3, 1f > @@ -784,8 +728,10 @@ __kvm_vcpu_return: > kern_hyp_va x2 > > bl __restore_sysregs > - bl __restore_fpsimd > - /* Clear FPSIMD and Trace trapping */ > + > + /* Save CPTR_EL2 between exits and clear FPSIMD and Trace trapping */ > + mrs x3, cptr_el2 > + str w3, [x0, VCPU_CPTR_EL2] > msr cptr_el2, xzr > > skip_debug_state x3, 1f > @@ -863,6 +809,34 @@ ENTRY(__kvm_flush_vm_context) > ret > ENDPROC(__kvm_flush_vm_context) > > +/** > + * void __kvm_enable_fpexc32(void) - > + * We may be entering the guest and set CPTR_EL2.TFP to trap all floating > + * point register accesses to EL2, however, the ARM manual clearly states > + * that traps are only taken to EL2 if the operation would not otherwise > + * trap to EL1. Therefore, always make sure that for 32-bit guests, > + * we set FPEXC.EN to prevent traps to EL1, when setting the TFP bit. > + */ > +ENTRY(__kvm_vcpu_enable_fpexc32) > + mov x3, #(1 << 30) > + msr fpexc32_el2, x3 > + isb this is only called via a hypercall so do you really need the ISB? > + ret > +ENDPROC(__kvm_vcpu_enable_fpexc32) > + > +/** > + * void __kvm_save_fpexc32(void) - > + * This function restores guest FPEXC to its vcpu context, we call this > + * function from vcpu_put. > + */ > +ENTRY(__kvm_vcpu_save_fpexc32) > + kern_hyp_va x0 > + add x2, x0, #VCPU_CONTEXT > + mrs x1, fpexc32_el2 > + str x1, [x2, #CPU_SYSREG_OFFSET(FPEXC32_EL2)] > + ret > +ENDPROC(__kvm_vcpu_save_fpexc32) > + > __kvm_hyp_panic: > // Guess the context by looking at VTTBR: > // If zero, then we're already a host. > diff --git a/arch/arm64/kvm/hyp_head.S b/arch/arm64/kvm/hyp_head.S > new file mode 100644 > index 0000000..bb32824 > --- /dev/null > +++ b/arch/arm64/kvm/hyp_head.S > @@ -0,0 +1,48 @@ > +/* > + * Copyright (C) 2012,2013 - ARM Ltd > + * Author: Marc Zyngier <marc.zyngier@xxxxxxx> > + * > + * This program is free software; you can redistribute it and/or modify > + * it under the terms of the GNU General Public License version 2 as > + * published by the Free Software Foundation. > + * > + * This program is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + * > + * You should have received a copy of the GNU General Public License > + * along with this program. If not, see <http://www.gnu.org/licenses/>. > + */ > + > +#include <asm/alternative.h> > +#include <asm/asm-offsets.h> > +#include <asm/assembler.h> > +#include <asm/cpufeature.h> > +#include <asm/debug-monitors.h> > +#include <asm/esr.h> > +#include <asm/fpsimdmacros.h> > +#include <asm/kvm.h> > +#include <asm/kvm_arm.h> > +#include <asm/kvm_asm.h> > +#include <asm/kvm_mmu.h> > +#include <asm/memory.h> > + > +#define CPU_GP_REG_OFFSET(x) (CPU_GP_REGS + x) > +#define CPU_XREG_OFFSET(x) CPU_GP_REG_OFFSET(CPU_USER_PT_REGS + 8*x) > +#define CPU_SPSR_OFFSET(x) CPU_GP_REG_OFFSET(CPU_SPSR + 8*x) > +#define CPU_SYSREG_OFFSET(x) (CPU_SYSREGS + 8*x) > + > +.macro save_fpsimd > + // x2: cpu context address > + // x3, x4: tmp regs > + add x3, x2, #CPU_GP_REG_OFFSET(CPU_FP_REGS) > + fpsimd_save x3, 4 > +.endm > + > +.macro restore_fpsimd > + // x2: cpu context address > + // x3, x4: tmp regs > + add x3, x2, #CPU_GP_REG_OFFSET(CPU_FP_REGS) > + fpsimd_restore x3, 4 > +.endm > -- > 1.9.1 > I'm not going to review the details of this, since we have to rebase it on the world-switch in C, sorry. The good news is that it should be much simpler to write in C-code. Let me know if you don't have the bandwidth to rebase this, in that case I'll be happy to help. Thanks, -Christoffer -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html