On Fri, Jan 18, 2019 at 05:15:07PM +0000, Marc Zyngier wrote: > On 17/01/2019 20:33, Dave Martin wrote: > > In order to give each vcpu its own view of the SVE registers, this > > patch adds context storage via a new sve_state pointer in struct > > vcpu_arch. An additional member sve_max_vl is also added for each > > vcpu, to determine the maximum vector length visible to the guest > > and thus the value to be configured in ZCR_EL2.LEN while the is > > active. This also determines the layout and size of the storage in > > sve_state, which is read and written by the same backend functions > > that are used for context-switching the SVE state for host tasks. > > > > On SVE-enabled vcpus, SVE access traps are now handled by switching > > in the vcpu's SVE context and disabling the trap before returning > > to the guest. On other vcpus, the trap is not handled and an exit > > back to the host occurs, where the handle_sve() fallback path > > reflects an undefined instruction exception back to the guest, > > consistently with the behaviour of non-SVE-capable hardware (as was > > done unconditionally prior to this patch). > > > > No SVE handling is added on non-VHE-only paths, since VHE is an > > architectural and Kconfig prerequisite of SVE. > > > > Signed-off-by: Dave Martin <Dave.Martin@xxxxxxx> > > > > --- > > > > Changes since RFC v3: > > > > * Ensure guest_has_sve is initialised before use. Previously the > > order of the code allowed this to be used before initilaisation > > on some paths. Mysteriously the compiler failed to warn. > > --- > > arch/arm64/include/asm/kvm_host.h | 6 ++++ > > arch/arm64/kvm/fpsimd.c | 5 +-- > > arch/arm64/kvm/hyp/switch.c | 70 ++++++++++++++++++++++++++++++--------- > > 3 files changed, 64 insertions(+), 17 deletions(-) [...] > > diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c > > index 9f07403..f7a8e3b 100644 > > --- a/arch/arm64/kvm/hyp/switch.c > > +++ b/arch/arm64/kvm/hyp/switch.c > > @@ -98,7 +98,10 @@ static void activate_traps_vhe(struct kvm_vcpu *vcpu) > > val = read_sysreg(cpacr_el1); > > val |= CPACR_EL1_TTA; > > val &= ~CPACR_EL1_ZEN; > > - if (!update_fp_enabled(vcpu)) { > > + if (update_fp_enabled(vcpu)) { > > + if (vcpu_has_sve(vcpu)) > > + val |= CPACR_EL1_ZEN; > > + } else { > > val &= ~CPACR_EL1_FPEN; > > __activate_traps_fpsimd32(vcpu); > > } > > @@ -313,26 +316,59 @@ static bool __hyp_text __populate_fault_info(struct kvm_vcpu *vcpu) > > return true; > > } > > > > -static bool __hyp_text __hyp_switch_fpsimd(struct kvm_vcpu *vcpu) > > +/* > > + * if () with a gating check for SVE support to minimise branch > > + * mispredictions in non-SVE systems. > > + * (system_supports_sve() is resolved at build time or via a static key.) > > + */ > > +#define if_sve(cond) if (system_supports_sve() && (cond)) > > I really hate this kind of construct, as while it provides an apparent > benefit, it actually leads to code that is harder to read. Do we have > any data showing that this adversely impacts any known workload on > non-SVE systems, given how rarely this exception is triggered? > > At any rate, I'd rather see system_supports_sve() as part of the condition. The alternative was to have a lot of if (system_supports_sve() && foo), which makes for more branchy code. But I'll agree that we don't have good evidence to justify this if_sve () thing either. I'll see if I can come up with something saner. Is it worth at least annotating paths with likely/unlikely, or is that again something we shouldn't do without some benchmarking that backs it up? > > + > > +/* Check for an FPSIMD/SVE trap and handle as appropriate */ > > +static bool __hyp_text __hyp_handle_fpsimd(struct kvm_vcpu *vcpu) > > { > > - struct user_fpsimd_state *host_fpsimd = vcpu->arch.host_fpsimd_state; > > + u8 trap_class; > > + bool guest_has_sve; > > > > - if (has_vhe()) > > - write_sysreg(read_sysreg(cpacr_el1) | CPACR_EL1_FPEN, > > - cpacr_el1); > > - else > > + if (!system_supports_fpsimd()) > > + return false; > > + > > + guest_has_sve = vcpu_has_sve(vcpu); > > + trap_class = kvm_vcpu_trap_get_class(vcpu); > > + > > + if (trap_class == ESR_ELx_EC_FP_ASIMD) > > + goto handle; > > + > > + if_sve (guest_has_sve && trap_class == ESR_ELx_EC_SVE) > > + goto handle; > > + > > + return false; > > In the end, this is the stuff we want to handle quickly: nothing to do > at all. Given that vcpu_has_sve() is already gated by True -- would it be worth renaming the function __try_handle_fpsimd() or similar, to make it clearer that the function may actually not handle FPSIMD (or do anything material) in the common case? I may have led myself astray with the current naming. > system_supports_sve(), I'm pretty sure that there is no difference in > code generation between your if_sve() construct and a simple if(). Maybe so. At least, if_sve () may make the code more obscure in practice for no good reason. > > + > > +handle: > > + /* The trap is an FPSIMD/SVE trap: switch the context */ > > + > > + if (has_vhe()) { > > + u64 reg = read_sysreg(cpacr_el1) | CPACR_EL1_FPEN; > > + > > + if_sve (guest_has_sve) > > + reg |= CPACR_EL1_ZEN; > > + > > + write_sysreg(reg, cpacr_el1); > > + } else { > > write_sysreg(read_sysreg(cptr_el2) & ~(u64)CPTR_EL2_TFP, > > cptr_el2); > > + } > > > > isb(); > > > > if (vcpu->arch.flags & KVM_ARM64_FP_HOST) { > > + struct user_fpsimd_state *host_fpsimd = > > + vcpu->arch.host_fpsimd_state; > > + > > Single line, please. Or split declaration and assignment. Can do. > > /* > > * In the SVE case, VHE is assumed: it is enforced by > > * Kconfig and kvm_arch_init(). > > */ > > - if (system_supports_sve() && > > - (vcpu->arch.flags & KVM_ARM64_HOST_SVE_IN_USE)) { > > + if_sve (vcpu->arch.flags & KVM_ARM64_HOST_SVE_IN_USE) { > > struct thread_struct *thread = container_of( > > host_fpsimd, > > struct thread_struct, uw.fpsimd_state); > > @@ -345,10 +381,14 @@ static bool __hyp_text __hyp_switch_fpsimd(struct kvm_vcpu *vcpu) > > vcpu->arch.flags &= ~KVM_ARM64_FP_HOST; > > } [...] Cheers ---Dave _______________________________________________ kvmarm mailing list kvmarm@xxxxxxxxxxxxxxxxxxxxx https://lists.cs.columbia.edu/mailman/listinfo/kvmarm