On Wed, Feb 20, 2019 at 11:12:49AM +0000, Julien Thierry wrote: > Hi Dave, > > On 18/02/2019 19:52, Dave Martin wrote: > > Due to the way the effective SVE vector length is controlled and > > trapped at different exception levels, certain mismatches in the > > sets of vector lengths supported by different physical CPUs in the > > system may prevent straightforward virtualisation of SVE at parity > > with the host. > > > > This patch analyses the extent to which SVE can be virtualised > > safely without interfering with migration of vcpus between physical > > CPUs, and rejects late secondary CPUs that would erode the > > situation further. > > > > It is left up to KVM to decide what to do with this information. > > > > Signed-off-by: Dave Martin <Dave.Martin@xxxxxxx> > > --- > > arch/arm64/include/asm/fpsimd.h | 1 + > > arch/arm64/kernel/cpufeature.c | 2 +- > > arch/arm64/kernel/fpsimd.c | 86 ++++++++++++++++++++++++++++++++++------- > > 3 files changed, 73 insertions(+), 16 deletions(-) > > > > diff --git a/arch/arm64/include/asm/fpsimd.h b/arch/arm64/include/asm/fpsimd.h > > index dd1ad39..964adc9 100644 > > --- a/arch/arm64/include/asm/fpsimd.h > > +++ b/arch/arm64/include/asm/fpsimd.h > > @@ -87,6 +87,7 @@ extern void sve_kernel_enable(const struct arm64_cpu_capabilities *__unused); > > extern u64 read_zcr_features(void); > > > > extern int __ro_after_init sve_max_vl; > > +extern int __ro_after_init sve_max_virtualisable_vl; > > > > #ifdef CONFIG_ARM64_SVE > > > > diff --git a/arch/arm64/kernel/cpufeature.c b/arch/arm64/kernel/cpufeature.c > > index f6d84e2..5eaacb4 100644 > > --- a/arch/arm64/kernel/cpufeature.c > > +++ b/arch/arm64/kernel/cpufeature.c > > @@ -1825,7 +1825,7 @@ static void verify_sve_features(void) > > unsigned int len = zcr & ZCR_ELx_LEN_MASK; > > > > if (len < safe_len || sve_verify_vq_map()) { > > - pr_crit("CPU%d: SVE: required vector length(s) missing\n", > > + pr_crit("CPU%d: SVE: vector length support mismatch\n", > > smp_processor_id()); > > cpu_die_early(); > > } > > diff --git a/arch/arm64/kernel/fpsimd.c b/arch/arm64/kernel/fpsimd.c > > index 62c37f0..64729e2 100644 > > --- a/arch/arm64/kernel/fpsimd.c > > +++ b/arch/arm64/kernel/fpsimd.c > > @@ -18,6 +18,7 @@ > > */ > > > > #include <linux/bitmap.h> > > +#include <linux/bitops.h> > > #include <linux/bottom_half.h> > > #include <linux/bug.h> > > #include <linux/cache.h> > > @@ -48,6 +49,7 @@ > > #include <asm/sigcontext.h> > > #include <asm/sysreg.h> > > #include <asm/traps.h> > > +#include <asm/virt.h> > > > > #define FPEXC_IOF (1 << 0) > > #define FPEXC_DZF (1 << 1) > > @@ -130,14 +132,18 @@ static int sve_default_vl = -1; > > > > /* Maximum supported vector length across all CPUs (initially poisoned) */ > > int __ro_after_init sve_max_vl = SVE_VL_MIN; > > +int __ro_after_init sve_max_virtualisable_vl = SVE_VL_MIN; > > /* Set of available vector lengths, as vq_to_bit(vq): */ > > static __ro_after_init DECLARE_BITMAP(sve_vq_map, SVE_VQ_MAX); > > +/* Set of vector lengths present on at least one cpu: */ > > +static __ro_after_init DECLARE_BITMAP(sve_vq_partial_map, SVE_VQ_MAX); > > static void __percpu *efi_sve_state; > > > > #else /* ! CONFIG_ARM64_SVE */ > > > > /* Dummy declaration for code that will be optimised out: */ > > extern __ro_after_init DECLARE_BITMAP(sve_vq_map, SVE_VQ_MAX); > > +extern __ro_after_init DECLARE_BITMAP(sve_vq_partial_map, SVE_VQ_MAX); > > extern void __percpu *efi_sve_state; > > > > #endif /* ! CONFIG_ARM64_SVE */ > > @@ -623,12 +629,6 @@ int sve_get_current_vl(void) > > return sve_prctl_status(0); > > } > > > > -/* > > - * Bitmap for temporary storage of the per-CPU set of supported vector lengths > > - * during secondary boot. > > - */ > > -static DECLARE_BITMAP(sve_secondary_vq_map, SVE_VQ_MAX); > > - > > static void sve_probe_vqs(DECLARE_BITMAP(map, SVE_VQ_MAX)) > > { > > unsigned int vq, vl; > > @@ -650,6 +650,7 @@ static void sve_probe_vqs(DECLARE_BITMAP(map, SVE_VQ_MAX)) > > void __init sve_init_vq_map(void) > > { > > sve_probe_vqs(sve_vq_map); > > + bitmap_copy(sve_vq_partial_map, sve_vq_map, SVE_VQ_MAX); > > } > > > > /* > > @@ -658,25 +659,58 @@ void __init sve_init_vq_map(void) > > */ > > void sve_update_vq_map(void) > > { > > - sve_probe_vqs(sve_secondary_vq_map); > > - bitmap_and(sve_vq_map, sve_vq_map, sve_secondary_vq_map, SVE_VQ_MAX); > > + DECLARE_BITMAP(tmp_map, SVE_VQ_MAX); > > + > > + sve_probe_vqs(tmp_map); > > + bitmap_and(sve_vq_map, sve_vq_map, tmp_map, SVE_VQ_MAX); > > + bitmap_or(sve_vq_partial_map, sve_vq_partial_map, tmp_map, SVE_VQ_MAX); > > } > > > > /* Check whether the current CPU supports all VQs in the committed set */ > > int sve_verify_vq_map(void) > > { > > - int ret = 0; > > + DECLARE_BITMAP(tmp_map, SVE_VQ_MAX); > > + unsigned long b; > > > > - sve_probe_vqs(sve_secondary_vq_map); > > - bitmap_andnot(sve_secondary_vq_map, sve_vq_map, sve_secondary_vq_map, > > - SVE_VQ_MAX); > > - if (!bitmap_empty(sve_secondary_vq_map, SVE_VQ_MAX)) { > > + sve_probe_vqs(tmp_map); > > + > > + bitmap_complement(tmp_map, tmp_map, SVE_VQ_MAX); > > + if (bitmap_intersects(tmp_map, sve_vq_map, SVE_VQ_MAX)) { > > pr_warn("SVE: cpu%d: Required vector length(s) missing\n", > > smp_processor_id()); > > - ret = -EINVAL; > > + return -EINVAL; > > } > > > > - return ret; > > + if (!IS_ENABLED(CONFIG_KVM) || !is_hyp_mode_available()) > > + return 0; > > + > > + /* > > + * For KVM, it is necessary to ensure that this CPU doesn't > > + * support any vector length that guests may have probed as > > + * unsupported. > > + */ > > + > > + /* Recover the set of supported VQs: */ > > + bitmap_complement(tmp_map, tmp_map, SVE_VQ_MAX); > > + /* Find VQs supported that are not globally supported: */ > > + bitmap_andnot(tmp_map, tmp_map, sve_vq_map, SVE_VQ_MAX); > > + > > + /* Find the lowest such VQ, if any: */ > > + b = find_last_bit(tmp_map, SVE_VQ_MAX); > > + if (b >= SVE_VQ_MAX) > > + return 0; /* no mismatches */ > > + > > + /* > > + * Mismatches above sve_max_virtualisable_vl are fine, since > > + * no guest is allowed to configure ZCR_EL2.LEN to exceed this: > > + */ > > + if (sve_vl_from_vq(bit_to_vq(b)) <= sve_max_virtualisable_vl) { > > + pr_warn("SVE: cpu%d: Unsupported vector length(s) present\n", > > Nit: might be good to specify that the vector length is unsupported for > virtualisation. > > Also, since KVM is the one deciding what to do with the information, > should we have a warning here? But I can understand that knowing which > CPUs are introducing unsupported vector length, maybe using pr_devel() > instead of pr_warn() These warnings are really for consumption by SoC vendors, not users: my aim is to flag up systems that we consider broken (or at least, unsuitable for running KVM). So I prefer to make this noisy and limit the amount of "useful" information people might be tempted to programmatically scrape from dmesg. cpufeatures uses pr_warn("SANITY CHECK: [...]") here. Maybe I should stick "SANITY CHECK" in here too? I will also try to make the commit message more explicit and/or add comments to make the intent of the code clearer. It may also make sense to make this noise even if KVM isn't enabled (which is a rare case anyhow). Thoughts? > In any case, the logic looks good to me: > > Reviewed-by: Julien Thierry <julien.thierry@xxxxxxx> Thanks ---Dave _______________________________________________ kvmarm mailing list kvmarm@xxxxxxxxxxxxxxxxxxxxx https://lists.cs.columbia.edu/mailman/listinfo/kvmarm