M2 MacBook Air has mismatched CCSIDR associativity bits, which makes the bits a KVM vCPU sees inconsistent when migrating. Record such mismatches so that KVM can use the information later to avoid the problem. Signed-off-by: Akihiko Odaki <akihiko.odaki@xxxxxxxxxx> --- arch/arm64/include/asm/cache.h | 3 ++ arch/arm64/include/asm/cpu.h | 1 + arch/arm64/include/asm/cpufeature.h | 8 +++++ arch/arm64/include/asm/sysreg.h | 7 ++++ arch/arm64/kernel/cacheinfo.c | 4 +-- arch/arm64/kernel/cpu_errata.c | 52 +++++++++++++++++++++++++++++ arch/arm64/kernel/cpufeature.c | 4 +++ arch/arm64/kernel/cpuinfo.c | 30 +++++++++++++++++ arch/arm64/kvm/sys_regs.c | 4 +-- arch/arm64/tools/cpucaps | 1 + 10 files changed, 110 insertions(+), 4 deletions(-) diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h index c0b178d1bb4f..eeab2b8c7e71 100644 --- a/arch/arm64/include/asm/cache.h +++ b/arch/arm64/include/asm/cache.h @@ -72,6 +72,8 @@ static inline u32 cache_type_cwg(void) #define __read_mostly __section(".data..read_mostly") +#define MAX_CACHE_LEVEL 7 /* Max 7 level supported */ + static inline int cache_line_size_of_cpu(void) { u32 cwg = cache_type_cwg(); @@ -80,6 +82,7 @@ static inline int cache_line_size_of_cpu(void) } int cache_line_size(void); +enum cache_type get_cache_type(int level); /* * Read the effective value of CTR_EL0. diff --git a/arch/arm64/include/asm/cpu.h b/arch/arm64/include/asm/cpu.h index fd7a92219eea..b8d4f31ed59b 100644 --- a/arch/arm64/include/asm/cpu.h +++ b/arch/arm64/include/asm/cpu.h @@ -41,6 +41,7 @@ struct cpuinfo_arm64 { struct cpu cpu; struct kobject kobj; u64 reg_ctr; + struct ccsidr reg_ccsidr[MAX_CACHE_LEVEL + 1]; u64 reg_cntfrq; u64 reg_dczid; u64 reg_midr; diff --git a/arch/arm64/include/asm/cpufeature.h b/arch/arm64/include/asm/cpufeature.h index f73f11b55042..104483151362 100644 --- a/arch/arm64/include/asm/cpufeature.h +++ b/arch/arm64/include/asm/cpufeature.h @@ -7,6 +7,7 @@ #define __ASM_CPUFEATURE_H #include <asm/alternative-macros.h> +#include <asm/cache.h> #include <asm/cpucaps.h> #include <asm/cputype.h> #include <asm/hwcap.h> @@ -917,6 +918,13 @@ extern struct arm64_ftr_override id_aa64isar2_override; u32 get_kvm_ipa_limit(void); void dump_cpu_features(void); +struct ccsidr { + u64 data; + u64 inst; +}; + +extern struct ccsidr ccsidr[MAX_CACHE_LEVEL + 1]; + #endif /* __ASSEMBLY__ */ #endif diff --git a/arch/arm64/include/asm/sysreg.h b/arch/arm64/include/asm/sysreg.h index 7d301700d1a9..e796f14fdc2a 100644 --- a/arch/arm64/include/asm/sysreg.h +++ b/arch/arm64/include/asm/sysreg.h @@ -941,6 +941,13 @@ #define HFGxTR_EL2_nSMPRI_EL1_SHIFT 54 #define HFGxTR_EL2_nSMPRI_EL1_MASK BIT_MASK(HFGxTR_EL2_nSMPRI_EL1_SHIFT) +/* CCSIDR_EL1 bit definitions */ +#define CCSIDR_ASSOCIATIVITY_BITS_MASK GENMASK(27, 3) + +/* CSSELR_EL1 */ +#define CSSELR_IN 1 +#define CSSELR_LEVEL_SHIFT 1 + #define ARM64_FEATURE_FIELD_BITS 4 /* Create a mask for the feature bits of the specified feature. */ diff --git a/arch/arm64/kernel/cacheinfo.c b/arch/arm64/kernel/cacheinfo.c index 97c42be71338..2e808ccc15bf 100644 --- a/arch/arm64/kernel/cacheinfo.c +++ b/arch/arm64/kernel/cacheinfo.c @@ -10,7 +10,6 @@ #include <linux/cacheinfo.h> #include <linux/of.h> -#define MAX_CACHE_LEVEL 7 /* Max 7 level supported */ /* Ctypen, bits[3(n - 1) + 2 : 3(n - 1)], for n = 1 to 7 */ #define CLIDR_CTYPE_SHIFT(level) (3 * (level - 1)) #define CLIDR_CTYPE_MASK(level) (7 << CLIDR_CTYPE_SHIFT(level)) @@ -26,7 +25,7 @@ int cache_line_size(void) } EXPORT_SYMBOL_GPL(cache_line_size); -static inline enum cache_type get_cache_type(int level) +enum cache_type get_cache_type(int level) { u64 clidr; @@ -35,6 +34,7 @@ static inline enum cache_type get_cache_type(int level) clidr = read_sysreg(clidr_el1); return CLIDR_CTYPE(clidr, level); } +EXPORT_SYMBOL_GPL(get_cache_type); static void ci_leaf_init(struct cacheinfo *this_leaf, enum cache_type type, unsigned int level) diff --git a/arch/arm64/kernel/cpu_errata.c b/arch/arm64/kernel/cpu_errata.c index 89ac00084f38..5caccf602fc0 100644 --- a/arch/arm64/kernel/cpu_errata.c +++ b/arch/arm64/kernel/cpu_errata.c @@ -8,6 +8,8 @@ #include <linux/arm-smccc.h> #include <linux/types.h> #include <linux/cpu.h> +#include <linux/cacheinfo.h> +#include <asm/cache.h> #include <asm/cpu.h> #include <asm/cputype.h> #include <asm/cpufeature.h> @@ -87,6 +89,50 @@ has_mismatched_cache_type(const struct arm64_cpu_capabilities *entry, return (ctr_real != sys) && (ctr_raw != sys); } +static bool +has_mismatched_cache_associativity(const struct arm64_cpu_capabilities *entry, + int scope) +{ + u64 mask = CCSIDR_ASSOCIATIVITY_BITS_MASK; + u64 real; + bool mismatched = false; + enum cache_type cache_type; + unsigned int i; + + WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible()); + + local_irq_disable(); + + for (i = 0; i <= MAX_CACHE_LEVEL; i++) { + cache_type = get_cache_type(i); + + if ((cache_type & (CACHE_TYPE_DATA | CACHE_TYPE_UNIFIED))) { + write_sysreg(i << CSSELR_LEVEL_SHIFT, csselr_el1); + isb(); + real = read_sysreg(ccsidr_el1); + if ((ccsidr[i].data & mask) != (real & mask)) { + mismatched = true; + break; + } + } + + if ((cache_type & CACHE_TYPE_INST)) { + write_sysreg((i << CSSELR_LEVEL_SHIFT) | CSSELR_IN, + csselr_el1); + isb(); + real = read_sysreg(ccsidr_el1); + if ((ccsidr[i].inst & mask) != (real & mask)) { + mismatched = true; + break; + } + } + } + + local_irq_enable(); + + return mismatched; +} + static void cpu_enable_trap_ctr_access(const struct arm64_cpu_capabilities *cap) { @@ -499,6 +545,12 @@ const struct arm64_cpu_capabilities arm64_errata[] = { ERRATA_MIDR_RANGE_LIST(cavium_erratum_30115_cpus), }, #endif + { + .desc = "Mismatched cache associativity", + .capability = ARM64_MISMATCHED_CACHE_ASSOCIATIVITY, + .matches = has_mismatched_cache_associativity, + .type = ARM64_CPUCAP_LOCAL_CPU_ERRATUM, + }, { .desc = "Mismatched cache type (CTR_EL0)", .capability = ARM64_MISMATCHED_CACHE_TYPE, diff --git a/arch/arm64/kernel/cpufeature.c b/arch/arm64/kernel/cpufeature.c index b3f37e2209ad..ef259396aa4c 100644 --- a/arch/arm64/kernel/cpufeature.c +++ b/arch/arm64/kernel/cpufeature.c @@ -930,6 +930,8 @@ static void init_cpu_ftr_reg(u32 sys_reg, u64 new) reg->user_mask = user_mask; } +struct ccsidr ccsidr[MAX_CACHE_LEVEL + 1]; + extern const struct arm64_cpu_capabilities arm64_errata[]; static const struct arm64_cpu_capabilities arm64_features[]; @@ -1039,6 +1041,8 @@ void __init init_cpu_features(struct cpuinfo_arm64 *info) * after we have initialised the CPU feature infrastructure. */ setup_boot_cpu_capabilities(); + + memcpy(ccsidr, info->reg_ccsidr, sizeof(ccsidr)); } static void update_cpu_ftr_reg(struct arm64_ftr_reg *reg, u64 new) diff --git a/arch/arm64/kernel/cpuinfo.c b/arch/arm64/kernel/cpuinfo.c index 28d4f442b0bc..b1ea276b9d10 100644 --- a/arch/arm64/kernel/cpuinfo.c +++ b/arch/arm64/kernel/cpuinfo.c @@ -13,6 +13,7 @@ #include <linux/bitops.h> #include <linux/bug.h> +#include <linux/cacheinfo.h> #include <linux/compat.h> #include <linux/elf.h> #include <linux/init.h> @@ -47,6 +48,34 @@ static inline const char *icache_policy_str(int l1ip) } } +static void read_ccsidr(struct ccsidr *ccsidr) +{ + enum cache_type cache_type; + unsigned int i; + + local_irq_disable(); + + for (i = 0; i <= MAX_CACHE_LEVEL; i++) { + cache_type = get_cache_type(i); + + if ((cache_type & (CACHE_TYPE_DATA | CACHE_TYPE_UNIFIED))) { + write_sysreg(i << CSSELR_LEVEL_SHIFT, csselr_el1); + isb(); + ccsidr[i].data = read_sysreg(ccsidr_el1); + break; + } + + if ((cache_type & CACHE_TYPE_INST)) { + write_sysreg((i << CSSELR_LEVEL_SHIFT) | CSSELR_IN, + csselr_el1); + isb(); + ccsidr[i].inst = read_sysreg(ccsidr_el1); + } + } + + local_irq_enable(); +} + unsigned long __icache_flags; static const char *const hwcap_str[] = { @@ -440,6 +469,7 @@ static void __cpuinfo_store_cpu(struct cpuinfo_arm64 *info) if (id_aa64pfr0_32bit_el0(info->reg_id_aa64pfr0)) __cpuinfo_store_cpu_32bit(&info->aarch32); + read_ccsidr(info->reg_ccsidr); cpuinfo_detect_icache_policy(info); } diff --git a/arch/arm64/kvm/sys_regs.c b/arch/arm64/kvm/sys_regs.c index 273ed1aaa6b3..1f0cb015e81c 100644 --- a/arch/arm64/kvm/sys_regs.c +++ b/arch/arm64/kvm/sys_regs.c @@ -111,8 +111,8 @@ static u32 get_ccsidr(struct kvm_vcpu *vcpu, u32 csselr) * geometry (which is not permitted by the architecture), they would * only do so for virtually indexed caches.] */ - if (vcpu_cache_overridden(vcpu) && !(csselr & 1)) // data or unified cache - ccsidr &= ~GENMASK(27, 3); + if (vcpu_cache_overridden(vcpu) && !(csselr & CSSELR_IN)) // data or unified cache + ccsidr &= ~CCSIDR_ASSOCIATIVITY_BITS_MASK; return ccsidr; } diff --git a/arch/arm64/tools/cpucaps b/arch/arm64/tools/cpucaps index f1c0347ec31a..061c93319295 100644 --- a/arch/arm64/tools/cpucaps +++ b/arch/arm64/tools/cpucaps @@ -44,6 +44,7 @@ HAS_VIRT_HOST_EXTN HAS_WFXT HW_DBM KVM_PROTECTED_MODE +MISMATCHED_CACHE_ASSOCIATIVITY MISMATCHED_CACHE_TYPE MTE MTE_ASYMM -- 2.38.1 _______________________________________________ kvmarm mailing list kvmarm@xxxxxxxxxxxxxxxxxxxxx https://lists.cs.columbia.edu/mailman/listinfo/kvmarm