On Tue, 27 Jul 2021 19:11:08 +0100, Will Deacon <will@xxxxxxxxxx> wrote: > > On Thu, Jul 15, 2021 at 05:31:47PM +0100, Marc Zyngier wrote: > > Introduce the infrastructure required to identify an IPA region > > that is expected to be used as an MMIO window. > > > > This include mapping, unmapping and checking the regions. Nothing > > calls into it yet, so no expected functional change. > > > > Signed-off-by: Marc Zyngier <maz@xxxxxxxxxx> > > --- > > arch/arm64/include/asm/kvm_host.h | 2 + > > arch/arm64/include/asm/kvm_mmu.h | 5 ++ > > arch/arm64/kvm/mmu.c | 115 ++++++++++++++++++++++++++++++ > > 3 files changed, 122 insertions(+) > > > > diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h > > index 4add6c27251f..914c1b7bb3ad 100644 > > --- a/arch/arm64/include/asm/kvm_host.h > > +++ b/arch/arm64/include/asm/kvm_host.h > > @@ -125,6 +125,8 @@ struct kvm_arch { > > #define KVM_ARCH_FLAG_RETURN_NISV_IO_ABORT_TO_USER 0 > > /* Memory Tagging Extension enabled for the guest */ > > #define KVM_ARCH_FLAG_MTE_ENABLED 1 > > + /* Gues has bought into the MMIO guard extension */ > > +#define KVM_ARCH_FLAG_MMIO_GUARD 2 > > unsigned long flags; > > > > /* > > diff --git a/arch/arm64/include/asm/kvm_mmu.h b/arch/arm64/include/asm/kvm_mmu.h > > index b52c5c4b9a3d..f6b8fc1671b3 100644 > > --- a/arch/arm64/include/asm/kvm_mmu.h > > +++ b/arch/arm64/include/asm/kvm_mmu.h > > @@ -170,6 +170,11 @@ phys_addr_t kvm_mmu_get_httbr(void); > > phys_addr_t kvm_get_idmap_vector(void); > > int kvm_mmu_init(u32 *hyp_va_bits); > > > > +/* MMIO guard */ > > +bool kvm_install_ioguard_page(struct kvm_vcpu *vcpu, gpa_t ipa); > > +bool kvm_remove_ioguard_page(struct kvm_vcpu *vcpu, gpa_t ipa); > > +bool kvm_check_ioguard_page(struct kvm_vcpu *vcpu, gpa_t ipa); > > + > > static inline void *__kvm_vector_slot2addr(void *base, > > enum arm64_hyp_spectre_vector slot) > > { > > diff --git a/arch/arm64/kvm/mmu.c b/arch/arm64/kvm/mmu.c > > index 3155c9e778f0..638827c8842b 100644 > > --- a/arch/arm64/kvm/mmu.c > > +++ b/arch/arm64/kvm/mmu.c > > @@ -1120,6 +1120,121 @@ static void handle_access_fault(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa) > > kvm_set_pfn_accessed(pte_pfn(pte)); > > } > > > > +#define MMIO_NOTE ('M' << 24 | 'M' << 16 | 'I' << 8 | '0') > > Although this made me smile, maybe we should carve up the bit space a bit > more carefully ;) Also, you know somebody clever will "fix" that typo to > 'O'! They'll get to keep the pieces when the whole thing breaks! More seriously, happy to have a more elaborate allocation scheme. For the purpose of this series, it really doesn't matter. > Quentin, as the other user of this stuff at the moment, how do you see the > annotation space being allocated? Feels like we should have some 'type' > bits which decide how to parse the rest of the entry. > > > + > > +bool kvm_install_ioguard_page(struct kvm_vcpu *vcpu, gpa_t ipa) > > +{ > > + struct kvm_mmu_memory_cache *memcache; > > + struct kvm_memory_slot *memslot; > > + int ret, idx; > > + > > + if (!test_bit(KVM_ARCH_FLAG_MMIO_GUARD, &vcpu->kvm->arch.flags)) > > + return false; > > + > > + /* Must be page-aligned */ > > + if (ipa & ~PAGE_MASK) > > + return false; > > + > > + /* > > + * The page cannot be in a memslot. At some point, this will > > + * have to deal with device mappings though. > > + */ > > + idx = srcu_read_lock(&vcpu->kvm->srcu); > > + memslot = gfn_to_memslot(vcpu->kvm, ipa >> PAGE_SHIFT); > > + srcu_read_unlock(&vcpu->kvm->srcu, idx); > > What does this memslot check achieve? A new memslot could be added after > you've checked, no? If you start allowing S2 annotations to coexist with potential memory mappings, you're in for trouble. The faulting logic will happily overwrite the annotation, and that's probably not what you want. As for new (or moving) memslots, I guess they should be checked against existing annotations. > > > +/* Assumes mmu_lock taken */ > > You can use a lockdep assertion for that! Sure. Thanks, M. -- Without deviation from the norm, progress is not possible.