Hi Fuad, On Monday 02 Aug 2021 at 11:49:28 (+0200), Fuad Tabba wrote: > On Thu, Jul 29, 2021 at 3:28 PM Quentin Perret <qperret@xxxxxxxxxx> wrote: > > > > Much of the stage-2 manipulation logic relies on being able to destroy > > block mappings if e.g. installing a smaller mapping in the range. The > > rationale for this behaviour is that stage-2 mappings can always be > > re-created lazily. However, this gets more complicated when the stage-2 > > page-table is used to store metadata about the underlying pages. In such > > cases, destroying a block mapping may lead to losing part of the state, > > and confuse the user of those metadata (such as the hypervisor in nVHE > > protected mode). > > > > To avoid this, introduce a callback function in the pgtable struct which > > is called during all map operations to determine whether the mappings > > can use blocks, or should be forced to page granularity. This is used by > > the hypervisor when creating the host stage-2 to force page-level > > mappings when using non-default protection attributes. > > > > Signed-off-by: Quentin Perret <qperret@xxxxxxxxxx> > > --- > > arch/arm64/include/asm/kvm_pgtable.h | 65 ++++++++++++++++----------- > > arch/arm64/kvm/hyp/nvhe/mem_protect.c | 30 +++++++++++-- > > arch/arm64/kvm/hyp/pgtable.c | 29 +++++++++--- > > 3 files changed, 91 insertions(+), 33 deletions(-) > > > > diff --git a/arch/arm64/include/asm/kvm_pgtable.h b/arch/arm64/include/asm/kvm_pgtable.h > > index 83c5c97d9eac..ba7dcade2798 100644 > > --- a/arch/arm64/include/asm/kvm_pgtable.h > > +++ b/arch/arm64/include/asm/kvm_pgtable.h > > @@ -115,25 +115,6 @@ enum kvm_pgtable_stage2_flags { > > KVM_PGTABLE_S2_IDMAP = BIT(1), > > }; > > > > -/** > > - * struct kvm_pgtable - KVM page-table. > > - * @ia_bits: Maximum input address size, in bits. > > - * @start_level: Level at which the page-table walk starts. > > - * @pgd: Pointer to the first top-level entry of the page-table. > > - * @mm_ops: Memory management callbacks. > > - * @mmu: Stage-2 KVM MMU struct. Unused for stage-1 page-tables. > > - */ > > -struct kvm_pgtable { > > - u32 ia_bits; > > - u32 start_level; > > - kvm_pte_t *pgd; > > - struct kvm_pgtable_mm_ops *mm_ops; > > - > > - /* Stage-2 only */ > > - struct kvm_s2_mmu *mmu; > > - enum kvm_pgtable_stage2_flags flags; > > -}; > > - > > /** > > * enum kvm_pgtable_prot - Page-table permissions and attributes. > > * @KVM_PGTABLE_PROT_X: Execute permission. > > @@ -149,11 +130,41 @@ enum kvm_pgtable_prot { > > KVM_PGTABLE_PROT_DEVICE = BIT(3), > > }; > > > > -#define PAGE_HYP (KVM_PGTABLE_PROT_R | KVM_PGTABLE_PROT_W) > > +#define KVM_PGTABLE_PROT_RW (KVM_PGTABLE_PROT_R | KVM_PGTABLE_PROT_W) > > +#define KVM_PGTABLE_PROT_RWX (KVM_PGTABLE_PROT_RW | KVM_PGTABLE_PROT_X) > > + > > +#define PAGE_HYP KVM_PGTABLE_PROT_RW > > #define PAGE_HYP_EXEC (KVM_PGTABLE_PROT_R | KVM_PGTABLE_PROT_X) > > #define PAGE_HYP_RO (KVM_PGTABLE_PROT_R) > > #define PAGE_HYP_DEVICE (PAGE_HYP | KVM_PGTABLE_PROT_DEVICE) > > I wonder if it would be useful to add a couple of other aliases for > default memory and default mmio protections, e.g., > > #define KVM_PGTABLE_PROT_MEM KVM_PGTABLE_PROT_RWX > #define KVM_PGTABLE_PROT_MMIO KVM_PGTABLE_PROT_RW > > I think that using these below, e.g., host_stage2_force_pte_cb(), > might make it clearer and answer comments you had in earlier patches > about why "RWX" for memory. Sure I can add something. I'll probably call them something else than KVM_PGTABLE_PROT_{MEM,MMIO} though, just to make it clear this is all specific to the host stage-2 stuff and not a general requirement of the pgtable code to map things like this. > > > > +typedef bool (*kvm_pgtable_force_pte_cb_t)(u64 addr, u64 end, > > + enum kvm_pgtable_prot prot); > > + > > +/** > > + * struct kvm_pgtable - KVM page-table. > > + * @ia_bits: Maximum input address size, in bits. > > + * @start_level: Level at which the page-table walk starts. > > + * @pgd: Pointer to the first top-level entry of the page-table. > > + * @mm_ops: Memory management callbacks. > > + * @mmu: Stage-2 KVM MMU struct. Unused for stage-1 page-tables. > > + * @flags: Stage-2 page-table flags. > > + * @force_pte_cb: Callback function used during map operations to decide > > + * whether block mappings can be used to map the given IPA > > + * range. > > + */ > > nit: I think it might be clearer (and probably not longer) to rephrase > to describe in terms of the return value of the callback, e.g., "... > function that returns true if page level mappings must be used instead > of block mappings." Works for me, thanks for the suggestion. > > +struct kvm_pgtable { > > + u32 ia_bits; > > + u32 start_level; > > + kvm_pte_t *pgd; > > + struct kvm_pgtable_mm_ops *mm_ops; > > + > > + /* Stage-2 only */ > > + struct kvm_s2_mmu *mmu; > > + enum kvm_pgtable_stage2_flags flags; > > + kvm_pgtable_force_pte_cb_t force_pte_cb; > > +}; > > + > > /** > > * enum kvm_pgtable_walk_flags - Flags to control a depth-first page-table walk. > > * @KVM_PGTABLE_WALK_LEAF: Visit leaf entries, including invalid > > @@ -246,21 +257,25 @@ int kvm_pgtable_hyp_map(struct kvm_pgtable *pgt, u64 addr, u64 size, u64 phys, > > u64 kvm_get_vtcr(u64 mmfr0, u64 mmfr1, u32 phys_shift); > > > > /** > > - * kvm_pgtable_stage2_init_flags() - Initialise a guest stage-2 page-table. > > + * __kvm_pgtable_stage2_init() - Initialise a guest stage-2 page-table. > > * @pgt: Uninitialised page-table structure to initialise. > > * @arch: Arch-specific KVM structure representing the guest virtual > > * machine. > > * @mm_ops: Memory management callbacks. > > * @flags: Stage-2 configuration flags. > > + * @force_pte_cb: Callback function used during map operations to decide > > + * whether block mappings can be used to map the given IPA > > + * range. > > nit: same nit as above with describing the callback in terms of its return value > > > * > > * Return: 0 on success, negative error code on failure. > > */ > > -int kvm_pgtable_stage2_init_flags(struct kvm_pgtable *pgt, struct kvm_arch *arch, > > - struct kvm_pgtable_mm_ops *mm_ops, > > - enum kvm_pgtable_stage2_flags flags); > > +int __kvm_pgtable_stage2_init(struct kvm_pgtable *pgt, struct kvm_arch *arch, > > + struct kvm_pgtable_mm_ops *mm_ops, > > + enum kvm_pgtable_stage2_flags flags, > > + kvm_pgtable_force_pte_cb_t force_pte_cb); > > > > #define kvm_pgtable_stage2_init(pgt, arch, mm_ops) \ > > - kvm_pgtable_stage2_init_flags(pgt, arch, mm_ops, 0) > > + __kvm_pgtable_stage2_init(pgt, arch, mm_ops, 0, NULL) > > > > /** > > * kvm_pgtable_stage2_destroy() - Destroy an unused guest stage-2 page-table. > > diff --git a/arch/arm64/kvm/hyp/nvhe/mem_protect.c b/arch/arm64/kvm/hyp/nvhe/mem_protect.c > > index 2148d3968aa5..70c57d2c3024 100644 > > --- a/arch/arm64/kvm/hyp/nvhe/mem_protect.c > > +++ b/arch/arm64/kvm/hyp/nvhe/mem_protect.c > > @@ -89,6 +89,7 @@ static void prepare_host_vtcr(void) > > id_aa64mmfr1_el1_sys_val, phys_shift); > > } > > > > +static bool host_stage2_force_pte_cb(u64 addr, u64 end, enum kvm_pgtable_prot prot); > > nit: newline > > > int kvm_host_prepare_stage2(void *pgt_pool_base) > > { > > struct kvm_s2_mmu *mmu = &host_kvm.arch.mmu; > > @@ -101,8 +102,9 @@ int kvm_host_prepare_stage2(void *pgt_pool_base) > > if (ret) > > return ret; > > > > - ret = kvm_pgtable_stage2_init_flags(&host_kvm.pgt, &host_kvm.arch, > > - &host_kvm.mm_ops, KVM_HOST_S2_FLAGS); > > + ret = __kvm_pgtable_stage2_init(&host_kvm.pgt, &host_kvm.arch, > > + &host_kvm.mm_ops, KVM_HOST_S2_FLAGS, > > + host_stage2_force_pte_cb); > > if (ret) > > return ret; > > > > @@ -270,9 +272,31 @@ static int host_stage2_adjust_range(u64 addr, struct kvm_mem_range *range) > > return 0; > > } > > > > +static bool host_stage2_force_pte_cb(u64 addr, u64 end, enum kvm_pgtable_prot prot) > > +{ > > + /* > > + * Block mappings must be used with care in the host stage-2 as a > > + * kvm_pgtable_stage2_map() operation targeting a page in the range of > > + * an existing block will delete the block under the assumption that > > + * mappings in the rest of the block range can always be rebuilt lazily. > > + * That assumption is correct for the host stage-2 with RWX mappings > > + * targeting memory or RW mappings targeting MMIO ranges (see > > + * host_stage2_idmap() below which implements some of the host memory > > + * abort logic). However, this is not safe for any other mappings where > > + * the host stage-2 page-table is in fact the only place where this > > + * state is stored. In all those cases, it is safer to use page-level > > + * mappings, hence avoiding to lose the state because of side-effects in > > + * kvm_pgtable_stage2_map(). > > + */ > > + if (range_is_memory(addr, end)) > > + return prot != KVM_PGTABLE_PROT_RWX; > > + else > > + return prot != KVM_PGTABLE_PROT_RW; > > +} > > Just checking, I don't think that it's possible for the range to be > big enough to somehow include both memory and mmio, neither now nor in > future use cases, is it? That really shouldn't be the case no -- the host_stage2_idmap() function tries hard to respect that, so I figured as long as these two are consistent we should be fine. > > + > > static int host_stage2_idmap(u64 addr) > > { > > - enum kvm_pgtable_prot prot = KVM_PGTABLE_PROT_R | KVM_PGTABLE_PROT_W; > > + enum kvm_pgtable_prot prot = KVM_PGTABLE_PROT_RW; > > struct kvm_mem_range range; > > bool is_memory = find_mem_range(addr, &range); > > int ret; > > diff --git a/arch/arm64/kvm/hyp/pgtable.c b/arch/arm64/kvm/hyp/pgtable.c > > index 2689fcb7901d..bcc02e6e0f62 100644 > > --- a/arch/arm64/kvm/hyp/pgtable.c > > +++ b/arch/arm64/kvm/hyp/pgtable.c > > @@ -452,6 +452,8 @@ int kvm_pgtable_hyp_init(struct kvm_pgtable *pgt, u32 va_bits, > > pgt->start_level = KVM_PGTABLE_MAX_LEVELS - levels; > > pgt->mm_ops = mm_ops; > > pgt->mmu = NULL; > > + pgt->force_pte_cb = NULL; > > + > > return 0; > > } > > > > @@ -489,6 +491,9 @@ struct stage2_map_data { > > void *memcache; > > > > struct kvm_pgtable_mm_ops *mm_ops; > > + > > + /* Force mappings to page granularity */ > > + bool force_pte; > > }; > > > > u64 kvm_get_vtcr(u64 mmfr0, u64 mmfr1, u32 phys_shift) > > @@ -602,6 +607,15 @@ static bool stage2_pte_executable(kvm_pte_t pte) > > return !(pte & KVM_PTE_LEAF_ATTR_HI_S2_XN); > > } > > > > +static bool stage2_block_mapping_allowed(u64 addr, u64 end, u32 level, > > + struct stage2_map_data *data) > > +{ > > + if (data->force_pte && (level < (KVM_PGTABLE_MAX_LEVELS - 1))) > > + return false; > > I'm not sure I understand why checking the level is necessary. Can > there be block mapping at the last possible level? That's probably just a matter of naming, but this function is in fact called at every level, just like kvm_block_mapping_supported() was before. And we rely on it returning true at the last level, so I need to do that check here. Maybe renaming this stage2_leaf_mapping_allowed() would clarify? Thanks, Quentin _______________________________________________ kvmarm mailing list kvmarm@xxxxxxxxxxxxxxxxxxxxx https://lists.cs.columbia.edu/mailman/listinfo/kvmarm