On Sat, May 21, 2022, Lai Jiangshan wrote: > From: Lai Jiangshan <jiangshan.ljs@xxxxxxxxxxxx> > > mmu->pae_root for non-PAE paging is allocated on-demand, but > mmu->pae_root for PAE paging is allocated early when struct kvm_mmu is > being created. > > Simplify the code to allocate mmu->pae_root for PAE paging and make > it on-demand. Hmm, I'm not convinced this simplifies things enough to justify the risk. There's a non-zero chance that the __GFP_DMA32 allocation was intentionally done during VM creation in order to avoid OOM on low memory. Maybe move this patch to the tail end of the series so that it has a higher chance of reverting cleanly if on-demand allocation breaks someone's setup? > Signed-off-by: Lai Jiangshan <jiangshan.ljs@xxxxxxxxxxxx> > --- > arch/x86/include/asm/kvm_host.h | 2 +- > arch/x86/kvm/mmu/mmu.c | 101 +++++++++++++------------------- > arch/x86/kvm/x86.c | 4 +- > 3 files changed, 44 insertions(+), 63 deletions(-) > > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h > index 9cdc5bbd721f..fb9751dfc1a7 100644 > --- a/arch/x86/include/asm/kvm_host.h > +++ b/arch/x86/include/asm/kvm_host.h > @@ -1615,7 +1615,7 @@ int kvm_mmu_vendor_module_init(void); > void kvm_mmu_vendor_module_exit(void); > > void kvm_mmu_destroy(struct kvm_vcpu *vcpu); > -int kvm_mmu_create(struct kvm_vcpu *vcpu); > +void kvm_mmu_create(struct kvm_vcpu *vcpu); > int kvm_mmu_init_vm(struct kvm *kvm); > void kvm_mmu_uninit_vm(struct kvm *kvm); > > diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c > index 90b715eefe6a..63c2b2c6122c 100644 > --- a/arch/x86/kvm/mmu/mmu.c > +++ b/arch/x86/kvm/mmu/mmu.c > @@ -668,6 +668,41 @@ static void walk_shadow_page_lockless_end(struct kvm_vcpu *vcpu) > } > } > > +static int mmu_alloc_pae_root(struct kvm_vcpu *vcpu) Now that pae_root isn't the "full" root, just the page table, I think we should rename pae_root to something else, and then name this accordingly. pae_root_backing_page and mmu_alloc_pae_root_backing_page()? Definitely don't love the name if someone has a better idea. > +{ > + struct page *page; > + > + if (vcpu->arch.mmu->root_role.level != PT32E_ROOT_LEVEL) > + return 0; I think I'd prefer to move this check to the caller, it's confusing to see an unconditional call to a PAE-specific helper. > + if (vcpu->arch.mmu->pae_root) > + return 0; > + > + /* > + * Allocate a page to hold the four PDPTEs for PAE paging when emulating > + * 32-bit mode. CR3 is only 32 bits even on x86_64 in this case. > + * Therefore we need to allocate the PDP table in the first 4GB of > + * memory, which happens to fit the DMA32 zone. > + */ > + page = alloc_page(GFP_KERNEL_ACCOUNT | __GFP_ZERO | __GFP_DMA32); Leave off __GFP_ZERO, it's unnecesary in both cases, and actively misleading in when TDP is disabled. KVM _must_ write the page after making it decrypted. And since I can't find any code that actually does initialize "pae_root", I suspect this series is buggy. But if there is a bug, it was introduced earlier in this series, either by KVM: X86/MMU: Add local shadow pages or by KVM: X86/MMU: Activate local shadow pages and remove old logic depending on whether you want to blame the function that is buggy, or the patch that uses the buggy function.. The right place to initialize the root is kvm_mmu_alloc_local_shadow_page(). KVM sets __GFP_ZERO for mmu_shadow_page_cache, i.e. relies on new sp->spt pages to be zeroed prior to "allocating" from the cache. The PAE root backing page on the other hand is allocated once and then reused over and over. if (role.level == PT32E_ROOT_LEVEL && !WARN_ON_ONCE(!vcpu->arch.mmu->pae_root)) { sp->spt = vcpu->arch.mmu->pae_root; kvm_mmu_initialize_pae_root(sp->spt): <==== something like this } else { sp->spt = kvm_mmu_memory_cache_alloc(&vcpu->arch.mmu_shadow_page_cache); } > - for (i = 0; i < 4; ++i) > - mmu->pae_root[i] = INVALID_PAE_ROOT; Please remove this code in a separate patch. I don't care if it is removed before or after (I'm pretty sure the existing behavior is paranoia), but I don't want multiple potentially-functional changes in this patch.