On Mon, Jun 27, 2022 at 07:17:09PM -0700, John Hubbard wrote: > On 6/22/22 14:36, Peter Xu wrote: > > Merge two boolean parameters into a bitmask flag called kvm_gtp_flag_t for > > __gfn_to_pfn_memslot(). This cleans the parameter lists, and also prepare > > for new boolean to be added to __gfn_to_pfn_memslot(). > > > > Signed-off-by: Peter Xu <peterx@xxxxxxxxxx> > > --- > > arch/arm64/kvm/mmu.c | 5 ++-- > > arch/powerpc/kvm/book3s_64_mmu_hv.c | 5 ++-- > > arch/powerpc/kvm/book3s_64_mmu_radix.c | 5 ++-- > > arch/x86/kvm/mmu/mmu.c | 10 +++---- > > include/linux/kvm_host.h | 9 ++++++- > > virt/kvm/kvm_main.c | 37 +++++++++++++++----------- > > virt/kvm/kvm_mm.h | 6 +++-- > > virt/kvm/pfncache.c | 2 +- > > 8 files changed, 49 insertions(+), 30 deletions(-) > > > > diff --git a/arch/arm64/kvm/mmu.c b/arch/arm64/kvm/mmu.c > > index f5651a05b6a8..ce1edb512b4e 100644 > > --- a/arch/arm64/kvm/mmu.c > > +++ b/arch/arm64/kvm/mmu.c > > @@ -1204,8 +1204,9 @@ static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa, > > */ > > smp_rmb(); > > - pfn = __gfn_to_pfn_memslot(memslot, gfn, false, NULL, > > - write_fault, &writable, NULL); > > + pfn = __gfn_to_pfn_memslot(memslot, gfn, > > + write_fault ? KVM_GTP_WRITE : 0, > > + NULL, &writable, NULL); > > if (pfn == KVM_PFN_ERR_HWPOISON) { > > kvm_send_hwpoison_signal(hva, vma_shift); > > return 0; > > diff --git a/arch/powerpc/kvm/book3s_64_mmu_hv.c b/arch/powerpc/kvm/book3s_64_mmu_hv.c > > index 514fd45c1994..e2769d58dd87 100644 > > --- a/arch/powerpc/kvm/book3s_64_mmu_hv.c > > +++ b/arch/powerpc/kvm/book3s_64_mmu_hv.c > > @@ -598,8 +598,9 @@ int kvmppc_book3s_hv_page_fault(struct kvm_vcpu *vcpu, > > write_ok = true; > > } else { > > /* Call KVM generic code to do the slow-path check */ > > - pfn = __gfn_to_pfn_memslot(memslot, gfn, false, NULL, > > - writing, &write_ok, NULL); > > + pfn = __gfn_to_pfn_memslot(memslot, gfn, > > + writing ? KVM_GTP_WRITE : 0, > > + NULL, &write_ok, NULL); > > if (is_error_noslot_pfn(pfn)) > > return -EFAULT; > > page = NULL; > > diff --git a/arch/powerpc/kvm/book3s_64_mmu_radix.c b/arch/powerpc/kvm/book3s_64_mmu_radix.c > > index 42851c32ff3b..232b17c75b83 100644 > > --- a/arch/powerpc/kvm/book3s_64_mmu_radix.c > > +++ b/arch/powerpc/kvm/book3s_64_mmu_radix.c > > @@ -845,8 +845,9 @@ int kvmppc_book3s_instantiate_page(struct kvm_vcpu *vcpu, > > unsigned long pfn; > > /* Call KVM generic code to do the slow-path check */ > > - pfn = __gfn_to_pfn_memslot(memslot, gfn, false, NULL, > > - writing, upgrade_p, NULL); > > + pfn = __gfn_to_pfn_memslot(memslot, gfn, > > + writing ? KVM_GTP_WRITE : 0, > > + NULL, upgrade_p, NULL); > > if (is_error_noslot_pfn(pfn)) > > return -EFAULT; > > page = NULL; > > diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c > > index f4653688fa6d..e92f1ab63d6a 100644 > > --- a/arch/x86/kvm/mmu/mmu.c > > +++ b/arch/x86/kvm/mmu/mmu.c > > @@ -3968,6 +3968,7 @@ void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, struct kvm_async_pf *work) > > static int kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault) > > { > > + kvm_gtp_flag_t flags = fault->write ? KVM_GTP_WRITE : 0; > > struct kvm_memory_slot *slot = fault->slot; > > bool async; > > @@ -3999,8 +4000,8 @@ static int kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault) > > } > > async = false; > > - fault->pfn = __gfn_to_pfn_memslot(slot, fault->gfn, false, &async, > > - fault->write, &fault->map_writable, > > + fault->pfn = __gfn_to_pfn_memslot(slot, fault->gfn, flags, > > + &async, &fault->map_writable, > > &fault->hva); > > if (!async) > > return RET_PF_CONTINUE; /* *pfn has correct page already */ > > @@ -4016,9 +4017,8 @@ static int kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault) > > } > > } > > - fault->pfn = __gfn_to_pfn_memslot(slot, fault->gfn, false, NULL, > > - fault->write, &fault->map_writable, > > - &fault->hva); > > + fault->pfn = __gfn_to_pfn_memslot(slot, fault->gfn, flags, NULL, > > + &fault->map_writable, &fault->hva); > > The flags arg does improve the situation, yes. Thanks for supporting a flag's existance. :) I'd say ultimately it could be a personal preference thing when the struct comes. > > > return RET_PF_CONTINUE; > > } > > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > > index c20f2d55840c..b646b6fcaec6 100644 > > --- a/include/linux/kvm_host.h > > +++ b/include/linux/kvm_host.h > > @@ -1146,8 +1146,15 @@ kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, > > bool *writable); > > kvm_pfn_t gfn_to_pfn_memslot(const struct kvm_memory_slot *slot, gfn_t gfn); > > kvm_pfn_t gfn_to_pfn_memslot_atomic(const struct kvm_memory_slot *slot, gfn_t gfn); > > + > > +/* gfn_to_pfn (gtp) flags */ > > +typedef unsigned int __bitwise kvm_gtp_flag_t; > > A minor naming problem: GTP and especially gtp_flags is way too close > to gfp_flags. It will make people either wonder if it's a typo, or > worse, *assume* that it's a typo. :) I'd try to argu with "I prefixed it with kvm_", but oh well.. yes they're a bit close :) > > Yes, "read the code", but if you can come up with a better TLA than GTP > here, let's consider using it. Could I ask what's TLA? Any suggestions on the abbrev, btw? > > Overall, the change looks like an improvement, even though > > write_fault ? KVM_GTP_WRITE : 0 > > is not wonderful. But improving *that* leads to a a big pile of diffs > that are rather beyond the scope here. Thanks, -- Peter Xu