+ Lorenzo, he really needs to be on this thread. On Tue, Dec 05, 2023 at 09:21:28AM +0000, Marc Zyngier wrote: > On Tue, 05 Dec 2023 03:30:15 +0000, > <ankita@xxxxxxxxxx> wrote: > > From: Ankit Agrawal <ankita@xxxxxxxxxx> > > > > Currently, KVM for ARM64 maps at stage 2 memory that is considered device > > (i.e. it is not RAM) with DEVICE_nGnRE memory attributes; this setting > > overrides (as per the ARM architecture [1]) any device MMIO mapping > > present at stage 1, resulting in a set-up whereby a guest operating > > system cannot determine device MMIO mapping memory attributes on its > > own but it is always overridden by the KVM stage 2 default. [...] > Despite the considerable increase in the commit message length, a > number of questions are left unanswered: > > - Shameer reported a regression on non-FWB systems, breaking device > assignment: > > https://lore.kernel.org/all/af13ed63dc9a4f26a6c958ebfa77d78a@xxxxxxxxxx/ This referred to the first patch in the old series which was trying to make the Stage 2 cacheable based on the vma attributes. That patch has been dropped for now. It would be good if Shameer confirms this was the problem. > - Will had unanswered questions in another part of the thread: > > https://lore.kernel.org/all/20231013092954.GB13524@willie-the-truck/ > > Can someone please help concluding it? Is this about reclaiming the device? I think we concluded that we can't generalise this beyond PCIe, though not sure there was any formal statement to that thread. The other point Will had was around stating in the commit message why we only relax this to Normal NC. I haven't checked the commit message yet, it needs careful reading ;). > > diff --git a/arch/arm64/kvm/mmu.c b/arch/arm64/kvm/mmu.c > > index d14504821b79..1cb302457d3f 100644 > > --- a/arch/arm64/kvm/mmu.c > > +++ b/arch/arm64/kvm/mmu.c > > @@ -1071,7 +1071,7 @@ int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa, > > struct kvm_mmu_memory_cache cache = { .gfp_zero = __GFP_ZERO }; > > struct kvm_s2_mmu *mmu = &kvm->arch.mmu; > > struct kvm_pgtable *pgt = mmu->pgt; > > - enum kvm_pgtable_prot prot = KVM_PGTABLE_PROT_DEVICE | > > + enum kvm_pgtable_prot prot = KVM_PGTABLE_PROT_NORMAL_NC | > > KVM_PGTABLE_PROT_R | > > (writable ? KVM_PGTABLE_PROT_W : 0); > > Doesn't this affect the GICv2 VCPU interface, which is effectively a > shared peripheral, now allowing a guest to affect another guest's > interrupt distribution? If that is the case, this needs to be fixed. > > In general, I don't think this should be a blanket statement, but be > limited to devices that we presume can deal with this (i.e. PCIe, and > not much else). Based on other on-list and off-line discussions, I came to the same conclusion that we can't relax this beyond PCIe. How we do this, it's up for debate. Some ideas: - something in the vfio-pci driver like a new VM_* flag that KVM can consume (hard sell for an arch-specific thing) - changing the vfio-pci driver to allow WC and NC mappings (x86 terminology) with additional knowledge about what it can safely map as NC. KVM would mimic the vma attributes (it doesn't eliminate the alias though, the guest can always go for Device while the VMM for Normal) - some per-SoC pfn ranges that are permitted as Normal NC. Can the arch code figure out where these PCIe BARs are or is it only the vfio-pci driver? I guess we can sort something out around the PCIe but I'm not familiar with this subsystem. The alternative is some "safe" ranges in firmware tables, assuming the firmware configures the PCIe BARs location -- Catalin