On Tue, 13 Aug 2019 10:04:41 -0700 Sean Christopherson <sean.j.christopherson@xxxxxxxxx> wrote: > On Tue, Aug 13, 2019 at 10:04:58AM -0600, Alex Williamson wrote: > > On Tue, 5 Feb 2019 13:01:21 -0800 > > Sean Christopherson <sean.j.christopherson@xxxxxxxxx> wrote: > > > > > Modify kvm_mmu_invalidate_zap_pages_in_memslot(), a.k.a. the x86 MMU's > > > handler for kvm_arch_flush_shadow_memslot(), to zap only the pages/PTEs > > > that actually belong to the memslot being removed. This improves > > > performance, especially why the deleted memslot has only a few shadow > > > entries, or even no entries. E.g. a microbenchmark to access regular > > > memory while concurrently reading PCI ROM to trigger memslot deletion > > > showed a 5% improvement in throughput. > > > > > > Cc: Xiao Guangrong <guangrong.xiao@xxxxxxxxx> > > > Signed-off-by: Sean Christopherson <sean.j.christopherson@xxxxxxxxx> > > > --- > > > arch/x86/kvm/mmu.c | 33 ++++++++++++++++++++++++++++++++- > > > 1 file changed, 32 insertions(+), 1 deletion(-) > > > > A number of vfio users are reporting VM instability issues since v5.1, > > some have traced it back to this commit 4e103134b862 ("KVM: x86/mmu: Zap > > only the relevant pages when removing a memslot"), which I've confirmed > > via bisection of the 5.1 merge window KVM pull (636deed6c0bc) and > > re-verified on current 5.3-rc4 using the below patch to toggle the > > broken behavior. > > > > My reproducer is a Windows 10 VM with assigned GeForce GPU running a > > variety of tests, including FurMark and PassMark Performance Test. > > With the code enabled as exists in upstream currently, PassMark will > > generally introduce graphics glitches or hangs. Sometimes it's > > necessary to reboot the VM to see these issues. > > As in, the issue only shows up when the VM is rebooted? Just want to > double check that that's not a typo. No, it can occur on the first boot as well, it's just that the recipe to induce a failure is not well understood and manifests itself in different ways. I generally run the tests, then if it still hasn't reproduced, I reboot the VM a couple times, running a couple apps in between to try to trigger/notice bad behavior. > > Flipping the 0/1 in the below patch appears to resolve the issue. > > > > I'd appreciate any insights into further debugging this block of code > > so that we can fix this regression. Thanks, > > If it's not too painful to reproduce, I'd say start by determining whether > it's a problem with the basic logic or if the cond_resched_lock() handling > is wrong. I.e. comment/ifdef out this chunk: > > if (need_resched() || spin_needbreak(&kvm->mmu_lock)) { > kvm_mmu_remote_flush_or_zap(kvm, &invalid_list, flush); > flush = false; > cond_resched_lock(&kvm->mmu_lock); > } If anything, removing this chunk seems to make things worse. Thanks, Alex