On 3/4/22 20:48, isaku.yamahata@xxxxxxxxx wrote:
From: Sean Christopherson <sean.j.christopherson@xxxxxxxxx>
Explicitly check for an MMIO spte in the fast page fault flow. TDX will
use a not-present entry for MMIO sptes, which can be mistaken for an
access-tracked spte since both have SPTE_SPECIAL_MASK set.
The fast page fault handles the case of changing access bits without
obtaining mmu_lock. For example, clear write protect bit for dirty page
tracking. MMIO emulation is handled in a slow path. So it doesn't affect
"MMIO sptes are handled in handle_mmio_page_fault for non-TDX VMs, so
this patch does not affect them. TDX will handle MMIO emulation through
a hypercall instead".
For this comment, it is not necessary to talk about the slow path, since
that is just where MMIO sptes are installed. If the slow path is
reached, fast_page_fault must not have seen is_mmio_spte(spte).
@@ -3167,7 +3167,7 @@ static int fast_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
break;
sp = sptep_to_sp(sptep);
- if (!is_last_spte(spte, sp->role.level))
+ if (!is_last_spte(spte, sp->role.level) || is_mmio_spte(spte))
break;
/*
I would include the check a couple lines before:
if (!is_shadow_present_pte(spte) || is_mmio_spte(spte))
This matches what is in the commit message: the problem is that MMIO
SPTEs are present in the TDX case, so you need to check them even if
is_shadow_present_pte(spte) returns true.
Paolo