On Fri, May 15, 2009 at 03:57:44PM +0800, Dong, Eddie wrote: > > And what if INTA already happened and CPU is ready to fetch IDT for > > interrupt vector and at this very moment CPU faults? > > If INTA happens, that means it is delivered. If its delivery triggers another > exception, that is what Table5-4 handles. > Is this documented behaviour? Why not execute instruction in parallel with INTA and IDT access? > My understanding is that it is 2 stage process. Table 5-2 talk about > events happening before delivery, so that HW needs to prioritize them. > Once a decision is make, the highest one is delivered but then it could > trigger another exception when fetching IDT etc. > > Current execption.pending/interrupt.pending/nmi_injected doesn't match > either of above, interrupt/nmi is only for failed event injection, and a strange > fixed priority check when it is really injected: > exception > failed NMI > failed IRQ > new NMI > new IRQ. The current code assumes that only one of failed NMI, failed IRQ or exception can be true, So there is no prioritization at all between them. Just checking which one is available. If there is no event that should be re-injected then exception indeed is checked first, but, as Avi said, it doesn't matter. It depends on timing and can't affect any guest code and that is what really important. This is not the only place where KVM timing is not the same as CPU timing BTW. Pending exception field has two meanings though. It will be true either for exception that failed injection or for newly generated exception, but KVM is not CPU emulator and the second case happens very rarely (only during emulation). Don't forget that most exceptions are handled by CPU directly without KVM even knowing it. I wrote a test case that generates #NP during NMI handling. I expected that after calling #NP handler that fixes segment descriptor NMI handler will be called, but this is not what happened on real HW. #NP was called, NMI was dropped, but table 5-2 states that they should be handled serially. The good thing it that the way KVM handles this situation now is the same as real HW does, but if we will change code to do what SDM says KVM will behave differently. I should do the same test with IRQ some day. > > Table 5-2 looks missed in current KVM IMO except a wrong (but minor) > exception > NMI > IRQ sequence. Doesn't matter. See above. > > > > >>> > >>>> We could have either: 1) A pure SW "queue" that will be flush to > >>>> HW register later (VM_ENTRY_INTR_INFO_FIELD), 2) Direct use HW > >>>> register. > >>>> > >>> We have three event sources 1) exceptions 2) IRQ 3) NMI. We should > >>> have queue of three elements sorted by priority. On each entry we > >>> should > >> > >> Table 5-4 alreadys says NMI/IRQ is BENIGN. > > Table 5-2 applies here not table 5-4 I think. > > > >> > >>> inject an event with highest priority. And remove it from queue on > >>> exit. > >> > >> The problem is that we have to decide to inject only one of above 3, > >> and discard the rest. Whether priority them or merge (to one event > >> as Table 5-4) is another story. > > Only a small number of event are merged into #DF. Most handled > > serially (SDM does not define what serially means unfortunately), so > > I don't understand where "discard the rest" is come from. We can > > vmx_complete_interrupts clear all of them at next EXIT. > Because it is assumed that only one of them is true at the same time. > Even from HW point of view, if there are pending NMI/IRQ/exception, > CPU pick highest one, NMI, ignore/discard IRQ (but LAPIC still holds > IRQ, thus it can be re-injected), completely discard exception. > For NMI/IRQ that what happens. If nmi_pending is true IRQ is held pending in LAPIC (we don't even check IRQ is pending, who cares). For exception ordering see above. > I don't say discarding has any problem, but unnecessary to keep all of 3. > the only difference is when to discard the rest 2, at queue_exception/irq/nmi > time or later on (even at next EXIT time), which is same to me. > > > discard exception since it will be regenerated anyway, but IRQ and > > NMI is another story. SDM says that IRQ should be held pending (once > > again not much explanation here), nothing about NMI. > > > >>> > >>>> > >>>> A potential benefit is that it can avoid duplicated code and > >>>> potential bugs in current code as following patch shows if I > >>>> understand correctly: > >>>> > >>>> --- a/arch/x86/kvm/vmx.c > >>>> +++ b/arch/x86/kvm/vmx.c > >>>> @@ -2599,7 +2599,7 @@ static int handle_exception(struct kvm_vcpu > >>>> *vcpu, struct kvm_run *kvm_run) cr2 = > >>>> vmcs_readl(EXIT_QUALIFICATION); > >>>> KVMTRACE_3D(PAGE_FAULT, vcpu, > >>>> error_code, (u32)cr2, (u32)((u64)cr2 >> 32), handler); - > >>>> if (vcpu->arch.interrupt.pending || vcpu->arch.exception.pending ) > >>>> + if (vcpu->arch.interrupt.pending || > >>>> vcpu->arch.exception.pending || > >>>> vcpu->arch.nmi_injected) kvm_mmu_unprotect_page_virt(vcpu, > >>>> cr2); return kvm_mmu_page_fault(vcpu, cr2, error_code); } > >>> This fix is already in Avi's tree (not yet pushed). > >>> > >>>> Either way are OK and up to you. BTW Xen uses HW register directly > >>>> to representing an pending event. > >>>> > >>> In this particular case I don't mind to use HW register either, but > >>> I don't see any advantage. > >>> > >>>>> > >>>>>> One example is that exception has higher priority > >>>>>> than NMI/IRQ injection in current code which is not true in > >>>>>> reality. > >>>>>> > >>>>> > >>>>> I don't think it matters in practice, since the guest will see it > >>>>> as a timing issue. NMIs and IRQs are asynchronous (even those > >>>>> generated by the guest through the local APIC). > >>>> > >>>> Yes. But also cause IRQ injection be delayed which may have side > >>>> effect. For example if guest exception handler is very longer or > >>>> if guest VCPU fall into recursive #GP. Within current logic, a > >>>> guest IRQ event from KDB (IPI) running on VCPU0, as an example, > >>>> can't force the dead loop VCPU1 into KDB since it is recursively > >>>> #GP. > >>> If one #GP causes another #GP this is a #DF. If CPU has a chance to > >> > >> Means another #GP in next instruction i.e. Beginning of #GP handler > >> in guest. > >> No #DF here. > >> > > In this case we will enter guest with "NMI windows open" request and > > should exit immediately before first instruction of #GP handler. At > > this moment KVM will be able to inject NMI. > > If the HW NMI windows is supported, it is fine, how about SW NMI case? There is no SW NMI case. There are buggy CPUs that does not provide proper NMI injection support. There is a hack to implement kinda working NMI on such CPUs, but there are enough scenarios that will not work correctly to not rely on this hack if your guest uses NMI for something serious. > The flow will then look like: > > Guest #GP instruction -> VM Exit -> Inject virtual #GP -> VMRESUME -> > try to execute 1st ins of guest #GP handler -> VM Exit again (#GP) -> > inject virtual #GP -> ..... No. This will look like: Guest #GP instruction -> guest #GP handler -> try to execute 1st ins of guest #GP handler -> Guest #GP -> guest #GP handler -> something sends NMI -> vcpu_kick() -> VM Exit due to kick() -> inject NMI or the next VM entry -> everything works event on cpus with "SW NMI". KVM is not emulator. Avi can you rename the project to KINE please. -- Gleb. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html