On Wed, Jun 23, 2021, Maxim Levitsky wrote: > On Wed, 2021-06-23 at 15:32 +0200, Vitaly Kuznetsov wrote: > > Maxim Levitsky <mlevitsk@xxxxxxxxxx> writes: > > > > > On Wed, 2021-06-23 at 16:01 +0300, Maxim Levitsky wrote: > > > > On Wed, 2021-06-23 at 11:39 +0200, Paolo Bonzini wrote: > > > > > On 23/06/21 09:44, Vitaly Kuznetsov wrote: > > > > > > - RFC: I'm not 100% sure my 'smart' idea to use currently- unused > > > > > > HSAVE area is that smart. Also, we don't even seem to check that L1 > > > > > > set it up upon nested VMRUN so hypervisors which don't do that may > > > > > > remain broken. Ya, per the APM, nested_svm_vmrun() needs to verify the MSR is non-zero, and svm_set_msr() needs to verify the incoming value is a legal, page-aligned address. Both conditions are #GP. > > > > > >A very much needed selftest is also missing. > > > > > > > > > > It's certainly a bit weird, but I guess it counts as smart too. It > > > > > needs a few more comments, but I think it's a good solution. > > > > > > > > > > One could delay the backwards memcpy until vmexit time, but that > > > > > would require a new flag so it's not worth it for what is a pretty > > > > > rare and already expensive case. And it's _almost_ architecturally legal, but the APM does clearly state that the HSAVE area is used on VRMUN and #VMEXIT. We'd definitely be taking a few liberties by accessing the area at SMI/RSM. > > We can resurrect 'hsave' and keep it internally indeed but to make this > > migratable, we'd have to add it to the nested state acquired through > > svm_get_nested_state(). Using L1's HSAVE area (ponted to by > > MSR_VM_HSAVE_PA) avoids that as we have everything in L1's memory. > I think I would prefer to avoid touching guest memory as much > as possible to avoid the shenanigans of accessing it: > > For example on nested state read we are not allowed to write guest > memory since at the point it is already migrated, and for setting > nested state we are not allowed to even read the guest memory since > the memory map might not be up to date. But that shouldn't conflict with using hsave, because hsave won't be accessed in this case until KVM_RUN, correct? > Then a malicious guest can always change its memory which also can cause > issues. The APM very clearly states that touching hsave is not allowed: Software must not attempt to read or write the host save-state area directly. > Since it didn't work before and both sides of migration need a fix, > adding a new flag and adding hsave area to nested state seems like a > very good thing. ... > This way we still avoid overhead of copying the hsave area > on each nested entry. > > What do you think? Hmm, I don't like allocating an extra page for every vCPU. And I believe this hackery is necessary only because nested_svm_vmexit() isn't following the architcture in the first place. I.e. using vmcb01 to restore host state is flat out wrong. If this the restore code (below) is instead converted to use the hsave area, then this bug is fixed _and_ we become (more) compliant with the spec. And the save/restore to HSAVE could be selective, i.e. only save/restore state that is actually consumed by nested_svm_vmexit(). It would require saving/restoring select segment _selectors_, but that is also a bug fix since the APM pseudocode clearly states that segments are reloading from the descriptor table (APM says "GDT", though I assume LDT is also legal). E.g. while software can't touch HSAVE, L2 can _legally_ change L1's GDT, so it's technically legal to have host segment registers magically change across VMRUN. There might also be side effects we're missing by not reloading segment registers, e.g. accessed bit updates? /* * Restore processor state that had been saved in vmcb01 <-- bad KVM */ kvm_set_rflags(vcpu, svm->vmcb->save.rflags); svm_set_efer(vcpu, svm->vmcb->save.efer); svm_set_cr0(vcpu, svm->vmcb->save.cr0 | X86_CR0_PE); svm_set_cr4(vcpu, svm->vmcb->save.cr4); kvm_rax_write(vcpu, svm->vmcb->save.rax); kvm_rsp_write(vcpu, svm->vmcb->save.rsp); kvm_rip_write(vcpu, svm->vmcb->save.rip); svm->vcpu.arch.dr7 = DR7_FIXED_1; kvm_update_dr7(&svm->vcpu); ... kvm_vcpu_unmap(vcpu, &map, true); nested_svm_transition_tlb_flush(vcpu); nested_svm_uninit_mmu_context(vcpu); rc = nested_svm_load_cr3(vcpu, svm->vmcb->save.cr3, false, true);