On 2020-01-10 12:12, Andrew Murray wrote:
On Fri, Jan 10, 2020 at 11:18:48AM +0000, Marc Zyngier wrote:
On 2020-01-10 10:54, Andrew Murray wrote:
> On Sat, Dec 21, 2019 at 02:13:25PM +0000, Marc Zyngier wrote:
> > On Fri, 20 Dec 2019 14:30:16 +0000
> > Andrew Murray <andrew.murray@xxxxxxx> wrote:
> >
> > [somehow managed not to do a reply all, re-sending]
> >
> > > From: Sudeep Holla <sudeep.holla@xxxxxxx>
> > >
> > > Now that we can save/restore the full SPE controls, we can enable it
> > > if SPE is setup and ready to use in KVM. It's supported in KVM only if
> > > all the CPUs in the system supports SPE.
> > >
> > > However to support heterogenous systems, we need to move the check if
> > > host supports SPE and do a partial save/restore.
> >
> > No. Let's just not go down that path. For now, KVM on heterogeneous
> > systems do not get SPE. If SPE has been enabled on a guest and a CPU
> > comes up without SPE, this CPU should fail to boot (same as exposing a
> > feature to userspace).
> >
> > >
> > > Signed-off-by: Sudeep Holla <sudeep.holla@xxxxxxx>
> > > Signed-off-by: Andrew Murray <andrew.murray@xxxxxxx>
> > > ---
> > > arch/arm64/kvm/hyp/debug-sr.c | 33 ++++++++++++++++-----------------
> > > include/kvm/arm_spe.h | 6 ++++++
> > > 2 files changed, 22 insertions(+), 17 deletions(-)
> > >
> > > diff --git a/arch/arm64/kvm/hyp/debug-sr.c b/arch/arm64/kvm/hyp/debug-sr.c
> > > index 12429b212a3a..d8d857067e6d 100644
> > > --- a/arch/arm64/kvm/hyp/debug-sr.c
> > > +++ b/arch/arm64/kvm/hyp/debug-sr.c
> > > @@ -86,18 +86,13 @@
> > > }
> > >
> > > static void __hyp_text
> > > -__debug_save_spe_nvhe(struct kvm_cpu_context *ctxt, bool full_ctxt)
> > > +__debug_save_spe_context(struct kvm_cpu_context *ctxt, bool full_ctxt)
> > > {
> > > u64 reg;
> > >
> > > /* Clear pmscr in case of early return */
> > > ctxt->sys_regs[PMSCR_EL1] = 0;
> > >
> > > - /* SPE present on this CPU? */
> > > - if (!cpuid_feature_extract_unsigned_field(read_sysreg(id_aa64dfr0_el1),
> > > - ID_AA64DFR0_PMSVER_SHIFT))
> > > - return;
> > > -
> > > /* Yes; is it owned by higher EL? */
> > > reg = read_sysreg_s(SYS_PMBIDR_EL1);
> > > if (reg & BIT(SYS_PMBIDR_EL1_P_SHIFT))
> > > @@ -142,7 +137,7 @@ __debug_save_spe_nvhe(struct kvm_cpu_context *ctxt, bool full_ctxt)
> > > }
> > >
> > > static void __hyp_text
> > > -__debug_restore_spe_nvhe(struct kvm_cpu_context *ctxt, bool full_ctxt)
> > > +__debug_restore_spe_context(struct kvm_cpu_context *ctxt, bool full_ctxt)
> > > {
> > > if (!ctxt->sys_regs[PMSCR_EL1])
> > > return;
> > > @@ -210,11 +205,14 @@ void __hyp_text __debug_restore_guest_context(struct kvm_vcpu *vcpu)
> > > struct kvm_guest_debug_arch *host_dbg;
> > > struct kvm_guest_debug_arch *guest_dbg;
> > >
> > > + host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
> > > + guest_ctxt = &vcpu->arch.ctxt;
> > > +
> > > + __debug_restore_spe_context(guest_ctxt, kvm_arm_spe_v1_ready(vcpu));
> > > +
> > > if (!(vcpu->arch.flags & KVM_ARM64_DEBUG_DIRTY))
> > > return;
> > >
> > > - host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
> > > - guest_ctxt = &vcpu->arch.ctxt;
> > > host_dbg = &vcpu->arch.host_debug_state.regs;
> > > guest_dbg = kern_hyp_va(vcpu->arch.debug_ptr);
> > >
> > > @@ -232,8 +230,7 @@ void __hyp_text __debug_restore_host_context(struct kvm_vcpu *vcpu)
> > > host_ctxt = kern_hyp_va(vcpu->arch.host_cpu_context);
> > > guest_ctxt = &vcpu->arch.ctxt;
> > >
> > > - if (!has_vhe())
> > > - __debug_restore_spe_nvhe(host_ctxt, false);
> > > + __debug_restore_spe_context(host_ctxt, kvm_arm_spe_v1_ready(vcpu));
> >
> > So you now do an unconditional save/restore on the exit path for VHE
> > as
> > well? Even if the host isn't using the SPE HW? That's not acceptable
> > as, in most cases, only the host /or/ the guest will use SPE. Here,
> > you
> > put a measurable overhead on each exit.
> >
> > If the host is not using SPE, then the restore/save should happen in
> > vcpu_load/vcpu_put. Only if the host is using SPE should you do
> > something in the run loop. Of course, this only applies to VHE and
> > non-VHE must switch eagerly.
> >
>
> On VHE where SPE is used in the guest only - we save/restore in
> vcpu_load/put.
Yes.
> On VHE where SPE is used in the host only - we save/restore in the run
> loop.
Why? If only the host is using SPE, why should we do *anything at
all*?
Oh yeah of course, we trap them in this case.
(Do I understand correctly that we don't/can't trap them for nVHE? -
and so
we should save/restore them for this use-case in nVHE)
We can always trap. Otherwise we wouldn't be able to hide the feature
from the guest.
> On VHE where SPE is used in guest and host - we save/restore in the run
> loop.
>
> As the guest can't trace EL2 it doesn't matter if we restore guest SPE
> early
> in the vcpu_load/put functions. (I assume it doesn't matter that we
> restore
> an EL0/EL1 profiling buffer address at this point and enable tracing
> given
> that there is nothing to trace until entering the guest).
As long as you do it after the EL1 sysregs have need restored so that
the
SPE
HW has a valid context, we should be fine. Don't restore it before
that
point
though (you have no idea whether the SPE HW can do speculative memory
accesses
that would use the wrong page tables).
Right, so don't enable tracing until SPE has a valid context. I
understand
that to mean at least the SPE buffer address registers (PMBPTR,
PMBLIMITR)
in the right context with respect to the E2PB bits (translation regime)
and having those tables mapped in (which I think relate to the
__activateX,
__sysreg_restore_guest_stateX type of calls in kvm_vcpu_run_X right?).
The full MM context has to be in place before you can do anything. This
means
at least TTBR*_EL1, TCR_EL1 and co. But maybe this note in the SPE
architecture
document would allow us to relax things:
"The Statistical Profiling Extension is always disabled if the owning
Exception
level is a lower Exception level than the current Exception level."
So as long as you restore the guest state from EL2, SPE should be
disabled.
I think that means we can restore the registers no earler than
vcpu_load/put
but we can't re-enable the tracing (PMSCR) until no earlier than just
before
__set_guest_arch_workaround_state. I think that applies to both VHE and
nVHE?
I'm sorry, but I don't understand what you mean.
M.
--
Jazz is not dead. It just smells funny...
_______________________________________________
kvmarm mailing list
kvmarm@xxxxxxxxxxxxxxxxxxxxx
https://lists.cs.columbia.edu/mailman/listinfo/kvmarm