On Mon, Nov 18, 2013 at 09:09:04PM +0200, Gleb Natapov wrote: > On Mon, Nov 18, 2013 at 09:04:51PM +0200, Michael S. Tsirkin wrote: > > On Tue, Nov 05, 2013 at 01:20:10PM +0200, Gleb Natapov wrote: > > > On Tue, Nov 05, 2013 at 12:22:49PM +0200, Gleb Natapov wrote: > > > > On Tue, Nov 05, 2013 at 12:18:57PM +0200, Michael S. Tsirkin wrote: > > > > > On Mon, Nov 04, 2013 at 10:44:43PM +0200, Gleb Natapov wrote: > > > > > > On Mon, Nov 04, 2013 at 10:33:57PM +0200, Michael S. Tsirkin wrote: > > > > > > > On Mon, Nov 04, 2013 at 10:13:39PM +0200, Gleb Natapov wrote: > > > > > > > > On Mon, Nov 04, 2013 at 10:11:33PM +0200, Michael S. Tsirkin wrote: > > > > > > > > > On Thu, Oct 31, 2013 at 09:48:08AM +0200, Gleb Natapov wrote: > > > > > > > > > > On Thu, Oct 31, 2013 at 02:21:46AM +0200, Michael S. Tsirkin wrote: > > > > > > > > > > > commit 8bf00a529967dafbbb210b377c38a15834d1e979: > > > > > > > > > > > " KVM: VMX: add support for switching of PERF_GLOBAL_CTRL " was > > > > > > > > > > > as far as I can tell supposed to bring about performance improvement > > > > > > > > > > > on hardware that supports it? > > > > > > > > > > No, it (and commits after it) supposed to fix a bug which it did. > > > > > > > > > > > > > > > > > > > > > Instead it seems to make the typical case (not running guest > > > > > > > > > > > under perf) a bit slower than it used to be. > > > > > > > > > > > the cost of VMexit goes up by about 50 cycles > > > > > > > > > > > on sandy bridge where the optimization in question > > > > > > > > > > > actually is activated. > > > > > > > > > > > > > > > > > > > > > You seams to be confused. 8bf00a529967dafbbb210 adds support for special > > > > > > > > > > PERF_GLOBAL_CTRL switching, but does not add code to switch anything, > > > > > > > > > > so the commit itself is a nop. > > > > > > > > > > > > > > > > > > It does add code to add_atomic_switch_msr. > > > > > > > > > > > > > > > > > So what? You do not read what I wrote. > > > > > > > > > > > > > > > > > > > > > It's simple: if I revert 8bf00a529967dafbbb210 then exit latency > > > > > > > is reduced. > > > > > > > You seem to tell me it should be a nop, but in practice it isn't. > > > > > > > > > > > > > > > > > > > No, if you read below I am saying that it looks like you are claiming that > > > > > > generic msr switch mechanism is faster and I am not buying that. If you > > > > > > believe this to be the case ask Intel for explanation. Your claim about > > > > > > "not running guest under perf" is even stranger since in this case no msr > > > > > > switch should happen regardless of the aforementioned commit (unless guest > > > > > > or host runs nmi watchdog, but then switch will happen no matter if perf > > > > > > is running, so again not running guest under perf" does not make sense). > > > > > > So, in short, you do not really know where the slow down is coming > > > > > > from. > > > > > > > > > > That's true. > > > > > > > > > Then dig dipper. > > > > > > > So quick and dirty patch to not needlessly write into VM_ENTRY_CONTROLS > > > when no PERF_GLOBAL_CTRL switching needed removes all the overhead. But we > > > probably need more generic code to shadow entire VM_ENTRY_CONTROLS/VM_EXIT_CONTROLS. > > > > > > > Do you plan to complete this? Merge as is? > I have a patch to shadow VM_ENTRY_CONTROLS/VM_EXIT_CONTROLS now, will > post it shortly, but it will have to wait for the end of a merge window > anyway. No rush. > > You said this saves about 50 cycles per exit for you, > > did you not? > Let's not exaggerate :) May be 20 after removing all the noise. Interesting. I remember reverting the feature gave me 50 cycles. I'll have to retest, but 20 cycles on all exits is nice too. > > > > > > > diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c > > > index e293a62..be64221 100644 > > > --- a/arch/x86/kvm/vmx.c > > > +++ b/arch/x86/kvm/vmx.c > > > @@ -413,6 +413,7 @@ struct vcpu_vmx { > > > struct shared_msr_entry *guest_msrs; > > > int nmsrs; > > > int save_nmsrs; > > > + bool core_perf_global_ctrl; > > > unsigned long host_idt_base; > > > #ifdef CONFIG_X86_64 > > > u64 msr_host_kernel_gs_base; > > > @@ -1432,9 +1433,12 @@ static void clear_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr) > > > break; > > > case MSR_CORE_PERF_GLOBAL_CTRL: > > > if (cpu_has_load_perf_global_ctrl) { > > > + if (vmx->core_perf_global_ctrl) { > > > clear_atomic_switch_msr_special( > > > VM_ENTRY_LOAD_IA32_PERF_GLOBAL_CTRL, > > > VM_EXIT_LOAD_IA32_PERF_GLOBAL_CTRL); > > > + vmx->core_perf_global_ctrl = false; > > > + } > > > return; > > > } > > > break; > > > @@ -1488,6 +1492,7 @@ static void add_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr, > > > GUEST_IA32_PERF_GLOBAL_CTRL, > > > HOST_IA32_PERF_GLOBAL_CTRL, > > > guest_val, host_val); > > > + vmx->core_perf_global_ctrl = true; > > > return; > > > } > > > break; > > > -- > > > Gleb. > > -- > Gleb. -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html