On Mon, Jun 14, 2021 at 03:53:23PM +0200, Joerg Roedel wrote: > From: Joerg Roedel <jroedel@xxxxxxx> > > The #VC handler only cares about IRQs being disabled while the GHCB is > active, as it must not be interrupted by something which could cause > another #VC while it holds the GHCB (NMI is the exception for which the > backup GHCB exits). > > Make sure nothing interrupts the code path while the GHCB is active by > disabling IRQs in sev_es_get_ghcb() and restoring the previous irq state > in sev_es_put_ghcb(). > > Signed-off-by: Joerg Roedel <jroedel@xxxxxxx> > --- > arch/x86/kernel/sev.c | 48 ++++++++++++++++++++++++++++++------------- > 1 file changed, 34 insertions(+), 14 deletions(-) Here's a diff ontop of yours with a couple of points: * I've named the low-level, interrupts-enabled workers __sev_get_ghcb()/__sev_put_ghcb() to mean a couple of things: ** underscored to mean, that callers need to disable local locks. There's also a lockdep_assert_irqs_disabled() to make sure, both in the get and put function. ** also only "sev" in the name because this code is not used for SEV-ES only anymore. * I've done it this way because you have a well-recognized code pattern where the caller disables interrupts, calls the low-level helpers and then enables interrupts again when done. VS passing a flags pointer back and forth which just looks weird. And as to being easy to use - users can botch flags too, when passing around so they can just as well do proper interrupts toggling like a gazillion other places in the kernel. Also, you have places like exc_vmm_communication() where you have to artifically pass in flags - I'm looking at your previous version - even if you already make sure interrupts are disabled with the BUG_ON assertion on entry. So in those cases you can simply call the interrupt-enabled, __-variants. Btw, while we're on exc_vmm_communication, it has a: BUG_ON(!irqs_disabled()); on entry and then later lockdep_assert_irqs_disabled(); and that second assertion is not really needed, methinks. So a hunk below removes it. Thoughts? diff --git a/arch/x86/kernel/sev.c b/arch/x86/kernel/sev.c index 7d70cddc38be..b85c4a2be9fa 100644 --- a/arch/x86/kernel/sev.c +++ b/arch/x86/kernel/sev.c @@ -192,11 +192,19 @@ void noinstr __sev_es_ist_exit(void) this_cpu_write(cpu_tss_rw.x86_tss.ist[IST_INDEX_VC], *(unsigned long *)ist); } -static __always_inline struct ghcb *__sev_es_get_ghcb(struct ghcb_state *state) +/* + * Nothing shall interrupt this code path while holding the per-CPU + * GHCB. The backup GHCB is only for NMIs interrupting this path. + * + * Callers must disable local interrupts around it. + */ +static __always_inline struct ghcb *__sev_get_ghcb(struct ghcb_state *state) { struct sev_es_runtime_data *data; struct ghcb *ghcb; + lockdep_assert_irqs_disabled(); + data = this_cpu_read(runtime_data); ghcb = &data->ghcb_page; @@ -231,18 +239,6 @@ static __always_inline struct ghcb *__sev_es_get_ghcb(struct ghcb_state *state) return ghcb; } -static __always_inline struct ghcb *sev_es_get_ghcb(struct ghcb_state *state, - unsigned long *flags) -{ - /* - * Nothing shall interrupt this code path while holding the per-cpu - * GHCB. The backup GHCB is only for NMIs interrupting this path. - */ - local_irq_save(*flags); - - return __sev_es_get_ghcb(state); -} - /* Needed in vc_early_forward_exception */ void do_early_exception(struct pt_regs *regs, int trapnr); @@ -491,11 +487,13 @@ static enum es_result vc_slow_virt_to_phys(struct ghcb *ghcb, struct es_em_ctxt /* Include code shared with pre-decompression boot stage */ #include "sev-shared.c" -static __always_inline void __sev_es_put_ghcb(struct ghcb_state *state) +static __always_inline void __sev_put_ghcb(struct ghcb_state *state) { struct sev_es_runtime_data *data; struct ghcb *ghcb; + lockdep_assert_irqs_disabled(); + data = this_cpu_read(runtime_data); ghcb = &data->ghcb_page; @@ -514,13 +512,6 @@ static __always_inline void __sev_es_put_ghcb(struct ghcb_state *state) } } -static __always_inline void sev_es_put_ghcb(struct ghcb_state *state, - unsigned long flags) -{ - __sev_es_put_ghcb(state); - local_irq_restore(flags); -} - void noinstr __sev_es_nmi_complete(void) { struct ghcb_state state; @@ -528,7 +519,7 @@ void noinstr __sev_es_nmi_complete(void) BUG_ON(!irqs_disabled()); - ghcb = __sev_es_get_ghcb(&state); + ghcb = __sev_get_ghcb(&state); vc_ghcb_invalidate(ghcb); ghcb_set_sw_exit_code(ghcb, SVM_VMGEXIT_NMI_COMPLETE); @@ -538,7 +529,7 @@ void noinstr __sev_es_nmi_complete(void) sev_es_wr_ghcb_msr(__pa_nodebug(ghcb)); VMGEXIT(); - __sev_es_put_ghcb(&state); + __sev_put_ghcb(&state); } static u64 get_jump_table_addr(void) @@ -548,7 +539,9 @@ static u64 get_jump_table_addr(void) struct ghcb *ghcb; u64 ret = 0; - ghcb = sev_es_get_ghcb(&state, &flags); + local_irq_save(flags); + + ghcb = __sev_get_ghcb(&state); vc_ghcb_invalidate(ghcb); ghcb_set_sw_exit_code(ghcb, SVM_VMGEXIT_AP_JUMP_TABLE); @@ -562,7 +555,9 @@ static u64 get_jump_table_addr(void) ghcb_sw_exit_info_2_is_valid(ghcb)) ret = ghcb->save.sw_exit_info_2; - sev_es_put_ghcb(&state, flags); + __sev_put_ghcb(&state); + + local_irq_restore(flags); return ret; } @@ -686,7 +681,9 @@ static void sev_es_ap_hlt_loop(void) unsigned long flags; struct ghcb *ghcb; - ghcb = sev_es_get_ghcb(&state, &flags); + local_irq_save(flags); + + ghcb = __sev_get_ghcb(&state); while (true) { vc_ghcb_invalidate(ghcb); @@ -703,7 +700,9 @@ static void sev_es_ap_hlt_loop(void) break; } - sev_es_put_ghcb(&state, flags); + __sev_put_ghcb(&state); + + local_irq_restore(flags); } /* @@ -1364,7 +1363,6 @@ DEFINE_IDTENTRY_VC_SAFE_STACK(exc_vmm_communication) } irq_state = irqentry_nmi_enter(regs); - lockdep_assert_irqs_disabled(); instrumentation_begin(); /* @@ -1373,7 +1371,7 @@ DEFINE_IDTENTRY_VC_SAFE_STACK(exc_vmm_communication) * keep the IRQs disabled to protect us against concurrent TLB flushes. */ - ghcb = __sev_es_get_ghcb(&state); + ghcb = __sev_get_ghcb(&state); vc_ghcb_invalidate(ghcb); result = vc_init_em_ctxt(&ctxt, regs, error_code); @@ -1381,7 +1379,7 @@ DEFINE_IDTENTRY_VC_SAFE_STACK(exc_vmm_communication) if (result == ES_OK) result = vc_handle_exitcode(&ctxt, ghcb, error_code); - __sev_es_put_ghcb(&state); + __sev_put_ghcb(&state); /* Done - now check the result */ switch (result) { -- Regards/Gruss, Boris. https://people.kernel.org/tglx/notes-about-netiquette _______________________________________________ Virtualization mailing list Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx https://lists.linuxfoundation.org/mailman/listinfo/virtualization