On Mon, Jan 22, 2024 at 03:52:50PM -0800, isaku.yamahata@xxxxxxxxx wrote: > From: Isaku Yamahata <isaku.yamahata@xxxxxxxxx> > > A VMM interacts with the TDX module using a new instruction (SEAMCALL). > For instance, a TDX VMM does not have full access to the VM control > structure corresponding to VMX VMCS. Instead, a VMM induces the TDX module > to act on behalf via SEAMCALLs. > > Export __seamcall and define C wrapper functions for SEAMCALLs for ^^^^^^^^^^^^^^^^^ It's not exported by this patch. Others LGTM. Reviewed-by: Yuan Yao <yuan.yao@xxxxxxxxx> > readability. > > Some SEAMCALL APIs donate host pages to TDX module or guest TD, and the > donated pages are encrypted. Those require the VMM to flush the cache > lines to avoid cache line alias. > > Signed-off-by: Sean Christopherson <sean.j.christopherson@xxxxxxxxx> > Signed-off-by: Isaku Yamahata <isaku.yamahata@xxxxxxxxx> > Reviewed-by: Binbin Wu <binbin.wu@xxxxxxxxxxxxxxx> > > --- > Changes > v18: > - removed stub functions for __seamcall{,_ret}() > - Added Reviewed-by Binbin > - Make tdx_seamcall() use struct tdx_module_args instead of taking > each inputs. > > v15 -> v16: > - use struct tdx_module_args instead of struct tdx_module_output > - Add tdh_mem_sept_rd() for SEPT_VE_DISABLE=1. > --- > arch/x86/kvm/vmx/tdx_ops.h | 360 +++++++++++++++++++++++++++++++++++++ > 1 file changed, 360 insertions(+) > create mode 100644 arch/x86/kvm/vmx/tdx_ops.h > > diff --git a/arch/x86/kvm/vmx/tdx_ops.h b/arch/x86/kvm/vmx/tdx_ops.h > new file mode 100644 > index 000000000000..0e26cf22240e > --- /dev/null > +++ b/arch/x86/kvm/vmx/tdx_ops.h > @@ -0,0 +1,360 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +/* constants/data definitions for TDX SEAMCALLs */ > + > +#ifndef __KVM_X86_TDX_OPS_H > +#define __KVM_X86_TDX_OPS_H > + > +#include <linux/compiler.h> > + > +#include <asm/cacheflush.h> > +#include <asm/asm.h> > +#include <asm/kvm_host.h> > + > +#include "tdx_errno.h" > +#include "tdx_arch.h" > +#include "x86.h" > + > +static inline u64 tdx_seamcall(u64 op, struct tdx_module_args *in, > + struct tdx_module_args *out) > +{ > + u64 ret; > + > + if (out) { > + *out = *in; > + ret = __seamcall_ret(op, out); > + } else > + ret = __seamcall(op, in); > + > + if (unlikely(ret == TDX_SEAMCALL_UD)) { > + /* > + * SEAMCALLs fail with TDX_SEAMCALL_UD returned when VMX is off. > + * This can happen when the host gets rebooted or live > + * updated. In this case, the instruction execution is ignored > + * as KVM is shut down, so the error code is suppressed. Other > + * than this, the error is unexpected and the execution can't > + * continue as the TDX features reply on VMX to be on. > + */ > + kvm_spurious_fault(); > + return 0; > + } > + return ret; > +} > + > +static inline u64 tdh_mng_addcx(hpa_t tdr, hpa_t addr) > +{ > + struct tdx_module_args in = { > + .rcx = addr, > + .rdx = tdr, > + }; > + > + clflush_cache_range(__va(addr), PAGE_SIZE); > + return tdx_seamcall(TDH_MNG_ADDCX, &in, NULL); > +} > + > +static inline u64 tdh_mem_page_add(hpa_t tdr, gpa_t gpa, hpa_t hpa, hpa_t source, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = gpa, > + .rdx = tdr, > + .r8 = hpa, > + .r9 = source, > + }; > + > + clflush_cache_range(__va(hpa), PAGE_SIZE); > + return tdx_seamcall(TDH_MEM_PAGE_ADD, &in, out); > +} > + > +static inline u64 tdh_mem_sept_add(hpa_t tdr, gpa_t gpa, int level, hpa_t page, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = gpa | level, > + .rdx = tdr, > + .r8 = page, > + }; > + > + clflush_cache_range(__va(page), PAGE_SIZE); > + return tdx_seamcall(TDH_MEM_SEPT_ADD, &in, out); > +} > + > +static inline u64 tdh_mem_sept_rd(hpa_t tdr, gpa_t gpa, int level, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = gpa | level, > + .rdx = tdr, > + }; > + > + return tdx_seamcall(TDH_MEM_SEPT_RD, &in, out); > +} > + > +static inline u64 tdh_mem_sept_remove(hpa_t tdr, gpa_t gpa, int level, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = gpa | level, > + .rdx = tdr, > + }; > + > + return tdx_seamcall(TDH_MEM_SEPT_REMOVE, &in, out); > +} > + > +static inline u64 tdh_vp_addcx(hpa_t tdvpr, hpa_t addr) > +{ > + struct tdx_module_args in = { > + .rcx = addr, > + .rdx = tdvpr, > + }; > + > + clflush_cache_range(__va(addr), PAGE_SIZE); > + return tdx_seamcall(TDH_VP_ADDCX, &in, NULL); > +} > + > +static inline u64 tdh_mem_page_relocate(hpa_t tdr, gpa_t gpa, hpa_t hpa, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = gpa, > + .rdx = tdr, > + .r8 = hpa, > + }; > + > + clflush_cache_range(__va(hpa), PAGE_SIZE); > + return tdx_seamcall(TDH_MEM_PAGE_RELOCATE, &in, out); > +} > + > +static inline u64 tdh_mem_page_aug(hpa_t tdr, gpa_t gpa, hpa_t hpa, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = gpa, > + .rdx = tdr, > + .r8 = hpa, > + }; > + > + clflush_cache_range(__va(hpa), PAGE_SIZE); > + return tdx_seamcall(TDH_MEM_PAGE_AUG, &in, out); > +} > + > +static inline u64 tdh_mem_range_block(hpa_t tdr, gpa_t gpa, int level, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = gpa | level, > + .rdx = tdr, > + }; > + > + return tdx_seamcall(TDH_MEM_RANGE_BLOCK, &in, out); > +} > + > +static inline u64 tdh_mng_key_config(hpa_t tdr) > +{ > + struct tdx_module_args in = { > + .rcx = tdr, > + }; > + > + return tdx_seamcall(TDH_MNG_KEY_CONFIG, &in, NULL); > +} > + > +static inline u64 tdh_mng_create(hpa_t tdr, int hkid) > +{ > + struct tdx_module_args in = { > + .rcx = tdr, > + .rdx = hkid, > + }; > + > + clflush_cache_range(__va(tdr), PAGE_SIZE); > + return tdx_seamcall(TDH_MNG_CREATE, &in, NULL); > +} > + > +static inline u64 tdh_vp_create(hpa_t tdr, hpa_t tdvpr) > +{ > + struct tdx_module_args in = { > + .rcx = tdvpr, > + .rdx = tdr, > + }; > + > + clflush_cache_range(__va(tdvpr), PAGE_SIZE); > + return tdx_seamcall(TDH_VP_CREATE, &in, NULL); > +} > + > +static inline u64 tdh_mng_rd(hpa_t tdr, u64 field, struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = tdr, > + .rdx = field, > + }; > + > + return tdx_seamcall(TDH_MNG_RD, &in, out); > +} > + > +static inline u64 tdh_mr_extend(hpa_t tdr, gpa_t gpa, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = gpa, > + .rdx = tdr, > + }; > + > + return tdx_seamcall(TDH_MR_EXTEND, &in, out); > +} > + > +static inline u64 tdh_mr_finalize(hpa_t tdr) > +{ > + struct tdx_module_args in = { > + .rcx = tdr, > + }; > + > + return tdx_seamcall(TDH_MR_FINALIZE, &in, NULL); > +} > + > +static inline u64 tdh_vp_flush(hpa_t tdvpr) > +{ > + struct tdx_module_args in = { > + .rcx = tdvpr, > + }; > + > + return tdx_seamcall(TDH_VP_FLUSH, &in, NULL); > +} > + > +static inline u64 tdh_mng_vpflushdone(hpa_t tdr) > +{ > + struct tdx_module_args in = { > + .rcx = tdr, > + }; > + > + return tdx_seamcall(TDH_MNG_VPFLUSHDONE, &in, NULL); > +} > + > +static inline u64 tdh_mng_key_freeid(hpa_t tdr) > +{ > + struct tdx_module_args in = { > + .rcx = tdr, > + }; > + > + return tdx_seamcall(TDH_MNG_KEY_FREEID, &in, NULL); > +} > + > +static inline u64 tdh_mng_init(hpa_t tdr, hpa_t td_params, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = tdr, > + .rdx = td_params, > + }; > + > + return tdx_seamcall(TDH_MNG_INIT, &in, out); > +} > + > +static inline u64 tdh_vp_init(hpa_t tdvpr, u64 rcx) > +{ > + struct tdx_module_args in = { > + .rcx = tdvpr, > + .rdx = rcx, > + }; > + > + return tdx_seamcall(TDH_VP_INIT, &in, NULL); > +} > + > +static inline u64 tdh_vp_rd(hpa_t tdvpr, u64 field, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = tdvpr, > + .rdx = field, > + }; > + > + return tdx_seamcall(TDH_VP_RD, &in, out); > +} > + > +static inline u64 tdh_mng_key_reclaimid(hpa_t tdr) > +{ > + struct tdx_module_args in = { > + .rcx = tdr, > + }; > + > + return tdx_seamcall(TDH_MNG_KEY_RECLAIMID, &in, NULL); > +} > + > +static inline u64 tdh_phymem_page_reclaim(hpa_t page, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = page, > + }; > + > + return tdx_seamcall(TDH_PHYMEM_PAGE_RECLAIM, &in, out); > +} > + > +static inline u64 tdh_mem_page_remove(hpa_t tdr, gpa_t gpa, int level, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = gpa | level, > + .rdx = tdr, > + }; > + > + return tdx_seamcall(TDH_MEM_PAGE_REMOVE, &in, out); > +} > + > +static inline u64 tdh_sys_lp_shutdown(void) > +{ > + struct tdx_module_args in = { > + }; > + > + return tdx_seamcall(TDH_SYS_LP_SHUTDOWN, &in, NULL); > +} > + > +static inline u64 tdh_mem_track(hpa_t tdr) > +{ > + struct tdx_module_args in = { > + .rcx = tdr, > + }; > + > + return tdx_seamcall(TDH_MEM_TRACK, &in, NULL); > +} > + > +static inline u64 tdh_mem_range_unblock(hpa_t tdr, gpa_t gpa, int level, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = gpa | level, > + .rdx = tdr, > + }; > + > + return tdx_seamcall(TDH_MEM_RANGE_UNBLOCK, &in, out); > +} > + > +static inline u64 tdh_phymem_cache_wb(bool resume) > +{ > + struct tdx_module_args in = { > + .rcx = resume ? 1 : 0, > + }; > + > + return tdx_seamcall(TDH_PHYMEM_CACHE_WB, &in, NULL); > +} > + > +static inline u64 tdh_phymem_page_wbinvd(hpa_t page) > +{ > + struct tdx_module_args in = { > + .rcx = page, > + }; > + > + return tdx_seamcall(TDH_PHYMEM_PAGE_WBINVD, &in, NULL); > +} > + > +static inline u64 tdh_vp_wr(hpa_t tdvpr, u64 field, u64 val, u64 mask, > + struct tdx_module_args *out) > +{ > + struct tdx_module_args in = { > + .rcx = tdvpr, > + .rdx = field, > + .r8 = val, > + .r9 = mask, > + }; > + > + return tdx_seamcall(TDH_VP_WR, &in, out); > +} > + > +#endif /* __KVM_X86_TDX_OPS_H */ > -- > 2.25.1 > >