Many of the integrity guarantees of SEV-SNP are enforced through the Reverse Map Table (RMP). Each RMP entry contains the GPA at which a particular page of DRAM should be mapped. The VMs can request the hypervisor to add pages in the RMP table via the Page State Change VMGEXIT defined in the GHCB specification section 2.5.1 and 4.1.6. Inside each RMP entry is a Validated flag; this flag is automatically cleared to 0 by the CPU hardware when a new RMP entry is created for a guest. Each VM page can be either validated or invalidated, as indicated by the Validated flag in the RMP entry. Memory access to a private page that is not validated generates a #VC. A VM can use PVALIDATE instruction to validate the private page before using it. To maintain the security guarantee of SEV-SNP guests, when transitioning a memory from private to shared, the guest must invalidate the memory range before asking the hypervisor to change the page state to shared in the RMP table. After the page is mapped private in the page table, the guest must issue a page state change VMGEXIT to make the memory private in the RMP table and validate it. If the memory is not validated after its added in the RMP table as private, then a VC exception (page-not-validated) will be raised. We do not support the page-not-validated exception yet, so it will crash the guest. On boot, BIOS should have validated the entire system memory. During the kernel decompression stage, the VC handler uses the set_memory_decrypted() to make the GHCB page shared (i.e clear encryption attribute). And while exiting from the decompression, it calls the set_memory_encyrpted() to make the page private. Cc: Thomas Gleixner <tglx@xxxxxxxxxxxxx> Cc: Ingo Molnar <mingo@xxxxxxxxxx> Cc: Borislav Petkov <bp@xxxxxxxxx> Cc: Joerg Roedel <jroedel@xxxxxxx> Cc: "H. Peter Anvin" <hpa@xxxxxxxxx> Cc: Tony Luck <tony.luck@xxxxxxxxx> Cc: Dave Hansen <dave.hansen@xxxxxxxxx> Cc: "Peter Zijlstra (Intel)" <peterz@xxxxxxxxxxxxx> Cc: Paolo Bonzini <pbonzini@xxxxxxxxxx> Cc: Tom Lendacky <thomas.lendacky@xxxxxxx> Cc: David Rientjes <rientjes@xxxxxxxxxx> Cc: Sean Christopherson <seanjc@xxxxxxxxxx> Cc: x86@xxxxxxxxxx Cc: kvm@xxxxxxxxxxxxxxx Signed-off-by: Brijesh Singh <brijesh.singh@xxxxxxx> --- arch/x86/boot/compressed/Makefile | 1 + arch/x86/boot/compressed/ident_map_64.c | 18 ++++ arch/x86/boot/compressed/sev-snp.c | 115 ++++++++++++++++++++++++ arch/x86/boot/compressed/sev-snp.h | 25 ++++++ 4 files changed, 159 insertions(+) create mode 100644 arch/x86/boot/compressed/sev-snp.c create mode 100644 arch/x86/boot/compressed/sev-snp.h diff --git a/arch/x86/boot/compressed/Makefile b/arch/x86/boot/compressed/Makefile index e0bc3988c3fa..4d422aae8a86 100644 --- a/arch/x86/boot/compressed/Makefile +++ b/arch/x86/boot/compressed/Makefile @@ -93,6 +93,7 @@ ifdef CONFIG_X86_64 vmlinux-objs-y += $(obj)/mem_encrypt.o vmlinux-objs-y += $(obj)/pgtable_64.o vmlinux-objs-$(CONFIG_AMD_MEM_ENCRYPT) += $(obj)/sev-es.o + vmlinux-objs-$(CONFIG_AMD_MEM_ENCRYPT) += $(obj)/sev-snp.o endif vmlinux-objs-$(CONFIG_ACPI) += $(obj)/acpi.o diff --git a/arch/x86/boot/compressed/ident_map_64.c b/arch/x86/boot/compressed/ident_map_64.c index f7213d0943b8..0a420ce5550f 100644 --- a/arch/x86/boot/compressed/ident_map_64.c +++ b/arch/x86/boot/compressed/ident_map_64.c @@ -37,6 +37,8 @@ #include <asm/setup.h> /* For COMMAND_LINE_SIZE */ #undef _SETUP +#include "sev-snp.h" + extern unsigned long get_cmd_line_ptr(void); /* Used by PAGE_KERN* macros: */ @@ -278,12 +280,28 @@ static int set_clr_page_flags(struct x86_mapping_info *info, if ((set | clr) & _PAGE_ENC) clflush_page(address); + /* + * If the encryption attribute is being cleared, then change the page state to + * shared in the RMP entry. Change of the page state must be done before the + * PTE updates. + */ + if (clr & _PAGE_ENC) + sev_snp_set_page_shared(pte_pfn(*ptep) << PAGE_SHIFT); + /* Update PTE */ pte = *ptep; pte = pte_set_flags(pte, set); pte = pte_clear_flags(pte, clr); set_pte(ptep, pte); + /* + * If the encryption attribute is being set, then change the page state to + * private in the RMP entry. The page state must be done after the PTE + * is updated. + */ + if (set & _PAGE_ENC) + sev_snp_set_page_private(pte_pfn(*ptep) << PAGE_SHIFT); + /* Flush TLB after changing encryption attribute */ write_cr3(top_level_pgt); diff --git a/arch/x86/boot/compressed/sev-snp.c b/arch/x86/boot/compressed/sev-snp.c new file mode 100644 index 000000000000..5c25103b0df1 --- /dev/null +++ b/arch/x86/boot/compressed/sev-snp.c @@ -0,0 +1,115 @@ +// SPDX-License-Identifier: GPL-2.0 +/* + * AMD SEV SNP support + * + * Author: Brijesh Singh <brijesh.singh@xxxxxxx> + * + */ + +#include "misc.h" +#include "error.h" + +#include <asm/msr-index.h> +#include <asm/sev-snp.h> +#include <asm/sev-es.h> + +#include "sev-snp.h" + +static bool sev_snp_enabled(void) +{ + unsigned long low, high; + u64 val; + + asm volatile("rdmsr\n" : "=a" (low), "=d" (high) : + "c" (MSR_AMD64_SEV)); + + val = (high << 32) | low; + + if (val & MSR_AMD64_SEV_SNP_ENABLED) + return true; + + return false; +} + +/* Provides sev_snp_{wr,rd}_ghcb_msr() */ +#include "sev-common.c" + +/* Provides sev_es_terminate() */ +#include "../../kernel/sev-common-shared.c" + +static void sev_snp_pages_state_change(unsigned long paddr, int op) +{ + u64 pfn = paddr >> PAGE_SHIFT; + u64 old, val; + + /* save the old GHCB MSR */ + old = sev_es_rd_ghcb_msr(); + + /* Issue VMGEXIT to change the page state */ + sev_es_wr_ghcb_msr(GHCB_SNP_PAGE_STATE_REQ_GFN(pfn, op)); + VMGEXIT(); + + /* Read the response of the VMGEXIT */ + val = sev_es_rd_ghcb_msr(); + if ((GHCB_SEV_GHCB_RESP_CODE(val) != GHCB_SNP_PAGE_STATE_CHANGE_RESP) || + (GHCB_SNP_PAGE_STATE_RESP_VAL(val) != 0)) + sev_es_terminate(GHCB_SEV_ES_REASON_GENERAL_REQUEST); + + /* Restore the GHCB MSR value */ + sev_es_wr_ghcb_msr(old); +} + +static void sev_snp_issue_pvalidate(unsigned long paddr, bool validate) +{ + unsigned long eflags; + int rc; + + rc = __pvalidate(paddr, RMP_PG_SIZE_4K, validate, &eflags); + if (rc) { + error("Failed to validate address"); + goto e_fail; + } + + /* Check for the double validation and assert on failure */ + if (eflags & X86_EFLAGS_CF) { + error("Double validation detected"); + goto e_fail; + } + + return; +e_fail: + sev_es_terminate(GHCB_SEV_ES_REASON_GENERAL_REQUEST); +} + +static void sev_snp_set_page_private_shared(unsigned long paddr, int op) +{ + if (!sev_snp_enabled()) + return; + + /* + * We are change the page state from private to shared, invalidate the pages before + * making the page state change in the RMP table. + */ + if (op == SNP_PAGE_STATE_SHARED) + sev_snp_issue_pvalidate(paddr, false); + + /* Request the page state change in the RMP table. */ + sev_snp_pages_state_change(paddr, op); + + /* + * Now that pages are added in the RMP table as a private memory, validate the + * memory range so that it is consistent with the RMP entry. + */ + if (op == SNP_PAGE_STATE_PRIVATE) + sev_snp_issue_pvalidate(paddr, true); +} + +void sev_snp_set_page_private(unsigned long paddr) +{ + sev_snp_set_page_private_shared(paddr, SNP_PAGE_STATE_PRIVATE); +} + +void sev_snp_set_page_shared(unsigned long paddr) +{ + sev_snp_set_page_private_shared(paddr, SNP_PAGE_STATE_SHARED); +} diff --git a/arch/x86/boot/compressed/sev-snp.h b/arch/x86/boot/compressed/sev-snp.h new file mode 100644 index 000000000000..12fe9581a255 --- /dev/null +++ b/arch/x86/boot/compressed/sev-snp.h @@ -0,0 +1,25 @@ +/* SPDX-License-Identifier: GPL-2.0 */ +/* + * AMD SEV Secure Nested Paging Support + * + * Copyright (C) 2021 Advanced Micro Devices, Inc. + * + * Author: Brijesh Singh <brijesh.singh@xxxxxxx> + */ + +#ifndef __COMPRESSED_SECURE_NESTED_PAGING_H +#define __COMPRESSED_SECURE_NESTED_PAGING_H + +#ifdef CONFIG_AMD_MEM_ENCRYPT + +void sev_snp_set_page_private(unsigned long paddr); +void sev_snp_set_page_shared(unsigned long paddr); + +#else + +static inline void sev_snp_set_page_private(unsigned long paddr) { } +static inline void sev_snp_set_page_shared(unsigned long paddr) { } + +#endif /* CONFIG_AMD_MEM_ENCRYPT */ + +#endif /* __COMPRESSED_SECURE_NESTED_PAGING_H */ -- 2.17.1