On 28 June 2016 at 18:05, Catalin Marinas <catalin.marinas@xxxxxxx> wrote: > (Restarting the thread before I forget the entire discussion) > > On Mon, Jun 06, 2016 at 11:18:14PM +0200, Ard Biesheuvel wrote: >> >> > On 31 May 2016 at 17:14, Catalin Marinas <catalin.marinas@xxxxxxx> wrote: >> >> > > diff --git a/arch/arm64/kernel/efi.c b/arch/arm64/kernel/efi.c >> >> > > index 78f52488f9ff..0d5753c31c7f 100644 >> >> > > --- a/arch/arm64/kernel/efi.c >> >> > > +++ b/arch/arm64/kernel/efi.c >> >> > > @@ -62,10 +62,26 @@ struct screen_info screen_info __section(.data); >> >> > > int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md) >> >> > > { >> >> > > pteval_t prot_val = create_mapping_protection(md); >> >> > > + phys_addr_t length = md->num_pages << EFI_PAGE_SHIFT; >> >> > > + efi_memory_desc_t *next = md; >> >> > > >> >> > > - create_pgd_mapping(mm, md->phys_addr, md->virt_addr, >> >> > > - md->num_pages << EFI_PAGE_SHIFT, >> >> > > - __pgprot(prot_val | PTE_NG)); >> >> > > + /* >> >> > > + * Search for the next EFI runtime map and check for any overlap with >> >> > > + * the current map when aligned to PAGE_SIZE. In such case, defer >> >> > > + * mapping the end of the current range until the next >> >> > > + * efi_create_mapping() call. >> >> > > + */ >> >> > > + for_each_efi_memory_desc_continue(next) { >> >> > > + if (!(next->attribute & EFI_MEMORY_RUNTIME)) >> >> > > + continue; >> >> > > + if (next->phys_addr < PAGE_ALIGN(md->phys_addr + length)) >> >> > > + length -= (md->phys_addr + length) & ~PAGE_MASK; > [...] >> Another thing I failed to mention is that the new Memory Attributes >> table support may map all of the RuntimeServicesCode regions a second >> time, but with a higher granularity, using RO for .text and .rodata >> and NX for .data and .bss (and the PE/COFF header). > > Can this not be done in a single go without multiple passes? That's what > we did for the core arm64 code, the only one left being EFI run-time > mappings. > Well, we probably could, but it is far from trivial. >> Due to the higher >> granularity, regions that were mapped using the contiguous bit the >> first time around may be split into smaller regions. Your current code >> does not address that case. > > If the above doesn't work, the only solution would be to permanently map > these ranges as individual pages, no large blocks. > That is not unreasonable, since regions >2MB are unusual. >> I wonder how the PT debug code deals with this, and whether there is >> anything we can reuse to inhibit cont and block mappings > > Do you mean DEBUG_PAGEALLOC? When enabled, it defaults to single page > mappings for the kernel, no sections. > Indeed. So we could reuse the logic that imposes that to prevent creating regions that are subject to splitting later on. That way, there is no need to worry about the code running twice in case the Memory Attributes table is exposed by the firmware. -- To unsubscribe from this list: send the line "unsubscribe linux-efi" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html