From: Nadav Amit <nadav.amit@xxxxxxxxx> Now that alloc_page() zeros the page, remove the redundant page zeroing. Suggested-by: Andrew Jones <drjones@xxxxxxxxxx> Signed-off-by: Nadav Amit <nadav.amit@xxxxxxxxx> --- lib/arm/asm/pgtable.h | 2 -- lib/arm/mmu.c | 1 - lib/arm64/asm/pgtable.h | 1 - 3 files changed, 4 deletions(-) diff --git a/lib/arm/asm/pgtable.h b/lib/arm/asm/pgtable.h index b614bce..241dff6 100644 --- a/lib/arm/asm/pgtable.h +++ b/lib/arm/asm/pgtable.h @@ -53,7 +53,6 @@ static inline pmd_t *pmd_alloc_one(void) { assert(PTRS_PER_PMD * sizeof(pmd_t) == PAGE_SIZE); pmd_t *pmd = alloc_page(); - memset(pmd, 0, PTRS_PER_PMD * sizeof(pmd_t)); return pmd; } static inline pmd_t *pmd_alloc(pgd_t *pgd, unsigned long addr) @@ -80,7 +79,6 @@ static inline pte_t *pte_alloc_one(void) { assert(PTRS_PER_PTE * sizeof(pte_t) == PAGE_SIZE); pte_t *pte = alloc_page(); - memset(pte, 0, PTRS_PER_PTE * sizeof(pte_t)); return pte; } static inline pte_t *pte_alloc(pmd_t *pmd, unsigned long addr) diff --git a/lib/arm/mmu.c b/lib/arm/mmu.c index 03f6622..3d38c83 100644 --- a/lib/arm/mmu.c +++ b/lib/arm/mmu.c @@ -166,7 +166,6 @@ void *setup_mmu(phys_addr_t phys_end) #endif mmu_idmap = alloc_page(); - memset(mmu_idmap, 0, PAGE_SIZE); /* * mach-virt I/O regions: diff --git a/lib/arm64/asm/pgtable.h b/lib/arm64/asm/pgtable.h index 5860abe..ee0a2c8 100644 --- a/lib/arm64/asm/pgtable.h +++ b/lib/arm64/asm/pgtable.h @@ -61,7 +61,6 @@ static inline pte_t *pte_alloc_one(void) { assert(PTRS_PER_PTE * sizeof(pte_t) == PAGE_SIZE); pte_t *pte = alloc_page(); - memset(pte, 0, PTRS_PER_PTE * sizeof(pte_t)); return pte; } static inline pte_t *pte_alloc(pmd_t *pmd, unsigned long addr) -- 2.17.1