Two new functions are introduced, that allow specific memory ranges to be reserved and freed. This is useful when a testcase needs memory at very specific addresses, with the guarantee that the page allocator will not touch those pages. Signed-off-by: Claudio Imbrenda <imbrenda@xxxxxxxxxxxxx> --- lib/alloc_page.h | 15 ++++++++++ lib/alloc_page.c | 78 ++++++++++++++++++++++++++++++++++++++++++++---- 2 files changed, 88 insertions(+), 5 deletions(-) diff --git a/lib/alloc_page.h b/lib/alloc_page.h index 26caefe..7010b20 100644 --- a/lib/alloc_page.h +++ b/lib/alloc_page.h @@ -76,4 +76,19 @@ static inline void free_pages_by_order(void *mem, unsigned int order) free_pages(mem); } +/* + * Allocates and reserves the specified memory range if possible. + * Returns NULL in case of failure. + */ +void *alloc_pages_special(uintptr_t addr, size_t npages); + +/* + * Frees a reserved memory range that had been reserved with + * alloc_pages_special. + * The memory range does not need to match a previous allocation + * exactly, it can also be a subset, in which case only the specified + * pages will be freed and unreserved. + */ +void free_pages_special(uintptr_t addr, size_t npages); + #endif diff --git a/lib/alloc_page.c b/lib/alloc_page.c index d3ade58..1ca905b 100644 --- a/lib/alloc_page.c +++ b/lib/alloc_page.c @@ -21,6 +21,7 @@ #define ORDER_MASK 0x3f #define ALLOC_MASK 0x40 +#define SPECIAL_MASK 0x80 struct free_list { struct free_list *prev; @@ -32,7 +33,7 @@ struct mem_area { uintptr_t base; /* Physical frame number of the first frame outside the area */ uintptr_t top; - /* Combination ALLOC_MASK and order */ + /* Combination of SPECIAL_MASK, ALLOC_MASK, and order */ u8 *page_states; /* One freelist for each possible block size, up to NLISTS */ struct free_list freelists[NLISTS]; @@ -166,6 +167,16 @@ static void *page_memalign_order(struct mem_area *a, u8 al, u8 sz) return res; } +static struct mem_area *get_area(uintptr_t pfn) +{ + uintptr_t i; + + for (i = 0; i < MAX_AREAS; i++) + if ((areas_mask & BIT(i)) && area_contains(areas + i, pfn)) + return areas + i; + return NULL; +} + /* * Try to merge two blocks into a bigger one. * Returns true in case of a successful merge. @@ -233,10 +244,7 @@ static void _free_pages(void *mem) if (!mem) return; assert(IS_ALIGNED((uintptr_t)mem, PAGE_SIZE)); - for (i = 0; !a && (i < MAX_AREAS); i++) { - if ((areas_mask & BIT(i)) && area_contains(areas + i, pfn)) - a = areas + i; - } + a = get_area(pfn); assert_msg(a, "memory does not belong to any area: %p", mem); p = pfn - a->base; @@ -267,6 +275,66 @@ void free_pages(void *mem) spin_unlock(&lock); } +static void *_alloc_page_special(uintptr_t addr) +{ + struct mem_area *a; + uintptr_t mask, i; + + a = get_area(PFN(addr)); + assert(a); + i = PFN(addr) - a->base; + if (a->page_states[i] & (ALLOC_MASK | SPECIAL_MASK)) + return NULL; + while (a->page_states[i]) { + mask = GENMASK_ULL(63, PAGE_SHIFT + a->page_states[i]); + split(a, (void *)(addr & mask)); + } + a->page_states[i] = SPECIAL_MASK; + return (void *)addr; +} + +static void _free_page_special(uintptr_t addr) +{ + struct mem_area *a; + uintptr_t i; + + a = get_area(PFN(addr)); + assert(a); + i = PFN(addr) - a->base; + assert(a->page_states[i] == SPECIAL_MASK); + a->page_states[i] = ALLOC_MASK; + _free_pages((void *)addr); +} + +void *alloc_pages_special(uintptr_t addr, size_t n) +{ + uintptr_t i; + + assert(IS_ALIGNED(addr, PAGE_SIZE)); + spin_lock(&lock); + for (i = 0; i < n; i++) + if (!_alloc_page_special(addr + i * PAGE_SIZE)) + break; + if (i < n) { + for (n = 0 ; n < i; n++) + _free_page_special(addr + n * PAGE_SIZE); + addr = 0; + } + spin_unlock(&lock); + return (void *)addr; +} + +void free_pages_special(uintptr_t addr, size_t n) +{ + uintptr_t i; + + assert(IS_ALIGNED(addr, PAGE_SIZE)); + spin_lock(&lock); + for (i = 0; i < n; i++) + _free_page_special(addr + i * PAGE_SIZE); + spin_unlock(&lock); +} + static void *page_memalign_order_area(unsigned area, u8 ord, u8 al) { void *res = NULL; -- 2.26.2