On 05/22/2018 09:41 AM, Reinette Chatre wrote: > On 5/21/2018 4:48 PM, Mike Kravetz wrote: >> On 05/21/2018 01:54 AM, Vlastimil Babka wrote: >>> On 05/04/2018 01:29 AM, Mike Kravetz wrote: >>>> +/** >>>> + * find_alloc_contig_pages() -- attempt to find and allocate a contiguous >>>> + * range of pages >>>> + * @nr_pages: number of pages to find/allocate >>>> + * @gfp: gfp mask used to limit search as well as during compaction >>>> + * @nid: target node >>>> + * @nodemask: mask of other possible nodes >>>> + * >>>> + * Pages can be freed with a call to free_contig_pages(), or by manually >>>> + * calling __free_page() for each page allocated. >>>> + * >>>> + * Return: pointer to 'order' pages on success, or NULL if not successful. >>>> + */ >>>> +struct page *find_alloc_contig_pages(unsigned long nr_pages, gfp_t gfp, >>>> + int nid, nodemask_t *nodemask) >>>> +{ >>>> + unsigned long i, alloc_order, order_pages; >>>> + struct page *pages; >>>> + >>>> + /* >>>> + * Underlying allocators perform page order sized allocations. >>>> + */ >>>> + alloc_order = get_count_order(nr_pages); >>> >>> So if takes arbitrary nr_pages but convert it to order anyway? I think >>> that's rather suboptimal and wasteful... e.g. a range could be skipped >>> because some of the pages added by rounding cannot be migrated away. >> >> Yes. My idea with this series was to use existing allocators which are >> all order based. Let me think about how to do allocation for arbitrary >> number of allocations. >> - For less than MAX_ORDER size we rely on the buddy allocator, so we are >> pretty much stuck with order sized allocation. However, allocations of >> this size are not really interesting as you can call existing routines >> directly. >> - For sizes greater than MAX_ORDER, we know that the allocation size will >> be at least pageblock sized. So, the isolate/migrate scheme can still >> be used for full pageblocks. We can then use direct migration for the >> remaining pages. This does complicate things a bit. >> >> I'm guessing that most (?all?) allocations will be order based. The use >> cases I am aware of (hugetlbfs, Intel Cache Pseudo-Locking, RDMA) are all >> order based. However, as commented in previous version taking arbitrary >> nr_pages makes interface more future proof. >> > > I noticed this Cache Pseudo-Locking statement and would like to clarify. > I have not been following this thread in detail so I would like to > apologize first if my comments are out of context. > > Currently the Cache Pseudo-Locking allocations are order based because I > assumed it was required by the allocator. The contiguous regions needed > by Cache Pseudo-Locking will not always be order based - instead it is > based on the granularity of the cache allocation. One example is a > platform with 55MB L3 cache that can be divided into 20 equal portions. > To support Cache Pseudo-Locking on this platform we need to be able to > allocate contiguous regions at increments of 2816KB (the size of each > portion). In support of this example platform regions needed would thus > be 2816KB, 5632KB, 8448KB, etc. Thank you Reinette. I was not aware of these details. Yours is the most concrete new use case. This certainly makes more of a case for arbitrary sized allocations. -- Mike Kravetz -- To unsubscribe from this list: send the line "unsubscribe linux-api" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html