I asked about this before and it still isn't covered in the description: You were specifically asked (maybe in person at LSF/MM?) not to modify allocator to pass the keyid around. Please specifically mention how this design addresses that feedback in the patch description. You were told, "don't change the core allocator", so I think you just added new functions that wrap the core allocator and called them from the majority of sites that call into the core allocator. Personally, I think that misses the point of the original request. Do I have a better way? Nope, not really. > +/* > + * Encrypted page has to be cleared once keyid is set, not on allocation. > + */ > +static inline bool encrypted_page_needs_zero(int keyid, gfp_t *gfp_mask) > +{ > + if (!keyid) > + return false; > + > + if (*gfp_mask & __GFP_ZERO) { > + *gfp_mask &= ~__GFP_ZERO; > + return true; > + } > + > + return false; > +} Shouldn't this be zero_page_at_alloc()? Otherwise, it gets confusing about whether the page needs zeroing at *all*, vs at alloc vs. free. > +static inline struct page *alloc_pages_node_keyid(int nid, int keyid, > + gfp_t gfp_mask, unsigned int order) > +{ > + if (nid == NUMA_NO_NODE) > + nid = numa_mem_id(); > + > + return __alloc_pages_node_keyid(nid, keyid, gfp_mask, order); > +} We have an innumerable number of (__)?alloc_pages* functions. This adds two more. I'm not a big fan of making this worse. Do I have a better idea? Not really. The best I have is to start being more careful about all of the arguments and actually formalize the list of things that we need to succeed in an allocation in a struct alloc_args or something. > #define alloc_page(gfp_mask) alloc_pages(gfp_mask, 0) > #define alloc_page_vma(gfp_mask, vma, addr) \ > diff --git a/include/linux/migrate.h b/include/linux/migrate.h > index f2b4abbca55e..fede9bfa89d9 100644 > --- a/include/linux/migrate.h > +++ b/include/linux/migrate.h > @@ -38,9 +38,15 @@ static inline struct page *new_page_nodemask(struct page *page, > unsigned int order = 0; > struct page *new_page = NULL; > > - if (PageHuge(page)) > + if (PageHuge(page)) { > + /* > + * HugeTLB doesn't support encryption. We shouldn't see > + * such pages. > + */ > + WARN_ON(page_keyid(page)); > return alloc_huge_page_nodemask(page_hstate(compound_head(page)), > preferred_nid, nodemask); > + } Shouldn't we be returning NULL? Seems like failing the allocation is much less likely to result in bad things happening. > if (PageTransHuge(page)) { > gfp_mask |= GFP_TRANSHUGE; > @@ -50,8 +56,8 @@ static inline struct page *new_page_nodemask(struct page *page, > if (PageHighMem(page) || (zone_idx(page_zone(page)) == ZONE_MOVABLE)) > gfp_mask |= __GFP_HIGHMEM; > > - new_page = __alloc_pages_nodemask(gfp_mask, order, > - preferred_nid, nodemask); > + new_page = __alloc_pages_nodemask_keyid(gfp_mask, order, > + preferred_nid, nodemask, page_keyid(page)); Needs a comment please. It's totally non-obvious that this is the migration case from the context, new_page_nodemask()'s name, or the name of 'page'. /* Allocate a page with the same KeyID as the source page */ > diff --git a/mm/compaction.c b/mm/compaction.c > index faca45ebe62d..fd51aa32ad96 100644 > --- a/mm/compaction.c > +++ b/mm/compaction.c > @@ -1187,6 +1187,7 @@ static struct page *compaction_alloc(struct page *migratepage, > list_del(&freepage->lru); > cc->nr_freepages--; > > + prep_encrypted_page(freepage, 0, page_keyid(migratepage), false); > return freepage; > } Comments, please. Why is this here? What other code might need prep_encrypted_page()? > diff --git a/mm/mempolicy.c b/mm/mempolicy.c > index 581b729e05a0..ce7b436444b5 100644 > --- a/mm/mempolicy.c > +++ b/mm/mempolicy.c > @@ -921,22 +921,28 @@ static void migrate_page_add(struct page *page, struct list_head *pagelist, > /* page allocation callback for NUMA node migration */ > struct page *alloc_new_node_page(struct page *page, unsigned long node) > { > - if (PageHuge(page)) > + if (PageHuge(page)) { > + /* > + * HugeTLB doesn't support encryption. We shouldn't see > + * such pages. > + */ > + WARN_ON(page_keyid(page)); > return alloc_huge_page_node(page_hstate(compound_head(page)), > node); > - else if (PageTransHuge(page)) { > + } else if (PageTransHuge(page)) { > struct page *thp; > > - thp = alloc_pages_node(node, > + thp = alloc_pages_node_keyid(node, page_keyid(page), > (GFP_TRANSHUGE | __GFP_THISNODE), > HPAGE_PMD_ORDER); > if (!thp) > return NULL; > prep_transhuge_page(thp); > return thp; > - } else > - return __alloc_pages_node(node, GFP_HIGHUSER_MOVABLE | > - __GFP_THISNODE, 0); > + } else { > + return __alloc_pages_node_keyid(node, page_keyid(page), > + GFP_HIGHUSER_MOVABLE | __GFP_THISNODE, 0); > + } > } > > /* > @@ -2013,9 +2019,16 @@ alloc_pages_vma(gfp_t gfp, int order, struct vm_area_struct *vma, > { > struct mempolicy *pol; > struct page *page; > - int preferred_nid; > + bool zero = false; > + int keyid, preferred_nid; > nodemask_t *nmask; > > + keyid = vma_keyid(vma); > + if (keyid && (gfp & __GFP_ZERO)) { > + zero = true; > + gfp &= ~__GFP_ZERO; > + } Comments, please. 'zero' should be 'deferred_zero', at least. Also, can't we hide this a _bit_ better? if (deferred_page_zero(vma)) gfp &= ~__GFP_ZERO; Then, later: deferred_page_prep(vma, page, order); and hide everything in deferred_page_zero() and deferred_page_prep(). > --- a/mm/page_alloc.c > +++ b/mm/page_alloc.c > @@ -3697,6 +3697,39 @@ should_compact_retry(struct alloc_context *ac, unsigned int order, int alloc_fla > } > #endif /* CONFIG_COMPACTION */ > > +#ifndef CONFIG_NUMA > +struct page *alloc_pages_vma(gfp_t gfp_mask, int order, > + struct vm_area_struct *vma, unsigned long addr, > + int node, bool hugepage) > +{ > + struct page *page; > + bool need_zero; > + int keyid = vma_keyid(vma); > + > + need_zero = encrypted_page_needs_zero(keyid, &gfp_mask); > + page = alloc_pages(gfp_mask, order); > + prep_encrypted_page(page, order, keyid, need_zero); > + > + return page; > +} > +#endif Is there *ever* a VMA-based allocation that doesn't need zeroing? > +struct page * __alloc_pages_node_keyid(int nid, int keyid, > + gfp_t gfp_mask, unsigned int order) > +{ > + struct page *page; > + bool need_zero; > + > + VM_BUG_ON(nid < 0 || nid >= MAX_NUMNODES); > + VM_WARN_ON(!node_online(nid)); > + > + need_zero = encrypted_page_needs_zero(keyid, &gfp_mask); > + page = __alloc_pages(gfp_mask, order, nid); > + prep_encrypted_page(page, order, keyid, need_zero); > + > + return page; > +} > + > #ifdef CONFIG_LOCKDEP > static struct lockdep_map __fs_reclaim_map = > STATIC_LOCKDEP_MAP_INIT("fs_reclaim", &__fs_reclaim_map); > @@ -4401,6 +4434,20 @@ __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, int preferred_nid, > } > EXPORT_SYMBOL(__alloc_pages_nodemask); > > +struct page * > +__alloc_pages_nodemask_keyid(gfp_t gfp_mask, unsigned int order, > + int preferred_nid, nodemask_t *nodemask, int keyid) > +{ > + struct page *page; > + bool need_zero; > + > + need_zero = encrypted_page_needs_zero(keyid, &gfp_mask); > + page = __alloc_pages_nodemask(gfp_mask, order, preferred_nid, nodemask); > + prep_encrypted_page(page, order, keyid, need_zero); > + return page; > +} > +EXPORT_SYMBOL(__alloc_pages_nodemask_keyid); That looks like three duplicates of the same code, wrapping three more allocator variants. Do we really have no other alternatives? Can you please go ask the folks that gave you the feedback about the allocator modifications and ask them if this is OK explicitly?