On Thu, 18 Jul 2019, Yang Shi wrote: > The transhuge_vma_suitable() was only available for shmem THP, but > anonymous THP has the same check except pgoff check. And, it will be > used for THP eligible check in the later patch, so make it available for > all kind of THPs. This also helps reduce code duplication slightly. > > Since anonymous THP doesn't have to check pgoff, so make pgoff check > shmem vma only. > > And regroup some functions in include/linux/mm.h to solve compile issue since > transhuge_vma_suitable() needs call vma_is_anonymous() which was defined > after huge_mm.h is included. > > Cc: Hugh Dickins <hughd@xxxxxxxxxx> Thanks! Acked-by: Hugh Dickins <hughd@xxxxxxxxxx> > Cc: Kirill A. Shutemov <kirill.shutemov@xxxxxxxxxxxxxxx> > Cc: Michal Hocko <mhocko@xxxxxxxx> > Cc: Vlastimil Babka <vbabka@xxxxxxx> > Cc: David Rientjes <rientjes@xxxxxxxxxx> > Signed-off-by: Yang Shi <yang.shi@xxxxxxxxxxxxxxxxx> > --- > include/linux/huge_mm.h | 23 +++++++++++++++++++++++ > include/linux/mm.h | 34 +++++++++++++++++----------------- > mm/huge_memory.c | 2 +- > mm/memory.c | 13 ------------- > 4 files changed, 41 insertions(+), 31 deletions(-) > > diff --git a/include/linux/huge_mm.h b/include/linux/huge_mm.h > index 7cd5c15..45ede62 100644 > --- a/include/linux/huge_mm.h > +++ b/include/linux/huge_mm.h > @@ -121,6 +121,23 @@ static inline bool __transparent_hugepage_enabled(struct vm_area_struct *vma) > > bool transparent_hugepage_enabled(struct vm_area_struct *vma); > > +#define HPAGE_CACHE_INDEX_MASK (HPAGE_PMD_NR - 1) > + > +static inline bool transhuge_vma_suitable(struct vm_area_struct *vma, > + unsigned long haddr) > +{ > + /* Don't have to check pgoff for anonymous vma */ > + if (!vma_is_anonymous(vma)) { > + if (((vma->vm_start >> PAGE_SHIFT) & HPAGE_CACHE_INDEX_MASK) != > + (vma->vm_pgoff & HPAGE_CACHE_INDEX_MASK)) > + return false; > + } > + > + if (haddr < vma->vm_start || haddr + HPAGE_PMD_SIZE > vma->vm_end) > + return false; > + return true; > +} > + > #define transparent_hugepage_use_zero_page() \ > (transparent_hugepage_flags & \ > (1<<TRANSPARENT_HUGEPAGE_USE_ZERO_PAGE_FLAG)) > @@ -271,6 +288,12 @@ static inline bool transparent_hugepage_enabled(struct vm_area_struct *vma) > return false; > } > > +static inline bool transhuge_vma_suitable(struct vm_area_struct *vma, > + unsigned long haddr) > +{ > + return false; > +} > + > static inline void prep_transhuge_page(struct page *page) {} > > #define transparent_hugepage_flags 0UL > diff --git a/include/linux/mm.h b/include/linux/mm.h > index 0389c34..beae0ae 100644 > --- a/include/linux/mm.h > +++ b/include/linux/mm.h > @@ -541,6 +541,23 @@ static inline void vma_set_anonymous(struct vm_area_struct *vma) > vma->vm_ops = NULL; > } > > +static inline bool vma_is_anonymous(struct vm_area_struct *vma) > +{ > + return !vma->vm_ops; > +} > + > +#ifdef CONFIG_SHMEM > +/* > + * The vma_is_shmem is not inline because it is used only by slow > + * paths in userfault. > + */ > +bool vma_is_shmem(struct vm_area_struct *vma); > +#else > +static inline bool vma_is_shmem(struct vm_area_struct *vma) { return false; } > +#endif > + > +int vma_is_stack_for_current(struct vm_area_struct *vma); > + > /* flush_tlb_range() takes a vma, not a mm, and can care about flags */ > #define TLB_FLUSH_VMA(mm,flags) { .vm_mm = (mm), .vm_flags = (flags) } > > @@ -1629,23 +1646,6 @@ static inline void cancel_dirty_page(struct page *page) > > int get_cmdline(struct task_struct *task, char *buffer, int buflen); > > -static inline bool vma_is_anonymous(struct vm_area_struct *vma) > -{ > - return !vma->vm_ops; > -} > - > -#ifdef CONFIG_SHMEM > -/* > - * The vma_is_shmem is not inline because it is used only by slow > - * paths in userfault. > - */ > -bool vma_is_shmem(struct vm_area_struct *vma); > -#else > -static inline bool vma_is_shmem(struct vm_area_struct *vma) { return false; } > -#endif > - > -int vma_is_stack_for_current(struct vm_area_struct *vma); > - > extern unsigned long move_page_tables(struct vm_area_struct *vma, > unsigned long old_addr, struct vm_area_struct *new_vma, > unsigned long new_addr, unsigned long len, > diff --git a/mm/huge_memory.c b/mm/huge_memory.c > index 885642c..782dd14 100644 > --- a/mm/huge_memory.c > +++ b/mm/huge_memory.c > @@ -689,7 +689,7 @@ vm_fault_t do_huge_pmd_anonymous_page(struct vm_fault *vmf) > struct page *page; > unsigned long haddr = vmf->address & HPAGE_PMD_MASK; > > - if (haddr < vma->vm_start || haddr + HPAGE_PMD_SIZE > vma->vm_end) > + if (!transhuge_vma_suitable(vma, haddr)) > return VM_FAULT_FALLBACK; > if (unlikely(anon_vma_prepare(vma))) > return VM_FAULT_OOM; > diff --git a/mm/memory.c b/mm/memory.c > index 89325f9..e2bb51b 100644 > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -3162,19 +3162,6 @@ static vm_fault_t pte_alloc_one_map(struct vm_fault *vmf) > } > > #ifdef CONFIG_TRANSPARENT_HUGE_PAGECACHE > - > -#define HPAGE_CACHE_INDEX_MASK (HPAGE_PMD_NR - 1) > -static inline bool transhuge_vma_suitable(struct vm_area_struct *vma, > - unsigned long haddr) > -{ > - if (((vma->vm_start >> PAGE_SHIFT) & HPAGE_CACHE_INDEX_MASK) != > - (vma->vm_pgoff & HPAGE_CACHE_INDEX_MASK)) > - return false; > - if (haddr < vma->vm_start || haddr + HPAGE_PMD_SIZE > vma->vm_end) > - return false; > - return true; > -} > - > static void deposit_prealloc_pte(struct vm_fault *vmf) > { > struct vm_area_struct *vma = vmf->vma; > -- > 1.8.3.1 > >