This will be used for recharging memory cgroup accounting. Signed-off-by: Konstantin Khlebnikov <khlebnikov@xxxxxxxxxxxxxx> --- mm/gup.c | 2 +- mm/huge_memory.c | 4 ++-- mm/internal.h | 4 ++-- mm/ksm.c | 2 +- mm/migrate.c | 2 +- mm/mlock.c | 2 +- mm/rmap.c | 2 +- 7 files changed, 9 insertions(+), 9 deletions(-) diff --git a/mm/gup.c b/mm/gup.c index 98f13ab37bac..f0accc229266 100644 --- a/mm/gup.c +++ b/mm/gup.c @@ -306,7 +306,7 @@ static struct page *follow_page_pte(struct vm_area_struct *vma, * know the page is still mapped, we don't even * need to check for file-cache page truncation. */ - mlock_vma_page(page); + mlock_vma_page(vma, page); unlock_page(page); } } diff --git a/mm/huge_memory.c b/mm/huge_memory.c index de1f15969e27..157faa231e26 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -1513,7 +1513,7 @@ struct page *follow_trans_huge_pmd(struct vm_area_struct *vma, goto skip_mlock; lru_add_drain(); if (page->mapping && !PageDoubleMap(page)) - mlock_vma_page(page); + mlock_vma_page(vma, page); unlock_page(page); } skip_mlock: @@ -3009,7 +3009,7 @@ void remove_migration_pmd(struct page_vma_mapped_walk *pvmw, struct page *new) page_add_file_rmap(new, true); set_pmd_at(mm, mmun_start, pvmw->pmd, pmde); if ((vma->vm_flags & VM_LOCKED) && !PageDoubleMap(new)) - mlock_vma_page(new); + mlock_vma_page(vma, new); update_mmu_cache_pmd(vma, address, pvmw->pmd); } #endif diff --git a/mm/internal.h b/mm/internal.h index e32390802fd3..9f91992ef281 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -305,7 +305,7 @@ static inline void munlock_vma_pages_all(struct vm_area_struct *vma) /* * must be called with vma's mmap_sem held for read or write, and page locked. */ -extern void mlock_vma_page(struct page *page); +extern void mlock_vma_page(struct vm_area_struct *vma, struct page *page); extern unsigned int munlock_vma_page(struct page *page); /* @@ -364,7 +364,7 @@ vma_address(struct page *page, struct vm_area_struct *vma) #else /* !CONFIG_MMU */ static inline void clear_page_mlock(struct page *page) { } -static inline void mlock_vma_page(struct page *page) { } +static inline void mlock_vma_page(struct vm_area_struct *, struct page *) { } static inline void mlock_migrate_page(struct page *new, struct page *old) { } #endif /* !CONFIG_MMU */ diff --git a/mm/ksm.c b/mm/ksm.c index 3dc4346411e4..cb5705d6f26c 100644 --- a/mm/ksm.c +++ b/mm/ksm.c @@ -1274,7 +1274,7 @@ static int try_to_merge_one_page(struct vm_area_struct *vma, if (!PageMlocked(kpage)) { unlock_page(page); lock_page(kpage); - mlock_vma_page(kpage); + mlock_vma_page(vma, kpage); page = kpage; /* for final unlock */ } } diff --git a/mm/migrate.c b/mm/migrate.c index a42858d8e00b..1f6151cb7310 100644 --- a/mm/migrate.c +++ b/mm/migrate.c @@ -269,7 +269,7 @@ static bool remove_migration_pte(struct page *page, struct vm_area_struct *vma, page_add_file_rmap(new, false); } if (vma->vm_flags & VM_LOCKED && !PageTransCompound(new)) - mlock_vma_page(new); + mlock_vma_page(vma, new); if (PageTransHuge(page) && PageMlocked(page)) clear_page_mlock(page); diff --git a/mm/mlock.c b/mm/mlock.c index a90099da4fb4..73d477aaa411 100644 --- a/mm/mlock.c +++ b/mm/mlock.c @@ -85,7 +85,7 @@ void clear_page_mlock(struct page *page) * Mark page as mlocked if not already. * If page on LRU, isolate and putback to move to unevictable list. */ -void mlock_vma_page(struct page *page) +void mlock_vma_page(struct vm_area_struct *vma, struct page *page) { /* Serialize with page migration */ BUG_ON(!PageLocked(page)); diff --git a/mm/rmap.c b/mm/rmap.c index 003377e24232..de88f4897c1d 100644 --- a/mm/rmap.c +++ b/mm/rmap.c @@ -1410,7 +1410,7 @@ static bool try_to_unmap_one(struct page *page, struct vm_area_struct *vma, * Holding pte lock, we do *not* need * mmap_sem here */ - mlock_vma_page(page); + mlock_vma_page(vma, page); } ret = false; page_vma_mapped_walk_done(&pvmw);