The patch titled Subject: mm: drop unneeded ->vm_ops checks has been added to the -mm tree. Its filename is mm-drop-unneeded-vm_ops-checks.patch This patch should soon appear at http://ozlabs.org/~akpm/mmots/broken-out/mm-drop-unneeded-vm_ops-checks.patch and later at http://ozlabs.org/~akpm/mmotm/broken-out/mm-drop-unneeded-vm_ops-checks.patch Before you just go and hit "reply", please: a) Consider who else should be cc'ed b) Prefer to cc a suitable mailing list as well c) Ideally: find the original patch on the mailing list and do a reply-to-all to that, adding suitable additional cc's *** Remember to use Documentation/process/submit-checklist.rst when testing your code *** The -mm tree is included into linux-next and is updated there every 3-4 working days ------------------------------------------------------ From: "Kirill A. Shutemov" <kirill.shutemov@xxxxxxxxxxxxxxx> Subject: mm: drop unneeded ->vm_ops checks We now have all VMAs with ->vm_ops set and don't need to check it for NULL everywhere. Link: http://lkml.kernel.org/r/20180710134821.84709-3-kirill.shutemov@xxxxxxxxxxxxxxx Signed-off-by: Kirill A. Shutemov <kirill.shutemov@xxxxxxxxxxxxxxx> Reviewed-by: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> Cc: Dmitry Vyukov <dvyukov@xxxxxxxxxx> Cc: Oleg Nesterov <oleg@xxxxxxxxxx> Cc: Andrea Arcangeli <aarcange@xxxxxxxxxx> Signed-off-by: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> --- fs/binfmt_elf.c | 2 +- fs/kernfs/file.c | 20 +------------------- fs/proc/task_mmu.c | 2 +- kernel/events/core.c | 2 +- kernel/fork.c | 2 +- mm/hugetlb.c | 2 +- mm/memory.c | 12 ++++++------ mm/mempolicy.c | 10 +++++----- mm/mmap.c | 14 +++++++------- mm/mremap.c | 2 +- mm/nommu.c | 4 ++-- 11 files changed, 27 insertions(+), 45 deletions(-) diff -puN fs/binfmt_elf.c~mm-drop-unneeded-vm_ops-checks fs/binfmt_elf.c --- a/fs/binfmt_elf.c~mm-drop-unneeded-vm_ops-checks +++ a/fs/binfmt_elf.c @@ -1301,7 +1301,7 @@ static bool always_dump_vma(struct vm_ar * Assume that all vmas with a .name op should always be dumped. * If this changes, a new vm_ops field can easily be added. */ - if (vma->vm_ops && vma->vm_ops->name && vma->vm_ops->name(vma)) + if (vma->vm_ops->name && vma->vm_ops->name(vma)) return true; /* diff -puN fs/kernfs/file.c~mm-drop-unneeded-vm_ops-checks fs/kernfs/file.c --- a/fs/kernfs/file.c~mm-drop-unneeded-vm_ops-checks +++ a/fs/kernfs/file.c @@ -336,9 +336,6 @@ static void kernfs_vma_open(struct vm_ar struct file *file = vma->vm_file; struct kernfs_open_file *of = kernfs_of(file); - if (!of->vm_ops) - return; - if (!kernfs_get_active(of->kn)) return; @@ -354,9 +351,6 @@ static vm_fault_t kernfs_vma_fault(struc struct kernfs_open_file *of = kernfs_of(file); vm_fault_t ret; - if (!of->vm_ops) - return VM_FAULT_SIGBUS; - if (!kernfs_get_active(of->kn)) return VM_FAULT_SIGBUS; @@ -374,9 +368,6 @@ static vm_fault_t kernfs_vma_page_mkwrit struct kernfs_open_file *of = kernfs_of(file); vm_fault_t ret; - if (!of->vm_ops) - return VM_FAULT_SIGBUS; - if (!kernfs_get_active(of->kn)) return VM_FAULT_SIGBUS; @@ -397,9 +388,6 @@ static int kernfs_vma_access(struct vm_a struct kernfs_open_file *of = kernfs_of(file); int ret; - if (!of->vm_ops) - return -EINVAL; - if (!kernfs_get_active(of->kn)) return -EINVAL; @@ -419,9 +407,6 @@ static int kernfs_vma_set_policy(struct struct kernfs_open_file *of = kernfs_of(file); int ret; - if (!of->vm_ops) - return 0; - if (!kernfs_get_active(of->kn)) return -EINVAL; @@ -440,9 +425,6 @@ static struct mempolicy *kernfs_vma_get_ struct kernfs_open_file *of = kernfs_of(file); struct mempolicy *pol; - if (!of->vm_ops) - return vma->vm_policy; - if (!kernfs_get_active(of->kn)) return vma->vm_policy; @@ -511,7 +493,7 @@ static int kernfs_fop_mmap(struct file * * So error if someone is trying to use close. */ rc = -EINVAL; - if (vma->vm_ops && vma->vm_ops->close) + if (vma->vm_ops->close) goto out_put; rc = 0; diff -puN fs/proc/task_mmu.c~mm-drop-unneeded-vm_ops-checks fs/proc/task_mmu.c --- a/fs/proc/task_mmu.c~mm-drop-unneeded-vm_ops-checks +++ a/fs/proc/task_mmu.c @@ -326,7 +326,7 @@ show_map_vma(struct seq_file *m, struct goto done; } - if (vma->vm_ops && vma->vm_ops->name) { + if (vma->vm_ops->name) { name = vma->vm_ops->name(vma); if (name) goto done; diff -puN kernel/events/core.c~mm-drop-unneeded-vm_ops-checks kernel/events/core.c --- a/kernel/events/core.c~mm-drop-unneeded-vm_ops-checks +++ a/kernel/events/core.c @@ -7269,7 +7269,7 @@ static void perf_event_mmap_event(struct goto got_name; } else { - if (vma->vm_ops && vma->vm_ops->name) { + if (vma->vm_ops->name) { name = (char *) vma->vm_ops->name(vma); if (name) goto cpy_name; diff -puN kernel/fork.c~mm-drop-unneeded-vm_ops-checks kernel/fork.c --- a/kernel/fork.c~mm-drop-unneeded-vm_ops-checks +++ a/kernel/fork.c @@ -519,7 +519,7 @@ static __latent_entropy int dup_mmap(str if (!(tmp->vm_flags & VM_WIPEONFORK)) retval = copy_page_range(mm, oldmm, mpnt); - if (tmp->vm_ops && tmp->vm_ops->open) + if (tmp->vm_ops->open) tmp->vm_ops->open(tmp); if (retval) diff -puN mm/hugetlb.c~mm-drop-unneeded-vm_ops-checks mm/hugetlb.c --- a/mm/hugetlb.c~mm-drop-unneeded-vm_ops-checks +++ a/mm/hugetlb.c @@ -637,7 +637,7 @@ EXPORT_SYMBOL_GPL(linear_hugepage_index) */ unsigned long vma_kernel_pagesize(struct vm_area_struct *vma) { - if (vma->vm_ops && vma->vm_ops->pagesize) + if (vma->vm_ops->pagesize) return vma->vm_ops->pagesize(vma); return PAGE_SIZE; } diff -puN mm/memory.c~mm-drop-unneeded-vm_ops-checks mm/memory.c --- a/mm/memory.c~mm-drop-unneeded-vm_ops-checks +++ a/mm/memory.c @@ -768,7 +768,7 @@ static void print_bad_pte(struct vm_area (void *)addr, vma->vm_flags, vma->anon_vma, mapping, index); pr_alert("file:%pD fault:%pf mmap:%pf readpage:%pf\n", vma->vm_file, - vma->vm_ops ? vma->vm_ops->fault : NULL, + vma->vm_ops->fault, vma->vm_file ? vma->vm_file->f_op->mmap : NULL, mapping ? mapping->a_ops->readpage : NULL); dump_stack(); @@ -825,7 +825,7 @@ struct page *_vm_normal_page(struct vm_a if (IS_ENABLED(CONFIG_ARCH_HAS_PTE_SPECIAL)) { if (likely(!pte_special(pte))) goto check_pfn; - if (vma->vm_ops && vma->vm_ops->find_special_page) + if (vma->vm_ops->find_special_page) return vma->vm_ops->find_special_page(vma, addr); if (vma->vm_flags & (VM_PFNMAP | VM_MIXEDMAP)) return NULL; @@ -2398,7 +2398,7 @@ static void fault_dirty_shared_page(stru { struct address_space *mapping; bool dirtied; - bool page_mkwrite = vma->vm_ops && vma->vm_ops->page_mkwrite; + bool page_mkwrite = vma->vm_ops->page_mkwrite; dirtied = set_page_dirty(page); VM_BUG_ON_PAGE(PageAnon(page), page); @@ -2642,7 +2642,7 @@ static int wp_pfn_shared(struct vm_fault { struct vm_area_struct *vma = vmf->vma; - if (vma->vm_ops && vma->vm_ops->pfn_mkwrite) { + if (vma->vm_ops->pfn_mkwrite) { int ret; pte_unmap_unlock(vmf->pte, vmf->ptl); @@ -2663,7 +2663,7 @@ static int wp_page_shared(struct vm_faul get_page(vmf->page); - if (vma->vm_ops && vma->vm_ops->page_mkwrite) { + if (vma->vm_ops->page_mkwrite) { int tmp; pte_unmap_unlock(vmf->pte, vmf->ptl); @@ -4433,7 +4433,7 @@ int __access_remote_vm(struct task_struc vma = find_vma(mm, addr); if (!vma || vma->vm_start > addr) break; - if (vma->vm_ops && vma->vm_ops->access) + if (vma->vm_ops->access) ret = vma->vm_ops->access(vma, addr, buf, len, write); if (ret <= 0) diff -puN mm/mempolicy.c~mm-drop-unneeded-vm_ops-checks mm/mempolicy.c --- a/mm/mempolicy.c~mm-drop-unneeded-vm_ops-checks +++ a/mm/mempolicy.c @@ -651,13 +651,13 @@ static int vma_replace_policy(struct vm_ pr_debug("vma %lx-%lx/%lx vm_ops %p vm_file %p set_policy %p\n", vma->vm_start, vma->vm_end, vma->vm_pgoff, vma->vm_ops, vma->vm_file, - vma->vm_ops ? vma->vm_ops->set_policy : NULL); + vma->vm_ops->set_policy); new = mpol_dup(pol); if (IS_ERR(new)) return PTR_ERR(new); - if (vma->vm_ops && vma->vm_ops->set_policy) { + if (vma->vm_ops->set_policy) { err = vma->vm_ops->set_policy(vma, new); if (err) goto err_out; @@ -845,7 +845,7 @@ static long do_get_mempolicy(int *policy up_read(&mm->mmap_sem); return -EFAULT; } - if (vma->vm_ops && vma->vm_ops->get_policy) + if (vma->vm_ops->get_policy) pol = vma->vm_ops->get_policy(vma, addr); else pol = vma->vm_policy; @@ -1617,7 +1617,7 @@ struct mempolicy *__get_vma_policy(struc struct mempolicy *pol = NULL; if (vma) { - if (vma->vm_ops && vma->vm_ops->get_policy) { + if (vma->vm_ops->get_policy) { pol = vma->vm_ops->get_policy(vma, addr); } else if (vma->vm_policy) { pol = vma->vm_policy; @@ -1663,7 +1663,7 @@ bool vma_policy_mof(struct vm_area_struc { struct mempolicy *pol; - if (vma->vm_ops && vma->vm_ops->get_policy) { + if (vma->vm_ops->get_policy) { bool ret = false; pol = vma->vm_ops->get_policy(vma, vma->vm_start); diff -puN mm/mmap.c~mm-drop-unneeded-vm_ops-checks mm/mmap.c --- a/mm/mmap.c~mm-drop-unneeded-vm_ops-checks +++ a/mm/mmap.c @@ -180,7 +180,7 @@ static struct vm_area_struct *remove_vma struct vm_area_struct *next = vma->vm_next; might_sleep(); - if (vma->vm_ops && vma->vm_ops->close) + if (vma->vm_ops->close) vma->vm_ops->close(vma); if (vma->vm_file) fput(vma->vm_file); @@ -1001,7 +1001,7 @@ static inline int is_mergeable_vma(struc return 0; if (vma->vm_file != file) return 0; - if (vma->vm_ops && vma->vm_ops->close) + if (vma->vm_ops->close) return 0; if (!is_mergeable_vm_userfaultfd_ctx(vma, vm_userfaultfd_ctx)) return 0; @@ -1641,7 +1641,7 @@ int vma_wants_writenotify(struct vm_area return 0; /* The backer wishes to know when pages are first written to? */ - if (vm_ops && (vm_ops->page_mkwrite || vm_ops->pfn_mkwrite)) + if (vm_ops->page_mkwrite || vm_ops->pfn_mkwrite) return 1; /* The open routine did something to the protections that pgprot_modify @@ -2627,7 +2627,7 @@ int __split_vma(struct mm_struct *mm, st struct vm_area_struct *new; int err; - if (vma->vm_ops && vma->vm_ops->split) { + if (vma->vm_ops->split) { err = vma->vm_ops->split(vma, addr); if (err) return err; @@ -2660,7 +2660,7 @@ int __split_vma(struct mm_struct *mm, st if (new->vm_file) get_file(new->vm_file); - if (new->vm_ops && new->vm_ops->open) + if (new->vm_ops->open) new->vm_ops->open(new); if (new_below) @@ -2674,7 +2674,7 @@ int __split_vma(struct mm_struct *mm, st return 0; /* Clean everything up if vma_adjust failed. */ - if (new->vm_ops && new->vm_ops->close) + if (new->vm_ops->close) new->vm_ops->close(new); if (new->vm_file) fput(new->vm_file); @@ -3234,7 +3234,7 @@ struct vm_area_struct *copy_vma(struct v goto out_free_mempol; if (new_vma->vm_file) get_file(new_vma->vm_file); - if (new_vma->vm_ops && new_vma->vm_ops->open) + if (new_vma->vm_ops->open) new_vma->vm_ops->open(new_vma); vma_link(mm, new_vma, prev, rb_link, rb_parent); *need_rmap_locks = false; diff -puN mm/mremap.c~mm-drop-unneeded-vm_ops-checks mm/mremap.c --- a/mm/mremap.c~mm-drop-unneeded-vm_ops-checks +++ a/mm/mremap.c @@ -302,7 +302,7 @@ static unsigned long move_vma(struct vm_ need_rmap_locks); if (moved_len < old_len) { err = -ENOMEM; - } else if (vma->vm_ops && vma->vm_ops->mremap) { + } else if (vma->vm_ops->mremap) { err = vma->vm_ops->mremap(new_vma); } diff -puN mm/nommu.c~mm-drop-unneeded-vm_ops-checks mm/nommu.c --- a/mm/nommu.c~mm-drop-unneeded-vm_ops-checks +++ a/mm/nommu.c @@ -760,7 +760,7 @@ static void delete_vma_from_mm(struct vm */ static void delete_vma(struct mm_struct *mm, struct vm_area_struct *vma) { - if (vma->vm_ops && vma->vm_ops->close) + if (vma->vm_ops->close) vma->vm_ops->close(vma); if (vma->vm_file) fput(vma->vm_file); @@ -1485,7 +1485,7 @@ int split_vma(struct mm_struct *mm, stru region->vm_pgoff = new->vm_pgoff += npages; } - if (new->vm_ops && new->vm_ops->open) + if (new->vm_ops->open) new->vm_ops->open(new); delete_vma_from_mm(vma); _ Patches currently in -mm which might be from kirill.shutemov@xxxxxxxxxxxxxxx are mm-fix-vma_is_anonymous-false-positives.patch mm-page_ext-drop-definition-of-unused-page_ext_debug_poison.patch mm-page_ext-constify-lookup_page_ext-argument.patch mm-drop-unneeded-vm_ops-checks.patch -- To unsubscribe from this list: send the line "unsubscribe mm-commits" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html