Hmmm. The intention was to use a non-specific name - I thought 'lock' would work as a generic way to describe any sort of locking primitive, not just mutexes. After all, the main push for the mmap locking API is to make it easier to change the details of how it's implemented... I'm not sure if there would be any better name that conveys that intent ? If we can't agree on a non-specific name, one option could be to leave the name as mmap_sem for now and sweep through new uses of mmap_sem every once in a while to convert these to the new API (which would mostly come down to re-running the coccinelle converter). On Mon, Apr 6, 2020 at 5:46 AM Laurent Dufour <ldufour@xxxxxxxxxxxxx> wrote: > > Le 26/03/2020 à 08:02, Michel Lespinasse a écrit : > > Rename the mmap_sem field to mmap_lock. Any new uses of this lock > > should now go through the new mmap locking api. The mmap_lock is > > still implemented as a rwsem, though this could change in the future. > > Since you don't change the type of the mmap_sem, it's still a semaphore, I don't > think mmap_lock is a good name. This is not a lock. > > I'm bad at naming, but what about __map_sem? > > > > > Signed-off-by: Michel Lespinasse <walken@xxxxxxxxxx> > > --- > > arch/ia64/mm/fault.c | 4 ++-- > > arch/x86/events/core.c | 4 ++-- > > arch/x86/kernel/tboot.c | 2 +- > > arch/x86/mm/fault.c | 2 +- > > drivers/firmware/efi/efi.c | 2 +- > > drivers/gpu/drm/etnaviv/etnaviv_gem.c | 2 +- > > include/linux/mm_types.h | 2 +- > > include/linux/mmap_lock.h | 30 +++++++++++++-------------- > > mm/hmm.c | 2 +- > > mm/init-mm.c | 2 +- > > mm/memory.c | 2 +- > > mm/mmap.c | 4 ++-- > > mm/mmu_notifier.c | 18 ++++++++-------- > > mm/pagewalk.c | 15 +++++++------- > > mm/util.c | 4 ++-- > > 15 files changed, 48 insertions(+), 47 deletions(-) > > > > diff --git a/arch/ia64/mm/fault.c b/arch/ia64/mm/fault.c > > index b423f0a970e4..70c7c7909cc5 100644 > > --- a/arch/ia64/mm/fault.c > > +++ b/arch/ia64/mm/fault.c > > @@ -70,8 +70,8 @@ ia64_do_page_fault (unsigned long address, unsigned long isr, struct pt_regs *re > > mask = ((((isr >> IA64_ISR_X_BIT) & 1UL) << VM_EXEC_BIT) > > | (((isr >> IA64_ISR_W_BIT) & 1UL) << VM_WRITE_BIT)); > > > > - /* mmap_sem is performance critical.... */ > > - prefetchw(&mm->mmap_sem); > > + /* mmap_lock is performance critical.... */ > > + prefetchw(&mm->mmap_lock); > > > > /* > > * If we're in an interrupt or have no user context, we must not take the fault.. > > diff --git a/arch/x86/events/core.c b/arch/x86/events/core.c > > index 3bb738f5a472..ad21924c575e 100644 > > --- a/arch/x86/events/core.c > > +++ b/arch/x86/events/core.c > > @@ -2179,10 +2179,10 @@ static void x86_pmu_event_mapped(struct perf_event *event, struct mm_struct *mm) > > * userspace with CR4.PCE clear while another task is still > > * doing on_each_cpu_mask() to propagate CR4.PCE. > > * > > - * For now, this can't happen because all callers hold mmap_sem > > + * For now, this can't happen because all callers hold mmap_lock > > * for write. If this changes, we'll need a different solution. > > */ > > - lockdep_assert_held_write(&mm->mmap_sem); > > + lockdep_assert_held_write(&mm->mmap_lock); > > > > if (atomic_inc_return(&mm->context.perf_rdpmc_allowed) == 1) > > on_each_cpu_mask(mm_cpumask(mm), refresh_pce, NULL, 1); > > diff --git a/arch/x86/kernel/tboot.c b/arch/x86/kernel/tboot.c > > index 4b79335624b1..4792e8778b28 100644 > > --- a/arch/x86/kernel/tboot.c > > +++ b/arch/x86/kernel/tboot.c > > @@ -90,7 +90,7 @@ static struct mm_struct tboot_mm = { > > .pgd = swapper_pg_dir, > > .mm_users = ATOMIC_INIT(2), > > .mm_count = ATOMIC_INIT(1), > > - .mmap_sem = MMAP_LOCK_INITIALIZER(init_mm.mmap_sem), > > + .mmap_lock = MMAP_LOCK_INITIALIZER(init_mm.mmap_lock), > > .page_table_lock = __SPIN_LOCK_UNLOCKED(init_mm.page_table_lock), > > .mmlist = LIST_HEAD_INIT(init_mm.mmlist), > > }; > > diff --git a/arch/x86/mm/fault.c b/arch/x86/mm/fault.c > > index 5bb97d2a7d3b..98d413e6fbb2 100644 > > --- a/arch/x86/mm/fault.c > > +++ b/arch/x86/mm/fault.c > > @@ -1526,7 +1526,7 @@ dotraplinkage void > > do_page_fault(struct pt_regs *regs, unsigned long hw_error_code, > > unsigned long address) > > { > > - prefetchw(¤t->mm->mmap_sem); > > + prefetchw(¤t->mm->mmap_lock); > > trace_page_fault_entries(regs, hw_error_code, address); > > > > if (unlikely(kmmio_fault(regs, address))) > > diff --git a/drivers/firmware/efi/efi.c b/drivers/firmware/efi/efi.c > > index 5bdfe698cd7f..d38e0e85eb0d 100644 > > --- a/drivers/firmware/efi/efi.c > > +++ b/drivers/firmware/efi/efi.c > > @@ -60,7 +60,7 @@ struct mm_struct efi_mm = { > > .mm_rb = RB_ROOT, > > .mm_users = ATOMIC_INIT(2), > > .mm_count = ATOMIC_INIT(1), > > - .mmap_sem = MMAP_LOCK_INITIALIZER(efi_mm.mmap_sem), > > + .mmap_lock = MMAP_LOCK_INITIALIZER(efi_mm.mmap_lock), > > .page_table_lock = __SPIN_LOCK_UNLOCKED(efi_mm.page_table_lock), > > .mmlist = LIST_HEAD_INIT(efi_mm.mmlist), > > .cpu_bitmap = { [BITS_TO_LONGS(NR_CPUS)] = 0}, > > diff --git a/drivers/gpu/drm/etnaviv/etnaviv_gem.c b/drivers/gpu/drm/etnaviv/etnaviv_gem.c > > index 6adea180d629..3470482d95bc 100644 > > --- a/drivers/gpu/drm/etnaviv/etnaviv_gem.c > > +++ b/drivers/gpu/drm/etnaviv/etnaviv_gem.c > > @@ -661,7 +661,7 @@ static int etnaviv_gem_userptr_get_pages(struct etnaviv_gem_object *etnaviv_obj) > > struct etnaviv_gem_userptr *userptr = &etnaviv_obj->userptr; > > int ret, pinned = 0, npages = etnaviv_obj->base.size >> PAGE_SHIFT; > > > > - might_lock_read(¤t->mm->mmap_sem); > > + might_lock_read(¤t->mm->mmap_lock); > > > > if (userptr->mm != current->mm) > > return -EPERM; > > diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h > > index c28911c3afa8..a168d13b5c44 100644 > > --- a/include/linux/mm_types.h > > +++ b/include/linux/mm_types.h > > @@ -431,7 +431,7 @@ struct mm_struct { > > spinlock_t page_table_lock; /* Protects page tables and some > > * counters > > */ > > - struct rw_semaphore mmap_sem; > > + struct rw_semaphore mmap_lock; > > > > struct list_head mmlist; /* List of maybe swapped mm's. These > > * are globally strung together off > > diff --git a/include/linux/mmap_lock.h b/include/linux/mmap_lock.h > > index 7474b15bba38..700dd297f2af 100644 > > --- a/include/linux/mmap_lock.h > > +++ b/include/linux/mmap_lock.h > > @@ -5,78 +5,78 @@ > > > > static inline void mmap_init_lock(struct mm_struct *mm) > > { > > - init_rwsem(&mm->mmap_sem); > > + init_rwsem(&mm->mmap_lock); > > } > > > > static inline void mmap_write_lock(struct mm_struct *mm) > > { > > - down_write(&mm->mmap_sem); > > + down_write(&mm->mmap_lock); > > } > > > > static inline void mmap_write_lock_nested(struct mm_struct *mm, int subclass) > > { > > - down_write_nested(&mm->mmap_sem, subclass); > > + down_write_nested(&mm->mmap_lock, subclass); > > } > > > > static inline int mmap_write_lock_killable(struct mm_struct *mm) > > { > > - return down_write_killable(&mm->mmap_sem); > > + return down_write_killable(&mm->mmap_lock); > > } > > > > static inline bool mmap_write_trylock(struct mm_struct *mm) > > { > > - return down_write_trylock(&mm->mmap_sem) != 0; > > + return down_write_trylock(&mm->mmap_lock) != 0; > > } > > > > static inline void mmap_write_unlock(struct mm_struct *mm) > > { > > - up_write(&mm->mmap_sem); > > + up_write(&mm->mmap_lock); > > } > > > > /* Pairs with mmap_write_lock_nested() */ > > static inline void mmap_write_unlock_nested(struct mm_struct *mm) > > { > > - up_write(&mm->mmap_sem); > > + up_write(&mm->mmap_lock); > > } > > > > static inline void mmap_downgrade_write_lock(struct mm_struct *mm) > > { > > - downgrade_write(&mm->mmap_sem); > > + downgrade_write(&mm->mmap_lock); > > } > > > > static inline void mmap_read_lock(struct mm_struct *mm) > > { > > - down_read(&mm->mmap_sem); > > + down_read(&mm->mmap_lock); > > } > > > > static inline int mmap_read_lock_killable(struct mm_struct *mm) > > { > > - return down_read_killable(&mm->mmap_sem); > > + return down_read_killable(&mm->mmap_lock); > > } > > > > static inline bool mmap_read_trylock(struct mm_struct *mm) > > { > > - return down_read_trylock(&mm->mmap_sem) != 0; > > + return down_read_trylock(&mm->mmap_lock) != 0; > > } > > > > static inline void mmap_read_unlock(struct mm_struct *mm) > > { > > - up_read(&mm->mmap_sem); > > + up_read(&mm->mmap_lock); > > } > > > > static inline void mmap_read_release(struct mm_struct *mm, unsigned long ip) > > { > > - rwsem_release(&mm->mmap_sem.dep_map, ip); > > + rwsem_release(&mm->mmap_lock.dep_map, ip); > > } > > > > static inline void mmap_read_unlock_non_owner(struct mm_struct *mm) > > { > > - up_read_non_owner(&mm->mmap_sem); > > + up_read_non_owner(&mm->mmap_lock); > > } > > > > static inline bool mmap_is_locked(struct mm_struct *mm) > > { > > - return rwsem_is_locked(&mm->mmap_sem) != 0; > > + return rwsem_is_locked(&mm->mmap_lock) != 0; > > } > > > > #endif /* _LINUX_MMAP_LOCK_H */ > > diff --git a/mm/hmm.c b/mm/hmm.c > > index 72e5a6d9a417..4d736a710910 100644 > > --- a/mm/hmm.c > > +++ b/mm/hmm.c > > @@ -681,7 +681,7 @@ long hmm_range_fault(struct hmm_range *range, unsigned int flags) > > struct mm_struct *mm = range->notifier->mm; > > int ret; > > > > - lockdep_assert_held(&mm->mmap_sem); > > + lockdep_assert_held(&mm->mmap_lock); > > > > do { > > /* If range is no longer valid force retry. */ > > diff --git a/mm/init-mm.c b/mm/init-mm.c > > index 3c128bd6a30c..2b16924419b4 100644 > > --- a/mm/init-mm.c > > +++ b/mm/init-mm.c > > @@ -31,7 +31,7 @@ struct mm_struct init_mm = { > > .pgd = swapper_pg_dir, > > .mm_users = ATOMIC_INIT(2), > > .mm_count = ATOMIC_INIT(1), > > - .mmap_sem = MMAP_LOCK_INITIALIZER(init_mm.mmap_sem), > > + .mmap_lock = MMAP_LOCK_INITIALIZER(init_mm.mmap_lock), > > .page_table_lock = __SPIN_LOCK_UNLOCKED(init_mm.page_table_lock), > > .arg_lock = __SPIN_LOCK_UNLOCKED(init_mm.arg_lock), > > .mmlist = LIST_HEAD_INIT(init_mm.mmlist), > > diff --git a/mm/memory.c b/mm/memory.c > > index 5f9a81b2507a..f146f58898f9 100644 > > --- a/mm/memory.c > > +++ b/mm/memory.c > > @@ -4648,7 +4648,7 @@ void __might_fault(const char *file, int line) > > __might_sleep(file, line, 0); > > #if defined(CONFIG_DEBUG_ATOMIC_SLEEP) > > if (current->mm) > > - might_lock_read(¤t->mm->mmap_sem); > > + might_lock_read(¤t->mm->mmap_lock); > > #endif > > } > > EXPORT_SYMBOL(__might_fault); > > diff --git a/mm/mmap.c b/mm/mmap.c > > index ba51ff516ec0..c81db16c8c23 100644 > > --- a/mm/mmap.c > > +++ b/mm/mmap.c > > @@ -3450,7 +3450,7 @@ static void vm_lock_anon_vma(struct mm_struct *mm, struct anon_vma *anon_vma) > > * The LSB of head.next can't change from under us > > * because we hold the mm_all_locks_mutex. > > */ > > - down_write_nest_lock(&anon_vma->root->rwsem, &mm->mmap_sem); > > + down_write_nest_lock(&anon_vma->root->rwsem, &mm->mmap_lock); > > /* > > * We can safely modify head.next after taking the > > * anon_vma->root->rwsem. If some other vma in this mm shares > > @@ -3480,7 +3480,7 @@ static void vm_lock_mapping(struct mm_struct *mm, struct address_space *mapping) > > */ > > if (test_and_set_bit(AS_MM_ALL_LOCKS, &mapping->flags)) > > BUG(); > > - down_write_nest_lock(&mapping->i_mmap_rwsem, &mm->mmap_sem); > > + down_write_nest_lock(&mapping->i_mmap_rwsem, &mm->mmap_lock); > > } > > } > > > > diff --git a/mm/mmu_notifier.c b/mm/mmu_notifier.c > > index cfd0a03bf5cc..6717278d6d49 100644 > > --- a/mm/mmu_notifier.c > > +++ b/mm/mmu_notifier.c > > @@ -599,7 +599,7 @@ void __mmu_notifier_invalidate_range(struct mm_struct *mm, > > } > > > > /* > > - * Same as mmu_notifier_register but here the caller must hold the mmap_sem in > > + * Same as mmu_notifier_register but here the caller must hold the mmap_lock in > > * write mode. A NULL mn signals the notifier is being registered for itree > > * mode. > > */ > > @@ -609,7 +609,7 @@ int __mmu_notifier_register(struct mmu_notifier *subscription, > > struct mmu_notifier_subscriptions *subscriptions = NULL; > > int ret; > > > > - lockdep_assert_held_write(&mm->mmap_sem); > > + lockdep_assert_held_write(&mm->mmap_lock); > > BUG_ON(atomic_read(&mm->mm_users) <= 0); > > > > if (IS_ENABLED(CONFIG_LOCKDEP)) { > > @@ -623,7 +623,7 @@ int __mmu_notifier_register(struct mmu_notifier *subscription, > > /* > > * kmalloc cannot be called under mm_take_all_locks(), but we > > * know that mm->notifier_subscriptions can't change while we > > - * hold the write side of the mmap_sem. > > + * hold the write side of the mmap_lock. > > */ > > subscriptions = kzalloc( > > sizeof(struct mmu_notifier_subscriptions), GFP_KERNEL); > > @@ -655,7 +655,7 @@ int __mmu_notifier_register(struct mmu_notifier *subscription, > > * readers. acquire can only be used while holding the mmgrab or > > * mmget, and is safe because once created the > > * mmu_notifier_subscriptions is not freed until the mm is destroyed. > > - * As above, users holding the mmap_sem or one of the > > + * As above, users holding the mmap_lock or one of the > > * mm_take_all_locks() do not need to use acquire semantics. > > */ > > if (subscriptions) > > @@ -689,7 +689,7 @@ EXPORT_SYMBOL_GPL(__mmu_notifier_register); > > * @mn: The notifier to attach > > * @mm: The mm to attach the notifier to > > * > > - * Must not hold mmap_sem nor any other VM related lock when calling > > + * Must not hold mmap_lock nor any other VM related lock when calling > > * this registration function. Must also ensure mm_users can't go down > > * to zero while this runs to avoid races with mmu_notifier_release, > > * so mm has to be current->mm or the mm should be pinned safely such > > @@ -750,7 +750,7 @@ find_get_mmu_notifier(struct mm_struct *mm, const struct mmu_notifier_ops *ops) > > * are the same. > > * > > * Each call to mmu_notifier_get() must be paired with a call to > > - * mmu_notifier_put(). The caller must hold the write side of mm->mmap_sem. > > + * mmu_notifier_put(). The caller must hold the write side of mm->mmap_lock. > > * > > * While the caller has a mmu_notifier get the mm pointer will remain valid, > > * and can be converted to an active mm pointer via mmget_not_zero(). > > @@ -761,7 +761,7 @@ struct mmu_notifier *mmu_notifier_get_locked(const struct mmu_notifier_ops *ops, > > struct mmu_notifier *subscription; > > int ret; > > > > - lockdep_assert_held_write(&mm->mmap_sem); > > + lockdep_assert_held_write(&mm->mmap_lock); > > > > if (mm->notifier_subscriptions) { > > subscription = find_get_mmu_notifier(mm, ops); > > @@ -983,7 +983,7 @@ int mmu_interval_notifier_insert(struct mmu_interval_notifier *interval_sub, > > struct mmu_notifier_subscriptions *subscriptions; > > int ret; > > > > - might_lock(&mm->mmap_sem); > > + might_lock(&mm->mmap_lock); > > > > subscriptions = smp_load_acquire(&mm->notifier_subscriptions); > > if (!subscriptions || !subscriptions->has_itree) { > > @@ -1006,7 +1006,7 @@ int mmu_interval_notifier_insert_locked( > > mm->notifier_subscriptions; > > int ret; > > > > - lockdep_assert_held_write(&mm->mmap_sem); > > + lockdep_assert_held_write(&mm->mmap_lock); > > > > if (!subscriptions || !subscriptions->has_itree) { > > ret = __mmu_notifier_register(NULL, mm); > > diff --git a/mm/pagewalk.c b/mm/pagewalk.c > > index 928df1638c30..d669a3146c0f 100644 > > --- a/mm/pagewalk.c > > +++ b/mm/pagewalk.c > > @@ -373,8 +373,9 @@ static int __walk_page_range(unsigned long start, unsigned long end, > > * caller-specific data to callbacks, @private should be helpful. > > * > > * Locking: > > - * Callers of walk_page_range() and walk_page_vma() should hold @mm->mmap_sem, > > - * because these function traverse vma list and/or access to vma's data. > > + * Callers of walk_page_range() and walk_page_vma() should hold > > + * @mm->mmap_lock, because these function traverse the vma list > > + * and/or access the vma's data. > > */ > > int walk_page_range(struct mm_struct *mm, unsigned long start, > > unsigned long end, const struct mm_walk_ops *ops, > > @@ -395,7 +396,7 @@ int walk_page_range(struct mm_struct *mm, unsigned long start, > > if (!walk.mm) > > return -EINVAL; > > > > - lockdep_assert_held(&walk.mm->mmap_sem); > > + lockdep_assert_held(&walk.mm->mmap_lock); > > > > vma = find_vma(walk.mm, start); > > do { > > @@ -453,7 +454,7 @@ int walk_page_range_novma(struct mm_struct *mm, unsigned long start, > > if (start >= end || !walk.mm) > > return -EINVAL; > > > > - lockdep_assert_held(&walk.mm->mmap_sem); > > + lockdep_assert_held(&walk.mm->mmap_lock); > > > > return __walk_page_range(start, end, &walk); > > } > > @@ -472,7 +473,7 @@ int walk_page_vma(struct vm_area_struct *vma, const struct mm_walk_ops *ops, > > if (!walk.mm) > > return -EINVAL; > > > > - lockdep_assert_held(&walk.mm->mmap_sem); > > + lockdep_assert_held(&walk.mm->mmap_lock); > > > > err = walk_page_test(vma->vm_start, vma->vm_end, &walk); > > if (err > 0) > > @@ -498,11 +499,11 @@ int walk_page_vma(struct vm_area_struct *vma, const struct mm_walk_ops *ops, > > * Also see walk_page_range() for additional information. > > * > > * Locking: > > - * This function can't require that the struct mm_struct::mmap_sem is held, > > + * This function can't require that the struct mm_struct::mmap_lock is held, > > * since @mapping may be mapped by multiple processes. Instead > > * @mapping->i_mmap_rwsem must be held. This might have implications in the > > * callbacks, and it's up tho the caller to ensure that the > > - * struct mm_struct::mmap_sem is not needed. > > + * struct mm_struct::mmap_lock is not needed. > > * > > * Also this means that a caller can't rely on the struct > > * vm_area_struct::vm_flags to be constant across a call, > > diff --git a/mm/util.c b/mm/util.c > > index ea2e15b21446..56c562f7ad19 100644 > > --- a/mm/util.c > > +++ b/mm/util.c > > @@ -425,7 +425,7 @@ void arch_pick_mmap_layout(struct mm_struct *mm, struct rlimit *rlim_stack) > > * @bypass_rlim: %true if checking RLIMIT_MEMLOCK should be skipped > > * > > * Assumes @task and @mm are valid (i.e. at least one reference on each), and > > - * that mmap_sem is held as writer. > > + * that mmap_lock is held as writer. > > * > > * Return: > > * * 0 on success > > @@ -437,7 +437,7 @@ int __account_locked_vm(struct mm_struct *mm, unsigned long pages, bool inc, > > unsigned long locked_vm, limit; > > int ret = 0; > > > > - lockdep_assert_held_write(&mm->mmap_sem); > > + lockdep_assert_held_write(&mm->mmap_lock); > > > > locked_vm = mm->locked_vm; > > if (inc) { > > > -- Michel "Walken" Lespinasse A program is never fully debugged until the last user dies.