vma_start_read() can temporarily raise vm_refcnt of a write-locked and detached vma: // vm_refcnt==1 (attached) vma_start_write() vma->vm_lock_seq = mm->mm_lock_seq vma_start_read() vm_refcnt++; // vm_refcnt==2 vma_mark_detached() vm_refcnt--; // vm_refcnt==1 // vma is detached but vm_refcnt!=0 temporarily if (vma->vm_lock_seq == mm->mm_lock_seq) vma_refcount_put() vm_refcnt--; // vm_refcnt==0 This is currently not a problem when freeing the vma because RCU grace period should pass before kmem_cache_free(vma) gets called and by that time vma_start_read() should be done and vm_refcnt is 0. However once we introduce possibility of vma reuse before RCU grace period is over, this will become a problem (reused vma might be in non-detached state). Introduce vma_ensure_detached() for the writer to wait for readers until they exit vma_start_read(). Signed-off-by: Suren Baghdasaryan <surenb@xxxxxxxxxx> --- include/linux/mm.h | 9 ++++++ mm/memory.c | 55 +++++++++++++++++++++++--------- tools/testing/vma/vma_internal.h | 8 +++++ 3 files changed, 57 insertions(+), 15 deletions(-) diff --git a/include/linux/mm.h b/include/linux/mm.h index b73cf64233a4..361f26dedab1 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -863,6 +863,15 @@ static inline bool is_vma_detached(struct vm_area_struct *vma) return refcount_read(&vma->vm_refcnt) == VMA_STATE_DETACHED; } +/* + * WARNING: to avoid racing with vma_mark_attached(), should be called either + * under mmap_write_lock or when the object has been isolated under + * mmap_write_lock, ensuring no competing writers. + * Should be called after marking vma as detached to wait for possible + * readers which temporarily raised vm_refcnt to drop it back and exit. + */ +void vma_ensure_detached(struct vm_area_struct *vma); + static inline void vma_mark_attached(struct vm_area_struct *vma) { vma_assert_write_locked(vma); diff --git a/mm/memory.c b/mm/memory.c index cff132003e24..534e279f98c1 100644 --- a/mm/memory.c +++ b/mm/memory.c @@ -6329,18 +6329,10 @@ struct vm_area_struct *lock_mm_and_find_vma(struct mm_struct *mm, #endif #ifdef CONFIG_PER_VMA_LOCK -void __vma_start_write(struct vm_area_struct *vma, unsigned int mm_lock_seq) +static inline bool __vma_enter_locked(struct vm_area_struct *vma) { - bool detached; - - /* - * If vma is detached then only vma_mark_attached() can raise the - * vm_refcnt. mmap_write_lock prevents racing with vma_mark_attached(). - */ - if (!refcount_inc_not_zero(&vma->vm_refcnt)) { - WRITE_ONCE(vma->vm_lock_seq, mm_lock_seq); - return; - } + if (!refcount_inc_not_zero(&vma->vm_refcnt)) + return false; rwsem_acquire(&vma->vmlock_dep_map, 0, 0, _RET_IP_); /* vma is attached, set the writer present bit */ @@ -6350,6 +6342,22 @@ void __vma_start_write(struct vm_area_struct *vma, unsigned int mm_lock_seq) refcount_read(&vma->vm_refcnt) == VMA_STATE_ATTACHED + (VMA_STATE_LOCKED + 1), TASK_UNINTERRUPTIBLE); lock_acquired(&vma->vmlock_dep_map, _RET_IP_); + + return true; +} + +static inline void __vma_exit_locked(struct vm_area_struct *vma, bool *is_detached) +{ + *is_detached = refcount_sub_and_test(VMA_STATE_LOCKED + 1, + &vma->vm_refcnt); + rwsem_release(&vma->vmlock_dep_map, _RET_IP_); +} + +void __vma_start_write(struct vm_area_struct *vma, unsigned int mm_lock_seq) +{ + bool locked; + + locked = __vma_enter_locked(vma); /* * We should use WRITE_ONCE() here because we can have concurrent reads * from the early lockless pessimistic check in vma_start_read(). @@ -6357,13 +6365,30 @@ void __vma_start_write(struct vm_area_struct *vma, unsigned int mm_lock_seq) * we should use WRITE_ONCE() for cleanliness and to keep KCSAN happy. */ WRITE_ONCE(vma->vm_lock_seq, mm_lock_seq); - detached = refcount_sub_and_test(VMA_STATE_LOCKED + 1, - &vma->vm_refcnt); - rwsem_release(&vma->vmlock_dep_map, _RET_IP_); - VM_BUG_ON_VMA(detached, vma); /* vma should remain attached */ + if (locked) { + bool detached; + + __vma_exit_locked(vma, &detached); + /* vma was originally attached and should remain so */ + VM_BUG_ON_VMA(detached, vma); + } } EXPORT_SYMBOL_GPL(__vma_start_write); +void vma_ensure_detached(struct vm_area_struct *vma) +{ + if (is_vma_detached(vma)) + return; + + if (__vma_enter_locked(vma)) { + bool detached; + + /* Wait for temporary readers to drop the vm_refcnt */ + __vma_exit_locked(vma, &detached); + VM_BUG_ON_VMA(!detached, vma); + } +} + /* * Lookup and lock a VMA under RCU protection. Returned VMA is guaranteed to be * stable and not isolated. If the VMA is not found or is being modified the diff --git a/tools/testing/vma/vma_internal.h b/tools/testing/vma/vma_internal.h index b55556b16060..ac0a59906fea 100644 --- a/tools/testing/vma/vma_internal.h +++ b/tools/testing/vma/vma_internal.h @@ -465,6 +465,14 @@ static inline bool is_vma_detached(struct vm_area_struct *vma) return refcount_read(&vma->vm_refcnt) == VMA_STATE_DETACHED; } +static inline void vma_ensure_detached(struct vm_area_struct *vma) +{ + if (is_vma_detached(vma)) + return; + + refcount_set(&vma->vm_refcnt, VMA_STATE_DETACHED); +} + static inline void vma_assert_write_locked(struct vm_area_struct *); static inline void vma_mark_attached(struct vm_area_struct *vma) { -- 2.47.1.613.gc27f4b7a9f-goog