From: Xiaogang Chen <xiaogang.chen@xxxxxxx> Current kfd/svm driver acquires mm's mmap write lock before update mm->notifier_subscriptions->itree. This tree is already protected by mm->notifier_subscriptions->lock at mmu notifier. Each process mm interval tree update from different components in kernel go to mmu interval notifier where they got serialized. This patch removes mmap write lock acquiring at kfd/svm driver when need updates process mm interval tree. It reduces chance of dead lock or warning from lockdev and simplifies the driver code. In addition, the patch adjusts some locks granularity to reduce the lock number that driver holds at same time which also reduces the chance of dead lock or warning from lockdev. Signed-off-by: Xiaogang Chen<Xiaogang.Chen@xxxxxxx> --- drivers/gpu/drm/amd/amdkfd/kfd_chardev.c | 3 +- drivers/gpu/drm/amd/amdkfd/kfd_migrate.c | 6 +- drivers/gpu/drm/amd/amdkfd/kfd_svm.c | 181 +++++++++++++++-------- drivers/gpu/drm/amd/amdkfd/kfd_svm.h | 2 +- 4 files changed, 122 insertions(+), 70 deletions(-) diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c b/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c index fdf171ad4a3c..b52588ded567 100644 --- a/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c +++ b/drivers/gpu/drm/amd/amdkfd/kfd_chardev.c @@ -1078,9 +1078,8 @@ static int kfd_ioctl_alloc_memory_of_gpu(struct file *filep, /* Flush pending deferred work to avoid racing with deferred actions * from previous memory map changes (e.g. munmap). */ - svm_range_list_lock_and_flush_work(&p->svms, current->mm); + svm_range_list_flush_work(&p->svms); mutex_lock(&p->svms.lock); - mmap_write_unlock(current->mm); if (interval_tree_iter_first(&p->svms.objects, args->va_addr >> PAGE_SHIFT, (args->va_addr + args->size - 1) >> PAGE_SHIFT)) { diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_migrate.c b/drivers/gpu/drm/amd/amdkfd/kfd_migrate.c index 8ee3d07ffbdf..eb46643d96b2 100644 --- a/drivers/gpu/drm/amd/amdkfd/kfd_migrate.c +++ b/drivers/gpu/drm/amd/amdkfd/kfd_migrate.c @@ -969,10 +969,12 @@ static vm_fault_t svm_migrate_to_ram(struct vm_fault *vmf) mutex_lock(&p->svms.lock); prange = svm_range_from_addr(&p->svms, addr, NULL); + + mutex_unlock(&p->svms.lock); if (!prange) { pr_debug("failed get range svms 0x%p addr 0x%lx\n", &p->svms, addr); r = -EFAULT; - goto out_unlock_svms; + goto out_unref_process; } mutex_lock(&prange->migrate_mutex); @@ -993,8 +995,6 @@ static vm_fault_t svm_migrate_to_ram(struct vm_fault *vmf) out_unlock_prange: mutex_unlock(&prange->migrate_mutex); -out_unlock_svms: - mutex_unlock(&p->svms.lock); out_unref_process: pr_debug("CPU fault svms 0x%p address 0x%lx done\n", &p->svms, addr); kfd_unref_process(p); diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c index 407636a68814..46f81c1215d9 100644 --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c @@ -106,12 +106,31 @@ static void svm_range_unlink(struct svm_range *prange) } static void -svm_range_add_notifier_locked(struct mm_struct *mm, struct svm_range *prange) +svm_range_add_notifier(struct mm_struct *mm, struct svm_range *prange, bool locked) { pr_debug("svms 0x%p prange 0x%p [0x%lx 0x%lx]\n", prange->svms, prange, prange->start, prange->last); - mmu_interval_notifier_insert_locked(&prange->notifier, mm, + /* mm->notifier_subscriptions should have been setup for this process + * ex: during kfd process creation + */ + WARN_ON_ONCE(!mm->notifier_subscriptions); + + /* not necessary hold mmap lock to update mm interval notifier tree as + * opeations on mm->notifier_subscriptions->itree are serialized by + * mm->notifier_subscriptions->lock + */ + if (locked) { + /* if mmap write lock has been hold use lock version to udpate + * mm interval notifier tree + */ + mmu_interval_notifier_insert_locked(&prange->notifier, mm, + prange->start << PAGE_SHIFT, + prange->npages << PAGE_SHIFT, + &svm_range_mn_ops); + } else + /* use no-mmap-lock version to update mm interval notifier tree */ + mmu_interval_notifier_insert(&prange->notifier, mm, prange->start << PAGE_SHIFT, prange->npages << PAGE_SHIFT, &svm_range_mn_ops); @@ -895,6 +914,7 @@ svm_range_copy_array(void *psrc, size_t size, uint64_t num_elements, *vram_pages = 0; for (i = 0; i < num_elements; i++) { dma_addr_t *temp; + temp = (dma_addr_t *)dst + i; *temp = *((dma_addr_t *)src + i); if (*temp&SVM_RANGE_VRAM_DOMAIN) @@ -1500,7 +1520,7 @@ static int svm_range_reserve_bos(struct svm_validate_context *ctx, bool intr) uint32_t gpuidx; int r; - drm_exec_init(&ctx->exec, intr ? DRM_EXEC_INTERRUPTIBLE_WAIT: 0, 0); + drm_exec_init(&ctx->exec, intr ? DRM_EXEC_INTERRUPTIBLE_WAIT : 0, 0); drm_exec_until_all_locked(&ctx->exec) { for_each_set_bit(gpuidx, ctx->bitmap, MAX_GPU_INSTANCE) { pdd = kfd_process_device_from_gpuidx(ctx->process, gpuidx); @@ -1751,25 +1771,20 @@ static int svm_range_validate_and_map(struct mm_struct *mm, } /** - * svm_range_list_lock_and_flush_work - flush pending deferred work + * svm_range_list_flush_work - flush pending deferred work * * @svms: the svm range list * @mm: the mm structure - * - * Context: Returns with mmap write lock held, pending deferred work flushed - * */ void -svm_range_list_lock_and_flush_work(struct svm_range_list *svms, - struct mm_struct *mm) +svm_range_list_flush_work(struct svm_range_list *svms) { retry_flush_work: flush_work(&svms->deferred_list_work); - mmap_write_lock(mm); if (list_empty(&svms->deferred_range_list)) return; - mmap_write_unlock(mm); + pr_debug("retry flush\n"); goto retry_flush_work; } @@ -1804,7 +1819,11 @@ static void svm_range_restore_work(struct work_struct *work) } mutex_lock(&process_info->lock); - svm_range_list_lock_and_flush_work(svms, mm); + + svm_range_list_flush_work(svms); + + mmap_read_lock(mm); + mutex_lock(&svms->lock); evicted_ranges = atomic_read(&svms->evicted_ranges); @@ -1854,8 +1873,9 @@ static void svm_range_restore_work(struct work_struct *work) pr_debug("restore svm ranges successfully\n"); out_reschedule: + mutex_unlock(&svms->lock); - mmap_write_unlock(mm); + mmap_read_unlock(mm); mutex_unlock(&process_info->lock); /* If validation failed, reschedule another attempt */ @@ -2186,8 +2206,8 @@ svm_range_add(struct kfd_process *p, uint64_t start, uint64_t size, } static void -svm_range_update_notifier_and_interval_tree(struct mm_struct *mm, - struct svm_range *prange) +svm_range_update_notifier_and_interval_tree(struct svm_range_list *svms, + struct mm_struct *mm, struct svm_range *prange) { unsigned long start; unsigned long last; @@ -2203,14 +2223,28 @@ svm_range_update_notifier_and_interval_tree(struct mm_struct *mm, prange->last); if (start != 0 && last != 0) { + + mutex_lock(&svms->lock); + mutex_lock(&prange->lock); interval_tree_remove(&prange->it_node, &prange->svms->objects); + mutex_unlock(&prange->lock); + mutex_unlock(&svms->lock); + svm_range_remove_notifier(prange); } + + mutex_lock(&prange->lock); prange->it_node.start = prange->start; prange->it_node.last = prange->last; + mutex_unlock(&prange->lock); + mutex_lock(&svms->lock); + mutex_lock(&prange->lock); interval_tree_insert(&prange->it_node, &prange->svms->objects); - svm_range_add_notifier_locked(mm, prange); + mutex_unlock(&prange->lock); + mutex_unlock(&svms->lock); + + svm_range_add_notifier(mm, prange, true); } static void @@ -2225,32 +2259,51 @@ svm_range_handle_list_op(struct svm_range_list *svms, struct svm_range *prange, case SVM_OP_UNMAP_RANGE: pr_debug("remove 0x%p prange 0x%p [0x%lx 0x%lx]\n", svms, prange, prange->start, prange->last); + + mutex_lock(&svms->lock); + svm_range_lock(prange); svm_range_unlink(prange); + svm_range_unlock(prange); + mutex_unlock(&svms->lock); + svm_range_remove_notifier(prange); svm_range_free(prange, true); break; case SVM_OP_UPDATE_RANGE_NOTIFIER: pr_debug("update notifier 0x%p prange 0x%p [0x%lx 0x%lx]\n", svms, prange, prange->start, prange->last); - svm_range_update_notifier_and_interval_tree(mm, prange); + svm_range_update_notifier_and_interval_tree(svms, mm, prange); break; case SVM_OP_UPDATE_RANGE_NOTIFIER_AND_MAP: pr_debug("update and map 0x%p prange 0x%p [0x%lx 0x%lx]\n", svms, prange, prange->start, prange->last); - svm_range_update_notifier_and_interval_tree(mm, prange); + svm_range_update_notifier_and_interval_tree(svms, mm, prange); + /* TODO: implement deferred validation and mapping */ break; case SVM_OP_ADD_RANGE: pr_debug("add 0x%p prange 0x%p [0x%lx 0x%lx]\n", svms, prange, prange->start, prange->last); + + mutex_lock(&svms->lock); + mutex_lock(&prange->lock); svm_range_add_to_svms(prange); - svm_range_add_notifier_locked(mm, prange); + mutex_unlock(&prange->lock); + mutex_unlock(&svms->lock); + + svm_range_add_notifier(mm, prange, true); break; case SVM_OP_ADD_RANGE_AND_MAP: pr_debug("add and map 0x%p prange 0x%p [0x%lx 0x%lx]\n", svms, prange, prange->start, prange->last); + + mutex_lock(&svms->lock); + mutex_lock(&prange->lock); svm_range_add_to_svms(prange); - svm_range_add_notifier_locked(mm, prange); + mutex_unlock(&prange->lock); + mutex_unlock(&svms->lock); + + svm_range_add_notifier(mm, prange, true); /* TODO: implement deferred validation and mapping */ break; default: @@ -2316,6 +2369,12 @@ static void svm_range_deferred_list_work(struct work_struct *work) mm = prange->work_item.mm; retry: + /* use mmap write lock to prevent unmap_from_cpu for same process + * happening at same time to prevent race conditions between + * svm_range_deferred_list_work and unmap_from_cpu, ex: + * work_item.op of prange may be changed by unmap_from_cpu + * during svm_range_deferred_list_work + */ mmap_write_lock(mm); /* Checking for the need to drain retry faults must be inside @@ -2327,20 +2386,11 @@ static void svm_range_deferred_list_work(struct work_struct *work) goto retry; } - /* Remove from deferred_list must be inside mmap write lock, for - * two race cases: - * 1. unmap_from_cpu may change work_item.op and add the range - * to deferred_list again, cause use after free bug. - * 2. svm_range_list_lock_and_flush_work may hold mmap write - * lock and continue because deferred_list is empty, but - * deferred_list work is actually waiting for mmap lock. - */ spin_lock(&svms->deferred_list_lock); list_del_init(&prange->deferred_list); spin_unlock(&svms->deferred_list_lock); - mutex_lock(&svms->lock); - mutex_lock(&prange->migrate_mutex); + svm_range_lock(prange); while (!list_empty(&prange->child_list)) { struct svm_range *pchild; @@ -2349,12 +2399,15 @@ static void svm_range_deferred_list_work(struct work_struct *work) pr_debug("child prange 0x%p op %d\n", pchild, pchild->work_item.op); list_del_init(&pchild->child_list); + svm_range_unlock(prange); + svm_range_handle_list_op(svms, pchild, mm); + svm_range_lock(prange); } - mutex_unlock(&prange->migrate_mutex); + svm_range_unlock(prange); svm_range_handle_list_op(svms, prange, mm); - mutex_unlock(&svms->lock); + mmap_write_unlock(mm); /* Pairs with mmget in svm_range_add_list_work. If dropping the @@ -2814,7 +2867,7 @@ svm_range *svm_range_create_unregistered_range(struct kfd_node *node, prange->preferred_loc = KFD_IOCTL_SVM_LOCATION_SYSMEM; svm_range_add_to_svms(prange); - svm_range_add_notifier_locked(mm, prange); + svm_range_add_notifier(mm, prange, false); return prange; } @@ -2911,7 +2964,6 @@ svm_range_restore_pages(struct amdgpu_device *adev, unsigned int pasid, struct kfd_node *node; int32_t best_loc; int32_t gpuidx = MAX_GPU_INSTANCE; - bool write_locked = false; struct vm_area_struct *vma; bool migration = false; int r = 0; @@ -2959,35 +3011,26 @@ svm_range_restore_pages(struct amdgpu_device *adev, unsigned int pasid, r = -EFAULT; goto out; } - mmap_read_lock(mm); -retry_write_locked: + mutex_lock(&svms->lock); prange = svm_range_from_addr(svms, addr, NULL); + mutex_unlock(&svms->lock); if (!prange) { pr_debug("failed to find prange svms 0x%p address [0x%llx]\n", svms, addr); - if (!write_locked) { - /* Need the write lock to create new range with MMU notifier. - * Also flush pending deferred work to make sure the interval - * tree is up to date before we add a new range - */ - mutex_unlock(&svms->lock); - mmap_read_unlock(mm); - mmap_write_lock(mm); - write_locked = true; - goto retry_write_locked; - } + /* not need to hold mmap write lock before create new range in + * mm interval notifier tree, mm has its own lock to protect it + */ prange = svm_range_create_unregistered_range(node, p, mm, addr); if (!prange) { pr_debug("failed to create unregistered range svms 0x%p address [0x%llx]\n", svms, addr); - mmap_write_downgrade(mm); r = -EFAULT; goto out_unlock_svms; } } - if (write_locked) - mmap_write_downgrade(mm); + + mmap_read_lock(mm); mutex_lock(&prange->migrate_mutex); @@ -3082,10 +3125,8 @@ svm_range_restore_pages(struct amdgpu_device *adev, unsigned int pasid, out_unlock_range: mutex_unlock(&prange->migrate_mutex); -out_unlock_svms: - mutex_unlock(&svms->lock); mmap_read_unlock(mm); - +out_unlock_svms: svm_range_count_fault(node, p, gpuidx); mmput(mm); @@ -3192,6 +3233,7 @@ int svm_range_list_init(struct kfd_process *p) { struct svm_range_list *svms = &p->svms; int i; + int ret; svms->objects = RB_ROOT_CACHED; mutex_init(&svms->lock); @@ -3208,6 +3250,14 @@ int svm_range_list_init(struct kfd_process *p) if (KFD_IS_SVM_API_SUPPORTED(p->pdds[i]->dev->adev)) bitmap_set(svms->bitmap_supported, i, 1); + /* signals the notifier is being registered for itree mode + * it does not register new mmu_notifier or increae mm->mm_count, + * so not need call mmu_notifier_unregister at svm_range_list_fini + */ + ret = mmu_notifier_register(NULL, p->mm); + if (ret) + return ret; + return 0; } @@ -3553,29 +3603,27 @@ svm_range_set_attr(struct kfd_process *p, struct mm_struct *mm, mutex_lock(&process_info->lock); - svm_range_list_lock_and_flush_work(svms, mm); + svm_range_list_flush_work(svms); r = svm_range_is_valid(p, start, size); if (r) { pr_debug("invalid range r=%d\n", r); - mmap_write_unlock(mm); goto out; } - mutex_lock(&svms->lock); - /* Add new range and split existing ranges as needed */ r = svm_range_add(p, start, size, nattr, attrs, &update_list, &insert_list, &remove_list, &remap_list); - if (r) { - mutex_unlock(&svms->lock); - mmap_write_unlock(mm); + if (r) goto out; - } + /* Apply changes as a transaction */ list_for_each_entry_safe(prange, next, &insert_list, list) { + mutex_lock(&svms->lock); svm_range_add_to_svms(prange); - svm_range_add_notifier_locked(mm, prange); + mutex_unlock(&svms->lock); + + svm_range_add_notifier(mm, prange, false); } list_for_each_entry(prange, &update_list, update_list) { svm_range_apply_attrs(p, prange, nattr, attrs, &update_mapping); @@ -3585,12 +3633,17 @@ svm_range_set_attr(struct kfd_process *p, struct mm_struct *mm, pr_debug("unlink old 0x%p prange 0x%p [0x%lx 0x%lx]\n", prange->svms, prange, prange->start, prange->last); + + mutex_lock(&svms->lock); svm_range_unlink(prange); + mutex_unlock(&svms->lock); + svm_range_remove_notifier(prange); svm_range_free(prange, false); } - mmap_write_downgrade(mm); + mmap_read_lock(mm); + /* Trigger migrations and revalidate and map to GPUs as needed. If * this fails we may be left with partially completed actions. There * is no clean way of rolling back to the previous state in such a @@ -3646,8 +3699,8 @@ svm_range_set_attr(struct kfd_process *p, struct mm_struct *mm, dynamic_svm_range_dump(svms); - mutex_unlock(&svms->lock); mmap_read_unlock(mm); + out: mutex_unlock(&process_info->lock); diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.h b/drivers/gpu/drm/amd/amdkfd/kfd_svm.h index 70c1776611c4..f7cee69ff191 100644 --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.h +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.h @@ -195,7 +195,7 @@ int kfd_criu_restore_svm(struct kfd_process *p, int kfd_criu_resume_svm(struct kfd_process *p); struct kfd_process_device * svm_range_get_pdd_by_node(struct svm_range *prange, struct kfd_node *node); -void svm_range_list_lock_and_flush_work(struct svm_range_list *svms, struct mm_struct *mm); +void svm_range_list_flush_work(struct svm_range_list *svms); /* SVM API and HMM page migration work together, device memory type * is initialized to not 0 when page migration register device memory. -- 2.25.1