On Tue, Sep 13, 2022 at 06:36:33AM +0000, Liam Howlett wrote: > > It sounds like the binder_alloc vma_vm_mm is being used unsafely as > well? I'd actually go the other way with this and try to add more > validation that are optimized out on production builds. Since binder is > saving a pointer to the mm struct and was saving the vma ponter, we > should be very careful around how we use them. Is the mutex in > binder_alloc protection enough for the vma binder buffers uses? How is > the close() not being called before the exit_mmap() path? The alloc->mutex is top-level so it can't be used under vm_ops or we risk a possible deadlock with mmap_lock unfortunately. > > When you look at the mmget_not_zero() stuff, have a look at > binder_alloc_new_buf_locked(). I think it is unsafely using the > vma_vm_mm pointer without calling mmget_not_zero(), but the calling > function is rather large so I'm not sure. We had used mm safely in places like binder_update_page_range() but not so in the recent changes to switch over to vma_lookup(). It seems this can be an issue if ->release() races with binder_alloc_print_allocated() so I'll have a closer look at this. So a fix for the initial BUG concern has landed in v5.15.70: https://git.kernel.org/stable/c/899f4160b140 However, after doing a deeper investigation it seems there is still an underlying problem. This requires a bit of context so please bear with me while I try to explain. It started with the maple tree patchset in linux-next which added a late allocation in mmap_region() in commit ("mm: start tracking VMAs with maple tree"). Syzbot failed this allocation and so mmap_region() unwinds, munmaps and frees the vma. This error path makes the cached alloc->vma in binder a dangling pointer. Liam explains the scenario here: https://lore.kernel.org/all/20220621020814.sjszxp3uz43rka62@revolver/ Also, Liam correctly points out that is safer to lookup the vma instead of caching a pointer to it. Such change is what eventually is proposed as the fix to the fuzzer report. However, I wonder why isn't ->mmap() being undone for this exit path in mmap_region()? If anything fails _after_ call_mmap() it seems we silently unmap and free the vma. What about allocations, atomic_incs, and anything done inside ->mmap()? Shouldn't a ->close() be needed here to undo these things as such: -- @@ -1872,6 +1889,9 @@ unsigned long mmap_region(struct file *file, unsigned long addr, return addr; +undo_mmap: + if (vma->vm_ops && vma->vm_ops->close) + vma->vm_ops->close(vma); unmap_and_free_vma: fput(vma->vm_file); vma->vm_file = NULL; -- I managed to reproduce the same syzbot issue in v5.15.41 by failing the arch_validate_flags() check by simply passing PROT_MTE flag to mmap(). I ran this in a "qemu-system-aarch64 -M virt,mte=on" system. Am I missing something? It looks scary to me all the memleaks, corrupt ref counts, etc. that could follow from this simple path. -- Carlos Llamas