On Wed, Jan 16, 2019 at 10:50:00AM -0700, Jens Axboe wrote: > If we have fixed user buffers, we can map them into the kernel when we > setup the io_context. That avoids the need to do get_user_pages() for > each and every IO. ..... > + return -ENOMEM; > + } while (atomic_long_cmpxchg(&ctx->user->locked_vm, cur_pages, > + new_pages) != cur_pages); > + > + return 0; > +} > + > +static int io_sqe_buffer_unregister(struct io_ring_ctx *ctx) > +{ > + int i, j; > + > + if (!ctx->user_bufs) > + return -EINVAL; > + > + for (i = 0; i < ctx->sq_entries; i++) { > + struct io_mapped_ubuf *imu = &ctx->user_bufs[i]; > + > + for (j = 0; j < imu->nr_bvecs; j++) { > + set_page_dirty_lock(imu->bvec[j].bv_page); > + put_page(imu->bvec[j].bv_page); > + } Hmmm, so we call set_page_dirty() when the gup reference is dropped... ..... > +static int io_sqe_buffer_register(struct io_ring_ctx *ctx, void __user *arg, > + unsigned nr_args) > +{ ..... > + down_write(¤t->mm->mmap_sem); > + pret = get_user_pages_longterm(ubuf, nr_pages, FOLL_WRITE, > + pages, NULL); > + up_write(¤t->mm->mmap_sem); Thought so. This has the same problem as RDMA w.r.t. using file-backed mappings for the user buffer. It is not synchronised against truncate, hole punches, async page writeback cleaning the page, etc, and so can lead to data corruption and/or kernel panics. It also can't be used with DAX because the above problems are actually a user-after-free of storage space, not just a dangling page reference that can be cleaned up after the gup pin is dropped. Perhaps, at least until we solve the GUP problems w.r.t. file backed pages and/or add and require file layout leases for these reference, we should error out if the user buffer pages are file-backed mappings? Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx