req->inflight_entry is used to track requests with ->files, and the only iopoll'ed requests (i.e. read/write) don't have it. Use req->inflight_entry for iopoll path, btw aliasing it in union with a more proper name for clarity. Signed-off-by: Pavel Begunkov <asml.silence@xxxxxxxxx> --- fs/io_uring.c | 31 +++++++++++++++++++------------ 1 file changed, 19 insertions(+), 12 deletions(-) diff --git a/fs/io_uring.c b/fs/io_uring.c index 669a131c22ec..bb92cc736afe 100644 --- a/fs/io_uring.c +++ b/fs/io_uring.c @@ -651,7 +651,14 @@ struct io_kiocb { struct list_head link_list; - struct list_head inflight_entry; + /* + * @inflight_entry is for reqs with ->files (see io_op_def::file_table) + * @iopoll_list is for read/write requests + */ + union { + struct list_head inflight_entry; + struct list_head iopoll_list; + }; struct percpu_ref *fixed_file_refs; @@ -1937,8 +1944,8 @@ static void io_iopoll_queue(struct list_head *again) struct io_kiocb *req; do { - req = list_first_entry(again, struct io_kiocb, list); - list_del(&req->list); + req = list_first_entry(again, struct io_kiocb, iopoll_list); + list_del(&req->iopoll_list); if (!io_rw_reissue(req, -EAGAIN)) io_complete_rw_common(&req->rw.kiocb, -EAGAIN, NULL); } while (!list_empty(again)); @@ -1961,13 +1968,13 @@ static void io_iopoll_complete(struct io_ring_ctx *ctx, unsigned int *nr_events, while (!list_empty(done)) { int cflags = 0; - req = list_first_entry(done, struct io_kiocb, list); + req = list_first_entry(done, struct io_kiocb, iopoll_list); if (READ_ONCE(req->result) == -EAGAIN) { req->iopoll_completed = 0; - list_move_tail(&req->list, &again); + list_move_tail(&req->iopoll_list, &again); continue; } - list_del(&req->list); + list_del(&req->iopoll_list); if (req->flags & REQ_F_BUFFER_SELECTED) cflags = io_put_kbuf(req); @@ -2003,7 +2010,7 @@ static int io_do_iopoll(struct io_ring_ctx *ctx, unsigned int *nr_events, spin = !ctx->poll_multi_file && *nr_events < min; ret = 0; - list_for_each_entry_safe(req, tmp, &ctx->iopoll_list, list) { + list_for_each_entry_safe(req, tmp, &ctx->iopoll_list, iopoll_list) { struct kiocb *kiocb = &req->rw.kiocb; /* @@ -2012,7 +2019,7 @@ static int io_do_iopoll(struct io_ring_ctx *ctx, unsigned int *nr_events, * and complete those lists first, if we have entries there. */ if (READ_ONCE(req->iopoll_completed)) { - list_move_tail(&req->list, &done); + list_move_tail(&req->iopoll_list, &done); continue; } if (!list_empty(&done)) @@ -2024,7 +2031,7 @@ static int io_do_iopoll(struct io_ring_ctx *ctx, unsigned int *nr_events, /* iopoll may have completed current req */ if (READ_ONCE(req->iopoll_completed)) - list_move_tail(&req->list, &done); + list_move_tail(&req->iopoll_list, &done); if (ret && spin) spin = false; @@ -2291,7 +2298,7 @@ static void io_iopoll_req_issued(struct io_kiocb *req) struct io_kiocb *list_req; list_req = list_first_entry(&ctx->iopoll_list, struct io_kiocb, - list); + iopoll_list); if (list_req->file != req->file) ctx->poll_multi_file = true; } @@ -2301,9 +2308,9 @@ static void io_iopoll_req_issued(struct io_kiocb *req) * it to the front so we find it first. */ if (READ_ONCE(req->iopoll_completed)) - list_add(&req->list, &ctx->iopoll_list); + list_add(&req->iopoll_list, &ctx->iopoll_list); else - list_add_tail(&req->list, &ctx->iopoll_list); + list_add_tail(&req->iopoll_list, &ctx->iopoll_list); if ((ctx->flags & IORING_SETUP_SQPOLL) && wq_has_sleeper(&ctx->sqo_wait)) -- 2.24.0