[PATCH 2/3] io_uring: fix failed linkchain code logic

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Given a linkchain like this:
req0(link_flag)-->req1(link_flag)-->...-->reqn(no link_flag)

There is a problem:
 - if some intermediate linked req like req1 's submittion fails, reqs
   after it won't be cancelled.

   - sqpoll disabled: maybe it's ok since users can get the error info
     of req1 and stop submitting the following sqes.

   - sqpoll enabled: definitely a problem, the following sqes will be
     submitted in the next round.

The solution is to refactor the code logic to:
 - link a linked req to the chain first, no matter its submittion fails
   or not.
 - if a linked req's submittion fails, just mark head as
   failed. leverage req->result to indicate whether the req is a failed
   one or cancelled one.
 - submit or fail the whole chain

Signed-off-by: Hao Xu <haoxu@xxxxxxxxxxxxxxxxx>
---
 fs/io_uring.c | 86 ++++++++++++++++++++++++++++++++++-----------------
 1 file changed, 58 insertions(+), 28 deletions(-)

diff --git a/fs/io_uring.c b/fs/io_uring.c
index c0b841506869..383668e07417 100644
--- a/fs/io_uring.c
+++ b/fs/io_uring.c
@@ -1920,11 +1920,13 @@ static void io_fail_links(struct io_kiocb *req)
 
 	req->link = NULL;
 	while (link) {
+		int res = link->result ? link->result : -ECANCELED;
+
 		nxt = link->link;
 		link->link = NULL;
 
 		trace_io_uring_fail_link(req, link);
-		io_cqring_fill_event(link->ctx, link->user_data, -ECANCELED, 0);
+		io_cqring_fill_event(link->ctx, link->user_data, res, 0);
 		io_put_req_deferred(link);
 		link = nxt;
 	}
@@ -5698,7 +5700,7 @@ static int io_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe,
 	if (is_timeout_link) {
 		struct io_submit_link *link = &req->ctx->submit_state.link;
 
-		if (!link->head)
+		if (!link->head || link->head == req)
 			return -EINVAL;
 		if (link->last->opcode == IORING_OP_LINK_TIMEOUT)
 			return -EINVAL;
@@ -6622,17 +6624,38 @@ static int io_submit_sqe(struct io_ring_ctx *ctx, struct io_kiocb *req,
 	__must_hold(&ctx->uring_lock)
 {
 	struct io_submit_link *link = &ctx->submit_state.link;
+	bool is_link = sqe->flags & (IOSQE_IO_LINK | IOSQE_IO_HARDLINK);
+	struct io_kiocb *head;
 	int ret;
 
+	/*
+	 * we don't update link->last until we've done io_req_prep()
+	 * since linked timeout uses old link->last
+	 */
+	if (link->head)
+		link->last->link = req;
+	else if (is_link)
+		link->head = req;
+	head = link->head;
+
 	ret = io_init_req(ctx, req, sqe);
 	if (unlikely(ret)) {
 fail_req:
-		if (link->head) {
-			/* fail even hard links since we don't submit */
-			io_req_complete_failed(link->head, -ECANCELED);
-			link->head = NULL;
+		req->result = ret;
+		if (head) {
+			link->last = req;
+			if (is_link) {
+				req_set_fail(head);
+			} else {
+				int res = head->result ? head->result : -ECANCELED;
+
+				link->head = NULL;
+				/* fail even hard links since we don't submit */
+				io_req_complete_failed(head, res);
+			}
+		} else {
+			io_req_complete_failed(req, ret);
 		}
-		io_req_complete_failed(req, ret);
 		return ret;
 	}
 
@@ -6652,28 +6675,26 @@ static int io_submit_sqe(struct io_ring_ctx *ctx, struct io_kiocb *req,
 	 * submitted sync once the chain is complete. If none of those
 	 * conditions are true (normal request), then just queue it.
 	 */
-	if (link->head) {
-		struct io_kiocb *head = link->head;
-
-		ret = io_req_prep_async(req);
-		if (unlikely(ret))
-			goto fail_req;
+	if (head) {
+		if (req != head) {
+			ret = io_req_prep_async(req);
+			if (unlikely(ret))
+				goto fail_req;
+		}
 		trace_io_uring_link(ctx, req, head);
-		link->last->link = req;
-		link->last = req;
 
+		link->last = req;
 		/* last request of a link, enqueue the link */
-		if (!(req->flags & (REQ_F_LINK | REQ_F_HARDLINK))) {
-			link->head = NULL;
-			io_queue_sqe(head);
+		if (!is_link) {
+			if (head->flags & REQ_F_FAIL) {
+				goto fail_req;
+			} else {
+				link->head = NULL;
+				io_queue_sqe(head);
+			}
 		}
 	} else {
-		if (req->flags & (REQ_F_LINK | REQ_F_HARDLINK)) {
-			link->head = req;
-			link->last = req;
-		} else {
-			io_queue_sqe(req);
-		}
+		io_queue_sqe(req);
 	}
 
 	return 0;
@@ -6685,8 +6706,17 @@ static int io_submit_sqe(struct io_ring_ctx *ctx, struct io_kiocb *req,
 static void io_submit_state_end(struct io_submit_state *state,
 				struct io_ring_ctx *ctx)
 {
-	if (state->link.head)
-		io_queue_sqe(state->link.head);
+	struct io_kiocb *head = state->link.head;
+
+	if (head) {
+		if (head->flags & REQ_F_FAIL) {
+			int res = head->result ? head->result : -ECANCELED;
+
+			io_req_complete_failed(head, res);
+		} else {
+			io_queue_sqe(head);
+		}
+	}
 	if (state->compl_nr)
 		io_submit_flush_completions(ctx);
 	if (state->plug_started)
@@ -6701,8 +6731,8 @@ static void io_submit_state_start(struct io_submit_state *state,
 {
 	state->plug_started = false;
 	state->ios_left = max_ios;
-	/* set only head, no need to init link_last in advance */
 	state->link.head = NULL;
+	state->link.last = NULL;
 }
 
 static void io_commit_sqring(struct io_ring_ctx *ctx)
@@ -6788,7 +6818,7 @@ static int io_submit_sqes(struct io_ring_ctx *ctx, unsigned int nr)
 		}
 		/* will complete beyond this point, count as submitted */
 		submitted++;
-		if (io_submit_sqe(ctx, req, sqe))
+		if (io_submit_sqe(ctx, req, sqe) && !(ctx->flags & IORING_SETUP_SQPOLL))
 			break;
 	}
 
-- 
2.24.4




[Index of Archives]     [Linux Samsung SoC]     [Linux Rockchip SoC]     [Linux Actions SoC]     [Linux for Synopsys ARC Processors]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]


  Powered by Linux