Re: [PATCH 2/3] io_uring: fix failed linkchain code logic

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



在 2021/8/18 下午6:20, Pavel Begunkov 写道:
On 8/18/21 8:43 AM, Hao Xu wrote:
Given a linkchain like this:
req0(link_flag)-->req1(link_flag)-->...-->reqn(no link_flag)

There is a problem:
  - if some intermediate linked req like req1 's submittion fails, reqs
    after it won't be cancelled.

    - sqpoll disabled: maybe it's ok since users can get the error info
      of req1 and stop submitting the following sqes.

    - sqpoll enabled: definitely a problem, the following sqes will be
      submitted in the next round.

The solution is to refactor the code logic to:
  - link a linked req to the chain first, no matter its submittion fails
    or not.
  - if a linked req's submittion fails, just mark head as
    failed. leverage req->result to indicate whether the req is a failed
    one or cancelled one.
  - submit or fail the whole chain

Signed-off-by: Hao Xu <haoxu@xxxxxxxxxxxxxxxxx>
---
  fs/io_uring.c | 86 ++++++++++++++++++++++++++++++++++-----------------
  1 file changed, 58 insertions(+), 28 deletions(-)

diff --git a/fs/io_uring.c b/fs/io_uring.c
index c0b841506869..383668e07417 100644
--- a/fs/io_uring.c
+++ b/fs/io_uring.c
@@ -1920,11 +1920,13 @@ static void io_fail_links(struct io_kiocb *req)
req->link = NULL;
  	while (link) {
+		int res = link->result ? link->result : -ECANCELED;

btw, we don't properly initialise req->result, and don't want to.
I see, req->result is cleaned to 0 in io_preinit_req() but not the same
when move it back to the free list.
Perhaps, can be more like

res = -ECANCELLED;
if (req->flags & FAIL)
	res = req->result;
Agree.



+
  		nxt = link->link;
  		link->link = NULL;
trace_io_uring_fail_link(req, link);
-		io_cqring_fill_event(link->ctx, link->user_data, -ECANCELED, 0);
+		io_cqring_fill_event(link->ctx, link->user_data, res, 0);
  		io_put_req_deferred(link);
  		link = nxt;
  	}
@@ -5698,7 +5700,7 @@ static int io_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe,
  	if (is_timeout_link) {
  		struct io_submit_link *link = &req->ctx->submit_state.link;
- if (!link->head)
+		if (!link->head || link->head == req)
  			return -EINVAL;
  		if (link->last->opcode == IORING_OP_LINK_TIMEOUT)
  			return -EINVAL;
@@ -6622,17 +6624,38 @@ static int io_submit_sqe(struct io_ring_ctx *ctx, struct io_kiocb *req,
  	__must_hold(&ctx->uring_lock)
  {
  	struct io_submit_link *link = &ctx->submit_state.link;
+	bool is_link = sqe->flags & (IOSQE_IO_LINK | IOSQE_IO_HARDLINK);
+	struct io_kiocb *head;
  	int ret;
+ /*
+	 * we don't update link->last until we've done io_req_prep()
+	 * since linked timeout uses old link->last
+	 */
+	if (link->head)
+		link->last->link = req;
+	else if (is_link)
+		link->head = req;
+	head = link->head;

It's a horrorsome amount of overhead. How about to set the fail flag
if failed early and actually fail on io_queue_sqe(), as below. It's
not tested and a couple more bits added, but hopefully gives the idea.
I get the idea, it's truely with less change. But why do you think the
above code bring in more overhead, since anyway we have to link the req
to the linkchain. I tested it with fio-direct-4k-read-with/without-sqpoll, didn't see performance regression.



diff --git a/fs/io_uring.c b/fs/io_uring.c
index ba087f395507..3fd0730655d0 100644
--- a/fs/io_uring.c
+++ b/fs/io_uring.c
@@ -6530,8 +6530,10 @@ static inline void io_queue_sqe(struct io_kiocb *req)
  	if (unlikely(req->ctx->drain_active) && io_drain_req(req))
  		return;
- if (likely(!(req->flags & REQ_F_FORCE_ASYNC))) {
+	if (likely(!(req->flags & (REQ_F_FORCE_ASYNC|REQ_F_FAIL)))) {
  		__io_queue_sqe(req);
+	} else if (req->flags & REQ_F_FAIL) {
+		io_req_complete_failed(req, ret);
  	} else {
  		int ret = io_req_prep_async(req);
@@ -6640,19 +6642,17 @@ static int io_submit_sqe(struct io_ring_ctx *ctx, struct io_kiocb *req,
  	ret = io_init_req(ctx, req, sqe);
  	if (unlikely(ret)) {
  fail_req:
-		if (link->head) {
-			/* fail even hard links since we don't submit */
+		/* fail even hard links since we don't submit */
+		if (link->head)
  			req_set_fail(link->head);
-			io_req_complete_failed(link->head, -ECANCELED);
-			link->head = NULL;
-		}
-		io_req_complete_failed(req, ret);
-		return ret;
+		req_set_fail(req);
+		req->result = ret;
+	} else {
+		ret = io_req_prep(req, sqe);
+		if (unlikely(ret))
+			goto fail_req;
  	}
- ret = io_req_prep(req, sqe);
-	if (unlikely(ret))
-		goto fail_req;
/* don't need @sqe from now on */
  	trace_io_uring_submit_sqe(ctx, req, req->opcode, req->user_data,
@@ -6670,8 +6670,10 @@ static int io_submit_sqe(struct io_ring_ctx *ctx, struct io_kiocb *req,
  		struct io_kiocb *head = link->head;
maybe better to add an if(head & FAIL) here, since we don't need to
prep_async if we know it will be cancelled.
  		ret = io_req_prep_async(req);
-		if (unlikely(ret))
-			goto fail_req;
+		if (unlikely(ret)) {
+			req->result = ret;
+			req_set_fail(link->head);
+		}
  		trace_io_uring_link(ctx, req, head);
  		link->last->link = req;
  		link->last = req;





[Index of Archives]     [Linux Samsung SoC]     [Linux Rockchip SoC]     [Linux Actions SoC]     [Linux for Synopsys ARC Processors]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]


  Powered by Linux