Re: [PATCH 2/6] io_uring: add a priority tw list for irq completion work

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



在 2021/11/18 上午7:03, Pavel Begunkov 写道:
On 10/29/21 13:22, Hao Xu wrote:
Now we have a lot of task_work users, some are just to complete a req
and generate a cqe. Let's put the work to a new tw list which has a
higher priority, so that it can be handled quickly and thus to reduce
avg req latency and users can issue next round of sqes earlier.
An explanatory case:

origin timeline:
     submit_sqe-->irq-->add completion task_work
     -->run heavy work0~n-->run completion task_work
now timeline:
     submit_sqe-->irq-->add completion task_work
     -->run completion task_work-->run heavy work0~n

Limitation: this optimization is only for those that submission and
reaping process are in different threads. Otherwise anyhow we have to
submit new sqes after returning to userspace, then the order of TWs
doesn't matter.

Tested this patch(and the following ones) by manually replace
__io_queue_sqe() in io_queue_sqe() by io_req_task_queue() to construct
'heavy' task works. Then test with fio:

ioengine=io_uring
sqpoll=1
thread=1
bs=4k
direct=1
rw=randread
time_based=1
runtime=600
randrepeat=0
group_reporting=1
filename=/dev/nvme0n1

Tried various iodepth.
The peak IOPS for this patch is 710K, while the old one is 665K.
For avg latency, difference shows when iodepth grow:
depth and avg latency(usec):
    depth      new          old
     1        7.05         7.10
     2        8.47         8.60
     4        10.42        10.42
     8        13.78        13.22
     16       27.41        24.33
     32       49.40        53.08
     64       102.53       103.36
     128      196.98       205.61
     256      372.99       414.88
          512      747.23       791.30
          1024     1472.59      1538.72
          2048     3153.49      3329.01
          4096     6387.86      6682.54
          8192     12150.25     12774.14
          16384    23085.58     26044.71

Signed-off-by: Hao Xu <haoxu@xxxxxxxxxxxxxxxxx>
---
  fs/io_uring.c | 38 +++++++++++++++++++++++++-------------
  1 file changed, 25 insertions(+), 13 deletions(-)

diff --git a/fs/io_uring.c b/fs/io_uring.c
index 17cb0e1b88f0..981794ee3f3f 100644
--- a/fs/io_uring.c
+++ b/fs/io_uring.c
@@ -467,6 +467,7 @@ struct io_uring_task {
      spinlock_t        task_lock;
      struct io_wq_work_list    task_list;
+    struct io_wq_work_list    prior_task_list;
      struct callback_head    task_work;
      bool            task_running;
  };
@@ -2148,13 +2149,17 @@ static void tctx_task_work(struct callback_head *cb)
      while (1) {
          struct io_wq_work_node *node;
+        struct io_wq_work_list *merged_list;
-        if (!tctx->task_list.first && locked)
+        if (!tctx->prior_task_list.first &&
+            !tctx->task_list.first && locked)
              io_submit_flush_completions(ctx);
          spin_lock_irq(&tctx->task_lock);
-        node = tctx->task_list.first;
+        merged_list = wq_list_merge(&tctx->prior_task_list, &tctx->task_list);
+        node = merged_list->first;
          INIT_WQ_LIST(&tctx->task_list);
+        INIT_WQ_LIST(&tctx->prior_task_list);
          if (!node)
              tctx->task_running = false;
          spin_unlock_irq(&tctx->task_lock);
@@ -2183,19 +2188,23 @@ static void tctx_task_work(struct callback_head *cb)
      ctx_flush_and_put(ctx, &locked);
  }
-static void io_req_task_work_add(struct io_kiocb *req)
+static void io_req_task_work_add(struct io_kiocb *req, bool priority)
  {
      struct task_struct *tsk = req->task;
      struct io_uring_task *tctx = tsk->io_uring;
      enum task_work_notify_mode notify;
      struct io_wq_work_node *node;
+    struct io_wq_work_list *merged_list;
      unsigned long flags;
      bool running;
      WARN_ON_ONCE(!tctx);
      spin_lock_irqsave(&tctx->task_lock, flags);
-    wq_list_add_tail(&req->io_task_work.node, &tctx->task_list);
+    if (priority)
+        wq_list_add_tail(&req->io_task_work.node, &tctx->prior_task_list);
+    else
+        wq_list_add_tail(&req->io_task_work.node, &tctx->task_list);
      running = tctx->task_running;
      if (!running)
          tctx->task_running = true;
@@ -2220,8 +2229,10 @@ static void io_req_task_work_add(struct io_kiocb *req)
      spin_lock_irqsave(&tctx->task_lock, flags);
      tctx->task_running = false;
-    node = tctx->task_list.first;
+    merged_list = wq_list_merge(&tctx->prior_task_list, &tctx->task_list);
+    node = merged_list->first;
      INIT_WQ_LIST(&tctx->task_list);
+    INIT_WQ_LIST(&tctx->prior_task_list);
      spin_unlock_irqrestore(&tctx->task_lock, flags);
      while (node) {
@@ -2258,19 +2269,19 @@ static void io_req_task_queue_fail(struct io_kiocb *req, int ret)
  {
      req->result = ret;
      req->io_task_work.func = io_req_task_cancel;
-    io_req_task_work_add(req);
+    io_req_task_work_add(req, false);
  }
  static void io_req_task_queue(struct io_kiocb *req)
  {
      req->io_task_work.func = io_req_task_submit;
-    io_req_task_work_add(req);
+    io_req_task_work_add(req, false);
  }
  static void io_req_task_queue_reissue(struct io_kiocb *req)
  {
      req->io_task_work.func = io_queue_async_work;
-    io_req_task_work_add(req);
+    io_req_task_work_add(req, false);
  }
  static inline void io_queue_next(struct io_kiocb *req)
@@ -2375,7 +2386,7 @@ static inline void io_put_req_deferred(struct io_kiocb *req)
  {
      if (req_ref_put_and_test(req)) {
          req->io_task_work.func = io_free_req_work;
-        io_req_task_work_add(req);
+        io_req_task_work_add(req, false);
      }
  }
@@ -2678,7 +2689,7 @@ static void io_complete_rw(struct kiocb *kiocb, long res, long res2)
          return;
      req->result = res;
      req->io_task_work.func = io_req_task_complete;
-    io_req_task_work_add(req);
+    io_req_task_work_add(req, !!(req->ctx->flags & IORING_SETUP_SQPOLL));

I'm not sure this special case makes sense. I remembered you mentioned
that you measured it, but what's the reason? Can it be related to my
comments on 6/6?
The discussion is here:

https://lore.kernel.org/io-uring/7a528ce1-a44e-3ee7-095c-1a92528ec441@xxxxxxxxxxxxxxxxx/

https://lore.kernel.org/io-uring/4bc45226-8b27-500a-58e7-36da2eb5f92e@xxxxxxxxxxxxxxxxx/



[Index of Archives]     [Linux Samsung SoC]     [Linux Rockchip SoC]     [Linux Actions SoC]     [Linux for Synopsys ARC Processors]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]


  Powered by Linux