Race between io_wqe_worker() and io_wqe_wake_worker()

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Jens,

I encountered an issue, which appears to be a race between io_wqe_worker() and io_wqe_wake_worker(). I am not sure how to address this issue and whether I am missing something, since this seems to occur in a common scenario. Your feedback (or fix ;-)) would be appreciated.

I run on 5.13 a workload that issues multiple async read operations that should run concurrently. Some read operations can not complete for unbounded time (e.g., read from a pipe that is never written to). The problem is that occasionally another read operation that should complete gets stuck. My understanding, based on debugging and the code is that the following race (or similar) occurs:


  cpu0					cpu1
  ----					----
					io_wqe_worker()
					 schedule_timeout()
					 // timed out
  io_wqe_enqueue()
   io_wqe_wake_worker()
    // work_flags & IO_WQ_WORK_CONCURRENT
    io_wqe_activate_free_worker()
					 io_worker_exit()


Basically, io_wqe_wake_worker() can find a worker, but this worker is about to exit and is not going to process further work. Once the worker exits, the concurrency level decreases and async work might be blocked by another work. I had a look at 5.14, but did not see anything that might address this issue.

Am I missing something?

If not, all my ideas for a solution are either complicated (track required concurrency-level) or relaxed (span another worker on io_worker_exit if work_list of unbounded work is not empty).

As said, feedback would be appreciated.

Thanks,
Nadav






[Index of Archives]     [Linux Samsung SoC]     [Linux Rockchip SoC]     [Linux Actions SoC]     [Linux for Synopsys ARC Processors]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]


  Powered by Linux