Re: KCSAN: data-race in __io_cqring_fill_event / io_uring_poll

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, Aug 10, 2020 at 11:58 AM syzbot
<syzbot+3020eb77f81ef0772fbe@xxxxxxxxxxxxxxxxxxxxxxxxx> wrote:
>
> Hello,
>
> syzbot found the following issue on:
>
> HEAD commit:    86cfccb6 Merge tag 'dlm-5.9' of git://git.kernel.org/pub/s..
> git tree:       upstream
> console output: https://syzkaller.appspot.com/x/log.txt?x=171cf11a900000
> kernel config:  https://syzkaller.appspot.com/x/.config?x=3757fa75ecfde8e0
> dashboard link: https://syzkaller.appspot.com/bug?extid=3020eb77f81ef0772fbe
> compiler:       clang version 11.0.0 (https://github.com/llvm/llvm-project.git ca2dcbd030eadbf0aa9b660efe864ff08af6e18b)
>
> Unfortunately, I don't have any reproducer for this issue yet.
>
> IMPORTANT: if you fix the issue, please add the following tag to the commit:
> Reported-by: syzbot+3020eb77f81ef0772fbe@xxxxxxxxxxxxxxxxxxxxxxxxx
>
> ==================================================================
> BUG: KCSAN: data-race in __io_cqring_fill_event / io_uring_poll
>
> write to 0xffff8880a04325c0 of 4 bytes by task 12088 on cpu 1:
>  io_get_cqring fs/io_uring.c:1282 [inline]
>  __io_cqring_fill_event+0x116/0x430 fs/io_uring.c:1386
>  io_cqring_add_event fs/io_uring.c:1420 [inline]
>  __io_req_complete+0xdb/0x1b0 fs/io_uring.c:1458
>  io_complete_rw_common fs/io_uring.c:2208 [inline]
>  __io_complete_rw+0x2c9/0x2e0 fs/io_uring.c:2289
>  kiocb_done fs/io_uring.c:2533 [inline]
>  io_write fs/io_uring.c:3199 [inline]
>  io_issue_sqe+0x4fb1/0x7140 fs/io_uring.c:5530
>  io_wq_submit_work+0x23e/0x340 fs/io_uring.c:5775
>  io_worker_handle_work+0xa69/0xcf0 fs/io-wq.c:527
>  io_wqe_worker+0x1f2/0x860 fs/io-wq.c:569
>  kthread+0x20d/0x230 kernel/kthread.c:292
>  ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:294
>
> read to 0xffff8880a04325c0 of 4 bytes by task 12086 on cpu 0:
>  io_cqring_events fs/io_uring.c:1937 [inline]
>  io_uring_poll+0x105/0x140 fs/io_uring.c:7751
>  vfs_poll include/linux/poll.h:90 [inline]
>  __io_arm_poll_handler+0x176/0x3f0 fs/io_uring.c:4735
>  io_arm_poll_handler+0x293/0x5c0 fs/io_uring.c:4792
>  __io_queue_sqe+0x413/0x760 fs/io_uring.c:5988
>  io_queue_sqe+0x81/0x2b0 fs/io_uring.c:6060
>  io_submit_sqe+0x333/0x560 fs/io_uring.c:6130
>  io_submit_sqes+0x8c6/0xfc0 fs/io_uring.c:6327
>  __do_sys_io_uring_enter fs/io_uring.c:8036 [inline]
>  __se_sys_io_uring_enter+0x1c2/0x720 fs/io_uring.c:7995
>  __x64_sys_io_uring_enter+0x74/0x80 fs/io_uring.c:7995
>  do_syscall_64+0x39/0x80 arch/x86/entry/common.c:46
>  entry_SYSCALL_64_after_hwframe+0x44/0xa9
>
> Reported by Kernel Concurrency Sanitizer on:
> CPU: 0 PID: 12086 Comm: syz-executor.5 Not tainted 5.8.0-syzkaller #0
> Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 01/01/2011
> ==================================================================
>
> ---
> This report is generated by a bot. It may contain errors.
> See https://goo.gl/tpsmEJ for more information about syzbot.
> syzbot engineers can be reached at syzkaller@xxxxxxxxxxxxxxxx.
>
> syzbot will keep track of this issue. See:
> https://goo.gl/tpsmEJ#status for how to communicate with syzbot.

Hi Jens,

I wonder if this read of cached_cq_tail needs READ_ONCE as all
concurrent access seems to be marked in io_uring.c overall.

I also have a concern for a potential bad scenario around
cached_cq_tail. I've seen it in other queues. I don't see that it can
happen in io_uring, but from a very brief look at the code io_uring
may be susceptible, so I thought it's better to ask.
The failure is around ordering of reads of head/tail when deciding if
the queue is empty or not.
Consider: initially queue is empty head==tail==0.
One thread adds an element and sets tail=1 and wakes poller.
Now, before the poller wakes, several threads push/pop from the queue
incrementing both head and tail, maybe also waking poller. But at no
point in time the queue is empty.
Now poller wakes and reads, say, tail=1 (at the time when head=0 and
tail=1 so the queue is not empty) and then reads head=1 (at the time
when head=1 and tail=3 so the queue is not empty again). But now the
poller looks at head==tail==1 and decides that the queue is empty.
This may lead to a missed epoll event and a deadlock.
This can be resolved by acquiring a consistent snapshot of head/tail in poller.
Is there anything that prevents such a scenario in io_uring? Or it's
not an issue for some other reason?



[Index of Archives]     [Linux Ext4 Filesystem]     [Union Filesystem]     [Filesystem Testing]     [Ceph Users]     [Ecryptfs]     [AutoFS]     [Kernel Newbies]     [Share Photos]     [Security]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux Cachefs]     [Reiser Filesystem]     [Linux RAID]     [Samba]     [Device Mapper]     [CEPH Development]

  Powered by Linux