Re: [bug-report] task info hung problem in fb_deferred_io_work()

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Patrik,

On 18/04/24 18:44, Patrik Jakobsson wrote:
On Thu, Apr 18, 2024 at 2:40 PM Harshit Mogalapalli
<harshit.m.mogalapalli@xxxxxxxxxx> wrote:

Hi,

While fuzzing 5.15.y kernel with Syzkaller, we noticed a INFO: task hung
bug in fb_deferred_io_work()


This is in 5.15.149 tag, and this is introduced by a set of commits:

2655757a3f10 fbdev: flush deferred IO before closing
15492bab7393 fbdev: Fix incorrect page mapping clearance at
fb_deferred_io_release()
87b9802ca824 fbdev: Fix invalid page access after closing deferred I/O
devices
9a95fc04261f fbdev: Rename pagelist to pagereflist for deferred I/O
186b89659c4c fbdev: Track deferred-I/O pages in pageref struct
e79b2b2aadef fbdev: defio: fix the pagelist corruption
0616b00a31d6 fbdev: Don't sort deferred-I/O pages by default
5d3aff76a316 fbdev/defio: Early-out if page is already enlisted

I then tried if it could be reproducible on upstream kernel as well, and
it did. below is the call trace with latest 6.9-rc4 based kernel. I have
enabled the LOCKDEP to get locking related information.

   INFO: task kworker/0:3:1077 blocked for more than 122 seconds.
         Not tainted 6.9.0-rc4+ #54
   "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
   task:kworker/0:3     state:D stack:0     pid:1077  tgid:1077  ppid:2
     flags:0x00004000
   Workqueue: events fb_deferred_io_work
   Call Trace:
    <TASK>
   __schedule (kernel/sched/core.c:5409 kernel/sched/core.c:6746)
   ? __pfx___schedule (kernel/sched/core.c:6617)
   ? schedule (kernel/sched/core.c:6804 kernel/sched/core.c:6837)
   ? __pfx___lock_release (kernel/locking/lockdep.c:5406)
   ? schedule (kernel/sched/core.c:6804 kernel/sched/core.c:6837)
   schedule (./arch/x86/include/asm/preempt.h:84 kernel/sched/core.c:6824
kernel/sched/core.c:6838)
   io_schedule (kernel/sched/core.c:9019 kernel/sched/core.c:9045)
   folio_wait_bit_common (mm/filemap.c:1272)
   ? __pfx_folio_wait_bit_common (mm/filemap.c:1206)
   ? __pfx___mutex_lock (kernel/locking/mutex.c:751)
   ? lock_acquire.part.0 (kernel/locking/lockdep.c:467
kernel/locking/lockdep.c:5756)
   ? __pfx_wake_page_function (mm/filemap.c:1091)
   ? __pfx_lock_acquire.part.0 (kernel/locking/lockdep.c:5719)
   fb_deferred_io_work (./include/linux/pagemap.h:1069
drivers/video/fbdev/core/fb_defio.c:248)
   ? lock_acquire (./include/trace/events/lock.h:24
kernel/locking/lockdep.c:5725)
   process_one_work (kernel/workqueue.c:3254)
   ? __pfx_lock_acquire.part.0 (kernel/locking/lockdep.c:5719)
   ? __pfx_process_one_work (kernel/workqueue.c:3156)
   ? assign_work (kernel/workqueue.c:1210)
   worker_thread (kernel/workqueue.c:3329 kernel/workqueue.c:3416)
   ? __kthread_parkme (kernel/kthread.c:293)
   ? __pfx_worker_thread (kernel/workqueue.c:3362)
   kthread (kernel/kthread.c:388)
   ? __pfx_kthread (kernel/kthread.c:341)
   ret_from_fork (arch/x86/kernel/process.c:147)
   ? __pfx_kthread (kernel/kthread.c:341)
   ret_from_fork_asm (arch/x86/entry/entry_64.S:257)
    </TASK>
   INFO: task r:2664 blocked for more than 123 seconds.
         Not tainted 6.9.0-rc4+ #54
   "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
   task:r               state:D stack:0     pid:2664  tgid:2664
ppid:2618   flags:0x00000002
   Call Trace:
    <TASK>
   __schedule (kernel/sched/core.c:5409 kernel/sched/core.c:6746)
   ? __pfx___schedule (kernel/sched/core.c:6617)
   ? schedule (kernel/sched/core.c:6804 kernel/sched/core.c:6837)
   ? __pfx___lock_release (kernel/locking/lockdep.c:5406)
   ? schedule (kernel/sched/core.c:6804 kernel/sched/core.c:6837)
   schedule (./arch/x86/include/asm/preempt.h:84 kernel/sched/core.c:6824
kernel/sched/core.c:6838)
   schedule_timeout (kernel/time/timer.c:2559)
   ? __pfx_schedule_timeout (kernel/time/timer.c:2544)
   ? mark_lock.part.0 (./arch/x86/include/asm/bitops.h:227
./arch/x86/include/asm/bitops.h:239
./include/asm-generic/bitops/instrumented-non-atomic.h:142
kernel/locking/lockdep.c:228 kernel/locking/lockdep.c:4656)
   ? mark_held_locks (kernel/locking/lockdep.c:4274)
   ? lockdep_hardirqs_on_prepare.part.0 (kernel/locking/lockdep.c:4300
kernel/locking/lockdep.c:4359)
   __wait_for_common (kernel/sched/completion.c:95
kernel/sched/completion.c:116)
   ? __pfx_schedule_timeout (kernel/time/timer.c:2544)
   ? __pfx___wait_for_common (kernel/sched/completion.c:110)
   ? start_flush_work (kernel/workqueue.c:4192)
   __flush_work (kernel/workqueue.c:4207)
   ? __pfx___flush_work (kernel/workqueue.c:4195)
   ? __pfx_wq_barrier_func (kernel/workqueue.c:3742)
   ? flush_delayed_work (kernel/workqueue.c:4249)
   ? __pfx_fb_release (drivers/video/fbdev/core/fb_chrdev.c:375)
   fb_deferred_io_lastclose (drivers/video/fbdev/core/fb_defio.c:321)
   ? __pfx_fb_release (drivers/video/fbdev/core/fb_chrdev.c:375)
   fb_deferred_io_release (drivers/video/fbdev/core/fb_defio.c:333)
   fb_release (drivers/video/fbdev/core/fb_chrdev.c:383)
   ? __pfx_fb_release (drivers/video/fbdev/core/fb_chrdev.c:375)
   __fput (fs/file_table.c:422)
   task_work_run (kernel/task_work.c:180 (discriminator 1))
   ? __pfx_task_work_run (kernel/task_work.c:148)
   ? do_raw_spin_unlock (./arch/x86/include/asm/paravirt.h:589
./arch/x86/include/asm/qspinlock.h:57 kernel/locking/spinlock_debug.c:142)
   do_exit (kernel/exit.c:879)
   ? do_raw_spin_lock (./arch/x86/include/asm/atomic.h:115
./include/linux/atomic/atomic-arch-fallback.h:2170
./include/linux/atomic/atomic-instrumented.h:1302
./include/asm-generic/qspinlock.h:111 kernel/locking/spinlock_debug.c:116)
   ? __pfx_do_exit (kernel/exit.c:819)
   ? __pfx_do_raw_spin_lock (kernel/locking/spinlock_debug.c:114)
   do_group_exit (kernel/exit.c:1008)
   __x64_sys_exit_group (kernel/exit.c:1036)
   do_syscall_64 (arch/x86/entry/common.c:52 arch/x86/entry/common.c:83)
   entry_SYSCALL_64_after_hwframe (arch/x86/entry/entry_64.S:130)
   RIP: 0033:0x7f9cfe71e5f1
   RSP: 002b:00007ffd636aec08 EFLAGS: 00000246 ORIG_RAX: 00000000000000e7
   RAX: ffffffffffffffda RBX: 00007f9cfe7fc510 RCX: 00007f9cfe71e5f1
   RDX: 000000000000003c RSI: 00000000000000e7 RDI: 0000000000000000
   RBP: 0000000000000000 R08: ffffffffffffff80 R09: 0000000000000000
   R10: 0000000000000020 R11: 0000000000000246 R12: 00007f9cfe7fc510
   R13: 0000000000000000 R14: 00007f9cfe7fc9e8 R15: 00007f9cfe7fca00
    </TASK>

   Showing all locks held in the system:
   1 lock held by khungtaskd/50:
   #0: ffffffff950219c0 (rcu_read_lock){....}-{1:2}, at:
debug_show_all_locks (./include/linux/rcupdate.h:329 (discriminator 1)
./include/linux/rcupdate.h:781 (discriminator 1)
kernel/locking/lockdep.c:6614 (discriminator 1))
   3 locks held by kworker/0:3/1077:
   #0: ffff88810005ad48 ((wq_completion)events){+.+.}-{0:0}, at:
process_one_work (kernel/workqueue.c:3229)
   #1: ffff8881080dfd98
((work_completion)(&(&info->deferred_work)->work)){+.+.}-{0:0}, at:
process_one_work (kernel/workqueue.c:3230)
   #2: ffff88811416db38 (&fbdefio->lock){+.+.}-{3:3}, at:
fb_deferred_io_work (drivers/video/fbdev/core/fb_defio.c:246)
   1 lock held by r/2664:
   #0: ffff8881141fe078 (&fb_info->lock){+.+.}-{3:3}, at: fb_release
(drivers/video/fbdev/core/fb_chrdev.c:380)



Syzkaller initially generated a long reproducer, Vegard simplified it to
the one below

C repro:

#define _GNU_SOURCE
#include <sys/mman.h>
#include <fcntl.h>
#include <stdlib.h>
#include <unistd.h>

int main(void)
{
          int fd = open("/dev/fb0", O_RDWR, 0);

          char *ptr = (char *) 0x20000000ul;

          mmap(ptr, 4096, PROT_WRITE, MAP_FIXED | MAP_SHARED, fd, 0xff000);
          *ptr = 0;

          mmap(ptr, 4096, PROT_WRITE, MAP_FIXED | MAP_SHARED, fd, 0);
          *ptr = 0;

          return 0;
}

This needs to be run as root to see the hang in 122 seconds.

238 static void fb_deferred_io_work(struct work_struct *work)
239 {
240         struct fb_info *info = container_of(work, struct fb_info,
deferred_work.work);
241         struct fb_deferred_io_pageref *pageref, *next;
242         struct fb_deferred_io *fbdefio = info->fbdefio;
243
244         /* here we mkclean the pages, then do all deferred IO */
245         mutex_lock(&fbdefio->lock);
246         list_for_each_entry(pageref, &fbdefio->pagereflist, list) {
247                 struct page *cur = pageref->page;
248                 lock_page(cur);
249                 page_mkclean(cur);
250                 unlock_page(cur);
251         }

It is stalling on lock_page() in fb_deferred_io_work()

I will be happy to try any patches and report back.

Hi,


Thanks for the quick response and sharing details.

This sounds similar to the SUSE bug [1]. We fixed it by reverting [2]
in the SUSE kernel. The problem seems to be that flush_delayed_work()
kills the timer and re-queues the work but doesn't guarantee that it
is finished when returning. So when the device is closed, the
fb_deferred_io_work() function can still be queued and tries to access
memory that's been freed.


I tried reverting that commit but saw some other similar problem,

here it is:

[  618.508726] INFO: task r:2682 blocked for more than 122 seconds.
[  618.513662]       Not tainted 6.9.0-rc4+ #55
[ 618.517105] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message. [ 618.523204] task:r state:D stack:0 pid:2682 tgid:2682 ppid:2636 flags:0x00000004
[  618.530325] Call Trace:
[  618.532470]  <TASK>
[  618.534394]  __schedule+0x787/0x1b10
[  618.537506]  ? __pfx___schedule+0x10/0x10
[  618.540784]  ? schedule+0x1e6/0x290
[  618.543698]  ? __lock_release+0x103/0x460
[  618.546982]  ? folio_wait_bit_common+0x22a/0x8b0
[  618.550697]  ? __pfx___lock_release+0x10/0x10
[  618.554268]  ? mark_lock.part.0+0x77/0x880
[  618.557614]  ? schedule+0x1e6/0x290
[  618.560545]  schedule+0xdd/0x290
[  618.563324]  io_schedule+0xc4/0x140
[  618.566251]  folio_wait_bit_common+0x32a/0x8b0
[  618.570062]  ? __pfx_folio_wait_bit_common+0x10/0x10
[  618.574231]  ? mark_lock.part.0+0x77/0x880
[  618.577575]  ? __pfx_wake_page_function+0x10/0x10
[  618.581430]  __do_fault+0x2fc/0x610
[  618.584329]  ? __thp_vma_allowable_orders+0x25e/0x9f0
[  618.588442]  do_fault+0x22e/0xb60
[  618.591222]  handle_pte_fault+0x55a/0x7e0
[  618.594543]  ? thp_vma_allowable_orders.constprop.0+0x66/0x1e0
[  618.599204]  __handle_mm_fault+0xf93/0x1580
[  618.602652]  ? __pfx___handle_mm_fault+0x10/0x10
[  618.606460]  ? find_vma+0xc5/0x150
[  618.609324]  ? __pfx_find_vma+0x10/0x10
[  618.612497]  ? lock_acquire+0x1bf/0x230
[  618.615639]  handle_mm_fault+0x288/0xa80
[  618.618882]  do_user_addr_fault+0x39a/0x1280
[  618.622453]  exc_page_fault+0xa0/0x1c0
[  618.625530]  asm_exc_page_fault+0x26/0x30
[  618.628807] RIP: 0033:0x40118d
[  618.631429] RSP: 002b:00007ffdae7ffa40 EFLAGS: 00010207
[ 618.635586] RAX: 0000000020000000 RBX: 0000000000000000 RCX: 00007f8c00b4d557 [ 618.641281] RDX: 0000000000000002 RSI: 0000000000001000 RDI: 0000000020000000 [ 618.646822] RBP: 00007ffdae7ffa50 R08: 0000000000000003 R09: 00000000000ff000 [ 618.652349] R10: 0000000000000011 R11: 0000000000000246 R12: 00007ffdae7ffb88 [ 618.657931] R13: 0000000000401136 R14: 00007f8c00cedaa0 R15: 0000000000403e08
[  618.663498]  </TASK>
[  618.665808]
[  618.665808] Showing all locks held in the system:
[  618.671090] 1 lock held by khungtaskd/51:
[ 618.674595] #0: ffffffff950219c0 (rcu_read_lock){....}-{1:2}, at: debug_show_all_locks+0x75/0x340
[  618.682063] 1 lock held by r/2682:
[ 618.685101] #0: ffff888100072fe0 (&mm->mmap_lock){++++}-{3:3}, at: lock_mm_and_find_vma+0x3a/0x6b0
[  618.692648]
[  618.694194] =============================================
[  618.694194]


Note that, reverting all 8 commits doesn't show any issue on 5.15.y, but that does not look like a feasible solution.

I think we can say that the issue is localized to the above pointed 8 commits.

Thanks,
Harshit


Patch [2] tries to solve the problem of not throwing away pending data
when closing the device. Perhaps calling cancel_delayed_work_sync()
and then follow up with a manual call to fb_deferred_io_work() would
be enough to flush the remaining data?

-Patrik

[1] https://bugzilla.suse.com/show_bug.cgi?id=1221814
[2] 33cd6ea9c067 fbdev: flush deferred IO before closing



Thanks,
Harshit





[Index of Archives]     [Video for Linux]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite Tourism]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux