On Wed, Jun 5, 2024 at 5:48 PM Ming Lei <ming.lei@xxxxxxxxxx> wrote: > > On Wed, Jun 05, 2024 at 03:20:34PM +0800, Changhui Zhong wrote: > > On Wed, Jun 5, 2024 at 9:41 AM Li Nan <linan666@xxxxxxxxxxxxxxx> wrote: > > > > > > > > > > > > 在 2024/6/4 9:32, Changhui Zhong 写道: > > > > On Mon, Jun 3, 2024 at 10:20 AM Li Nan <linan666@xxxxxxxxxxxxxxx> wrote: > > > >> > > > >> > > > >> > > > >> 在 2024/6/3 8:39, Ming Lei 写道: > > > >> > > > >> [...] > > > >> > > > >>>> diff --git a/drivers/block/ublk_drv.c b/drivers/block/ublk_drv.c > > > >>>> index 4e159948c912..99b621b2d40f 100644 > > > >>>> --- a/drivers/block/ublk_drv.c > > > >>>> +++ b/drivers/block/ublk_drv.c > > > >>>> @@ -2630,7 +2630,8 @@ static void ublk_queue_reinit(struct ublk_device *ub, struct ublk_queue *ubq) > > > >>>> { > > > >>>> int i; > > > >>>> > > > >>>> - WARN_ON_ONCE(!(ubq->ubq_daemon && ubq_daemon_is_dying(ubq))); > > > >>>> + if (WARN_ON_ONCE(!(ubq->ubq_daemon && ubq_daemon_is_dying(ubq)))) > > > >>>> + return; > > > >>> > > > >>> Yeah, it is one bug. However, it could be addressed by adding the check in > > > >>> ublk_ctrl_start_recovery() and return immediately in case of NULL ubq->ubq_daemon, > > > >>> what do you think about this way? > > > >>> > > > >> > > > >> Check ub->nr_queues_ready seems better. How about: > > > >> > > > >> @@ -2662,6 +2662,8 @@ static int ublk_ctrl_start_recovery(struct > > > >> ublk_device *ub, > > > >> mutex_lock(&ub->mutex); > > > >> if (!ublk_can_use_recovery(ub)) > > > >> goto out_unlock; > > > >> + if (!ub->nr_queues_ready) > > > >> + goto out_unlock; > > > >> /* > > > >> * START_RECOVERY is only allowd after: > > > >> * > > > >> > > > >>> > > > >>> Thanks, > > > >>> Ming > > > >> > > > >> -- > > > >> Thanks, > > > >> Nan > > > >> > > > > > > > > > > > > Hi,Nan > > > > > > > > After applying your new patch, I did not trigger "NULL pointer > > > > dereference" and "Warning", > > > > but hit task hung "Call Trace" info, please check > > > > > > > > [13617.812306] running generic/004 > > > > [13622.293674] blk_print_req_error: 91 callbacks suppressed > > > > [13622.293681] I/O error, dev ublkb4, sector 233256 op 0x1:(WRITE) > > > > flags 0x8800 phys_seg 1 prio class 0 > > > > [13622.308145] I/O error, dev ublkb4, sector 233256 op 0x0:(READ) > > > > flags 0x0 phys_seg 2 prio class 0 > > > > [13622.316923] I/O error, dev ublkb4, sector 233264 op 0x1:(WRITE) > > > > flags 0x8800 phys_seg 1 prio class 0 > > > > [13622.326048] I/O error, dev ublkb4, sector 233272 op 0x0:(READ) > > > > flags 0x0 phys_seg 1 prio class 0 > > > > [13622.334828] I/O error, dev ublkb4, sector 233272 op 0x1:(WRITE) > > > > flags 0x8800 phys_seg 1 prio class 0 > > > > [13622.343954] I/O error, dev ublkb4, sector 233312 op 0x0:(READ) > > > > flags 0x0 phys_seg 1 prio class 0 > > > > [13622.352733] I/O error, dev ublkb4, sector 233008 op 0x0:(READ) > > > > flags 0x0 phys_seg 1 prio class 0 > > > > [13622.361514] I/O error, dev ublkb4, sector 233112 op 0x0:(READ) > > > > flags 0x0 phys_seg 1 prio class 0 > > > > [13622.370292] I/O error, dev ublkb4, sector 233192 op 0x1:(WRITE) > > > > flags 0x8800 phys_seg 1 prio class 0 > > > > [13622.379419] I/O error, dev ublkb4, sector 233120 op 0x0:(READ) > > > > flags 0x0 phys_seg 1 prio class 0 > > > > [13641.069695] INFO: task fio:174413 blocked for more than 122 seconds. > > > > [13641.076061] Not tainted 6.10.0-rc1+ #1 > > > > [13641.080338] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" > > > > disables this message. > > > > [13641.088164] task:fio state:D stack:0 pid:174413 > > > > tgid:174413 ppid:174386 flags:0x00004002 > > > > [13641.088168] Call Trace: > > > > [13641.088170] <TASK> > > > > [13641.088171] __schedule+0x221/0x670 > > > > [13641.088177] schedule+0x23/0xa0 > > > > [13641.088179] io_schedule+0x42/0x70 > > > > [13641.088181] blk_mq_get_tag+0x118/0x2b0 > > > > [13641.088185] ? gup_fast_pgd_range+0x280/0x370 > > > > [13641.088188] ? __pfx_autoremove_wake_function+0x10/0x10 > > > > [13641.088192] __blk_mq_alloc_requests+0x194/0x3a0 > > > > [13641.088194] blk_mq_submit_bio+0x241/0x6c0 > > > > [13641.088196] __submit_bio+0x8a/0x1f0 > > > > [13641.088199] submit_bio_noacct_nocheck+0x168/0x250 > > > > [13641.088201] ? submit_bio_noacct+0x45/0x560 > > > > [13641.088203] __blkdev_direct_IO_async+0x167/0x1a0 > > > > [13641.088206] blkdev_write_iter+0x1c8/0x270 > > > > [13641.088208] aio_write+0x11c/0x240 > > > > [13641.088212] ? __rq_qos_issue+0x21/0x40 > > > > [13641.088214] ? blk_mq_start_request+0x34/0x1a0 > > > > [13641.088216] ? io_submit_one+0x68/0x380 > > > > [13641.088218] ? kmem_cache_alloc_noprof+0x4e/0x320 > > > > [13641.088221] ? fget+0x7c/0xc0 > > > > [13641.088224] ? io_submit_one+0xde/0x380 > > > > [13641.088226] io_submit_one+0xde/0x380 > > > > [13641.088228] __x64_sys_io_submit+0x80/0x160 > > > > [13641.088229] do_syscall_64+0x79/0x150 > > > > [13641.088233] ? syscall_exit_to_user_mode+0x6c/0x1f0 > > > > [13641.088237] ? do_io_getevents+0x8b/0xe0 > > > > [13641.088238] ? syscall_exit_work+0xf3/0x120 > > > > [13641.088241] ? syscall_exit_to_user_mode+0x6c/0x1f0 > > > > [13641.088243] ? do_syscall_64+0x85/0x150 > > > > [13641.088245] ? do_syscall_64+0x85/0x150 > > > > [13641.088247] ? blk_mq_flush_plug_list.part.0+0x108/0x160 > > > > [13641.088249] ? rseq_get_rseq_cs+0x1d/0x220 > > > > [13641.088252] ? rseq_ip_fixup+0x6d/0x1d0 > > > > [13641.088254] ? blk_finish_plug+0x24/0x40 > > > > [13641.088256] ? syscall_exit_to_user_mode+0x6c/0x1f0 > > > > [13641.088258] ? do_syscall_64+0x85/0x150 > > > > [13641.088260] ? syscall_exit_to_user_mode+0x6c/0x1f0 > > > > [13641.088262] ? do_syscall_64+0x85/0x150 > > > > [13641.088264] ? syscall_exit_to_user_mode+0x6c/0x1f0 > > > > [13641.088266] ? do_syscall_64+0x85/0x150 > > > > [13641.088268] ? do_syscall_64+0x85/0x150 > > > > [13641.088270] ? do_syscall_64+0x85/0x150 > > > > [13641.088272] ? clear_bhb_loop+0x45/0xa0 > > > > [13641.088275] ? clear_bhb_loop+0x45/0xa0 > > > > [13641.088277] ? clear_bhb_loop+0x45/0xa0 > > > > [13641.088279] entry_SYSCALL_64_after_hwframe+0x76/0x7e > > > > [13641.088281] RIP: 0033:0x7ff92150713d > > > > [13641.088283] RSP: 002b:00007ffca1ef81f8 EFLAGS: 00000246 ORIG_RAX: > > > > 00000000000000d1 > > > > [13641.088285] RAX: ffffffffffffffda RBX: 00007ff9217e2f70 RCX: 00007ff92150713d > > > > [13641.088286] RDX: 000055863b694fe0 RSI: 0000000000000010 RDI: 00007ff92164d000 > > > > [13641.088287] RBP: 00007ff92164d000 R08: 00007ff91936d000 R09: 0000000000000180 > > > > [13641.088288] R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000000010 > > > > [13641.088289] R13: 0000000000000000 R14: 000055863b694fe0 R15: 000055863b6970c0 > > > > [13641.088291] </TASK> > > > > > > > > Thanks, > > > > Changhui > > > > > > > > > > After applying the previous patch, will the test environment continue to > > > execute test cases after WARN? > > > > a few days ago,test with the previous patch, the test environment > > continued to execute test cases after WARN, > > and I terminated the test when I observed a WARN,so I did not observe > > the subsequent situation. > > > > > I am not sure whether this issue has always > > > existed but was not tested becasue of WARN, or whether the new patch > > > introduced it. > > > > today, I re-test previous patch, and let it run for a long time,I > > observed WARN and task hung, > > looks this issue already existed and not introduced by new patch. > > Hi Changhui, > > The hang is actually expected because recovery fails. > > Please pull the latest ublksrv and check if the issue can still be > reproduced: > > https://github.com/ublk-org/ublksrv > > BTW, one ublksrv segfault and two test cleanup issues are fixed. > > Thanks, > Ming > Hi,Ming and Nan after applying the new patch and pulling the latest ublksrv, I ran the test for 4 hours and did not observe any task hang. the test results looks good! Thanks, Changhui