Re: [f2fs-dev] [PATCH v4] f2fs: fix to do sanity check on total_data_blocks

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 2022/7/13 7:22, Matteo Croce wrote:
On Fri,  6 May 2022 09:33:06 +0800
Chao Yu <chao@xxxxxxxxxx> wrote:

As Yanming reported in bugzilla:

https://bugzilla.kernel.org/show_bug.cgi?id=215916

The kernel message is shown below:

kernel BUG at fs/f2fs/segment.c:2560!
Call Trace:
  allocate_segment_by_default+0x228/0x440
  f2fs_allocate_data_block+0x13d1/0x31f0
  do_write_page+0x18d/0x710
  f2fs_outplace_write_data+0x151/0x250
  f2fs_do_write_data_page+0xef9/0x1980
  move_data_page+0x6af/0xbc0
  do_garbage_collect+0x312f/0x46f0
  f2fs_gc+0x6b0/0x3bc0
  f2fs_balance_fs+0x921/0x2260
  f2fs_write_single_data_page+0x16be/0x2370
  f2fs_write_cache_pages+0x428/0xd00
  f2fs_write_data_pages+0x96e/0xd50
  do_writepages+0x168/0x550
  __writeback_single_inode+0x9f/0x870
  writeback_sb_inodes+0x47d/0xb20
  __writeback_inodes_wb+0xb2/0x200
  wb_writeback+0x4bd/0x660
  wb_workfn+0x5f3/0xab0
  process_one_work+0x79f/0x13e0
  worker_thread+0x89/0xf60
  kthread+0x26a/0x300
  ret_from_fork+0x22/0x30
RIP: 0010:new_curseg+0xe8d/0x15f0

The root cause is: ckpt.valid_block_count is inconsistent with SIT
table, stat info indicates filesystem has free blocks, but SIT table
indicates filesystem has no free segment.

So that during garbage colloection, it triggers panic when LFS
allocator fails to find free segment.

This patch tries to fix this issue by checking consistency in between
ckpt.valid_block_count and block accounted from SIT.

Cc: stable@xxxxxxxxxxxxxxx
Reported-by: Ming Yan <yanming@xxxxxxxxxx>
Signed-off-by: Chao Yu <chao.yu@xxxxxxxx>
---
v4:
- fix to set data/node type correctly.
  fs/f2fs/segment.c | 37 ++++++++++++++++++++++++++-----------
  1 file changed, 26 insertions(+), 11 deletions(-)

diff --git a/fs/f2fs/segment.c b/fs/f2fs/segment.c
index 3a3e2cec2ac4..4735d477059d 100644
--- a/fs/f2fs/segment.c
+++ b/fs/f2fs/segment.c
@@ -4461,7 +4461,8 @@ static int build_sit_entries(struct
f2fs_sb_info *sbi) unsigned int i, start, end;
  	unsigned int readed, start_blk = 0;
  	int err = 0;
-	block_t total_node_blocks = 0;
+	block_t sit_valid_blocks[2] = {0, 0};
+	int type;
do {
  		readed = f2fs_ra_meta_pages(sbi, start_blk,
BIO_MAX_VECS, @@ -4486,8 +4487,9 @@ static int
build_sit_entries(struct f2fs_sb_info *sbi) if (err)
  				return err;
  			seg_info_from_raw_sit(se, &sit);
-			if (IS_NODESEG(se->type))
-				total_node_blocks +=
se->valid_blocks; +
+			type = IS_NODESEG(se->type) ? NODE : DATA;
+			sit_valid_blocks[type] += se->valid_blocks;
if (f2fs_block_unit_discard(sbi)) {
  				/* build discard map only one time */
@@ -4527,15 +4529,17 @@ static int build_sit_entries(struct
f2fs_sb_info *sbi) sit = sit_in_journal(journal, i);
old_valid_blocks = se->valid_blocks;
-		if (IS_NODESEG(se->type))
-			total_node_blocks -= old_valid_blocks;
+
+		type = IS_NODESEG(se->type) ? NODE : DATA;
+		sit_valid_blocks[type] -= old_valid_blocks;
err = check_block_count(sbi, start, &sit);
  		if (err)
  			break;
  		seg_info_from_raw_sit(se, &sit);
-		if (IS_NODESEG(se->type))
-			total_node_blocks += se->valid_blocks;
+
+		type = IS_NODESEG(se->type) ? NODE : DATA;
+		sit_valid_blocks[type] += se->valid_blocks;
if (f2fs_block_unit_discard(sbi)) {
  			if (is_set_ckpt_flags(sbi, CP_TRIMMED_FLAG))
{ @@ -4557,13 +4561,24 @@ static int build_sit_entries(struct
f2fs_sb_info *sbi) }
  	up_read(&curseg->journal_rwsem);
- if (!err && total_node_blocks != valid_node_count(sbi)) {
+	if (err)
+		return err;
+
+	if (sit_valid_blocks[NODE] != valid_node_count(sbi)) {
  		f2fs_err(sbi, "SIT is corrupted node# %u vs %u",
-			 total_node_blocks, valid_node_count(sbi));
-		err = -EFSCORRUPTED;
+			 sit_valid_blocks[NODE],
valid_node_count(sbi));
+		return -EFSCORRUPTED;
  	}
- return err;
+	if (sit_valid_blocks[DATA] + sit_valid_blocks[NODE] >
+				valid_user_blocks(sbi)) {
+		f2fs_err(sbi, "SIT is corrupted data# %u %u vs %u",
+			 sit_valid_blocks[DATA],
sit_valid_blocks[NODE],
+			 valid_user_blocks(sbi));
+		return -EFSCORRUPTED;
+	}
+
+	return 0;
  }
static void init_free_segmap(struct f2fs_sb_info *sbi)

Hi all,

I'm experiencing this with kernel 5.18.5:

Hi Matteo,

It looks in your attached log there is no similar stack as below?

>> kernel BUG at fs/f2fs/segment.c:2560!
>> Call Trace:
>>   allocate_segment_by_default+0x228/0x440
>>   f2fs_allocate_data_block+0x13d1/0x31f0
>>   do_write_page+0x18d/0x710

Could you please check cpu usage when the bug reproduces?

Thanks,


[1774085.200363] rcu: INFO: rcu_sched self-detected stall on CPU
[1774085.206139] rcu:   1-....: (2099 ticks this GP) idle=a7f/1/0x4000000000000000 softirq=6435692/6435692 fqs=977
[1774085.216272]        (t=2101 jiffies g=30461565 q=289)
[1774085.216276] Task dump for CPU 1:
[1774085.216278] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
[1774085.216285] Call trace:
[1774085.216287]  dump_backtrace.part.0+0xb4/0xc0
[1774085.216299]  show_stack+0x14/0x30
[1774085.216304]  sched_show_task+0x130/0x160
[1774085.216309]  dump_cpu_task+0x40/0x4c
[1774085.216313]  rcu_dump_cpu_stacks+0xec/0x130
[1774085.216317]  rcu_sched_clock_irq+0x8f8/0xaa0
[1774085.216321]  update_process_times+0x98/0x180
[1774085.216324]  tick_sched_timer+0x54/0xd0
[1774085.216329]  __hrtimer_run_queues+0x134/0x2d0
[1774085.216333]  hrtimer_interrupt+0x110/0x2c0
[1774085.216336]  arch_timer_handler_phys+0x28/0x40
[1774085.216342]  handle_percpu_devid_irq+0x84/0x1c0
[1774085.216348]  generic_handle_domain_irq+0x28/0x40
[1774085.216352]  gic_handle_irq+0x58/0x80
[1774085.216357]  call_on_irq_stack+0x2c/0x3c
[1774085.216360]  do_interrupt_handler+0x78/0x84
[1774085.216364]  el1_interrupt+0x30/0x50
[1774085.216368]  el1h_64_irq_handler+0x14/0x20
[1774085.216371]  el1h_64_irq+0x64/0x68
[1774085.216374]  __filemap_get_folio+0xa0/0x3b0
[1774085.216378]  pagecache_get_page+0x18/0x70
[1774085.216383]  grab_cache_page_write_begin+0x20/0x30
[1774085.216388]  f2fs_get_read_data_page+0x3c/0x480
[1774085.216392]  f2fs_get_lock_data_page+0x3c/0x260
[1774085.216395]  move_data_page+0x34/0x530
[1774085.216398]  do_garbage_collect+0xc54/0x12e0
[1774085.216401]  f2fs_gc+0x3b4/0x800
[1774085.216404]  gc_thread_func+0x4c8/0x640
[1774085.216407]  kthread+0xd0/0xe0
[1774085.216411]  ret_from_fork+0x10/0x20
[1774115.558533] rcu: INFO: rcu_sched self-detected stall on CPU
[1774115.564309] rcu:   1-....: (2099 ticks this GP) idle=be7/1/0x4000000000000000 softirq=6435694/6435694 fqs=1016
[1774115.574528]        (t=2101 jiffies g=30461569 q=847)
[1774115.574532] Task dump for CPU 1:
[1774115.574534] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
[1774115.574541] Call trace:
[1774115.574543]  dump_backtrace.part.0+0xb4/0xc0
[1774115.574555]  show_stack+0x14/0x30
[1774115.574559]  sched_show_task+0x130/0x160
[1774115.574564]  dump_cpu_task+0x40/0x4c
[1774115.574569]  rcu_dump_cpu_stacks+0xec/0x130
[1774115.574572]  rcu_sched_clock_irq+0x8f8/0xaa0
[1774115.574576]  update_process_times+0x98/0x180
[1774115.574580]  tick_sched_timer+0x54/0xd0
[1774115.574584]  __hrtimer_run_queues+0x134/0x2d0
[1774115.574587]  hrtimer_interrupt+0x110/0x2c0
[1774115.574590]  arch_timer_handler_phys+0x28/0x40
[1774115.574596]  handle_percpu_devid_irq+0x84/0x1c0
[1774115.574603]  generic_handle_domain_irq+0x28/0x40
[1774115.574606]  gic_handle_irq+0x58/0x80
[1774115.574611]  call_on_irq_stack+0x2c/0x3c
[1774115.574615]  do_interrupt_handler+0x78/0x84
[1774115.574618]  el1_interrupt+0x30/0x50
[1774115.574623]  el1h_64_irq_handler+0x14/0x20
[1774115.574626]  el1h_64_irq+0x64/0x68
[1774115.574629]  f2fs_is_valid_blkaddr+0x184/0x310
[1774115.574633]  f2fs_get_read_data_page+0x80/0x480
[1774115.574637]  f2fs_get_lock_data_page+0x3c/0x260
[1774115.574640]  move_data_page+0x34/0x530
[1774115.574643]  do_garbage_collect+0xc54/0x12e0
[1774115.574646]  f2fs_gc+0x3b4/0x800
[1774115.574649]  gc_thread_func+0x4c8/0x640
[1774115.574652]  kthread+0xd0/0xe0
[1774115.574655]  ret_from_fork+0x10/0x20
[1774178.604732] rcu: INFO: rcu_sched self-detected stall on CPU
[1774178.610507] rcu:   1-....: (8404 ticks this GP) idle=be7/1/0x4000000000000000 softirq=6435694/6435694 fqs=3918
[1774178.620727]        (t=8406 jiffies g=30461569 q=2117)
[1774178.620730] Task dump for CPU 1:
[1774178.620732] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
[1774178.620739] Call trace:
[1774178.620741]  dump_backtrace.part.0+0xb4/0xc0
[1774178.620750]  show_stack+0x14/0x30
[1774178.620754]  sched_show_task+0x130/0x160
[1774178.620762]  dump_cpu_task+0x40/0x4c
[1774178.620766]  rcu_dump_cpu_stacks+0xec/0x130
[1774178.620769]  rcu_sched_clock_irq+0x8f8/0xaa0
[1774178.620774]  update_process_times+0x98/0x180
[1774178.620777]  tick_sched_timer+0x54/0xd0
[1774178.620781]  __hrtimer_run_queues+0x134/0x2d0
[1774178.620785]  hrtimer_interrupt+0x110/0x2c0
[1774178.620788]  arch_timer_handler_phys+0x28/0x40
[1774178.620793]  handle_percpu_devid_irq+0x84/0x1c0
[1774178.620798]  generic_handle_domain_irq+0x28/0x40
[1774178.620801]  gic_handle_irq+0x58/0x80
[1774178.620806]  call_on_irq_stack+0x2c/0x3c
[1774178.620810]  do_interrupt_handler+0x78/0x84
[1774178.620813]  el1_interrupt+0x30/0x50
[1774178.620818]  el1h_64_irq_handler+0x14/0x20
[1774178.620821]  el1h_64_irq+0x64/0x68
[1774178.620823]  f2fs_lookup_extent_cache+0x18c/0x310
[1774178.620828]  f2fs_get_read_data_page+0x54/0x480
[1774178.620832]  f2fs_get_lock_data_page+0x3c/0x260
[1774178.620835]  move_data_page+0x34/0x530
[1774178.620838]  do_garbage_collect+0xc54/0x12e0
[1774178.620842]  f2fs_gc+0x3b4/0x800
[1774178.620844]  gc_thread_func+0x4c8/0x640
[1774178.620848]  kthread+0xd0/0xe0
[1774178.620851]  ret_from_fork+0x10/0x20
[1774260.049828] rcu: INFO: rcu_sched self-detected stall on CPU
[1774260.055605] rcu:   0-....: (2099 ticks this GP) idle=915/1/0x4000000000000000 softirq=6031130/6031130 fqs=1015
[1774260.065826]        (t=2101 jiffies g=30461577 q=931)
[1774260.065830] Task dump for CPU 0:
[1774260.065832] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
[1774260.065839] Call trace:
[1774260.065841]  dump_backtrace.part.0+0xb4/0xc0
[1774260.065854]  show_stack+0x14/0x30
[1774260.065858]  sched_show_task+0x130/0x160
[1774260.065863]  dump_cpu_task+0x40/0x4c
[1774260.065868]  rcu_dump_cpu_stacks+0xec/0x130
[1774260.065871]  rcu_sched_clock_irq+0x8f8/0xaa0
[1774260.065875]  update_process_times+0x98/0x180
[1774260.065879]  tick_sched_timer+0x54/0xd0
[1774260.065883]  __hrtimer_run_queues+0x134/0x2d0
[1774260.065887]  hrtimer_interrupt+0x110/0x2c0
[1774260.065890]  arch_timer_handler_phys+0x28/0x40
[1774260.065896]  handle_percpu_devid_irq+0x84/0x1c0
[1774260.065901]  generic_handle_domain_irq+0x28/0x40
[1774260.065904]  gic_handle_irq+0x58/0x80
[1774260.065909]  call_on_irq_stack+0x2c/0x3c
[1774260.065913]  do_interrupt_handler+0x78/0x84
[1774260.065916]  el1_interrupt+0x30/0x50
[1774260.065922]  el1h_64_irq_handler+0x14/0x20
[1774260.065925]  el1h_64_irq+0x64/0x68
[1774260.065928]  folio_wait_stable+0x8/0x30
[1774260.065933]  pagecache_get_page+0x18/0x70
[1774260.065938]  grab_cache_page_write_begin+0x20/0x30
[1774260.065943]  f2fs_get_read_data_page+0x3c/0x480
[1774260.065947]  f2fs_get_lock_data_page+0x3c/0x260
[1774260.065950]  move_data_page+0x34/0x530
[1774260.065954]  do_garbage_collect+0xc54/0x12e0
[1774260.065957]  f2fs_gc+0x3b4/0x800
[1774260.065960]  gc_thread_func+0x4c8/0x640
[1774260.065963]  kthread+0xd0/0xe0
[1774260.065966]  ret_from_fork+0x10/0x20
[1774323.086033] rcu: INFO: rcu_sched self-detected stall on CPU
[1774323.091808] rcu:   0-....: (8403 ticks this GP) idle=915/1/0x4000000000000000 softirq=6031130/6031130 fqs=4088
[1774323.102028]        (t=8405 jiffies g=30461577 q=1422)
[1774323.102031] Task dump for CPU 0:
[1774323.102033] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
[1774323.102040] Call trace:
[1774323.102042]  dump_backtrace.part.0+0xb4/0xc0
[1774323.102052]  show_stack+0x14/0x30
[1774323.102055]  sched_show_task+0x130/0x160
[1774323.102060]  dump_cpu_task+0x40/0x4c
[1774323.102065]  rcu_dump_cpu_stacks+0xec/0x130
[1774323.102068]  rcu_sched_clock_irq+0x8f8/0xaa0
[1774323.102073]  update_process_times+0x98/0x180
[1774323.102076]  tick_sched_timer+0x54/0xd0
[1774323.102080]  __hrtimer_run_queues+0x134/0x2d0
[1774323.102084]  hrtimer_interrupt+0x110/0x2c0
[1774323.102087]  arch_timer_handler_phys+0x28/0x40
[1774323.102092]  handle_percpu_devid_irq+0x84/0x1c0
[1774323.102097]  generic_handle_domain_irq+0x28/0x40
[1774323.102100]  gic_handle_irq+0x58/0x80
[1774323.102105]  call_on_irq_stack+0x2c/0x3c
[1774323.102109]  do_interrupt_handler+0x78/0x84
[1774323.102112]  el1_interrupt+0x30/0x50
[1774323.102118]  el1h_64_irq_handler+0x14/0x20
[1774323.102120]  el1h_64_irq+0x64/0x68
[1774323.102123]  folio_unlock+0x8/0x40
[1774323.102127]  f2fs_get_lock_data_page+0xf4/0x260
[1774323.102131]  move_data_page+0x34/0x530
[1774323.102135]  do_garbage_collect+0xc54/0x12e0
[1774323.102138]  f2fs_gc+0x3b4/0x800
[1774323.102141]  gc_thread_func+0x4c8/0x640
[1774323.102144]  kthread+0xd0/0xe0
[1774323.102147]  ret_from_fork+0x10/0x20
[1774386.122242] rcu: INFO: rcu_sched self-detected stall on CPU
[1774386.128015] rcu:   0-....: (14707 ticks this GP) idle=915/1/0x4000000000000000 softirq=6031130/6031130 fqs=7133
[1774386.138322]        (t=14709 jiffies g=30461577 q=1895)
[1774386.138325] Task dump for CPU 0:
[1774386.138327] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
[1774386.138334] Call trace:
[1774386.138335]  dump_backtrace.part.0+0xb4/0xc0
[1774386.138345]  show_stack+0x14/0x30
[1774386.138349]  sched_show_task+0x130/0x160
[1774386.138354]  dump_cpu_task+0x40/0x4c
[1774386.138358]  rcu_dump_cpu_stacks+0xec/0x130
[1774386.138361]  rcu_sched_clock_irq+0x8f8/0xaa0
[1774386.138365]  update_process_times+0x98/0x180
[1774386.138369]  tick_sched_timer+0x54/0xd0
[1774386.138373]  __hrtimer_run_queues+0x134/0x2d0
[1774386.138376]  hrtimer_interrupt+0x110/0x2c0
[1774386.138380]  arch_timer_handler_phys+0x28/0x40
[1774386.138384]  handle_percpu_devid_irq+0x84/0x1c0
[1774386.138389]  generic_handle_domain_irq+0x28/0x40
[1774386.138393]  gic_handle_irq+0x58/0x80
[1774386.138397]  call_on_irq_stack+0x2c/0x3c
[1774386.138401]  do_interrupt_handler+0x78/0x84
[1774386.138405]  el1_interrupt+0x30/0x50
[1774386.138410]  el1h_64_irq_handler+0x14/0x20
[1774386.138413]  el1h_64_irq+0x64/0x68
[1774386.138416]  f2fs_lookup_extent_cache+0x18c/0x310
[1774386.138420]  f2fs_get_read_data_page+0x54/0x480
[1774386.138424]  f2fs_get_lock_data_page+0x3c/0x260
[1774386.138427]  move_data_page+0x34/0x530
[1774386.138431]  do_garbage_collect+0xc54/0x12e0
[1774386.138434]  f2fs_gc+0x3b4/0x800
[1774386.138437]  gc_thread_func+0x4c8/0x640
[1774386.138440]  kthread+0xd0/0xe0
[1774386.138443]  ret_from_fork+0x10/0x20
[1774449.158452] rcu: INFO: rcu_sched self-detected stall on CPU
[1774449.164226] rcu:   0-....: (21011 ticks this GP) idle=915/1/0x4000000000000000 softirq=6031130/6031130 fqs=10163
[1774449.174619]        (t=21013 jiffies g=30461577 q=2392)
[1774449.174622] Task dump for CPU 0:
[1774449.174624] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
[1774449.174630] Call trace:
[1774449.174632]  dump_backtrace.part.0+0xb4/0xc0
[1774449.174641]  show_stack+0x14/0x30
[1774449.174645]  sched_show_task+0x130/0x160
[1774449.174650]  dump_cpu_task+0x40/0x4c
[1774449.174654]  rcu_dump_cpu_stacks+0xec/0x130
[1774449.174657]  rcu_sched_clock_irq+0x8f8/0xaa0
[1774449.174661]  update_process_times+0x98/0x180
[1774449.174664]  tick_sched_timer+0x54/0xd0
[1774449.174668]  __hrtimer_run_queues+0x134/0x2d0
[1774449.174672]  hrtimer_interrupt+0x110/0x2c0
[1774449.174675]  arch_timer_handler_phys+0x28/0x40
[1774449.174680]  handle_percpu_devid_irq+0x84/0x1c0
[1774449.174685]  generic_handle_domain_irq+0x28/0x40
[1774449.174688]  gic_handle_irq+0x58/0x80
[1774449.174693]  call_on_irq_stack+0x2c/0x3c
[1774449.174696]  do_interrupt_handler+0x78/0x84
[1774449.174700]  el1_interrupt+0x30/0x50
[1774449.174705]  el1h_64_irq_handler+0x14/0x20
[1774449.174708]  el1h_64_irq+0x64/0x68
[1774449.174711]  __filemap_get_folio+0x20/0x3b0
[1774449.174715]  pagecache_get_page+0x18/0x70
[1774449.174720]  grab_cache_page_write_begin+0x20/0x30
[1774449.174725]  f2fs_get_read_data_page+0x3c/0x480
[1774449.174729]  f2fs_get_lock_data_page+0x3c/0x260
[1774449.174732]  move_data_page+0x34/0x530
[1774449.174735]  do_garbage_collect+0xc54/0x12e0
[1774449.174738]  f2fs_gc+0x3b4/0x800
[1774449.174741]  gc_thread_func+0x4c8/0x640
[1774449.174744]  kthread+0xd0/0xe0
[1774449.174748]  ret_from_fork+0x10/0x20
[1774512.194663] rcu: INFO: rcu_sched self-detected stall on CPU
[1774512.200436] rcu:   0-....: (27315 ticks this GP) idle=915/1/0x4000000000000000 softirq=6031130/6031130 fqs=13122
[1774512.210828]        (t=27317 jiffies g=30461577 q=2781)
[1774512.210831] Task dump for CPU 0:
[1774512.210834] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
[1774512.210841] Call trace:
[1774512.210842]  dump_backtrace.part.0+0xb4/0xc0
[1774512.210852]  show_stack+0x14/0x30
[1774512.210856]  sched_show_task+0x130/0x160
[1774512.210860]  dump_cpu_task+0x40/0x4c
[1774512.210864]  rcu_dump_cpu_stacks+0xec/0x130
[1774512.210868]  rcu_sched_clock_irq+0x8f8/0xaa0
[1774512.210872]  update_process_times+0x98/0x180
[1774512.210875]  tick_sched_timer+0x54/0xd0
[1774512.210879]  __hrtimer_run_queues+0x134/0x2d0
[1774512.210882]  hrtimer_interrupt+0x110/0x2c0
[1774512.210886]  arch_timer_handler_phys+0x28/0x40
[1774512.210891]  handle_percpu_devid_irq+0x84/0x1c0
[1774512.210896]  generic_handle_domain_irq+0x28/0x40
[1774512.210899]  gic_handle_irq+0x58/0x80
[1774512.210904]  call_on_irq_stack+0x2c/0x3c
[1774512.210907]  do_interrupt_handler+0x78/0x84
[1774512.210911]  el1_interrupt+0x30/0x50
[1774512.210916]  el1h_64_irq_handler+0x14/0x20
[1774512.210918]  el1h_64_irq+0x64/0x68
[1774512.210921]  __filemap_get_folio+0x8c/0x3b0
[1774512.210925]  pagecache_get_page+0x18/0x70
[1774512.210930]  grab_cache_page_write_begin+0x20/0x30
[1774512.210935]  f2fs_get_read_data_page+0x3c/0x480
[1774512.210939]  f2fs_get_lock_data_page+0x3c/0x260
[1774512.210942]  move_data_page+0x34/0x530
[1774512.210945]  do_garbage_collect+0xc54/0x12e0
[1774512.210948]  f2fs_gc+0x3b4/0x800
[1774512.210951]  gc_thread_func+0x4c8/0x640
[1774512.210954]  kthread+0xd0/0xe0
[1774512.210957]  ret_from_fork+0x10/0x20
[1774560.785116] INFO: task f2fs_ckpt-179:1:401 blocked for more than 404 seconds.
[1774560.792572]       Not tainted 5.18.5-matteo #107
[1774560.797401] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[1774560.805445] task:f2fs_ckpt-179:1 state:D stack:    0 pid:  401 ppid:     2 flags:0x00000008
[1774560.805453] Call trace:
[1774560.805456]  __switch_to+0xc0/0x100
[1774560.805465]  __schedule+0x258/0x620
[1774560.805469]  schedule+0x44/0xb0
[1774560.805472]  rwsem_down_write_slowpath+0x314/0x5a0
[1774560.805478]  down_write+0x44/0x50
[1774560.805482]  __checkpoint_and_complete_reqs+0x6c/0x1c0
[1774560.805487]  issue_checkpoint_thread+0x34/0xc0
[1774560.805490]  kthread+0xd0/0xe0
[1774560.805494]  ret_from_fork+0x10/0x20
[1774560.805504] INFO: task NetworkManager:1061 blocked for more than 404 seconds.
[1774560.812854]       Not tainted 5.18.5-matteo #107
[1774560.817667] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[1774560.825708] task:NetworkManager  state:D stack:    0 pid: 1061 ppid:     1 flags:0x00000000
[1774560.825714] Call trace:
[1774560.825716]  __switch_to+0xc0/0x100
[1774560.825720]  __schedule+0x258/0x620
[1774560.825723]  schedule+0x44/0xb0
[1774560.825726]  schedule_timeout+0x114/0x150
[1774560.825730]  __wait_for_common+0xc8/0x200
[1774560.825733]  wait_for_completion+0x1c/0x30
[1774560.825736]  f2fs_issue_checkpoint+0xd0/0x190
[1774560.825740]  f2fs_sync_fs+0x48/0xd0
[1774560.825745]  f2fs_do_sync_file+0x178/0x8a0
[1774560.825749]  f2fs_sync_file+0x28/0x40
[1774560.825753]  vfs_fsync_range+0x30/0x80
[1774560.825758]  do_fsync+0x38/0x80
[1774560.825762]  __arm64_sys_fsync+0x14/0x20
[1774560.825767]  invoke_syscall.constprop.0+0x4c/0xe0
[1774560.825771]  do_el0_svc+0x40/0xd0
[1774560.825775]  el0_svc+0x14/0x50
[1774560.825779]  el0t_64_sync_handler+0xa8/0xb0
[1774560.825782]  el0t_64_sync+0x148/0x14c
[1774560.825803] INFO: task kworker/u8:4:335638 blocked for more than 404 seconds.
[1774560.833149]       Not tainted 5.18.5-matteo #107
[1774560.837962] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[1774560.846002] task:kworker/u8:4    state:D stack:    0 pid:335638 ppid:     2 flags:0x00000008
[1774560.846009] Workqueue: writeback wb_workfn (flush-179:128)
[1774560.846016] Call trace:
[1774560.846017]  __switch_to+0xc0/0x100
[1774560.846021]  __schedule+0x258/0x620
[1774560.846024]  schedule+0x44/0xb0
[1774560.846027]  schedule_timeout+0x114/0x150
[1774560.846031]  __wait_for_common+0xc8/0x200
[1774560.846034]  wait_for_completion+0x1c/0x30
[1774560.846037]  f2fs_issue_checkpoint+0xd0/0x190
[1774560.846040]  f2fs_sync_fs+0x48/0xd0
[1774560.846044]  f2fs_balance_fs_bg+0x114/0x2b0
[1774560.846048]  f2fs_write_node_pages+0x38/0x1f0
[1774560.846053]  do_writepages+0x68/0x1e0
[1774560.846058]  __writeback_single_inode+0x3c/0x3c0
[1774560.846062]  writeback_sb_inodes+0x230/0x520
[1774560.846065]  __writeback_inodes_wb+0x50/0x130
[1774560.846069]  wb_writeback+0x244/0x340
[1774560.846073]  wb_workfn+0x308/0x560
[1774560.846077]  process_one_work+0x1dc/0x420
[1774560.846081]  worker_thread+0x170/0x4c0
[1774560.846084]  kthread+0xd0/0xe0
[1774560.846086]  ret_from_fork+0x10/0x20
[1774623.467979] rcu: INFO: rcu_sched self-detected stall on CPU
[1774623.473755] rcu:   0-....: (2099 ticks this GP) idle=643/1/0x4000000000000000 softirq=6031246/6031247 fqs=1028
[1774623.483976]        (t=2101 jiffies g=30461613 q=633)
[1774623.483979] Task dump for CPU 0:
[1774623.483982] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
[1774623.483989] Call trace:
[1774623.483991]  dump_backtrace.part.0+0xb4/0xc0
[1774623.484003]  show_stack+0x14/0x30
[1774623.484007]  sched_show_task+0x130/0x160
[1774623.484013]  dump_cpu_task+0x40/0x4c
[1774623.484018]  rcu_dump_cpu_stacks+0xec/0x130
[1774623.484021]  rcu_sched_clock_irq+0x8f8/0xaa0
[1774623.484025]  update_process_times+0x98/0x180
[1774623.484028]  tick_sched_timer+0x54/0xd0
[1774623.484033]  __hrtimer_run_queues+0x134/0x2d0
[1774623.484036]  hrtimer_interrupt+0x110/0x2c0
[1774623.484040]  arch_timer_handler_phys+0x28/0x40
[1774623.484045]  handle_percpu_devid_irq+0x84/0x1c0
[1774623.484050]  generic_handle_domain_irq+0x28/0x40
[1774623.484054]  gic_handle_irq+0x58/0x80
[1774623.484059]  call_on_irq_stack+0x2c/0x3c
[1774623.484062]  do_interrupt_handler+0x78/0x84
[1774623.484066]  el1_interrupt+0x30/0x50
[1774623.484071]  el1h_64_irq_handler+0x14/0x20
[1774623.484074]  el1h_64_irq+0x64/0x68
[1774623.484077]  folio_unlock+0x18/0x40
[1774623.484081]  f2fs_get_lock_data_page+0xf4/0x260
[1774623.484085]  move_data_page+0x34/0x530
[1774623.484089]  do_garbage_collect+0xc54/0x12e0
[1774623.484092]  f2fs_gc+0x3b4/0x800
[1774623.484095]  gc_thread_func+0x4c8/0x640
[1774623.484098]  kthread+0xd0/0xe0
[1774623.484101]  ret_from_fork+0x10/0x20

root@macchiatobin:~# uname -a
Linux macchiatobin 5.18.5-matteo #107 SMP Fri Jun 17 17:58:26 CEST 2022 aarch64 GNU/Linux





[Index of Archives]     [Linux Kernel]     [Kernel Development Newbies]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite Hiking]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux