lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <CAFnufp0-3=-=_TkQLEV33u3R4AZJu5V42jYy7=5pB9eCqo1JgA@mail.gmail.com>
Date:   Mon, 18 Jul 2022 20:01:51 +0200
From:   Matteo Croce <mcroce@...ux.microsoft.com>
To:     Chao Yu <chao@...nel.org>
Cc:     jaegeuk@...nel.org, linux-kernel@...r.kernel.org,
        stable@...r.kernel.org, linux-f2fs-devel@...ts.sourceforge.net,
        Ming Yan <yanming@....edu.cn>
Subject: Re: [f2fs-dev] [PATCH v4] f2fs: fix to do sanity check on total_data_blocks

On Wed, Jul 13, 2022 at 4:15 AM Chao Yu <chao@...nel.org> wrote:
>
> On 2022/7/13 7:22, Matteo Croce wrote:
> > On Fri,  6 May 2022 09:33:06 +0800
> > Chao Yu <chao@...nel.org> wrote:
> >
> >> As Yanming reported in bugzilla:
> >>
> >> https://bugzilla.kernel.org/show_bug.cgi?id=215916
> >>
> >> The kernel message is shown below:
> >>
> >> kernel BUG at fs/f2fs/segment.c:2560!
> >> Call Trace:
> >>   allocate_segment_by_default+0x228/0x440
> >>   f2fs_allocate_data_block+0x13d1/0x31f0
> >>   do_write_page+0x18d/0x710
> >>   f2fs_outplace_write_data+0x151/0x250
> >>   f2fs_do_write_data_page+0xef9/0x1980
> >>   move_data_page+0x6af/0xbc0
> >>   do_garbage_collect+0x312f/0x46f0
> >>   f2fs_gc+0x6b0/0x3bc0
> >>   f2fs_balance_fs+0x921/0x2260
> >>   f2fs_write_single_data_page+0x16be/0x2370
> >>   f2fs_write_cache_pages+0x428/0xd00
> >>   f2fs_write_data_pages+0x96e/0xd50
> >>   do_writepages+0x168/0x550
> >>   __writeback_single_inode+0x9f/0x870
> >>   writeback_sb_inodes+0x47d/0xb20
> >>   __writeback_inodes_wb+0xb2/0x200
> >>   wb_writeback+0x4bd/0x660
> >>   wb_workfn+0x5f3/0xab0
> >>   process_one_work+0x79f/0x13e0
> >>   worker_thread+0x89/0xf60
> >>   kthread+0x26a/0x300
> >>   ret_from_fork+0x22/0x30
> >> RIP: 0010:new_curseg+0xe8d/0x15f0
> >>
> >> The root cause is: ckpt.valid_block_count is inconsistent with SIT
> >> table, stat info indicates filesystem has free blocks, but SIT table
> >> indicates filesystem has no free segment.
> >>
> >> So that during garbage colloection, it triggers panic when LFS
> >> allocator fails to find free segment.
> >>
> >> This patch tries to fix this issue by checking consistency in between
> >> ckpt.valid_block_count and block accounted from SIT.
> >>
> >> Cc: stable@...r.kernel.org
> >> Reported-by: Ming Yan <yanming@....edu.cn>
> >> Signed-off-by: Chao Yu <chao.yu@...o.com>
> >> ---
> >> v4:
> >> - fix to set data/node type correctly.
> >>   fs/f2fs/segment.c | 37 ++++++++++++++++++++++++++-----------
> >>   1 file changed, 26 insertions(+), 11 deletions(-)
> >>
> >> diff --git a/fs/f2fs/segment.c b/fs/f2fs/segment.c
> >> index 3a3e2cec2ac4..4735d477059d 100644
> >> --- a/fs/f2fs/segment.c
> >> +++ b/fs/f2fs/segment.c
> >> @@ -4461,7 +4461,8 @@ static int build_sit_entries(struct
> >> f2fs_sb_info *sbi) unsigned int i, start, end;
> >>      unsigned int readed, start_blk = 0;
> >>      int err = 0;
> >> -    block_t total_node_blocks = 0;
> >> +    block_t sit_valid_blocks[2] = {0, 0};
> >> +    int type;
> >>
> >>      do {
> >>              readed = f2fs_ra_meta_pages(sbi, start_blk,
> >> BIO_MAX_VECS, @@ -4486,8 +4487,9 @@ static int
> >> build_sit_entries(struct f2fs_sb_info *sbi) if (err)
> >>                              return err;
> >>                      seg_info_from_raw_sit(se, &sit);
> >> -                    if (IS_NODESEG(se->type))
> >> -                            total_node_blocks +=
> >> se->valid_blocks; +
> >> +                    type = IS_NODESEG(se->type) ? NODE : DATA;
> >> +                    sit_valid_blocks[type] += se->valid_blocks;
> >>
> >>                      if (f2fs_block_unit_discard(sbi)) {
> >>                              /* build discard map only one time */
> >> @@ -4527,15 +4529,17 @@ static int build_sit_entries(struct
> >> f2fs_sb_info *sbi) sit = sit_in_journal(journal, i);
> >>
> >>              old_valid_blocks = se->valid_blocks;
> >> -            if (IS_NODESEG(se->type))
> >> -                    total_node_blocks -= old_valid_blocks;
> >> +
> >> +            type = IS_NODESEG(se->type) ? NODE : DATA;
> >> +            sit_valid_blocks[type] -= old_valid_blocks;
> >>
> >>              err = check_block_count(sbi, start, &sit);
> >>              if (err)
> >>                      break;
> >>              seg_info_from_raw_sit(se, &sit);
> >> -            if (IS_NODESEG(se->type))
> >> -                    total_node_blocks += se->valid_blocks;
> >> +
> >> +            type = IS_NODESEG(se->type) ? NODE : DATA;
> >> +            sit_valid_blocks[type] += se->valid_blocks;
> >>
> >>              if (f2fs_block_unit_discard(sbi)) {
> >>                      if (is_set_ckpt_flags(sbi, CP_TRIMMED_FLAG))
> >> { @@ -4557,13 +4561,24 @@ static int build_sit_entries(struct
> >> f2fs_sb_info *sbi) }
> >>      up_read(&curseg->journal_rwsem);
> >>
> >> -    if (!err && total_node_blocks != valid_node_count(sbi)) {
> >> +    if (err)
> >> +            return err;
> >> +
> >> +    if (sit_valid_blocks[NODE] != valid_node_count(sbi)) {
> >>              f2fs_err(sbi, "SIT is corrupted node# %u vs %u",
> >> -                     total_node_blocks, valid_node_count(sbi));
> >> -            err = -EFSCORRUPTED;
> >> +                     sit_valid_blocks[NODE],
> >> valid_node_count(sbi));
> >> +            return -EFSCORRUPTED;
> >>      }
> >>
> >> -    return err;
> >> +    if (sit_valid_blocks[DATA] + sit_valid_blocks[NODE] >
> >> +                            valid_user_blocks(sbi)) {
> >> +            f2fs_err(sbi, "SIT is corrupted data# %u %u vs %u",
> >> +                     sit_valid_blocks[DATA],
> >> sit_valid_blocks[NODE],
> >> +                     valid_user_blocks(sbi));
> >> +            return -EFSCORRUPTED;
> >> +    }
> >> +
> >> +    return 0;
> >>   }
> >>
> >>   static void init_free_segmap(struct f2fs_sb_info *sbi)
> >
> > Hi all,
> >
> > I'm experiencing this with kernel 5.18.5:
>
> Hi Matteo,
>
> It looks in your attached log there is no similar stack as below?
>
>  >> kernel BUG at fs/f2fs/segment.c:2560!
>  >> Call Trace:
>  >>   allocate_segment_by_default+0x228/0x440
>  >>   f2fs_allocate_data_block+0x13d1/0x31f0
>  >>   do_write_page+0x18d/0x710
>
> Could you please check cpu usage when the bug reproduces?
>
> Thanks,

Hi,

This happens when the bug happens:

  PID USER   PR  NI    S  %CPU  %MEM     TIME+ COMMAND
  398 root   20   0    R 100,0   0,0   2:05.80 f2fs_gc-179:130



>
> >
> > [1774085.200363] rcu: INFO: rcu_sched self-detected stall on CPU
> > [1774085.206139] rcu:   1-....: (2099 ticks this GP) idle=a7f/1/0x4000000000000000 softirq=6435692/6435692 fqs=977
> > [1774085.216272]        (t=2101 jiffies g=30461565 q=289)
> > [1774085.216276] Task dump for CPU 1:
> > [1774085.216278] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
> > [1774085.216285] Call trace:
> > [1774085.216287]  dump_backtrace.part.0+0xb4/0xc0
> > [1774085.216299]  show_stack+0x14/0x30
> > [1774085.216304]  sched_show_task+0x130/0x160
> > [1774085.216309]  dump_cpu_task+0x40/0x4c
> > [1774085.216313]  rcu_dump_cpu_stacks+0xec/0x130
> > [1774085.216317]  rcu_sched_clock_irq+0x8f8/0xaa0
> > [1774085.216321]  update_process_times+0x98/0x180
> > [1774085.216324]  tick_sched_timer+0x54/0xd0
> > [1774085.216329]  __hrtimer_run_queues+0x134/0x2d0
> > [1774085.216333]  hrtimer_interrupt+0x110/0x2c0
> > [1774085.216336]  arch_timer_handler_phys+0x28/0x40
> > [1774085.216342]  handle_percpu_devid_irq+0x84/0x1c0
> > [1774085.216348]  generic_handle_domain_irq+0x28/0x40
> > [1774085.216352]  gic_handle_irq+0x58/0x80
> > [1774085.216357]  call_on_irq_stack+0x2c/0x3c
> > [1774085.216360]  do_interrupt_handler+0x78/0x84
> > [1774085.216364]  el1_interrupt+0x30/0x50
> > [1774085.216368]  el1h_64_irq_handler+0x14/0x20
> > [1774085.216371]  el1h_64_irq+0x64/0x68
> > [1774085.216374]  __filemap_get_folio+0xa0/0x3b0
> > [1774085.216378]  pagecache_get_page+0x18/0x70
> > [1774085.216383]  grab_cache_page_write_begin+0x20/0x30
> > [1774085.216388]  f2fs_get_read_data_page+0x3c/0x480
> > [1774085.216392]  f2fs_get_lock_data_page+0x3c/0x260
> > [1774085.216395]  move_data_page+0x34/0x530
> > [1774085.216398]  do_garbage_collect+0xc54/0x12e0
> > [1774085.216401]  f2fs_gc+0x3b4/0x800
> > [1774085.216404]  gc_thread_func+0x4c8/0x640
> > [1774085.216407]  kthread+0xd0/0xe0
> > [1774085.216411]  ret_from_fork+0x10/0x20
> > [1774115.558533] rcu: INFO: rcu_sched self-detected stall on CPU
> > [1774115.564309] rcu:   1-....: (2099 ticks this GP) idle=be7/1/0x4000000000000000 softirq=6435694/6435694 fqs=1016
> > [1774115.574528]        (t=2101 jiffies g=30461569 q=847)
> > [1774115.574532] Task dump for CPU 1:
> > [1774115.574534] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
> > [1774115.574541] Call trace:
> > [1774115.574543]  dump_backtrace.part.0+0xb4/0xc0
> > [1774115.574555]  show_stack+0x14/0x30
> > [1774115.574559]  sched_show_task+0x130/0x160
> > [1774115.574564]  dump_cpu_task+0x40/0x4c
> > [1774115.574569]  rcu_dump_cpu_stacks+0xec/0x130
> > [1774115.574572]  rcu_sched_clock_irq+0x8f8/0xaa0
> > [1774115.574576]  update_process_times+0x98/0x180
> > [1774115.574580]  tick_sched_timer+0x54/0xd0
> > [1774115.574584]  __hrtimer_run_queues+0x134/0x2d0
> > [1774115.574587]  hrtimer_interrupt+0x110/0x2c0
> > [1774115.574590]  arch_timer_handler_phys+0x28/0x40
> > [1774115.574596]  handle_percpu_devid_irq+0x84/0x1c0
> > [1774115.574603]  generic_handle_domain_irq+0x28/0x40
> > [1774115.574606]  gic_handle_irq+0x58/0x80
> > [1774115.574611]  call_on_irq_stack+0x2c/0x3c
> > [1774115.574615]  do_interrupt_handler+0x78/0x84
> > [1774115.574618]  el1_interrupt+0x30/0x50
> > [1774115.574623]  el1h_64_irq_handler+0x14/0x20
> > [1774115.574626]  el1h_64_irq+0x64/0x68
> > [1774115.574629]  f2fs_is_valid_blkaddr+0x184/0x310
> > [1774115.574633]  f2fs_get_read_data_page+0x80/0x480
> > [1774115.574637]  f2fs_get_lock_data_page+0x3c/0x260
> > [1774115.574640]  move_data_page+0x34/0x530
> > [1774115.574643]  do_garbage_collect+0xc54/0x12e0
> > [1774115.574646]  f2fs_gc+0x3b4/0x800
> > [1774115.574649]  gc_thread_func+0x4c8/0x640
> > [1774115.574652]  kthread+0xd0/0xe0
> > [1774115.574655]  ret_from_fork+0x10/0x20
> > [1774178.604732] rcu: INFO: rcu_sched self-detected stall on CPU
> > [1774178.610507] rcu:   1-....: (8404 ticks this GP) idle=be7/1/0x4000000000000000 softirq=6435694/6435694 fqs=3918
> > [1774178.620727]        (t=8406 jiffies g=30461569 q=2117)
> > [1774178.620730] Task dump for CPU 1:
> > [1774178.620732] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
> > [1774178.620739] Call trace:
> > [1774178.620741]  dump_backtrace.part.0+0xb4/0xc0
> > [1774178.620750]  show_stack+0x14/0x30
> > [1774178.620754]  sched_show_task+0x130/0x160
> > [1774178.620762]  dump_cpu_task+0x40/0x4c
> > [1774178.620766]  rcu_dump_cpu_stacks+0xec/0x130
> > [1774178.620769]  rcu_sched_clock_irq+0x8f8/0xaa0
> > [1774178.620774]  update_process_times+0x98/0x180
> > [1774178.620777]  tick_sched_timer+0x54/0xd0
> > [1774178.620781]  __hrtimer_run_queues+0x134/0x2d0
> > [1774178.620785]  hrtimer_interrupt+0x110/0x2c0
> > [1774178.620788]  arch_timer_handler_phys+0x28/0x40
> > [1774178.620793]  handle_percpu_devid_irq+0x84/0x1c0
> > [1774178.620798]  generic_handle_domain_irq+0x28/0x40
> > [1774178.620801]  gic_handle_irq+0x58/0x80
> > [1774178.620806]  call_on_irq_stack+0x2c/0x3c
> > [1774178.620810]  do_interrupt_handler+0x78/0x84
> > [1774178.620813]  el1_interrupt+0x30/0x50
> > [1774178.620818]  el1h_64_irq_handler+0x14/0x20
> > [1774178.620821]  el1h_64_irq+0x64/0x68
> > [1774178.620823]  f2fs_lookup_extent_cache+0x18c/0x310
> > [1774178.620828]  f2fs_get_read_data_page+0x54/0x480
> > [1774178.620832]  f2fs_get_lock_data_page+0x3c/0x260
> > [1774178.620835]  move_data_page+0x34/0x530
> > [1774178.620838]  do_garbage_collect+0xc54/0x12e0
> > [1774178.620842]  f2fs_gc+0x3b4/0x800
> > [1774178.620844]  gc_thread_func+0x4c8/0x640
> > [1774178.620848]  kthread+0xd0/0xe0
> > [1774178.620851]  ret_from_fork+0x10/0x20
> > [1774260.049828] rcu: INFO: rcu_sched self-detected stall on CPU
> > [1774260.055605] rcu:   0-....: (2099 ticks this GP) idle=915/1/0x4000000000000000 softirq=6031130/6031130 fqs=1015
> > [1774260.065826]        (t=2101 jiffies g=30461577 q=931)
> > [1774260.065830] Task dump for CPU 0:
> > [1774260.065832] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
> > [1774260.065839] Call trace:
> > [1774260.065841]  dump_backtrace.part.0+0xb4/0xc0
> > [1774260.065854]  show_stack+0x14/0x30
> > [1774260.065858]  sched_show_task+0x130/0x160
> > [1774260.065863]  dump_cpu_task+0x40/0x4c
> > [1774260.065868]  rcu_dump_cpu_stacks+0xec/0x130
> > [1774260.065871]  rcu_sched_clock_irq+0x8f8/0xaa0
> > [1774260.065875]  update_process_times+0x98/0x180
> > [1774260.065879]  tick_sched_timer+0x54/0xd0
> > [1774260.065883]  __hrtimer_run_queues+0x134/0x2d0
> > [1774260.065887]  hrtimer_interrupt+0x110/0x2c0
> > [1774260.065890]  arch_timer_handler_phys+0x28/0x40
> > [1774260.065896]  handle_percpu_devid_irq+0x84/0x1c0
> > [1774260.065901]  generic_handle_domain_irq+0x28/0x40
> > [1774260.065904]  gic_handle_irq+0x58/0x80
> > [1774260.065909]  call_on_irq_stack+0x2c/0x3c
> > [1774260.065913]  do_interrupt_handler+0x78/0x84
> > [1774260.065916]  el1_interrupt+0x30/0x50
> > [1774260.065922]  el1h_64_irq_handler+0x14/0x20
> > [1774260.065925]  el1h_64_irq+0x64/0x68
> > [1774260.065928]  folio_wait_stable+0x8/0x30
> > [1774260.065933]  pagecache_get_page+0x18/0x70
> > [1774260.065938]  grab_cache_page_write_begin+0x20/0x30
> > [1774260.065943]  f2fs_get_read_data_page+0x3c/0x480
> > [1774260.065947]  f2fs_get_lock_data_page+0x3c/0x260
> > [1774260.065950]  move_data_page+0x34/0x530
> > [1774260.065954]  do_garbage_collect+0xc54/0x12e0
> > [1774260.065957]  f2fs_gc+0x3b4/0x800
> > [1774260.065960]  gc_thread_func+0x4c8/0x640
> > [1774260.065963]  kthread+0xd0/0xe0
> > [1774260.065966]  ret_from_fork+0x10/0x20
> > [1774323.086033] rcu: INFO: rcu_sched self-detected stall on CPU
> > [1774323.091808] rcu:   0-....: (8403 ticks this GP) idle=915/1/0x4000000000000000 softirq=6031130/6031130 fqs=4088
> > [1774323.102028]        (t=8405 jiffies g=30461577 q=1422)
> > [1774323.102031] Task dump for CPU 0:
> > [1774323.102033] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
> > [1774323.102040] Call trace:
> > [1774323.102042]  dump_backtrace.part.0+0xb4/0xc0
> > [1774323.102052]  show_stack+0x14/0x30
> > [1774323.102055]  sched_show_task+0x130/0x160
> > [1774323.102060]  dump_cpu_task+0x40/0x4c
> > [1774323.102065]  rcu_dump_cpu_stacks+0xec/0x130
> > [1774323.102068]  rcu_sched_clock_irq+0x8f8/0xaa0
> > [1774323.102073]  update_process_times+0x98/0x180
> > [1774323.102076]  tick_sched_timer+0x54/0xd0
> > [1774323.102080]  __hrtimer_run_queues+0x134/0x2d0
> > [1774323.102084]  hrtimer_interrupt+0x110/0x2c0
> > [1774323.102087]  arch_timer_handler_phys+0x28/0x40
> > [1774323.102092]  handle_percpu_devid_irq+0x84/0x1c0
> > [1774323.102097]  generic_handle_domain_irq+0x28/0x40
> > [1774323.102100]  gic_handle_irq+0x58/0x80
> > [1774323.102105]  call_on_irq_stack+0x2c/0x3c
> > [1774323.102109]  do_interrupt_handler+0x78/0x84
> > [1774323.102112]  el1_interrupt+0x30/0x50
> > [1774323.102118]  el1h_64_irq_handler+0x14/0x20
> > [1774323.102120]  el1h_64_irq+0x64/0x68
> > [1774323.102123]  folio_unlock+0x8/0x40
> > [1774323.102127]  f2fs_get_lock_data_page+0xf4/0x260
> > [1774323.102131]  move_data_page+0x34/0x530
> > [1774323.102135]  do_garbage_collect+0xc54/0x12e0
> > [1774323.102138]  f2fs_gc+0x3b4/0x800
> > [1774323.102141]  gc_thread_func+0x4c8/0x640
> > [1774323.102144]  kthread+0xd0/0xe0
> > [1774323.102147]  ret_from_fork+0x10/0x20
> > [1774386.122242] rcu: INFO: rcu_sched self-detected stall on CPU
> > [1774386.128015] rcu:   0-....: (14707 ticks this GP) idle=915/1/0x4000000000000000 softirq=6031130/6031130 fqs=7133
> > [1774386.138322]        (t=14709 jiffies g=30461577 q=1895)
> > [1774386.138325] Task dump for CPU 0:
> > [1774386.138327] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
> > [1774386.138334] Call trace:
> > [1774386.138335]  dump_backtrace.part.0+0xb4/0xc0
> > [1774386.138345]  show_stack+0x14/0x30
> > [1774386.138349]  sched_show_task+0x130/0x160
> > [1774386.138354]  dump_cpu_task+0x40/0x4c
> > [1774386.138358]  rcu_dump_cpu_stacks+0xec/0x130
> > [1774386.138361]  rcu_sched_clock_irq+0x8f8/0xaa0
> > [1774386.138365]  update_process_times+0x98/0x180
> > [1774386.138369]  tick_sched_timer+0x54/0xd0
> > [1774386.138373]  __hrtimer_run_queues+0x134/0x2d0
> > [1774386.138376]  hrtimer_interrupt+0x110/0x2c0
> > [1774386.138380]  arch_timer_handler_phys+0x28/0x40
> > [1774386.138384]  handle_percpu_devid_irq+0x84/0x1c0
> > [1774386.138389]  generic_handle_domain_irq+0x28/0x40
> > [1774386.138393]  gic_handle_irq+0x58/0x80
> > [1774386.138397]  call_on_irq_stack+0x2c/0x3c
> > [1774386.138401]  do_interrupt_handler+0x78/0x84
> > [1774386.138405]  el1_interrupt+0x30/0x50
> > [1774386.138410]  el1h_64_irq_handler+0x14/0x20
> > [1774386.138413]  el1h_64_irq+0x64/0x68
> > [1774386.138416]  f2fs_lookup_extent_cache+0x18c/0x310
> > [1774386.138420]  f2fs_get_read_data_page+0x54/0x480
> > [1774386.138424]  f2fs_get_lock_data_page+0x3c/0x260
> > [1774386.138427]  move_data_page+0x34/0x530
> > [1774386.138431]  do_garbage_collect+0xc54/0x12e0
> > [1774386.138434]  f2fs_gc+0x3b4/0x800
> > [1774386.138437]  gc_thread_func+0x4c8/0x640
> > [1774386.138440]  kthread+0xd0/0xe0
> > [1774386.138443]  ret_from_fork+0x10/0x20
> > [1774449.158452] rcu: INFO: rcu_sched self-detected stall on CPU
> > [1774449.164226] rcu:   0-....: (21011 ticks this GP) idle=915/1/0x4000000000000000 softirq=6031130/6031130 fqs=10163
> > [1774449.174619]        (t=21013 jiffies g=30461577 q=2392)
> > [1774449.174622] Task dump for CPU 0:
> > [1774449.174624] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
> > [1774449.174630] Call trace:
> > [1774449.174632]  dump_backtrace.part.0+0xb4/0xc0
> > [1774449.174641]  show_stack+0x14/0x30
> > [1774449.174645]  sched_show_task+0x130/0x160
> > [1774449.174650]  dump_cpu_task+0x40/0x4c
> > [1774449.174654]  rcu_dump_cpu_stacks+0xec/0x130
> > [1774449.174657]  rcu_sched_clock_irq+0x8f8/0xaa0
> > [1774449.174661]  update_process_times+0x98/0x180
> > [1774449.174664]  tick_sched_timer+0x54/0xd0
> > [1774449.174668]  __hrtimer_run_queues+0x134/0x2d0
> > [1774449.174672]  hrtimer_interrupt+0x110/0x2c0
> > [1774449.174675]  arch_timer_handler_phys+0x28/0x40
> > [1774449.174680]  handle_percpu_devid_irq+0x84/0x1c0
> > [1774449.174685]  generic_handle_domain_irq+0x28/0x40
> > [1774449.174688]  gic_handle_irq+0x58/0x80
> > [1774449.174693]  call_on_irq_stack+0x2c/0x3c
> > [1774449.174696]  do_interrupt_handler+0x78/0x84
> > [1774449.174700]  el1_interrupt+0x30/0x50
> > [1774449.174705]  el1h_64_irq_handler+0x14/0x20
> > [1774449.174708]  el1h_64_irq+0x64/0x68
> > [1774449.174711]  __filemap_get_folio+0x20/0x3b0
> > [1774449.174715]  pagecache_get_page+0x18/0x70
> > [1774449.174720]  grab_cache_page_write_begin+0x20/0x30
> > [1774449.174725]  f2fs_get_read_data_page+0x3c/0x480
> > [1774449.174729]  f2fs_get_lock_data_page+0x3c/0x260
> > [1774449.174732]  move_data_page+0x34/0x530
> > [1774449.174735]  do_garbage_collect+0xc54/0x12e0
> > [1774449.174738]  f2fs_gc+0x3b4/0x800
> > [1774449.174741]  gc_thread_func+0x4c8/0x640
> > [1774449.174744]  kthread+0xd0/0xe0
> > [1774449.174748]  ret_from_fork+0x10/0x20
> > [1774512.194663] rcu: INFO: rcu_sched self-detected stall on CPU
> > [1774512.200436] rcu:   0-....: (27315 ticks this GP) idle=915/1/0x4000000000000000 softirq=6031130/6031130 fqs=13122
> > [1774512.210828]        (t=27317 jiffies g=30461577 q=2781)
> > [1774512.210831] Task dump for CPU 0:
> > [1774512.210834] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
> > [1774512.210841] Call trace:
> > [1774512.210842]  dump_backtrace.part.0+0xb4/0xc0
> > [1774512.210852]  show_stack+0x14/0x30
> > [1774512.210856]  sched_show_task+0x130/0x160
> > [1774512.210860]  dump_cpu_task+0x40/0x4c
> > [1774512.210864]  rcu_dump_cpu_stacks+0xec/0x130
> > [1774512.210868]  rcu_sched_clock_irq+0x8f8/0xaa0
> > [1774512.210872]  update_process_times+0x98/0x180
> > [1774512.210875]  tick_sched_timer+0x54/0xd0
> > [1774512.210879]  __hrtimer_run_queues+0x134/0x2d0
> > [1774512.210882]  hrtimer_interrupt+0x110/0x2c0
> > [1774512.210886]  arch_timer_handler_phys+0x28/0x40
> > [1774512.210891]  handle_percpu_devid_irq+0x84/0x1c0
> > [1774512.210896]  generic_handle_domain_irq+0x28/0x40
> > [1774512.210899]  gic_handle_irq+0x58/0x80
> > [1774512.210904]  call_on_irq_stack+0x2c/0x3c
> > [1774512.210907]  do_interrupt_handler+0x78/0x84
> > [1774512.210911]  el1_interrupt+0x30/0x50
> > [1774512.210916]  el1h_64_irq_handler+0x14/0x20
> > [1774512.210918]  el1h_64_irq+0x64/0x68
> > [1774512.210921]  __filemap_get_folio+0x8c/0x3b0
> > [1774512.210925]  pagecache_get_page+0x18/0x70
> > [1774512.210930]  grab_cache_page_write_begin+0x20/0x30
> > [1774512.210935]  f2fs_get_read_data_page+0x3c/0x480
> > [1774512.210939]  f2fs_get_lock_data_page+0x3c/0x260
> > [1774512.210942]  move_data_page+0x34/0x530
> > [1774512.210945]  do_garbage_collect+0xc54/0x12e0
> > [1774512.210948]  f2fs_gc+0x3b4/0x800
> > [1774512.210951]  gc_thread_func+0x4c8/0x640
> > [1774512.210954]  kthread+0xd0/0xe0
> > [1774512.210957]  ret_from_fork+0x10/0x20
> > [1774560.785116] INFO: task f2fs_ckpt-179:1:401 blocked for more than 404 seconds.
> > [1774560.792572]       Not tainted 5.18.5-matteo #107
> > [1774560.797401] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
> > [1774560.805445] task:f2fs_ckpt-179:1 state:D stack:    0 pid:  401 ppid:     2 flags:0x00000008
> > [1774560.805453] Call trace:
> > [1774560.805456]  __switch_to+0xc0/0x100
> > [1774560.805465]  __schedule+0x258/0x620
> > [1774560.805469]  schedule+0x44/0xb0
> > [1774560.805472]  rwsem_down_write_slowpath+0x314/0x5a0
> > [1774560.805478]  down_write+0x44/0x50
> > [1774560.805482]  __checkpoint_and_complete_reqs+0x6c/0x1c0
> > [1774560.805487]  issue_checkpoint_thread+0x34/0xc0
> > [1774560.805490]  kthread+0xd0/0xe0
> > [1774560.805494]  ret_from_fork+0x10/0x20
> > [1774560.805504] INFO: task NetworkManager:1061 blocked for more than 404 seconds.
> > [1774560.812854]       Not tainted 5.18.5-matteo #107
> > [1774560.817667] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
> > [1774560.825708] task:NetworkManager  state:D stack:    0 pid: 1061 ppid:     1 flags:0x00000000
> > [1774560.825714] Call trace:
> > [1774560.825716]  __switch_to+0xc0/0x100
> > [1774560.825720]  __schedule+0x258/0x620
> > [1774560.825723]  schedule+0x44/0xb0
> > [1774560.825726]  schedule_timeout+0x114/0x150
> > [1774560.825730]  __wait_for_common+0xc8/0x200
> > [1774560.825733]  wait_for_completion+0x1c/0x30
> > [1774560.825736]  f2fs_issue_checkpoint+0xd0/0x190
> > [1774560.825740]  f2fs_sync_fs+0x48/0xd0
> > [1774560.825745]  f2fs_do_sync_file+0x178/0x8a0
> > [1774560.825749]  f2fs_sync_file+0x28/0x40
> > [1774560.825753]  vfs_fsync_range+0x30/0x80
> > [1774560.825758]  do_fsync+0x38/0x80
> > [1774560.825762]  __arm64_sys_fsync+0x14/0x20
> > [1774560.825767]  invoke_syscall.constprop.0+0x4c/0xe0
> > [1774560.825771]  do_el0_svc+0x40/0xd0
> > [1774560.825775]  el0_svc+0x14/0x50
> > [1774560.825779]  el0t_64_sync_handler+0xa8/0xb0
> > [1774560.825782]  el0t_64_sync+0x148/0x14c
> > [1774560.825803] INFO: task kworker/u8:4:335638 blocked for more than 404 seconds.
> > [1774560.833149]       Not tainted 5.18.5-matteo #107
> > [1774560.837962] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
> > [1774560.846002] task:kworker/u8:4    state:D stack:    0 pid:335638 ppid:     2 flags:0x00000008
> > [1774560.846009] Workqueue: writeback wb_workfn (flush-179:128)
> > [1774560.846016] Call trace:
> > [1774560.846017]  __switch_to+0xc0/0x100
> > [1774560.846021]  __schedule+0x258/0x620
> > [1774560.846024]  schedule+0x44/0xb0
> > [1774560.846027]  schedule_timeout+0x114/0x150
> > [1774560.846031]  __wait_for_common+0xc8/0x200
> > [1774560.846034]  wait_for_completion+0x1c/0x30
> > [1774560.846037]  f2fs_issue_checkpoint+0xd0/0x190
> > [1774560.846040]  f2fs_sync_fs+0x48/0xd0
> > [1774560.846044]  f2fs_balance_fs_bg+0x114/0x2b0
> > [1774560.846048]  f2fs_write_node_pages+0x38/0x1f0
> > [1774560.846053]  do_writepages+0x68/0x1e0
> > [1774560.846058]  __writeback_single_inode+0x3c/0x3c0
> > [1774560.846062]  writeback_sb_inodes+0x230/0x520
> > [1774560.846065]  __writeback_inodes_wb+0x50/0x130
> > [1774560.846069]  wb_writeback+0x244/0x340
> > [1774560.846073]  wb_workfn+0x308/0x560
> > [1774560.846077]  process_one_work+0x1dc/0x420
> > [1774560.846081]  worker_thread+0x170/0x4c0
> > [1774560.846084]  kthread+0xd0/0xe0
> > [1774560.846086]  ret_from_fork+0x10/0x20
> > [1774623.467979] rcu: INFO: rcu_sched self-detected stall on CPU
> > [1774623.473755] rcu:   0-....: (2099 ticks this GP) idle=643/1/0x4000000000000000 softirq=6031246/6031247 fqs=1028
> > [1774623.483976]        (t=2101 jiffies g=30461613 q=633)
> > [1774623.483979] Task dump for CPU 0:
> > [1774623.483982] task:f2fs_gc-179:130 state:R  running task     stack:    0 pid:  400 ppid:     2 flags:0x0000000a
> > [1774623.483989] Call trace:
> > [1774623.483991]  dump_backtrace.part.0+0xb4/0xc0
> > [1774623.484003]  show_stack+0x14/0x30
> > [1774623.484007]  sched_show_task+0x130/0x160
> > [1774623.484013]  dump_cpu_task+0x40/0x4c
> > [1774623.484018]  rcu_dump_cpu_stacks+0xec/0x130
> > [1774623.484021]  rcu_sched_clock_irq+0x8f8/0xaa0
> > [1774623.484025]  update_process_times+0x98/0x180
> > [1774623.484028]  tick_sched_timer+0x54/0xd0
> > [1774623.484033]  __hrtimer_run_queues+0x134/0x2d0
> > [1774623.484036]  hrtimer_interrupt+0x110/0x2c0
> > [1774623.484040]  arch_timer_handler_phys+0x28/0x40
> > [1774623.484045]  handle_percpu_devid_irq+0x84/0x1c0
> > [1774623.484050]  generic_handle_domain_irq+0x28/0x40
> > [1774623.484054]  gic_handle_irq+0x58/0x80
> > [1774623.484059]  call_on_irq_stack+0x2c/0x3c
> > [1774623.484062]  do_interrupt_handler+0x78/0x84
> > [1774623.484066]  el1_interrupt+0x30/0x50
> > [1774623.484071]  el1h_64_irq_handler+0x14/0x20
> > [1774623.484074]  el1h_64_irq+0x64/0x68
> > [1774623.484077]  folio_unlock+0x18/0x40
> > [1774623.484081]  f2fs_get_lock_data_page+0xf4/0x260
> > [1774623.484085]  move_data_page+0x34/0x530
> > [1774623.484089]  do_garbage_collect+0xc54/0x12e0
> > [1774623.484092]  f2fs_gc+0x3b4/0x800
> > [1774623.484095]  gc_thread_func+0x4c8/0x640
> > [1774623.484098]  kthread+0xd0/0xe0
> > [1774623.484101]  ret_from_fork+0x10/0x20
> >
> > root@...chiatobin:~# uname -a
> > Linux macchiatobin 5.18.5-matteo #107 SMP Fri Jun 17 17:58:26 CEST 2022 aarch64 GNU/Linux
> >
> >



--
per aspera ad upstream

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ