[<prev] [next>] [<thread-prev] [day] [month] [year] [list]
Message-ID: <93e6cc32-c894-0e73-697c-6d950c8063a1@mellanox.com>
Date: Sun, 10 May 2020 11:52:56 +0300
From: Roi Dayan <roid@...lanox.com>
To: Marcelo Ricardo Leitner <mleitner@...hat.com>,
Paul Blakey <paulb@...lanox.com>
Cc: Oz Shlomo <ozsh@...lanox.com>, netdev@...r.kernel.org,
netfilter-devel@...r.kernel.org,
Pablo Neira Ayuso <pablo@...filter.org>
Subject: Re: mlx5: Panic with conntrack offload
On 2020-04-24 7:10 AM, Marcelo Ricardo Leitner wrote:
> Hi Paul,
>
> I'm triggering this panic out of 1802136023c01075, net-next today
> (disregard the hash at the end of the kernel version).
> I have a dual-port CX5 with VF LAG, with 1 guest and 2 VFs on it, with
> different subnets. Ovs OF flows steering each subnet to each VF and
> doing conntrack. (flows at the bottom here)
>
> This happens after 2 or 5 netperf runs. I run netperf, wait for the
> flows to expire, then run it again.
>
> I'm suspecting this was introduced in 9808dd0a2aee ("net/mlx5e: CT: Use
> rhashtable's ct entries instead of a separate list"), but I didn't
> bisect it yet. I know it also happens with 2fcd80144b93ff908, FWIW.
>
> Ideas?
+pablo
Hi Marcelo,
I investigated this and see the root.
In nf flow table offload we create a workqueue with flag WQ_MEM_RECLAIM.
Each CT add/del/stats is queued as work to that workqueue.
In del flow, when we reach mlx5 driver, we reach del_sw_flow_group
and want to release an rhashtable (&fg->ftes_hash).
The rhashtable implementation could also queue a work if rehash is needed.
that work is queued on system_wq which is not created with flag
WQ_MEM_RECLAIM. So if we are now in a del flow work and flush and destroy
the ftes_hash rhashtable and there is a rehash work queue to be flushed
first then we reach in check_flush_dependency() and see target_wq is
system_wq which doesn't have WQ_MEM_RECLAIM flag so we dont return right
away and now checking current_wq_worker() is the nf flow table offload
workqueue which has flag WQ_MEM_RECLAIM.
This generates the warn_once() you see.
WARN_ONCE(worker && ((worker->current_pwq->wq->flags &
(WQ_MEM_RECLAIM | __WQ_LEGACY)) == WQ_MEM_RECLAIM),
"workqueue: WQ_MEM_RECLAIM %s:%ps is flushing !WQ_MEM_RECLAIM %s:%ps",
worker->current_pwq->wq->name, worker->current_func,
target_wq->name, target_func);
It's only avoided if there is also internal flag __WQ_LEGACY which is being
added only for create workqueue wrappers that creates single thread
workqueue. but we do want multi threaded workqueue.
Using WQ_MEM_RECLAIM means all other wqs being used inside a work should
also have WQ_MEM_RECLAIM. because of rhashtable this is not the case.
WQ_MEM_RECLAIM means the wq is guaranteed to have at least one execution
context regardless of memory pressure.
I think we should maybe remove WQ_MEM_RECLAIM from nf flow table offload
workqueue. Pablo, Oz, Paul ?
Thanks,
Roi
>
> Thanks,
> Marcelo
>
> [ 485.557189] ------------[ cut here ]------------
> [ 485.562976] workqueue: WQ_MEM_RECLAIM nf_flow_table_offload:flow_offload_worr
> [ 485.562985] WARNING: CPU: 7 PID: 3731 at kernel/workqueue.c:2610 check_flush0
> [ 485.590191] Kernel panic - not syncing: panic_on_warn set ...
> [ 485.597100] CPU: 7 PID: 3731 Comm: kworker/u112:8 Not tainted 5.7.0-rc1.21802
> [ 485.606629] Hardware name: Dell Inc. PowerEdge R730/072T6D, BIOS 2.4.3 01/177
> [ 485.615487] Workqueue: nf_flow_table_offload flow_offload_work_handler [nf_f]
> [ 485.624834] Call Trace:
> [ 485.628077] dump_stack+0x50/0x70
> [ 485.632280] panic+0xfb/0x2d7
> [ 485.636083] ? check_flush_dependency+0x110/0x130
> [ 485.641830] __warn.cold.12+0x20/0x2a
> [ 485.646405] ? check_flush_dependency+0x110/0x130
> [ 485.652154] ? check_flush_dependency+0x110/0x130
> [ 485.657900] report_bug+0xb8/0x100
> [ 485.662187] ? sched_clock_cpu+0xc/0xb0
> [ 485.666974] do_error_trap+0x9f/0xc0
> [ 485.671464] do_invalid_op+0x36/0x40
> [ 485.675950] ? check_flush_dependency+0x110/0x130
> [ 485.681699] invalid_op+0x28/0x30
> [ 485.685891] RIP: 0010:check_flush_dependency+0x110/0x130
> [ 485.692324] Code: ff ff 48 8b 50 18 48 8d 8b b0 00 00 00 49 89 e8 48 81 c6 b0
> [ 485.714353] RSP: 0018:ffffa9474aea7a48 EFLAGS: 00010086
> [ 485.720724] RAX: 0000000000000000 RBX: ffff912c07c19400 RCX: 0000000000000000
> [ 485.729232] RDX: 0000000000000090 RSI: ffffffffaf67e1f0 RDI: ffffffffaf67bd2c
> [ 485.737737] RBP: ffffffffade8f8d0 R08: ffffffffaf67e160 R09: 000000000002b6c0
> [ 485.746240] R10: 0000017f622837fe R11: 0000000000000e93 R12: ffff9148ad011780
> [ 485.754751] R13: ffff914b3f771700 R14: 0000000000000001 R15: ffff914b30f1c1d0
> [ 485.763261] ? rhashtable_insert_slow+0x470/0x470
> [ 485.769056] ? check_flush_dependency+0x110/0x130
> [ 485.774856] __flush_work+0x96/0x1d0
> [ 485.779376] ? work_busy+0x80/0x80
> [ 485.783681] __cancel_work_timer+0x103/0x190
> [ 485.788950] ? _cond_resched+0x15/0x30
> [ 485.793634] ? _cond_resched+0x15/0x30
> [ 485.798321] ? _cond_resched+0x15/0x30
> [ 485.803008] rhashtable_free_and_destroy+0x20/0x140
> [ 485.808979] del_sw_flow_group+0x45/0x2c0 [mlx5_core]
> [ 485.815119] tree_put_node+0xc3/0x150 [mlx5_core]
> [ 485.820893] mlx5_del_flow_rules+0x11c/0x240 [mlx5_core]
> [ 485.827344] __mlx5_eswitch_del_rule+0x20/0xf0 [mlx5_core]
> [ 485.833978] mlx5_tc_ct_entry_del_rule+0x46/0x90 [mlx5_core]
> [ 485.840813] mlx5_tc_ct_entry_del_rules+0x24/0x40 [mlx5_core]
> [ 485.847750] mlx5_tc_ct_block_flow_offload+0x336/0x9f0 [mlx5_core]
> [ 485.855153] ? __wake_up_common+0x7a/0x180
> [ 485.860230] nf_flow_offload_tuple.isra.23+0xca/0x170 [nf_flow_table]
> [ 485.867929] ? __switch_to_asm+0x40/0x70
> [ 485.872799] ? __switch_to_asm+0x34/0x70
> [ 485.877646] ? __switch_to_asm+0x40/0x70
> [ 485.882466] ? __switch_to_asm+0x34/0x70
> [ 485.887283] ? __switch_to_asm+0x40/0x70
> [ 485.892082] flow_offload_tuple_del+0x30/0x40 [nf_flow_table]
> [ 485.898917] flow_offload_work_handler+0x7f/0x270 [nf_flow_table]
> [ 485.906139] ? finish_task_switch+0x19c/0x2a0
> [ 485.911412] process_one_work+0x1a7/0x370
> [ 485.916299] worker_thread+0x30/0x380
> [ 485.920797] ? process_one_work+0x370/0x370
> [ 485.925876] kthread+0x112/0x130
> [ 485.929883] ? kthread_park+0x80/0x80
> [ 485.934377] ret_from_fork+0x35/0x40
> [ 486.975616] Shutting down cpus with NMI
> [ 486.980297] Kernel Offset: 0x2ca00000 from 0xffffffff81000000 (relocation ra)
> [ 486.997713] ---[ end Kernel panic - not syncing: panic_on_warn set ... ]---
>
> OF flows:
> ovs-ofctl add-flow br0 'in_port=eth4,ip,action=ct(table=10)'
> ovs-ofctl add-flow br0 'in_port=eth4,action=output=bond9'
> ovs-ofctl add-flow br0 'table=10,ct_state=+trk+est,ip,action=output=bond9'
> ovs-ofctl add-flow br0 'table=10,ct_state=+trk+new,tcp,action=ct(commit,table=12)'
> ovs-ofctl add-flow br0 'table=12,action=output=bond9'
>
> ovs-ofctl add-flow br0 'in_port=eth0,ip,action=ct(table=20)'
> ovs-ofctl add-flow br0 'in_port=eth0,action=output=bond9'
> ovs-ofctl add-flow br0 'table=20,ct_state=+trk+est,ip,action=output=bond9'
> ovs-ofctl add-flow br0 'table=20,ct_state=+trk+new,tcp,action=ct(commit,table=22)'
> ovs-ofctl add-flow br0 'table=22,action=output=bond9'
>
> ovs-ofctl add-flow br0 'in_port=bond9,ip,action=ct(table=30)'
> ovs-ofctl add-flow br0 'table=30,ct_state=+trk+est,ip,nw_src=172.0.0.1,action=output=eth0'
> ovs-ofctl add-flow br0 'table=30,ct_state=+trk+est,ip,nw_src=172.0.1.1,action=output=eth4'
> ovs-ofctl add-flow br0 'table=30,ct_state=+trk+new,ip,action=ct(commit,table=32)'
> ovs-ofctl add-flow br0 'table=32,ct_state=+trk,ip,nw_src=172.0.0.1,action=output=eth0'
> ovs-ofctl add-flow br0 'table=32,ct_state=+trk,ip,nw_src=172.0.1.1,action=output=eth4'
>
> ovs-ofctl add-flow br0 'arp,action=NORMAL'
>
Powered by blists - more mailing lists