[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <69516f245575e5ed09b3e291bcd784e2@matoro.tk>
Date: Sun, 02 Oct 2022 15:24:57 -0400
From: matoro <matoro_mailinglist_kernel@...oro.tk>
To: stefanc@...vell.com
Cc: netdev@...r.kernel.org, thomas.petazzoni@...tlin.com,
davem@...emloft.net, nadavh@...vell.com, ymarkman@...vell.com,
linux-kernel@...r.kernel.org, kuba@...nel.org,
linux@...linux.org.uk, mw@...ihalf.com, andrew@...n.ch,
rmk+kernel@...linux.org.uk, atenart@...nel.org, jon@...id-run.com
Subject: Re: [net-next] net: mvpp2: Add TX flow control support for jumbo
frames
Hi all, I know this is kind of an old change but I have got an issue
with this driver that might be related. Whenever I change the MTU of
any interfacing using mvpp2, it immediately results in the following
crash. Is this change related? Is this a known issue with this driver?
[ 1725.925804] mvpp2 f2000000.ethernet eth0: mtu 9000 too high,
switching to shared buffers
[ 1725.9258[ 1725.935611] mvpp2 f2000000.ethernet eth0: Link is Down
04] mvpp2 f2000000.ethernet eth0: mtu 9000 too high, switching to shared
buffers
[ 17[ 1725.950079] Unable to handle kernel NULL pointer dereference at
virtual address 0000000000000000
25.935611] Mem abort info:
[33mmvpp2 f20000[ 1725.963804] ESR = 0x0000000096000004
00.ethernet eth0[ 1725.968960] EC = 0x25: DABT (current EL), IL = 32
bits
: Link is Do[ 1725.975685] SET = 0, FnV = 0
wn
[ 1725.980143] EA = 0, S1PTW = 0
[ 1725.983643] FSC = 0x04: level 0 translation fault
[ 1725.988539] Data abort info:
[ 1725.991430] ISV = 0, ISS = 0x00000004
[ 1725.995279] CM = 0, WnR = 0
[ 1725.998256] user pgtable: 4k pages, 48-bit VAs, pgdp=0000000104b83000
[ 1726.004724] [0000000000000000] pgd=0000000000000000,
p4d=0000000000000000
[ 1726.011543] Internal error: Oops: 96000004 [#1] PREEMPT SMP
[ 1726.017137] Modules linked in: sfp mdio_i2c marvell mcp3021 mvmdio
at24 mvpp2 armada_thermal phylink sbsa_gwdt cfg80211 rfkill fuse
[ 1726.029032] CPU: 2 PID: 16253 Comm: ip Not tainted
5.19.8-1-aarch64-ARCH #1
[ 1726.036024] Hardware name: SolidRun CN9130 based SOM Clearfog Base
(DT)
[ 1726.042665] pstate: 800000c5 (Nzcv daIF -PAN -UAO -TCO -DIT -SSBS
BTYPE=--)
[ 1726.049656] pc : mvpp2_cm3_read.isra.0+0x8/0x2c [mvpp2]
[ 1726.054915] lr : mvpp2_bm_pool_update_fc+0x40/0x154 [mvpp2]
[ 1726.060515] sp : ffff80000b17b580
[ 1726.063842] x29: ffff80000b17b580 x28: 0000000000000000 x27:
0000000000000000
[ 1726.071010] x26: ffff8000013ceb60 x25: 0000000000000008 x24:
ffff0001054b5980
[ 1726.078177] x23: ffff0001021e2480 x22: 0000000000000038 x21:
0000000000000000
[ 1726.085346] x20: ffff0001049dac80 x19: ffff0001054b4980 x18:
0000000000000000
[ 1726.092513] x17: 0000000000000000 x16: 0000000000000000 x15:
0000000000000000
[ 1726.099680] x14: 0000000000000109 x13: 0000000000000109 x12:
0000000000000000
[ 1726.106847] x11: 0000000000000040 x10: ffff80000a3471b8 x9 :
ffff80000a3471b0
[ 1726.114015] x8 : ffff000100401b88 x7 : 0000000000000000 x6 :
0000000000000000
[ 1726.121182] x5 : ffff80000b17b4e0 x4 : 0000000000000000 x3 :
0000000000000000
[ 1726.128348] x2 : ffff0001021e2480 x1 : 0000000000000000 x0 :
0000000000000000
[ 1726.135514] Call trace:
[ 1726.137968] mvpp2_cm3_read.isra.0+0x8/0x2c [mvpp2]
[ 1726.142871] mvpp2_bm_pool_update_priv_fc+0xc0/0x100 [mvpp2]
[ 1726.148558] mvpp2_bm_switch_buffers.isra.0+0x1c0/0x1e0 [mvpp2]
[ 1726.154506] mvpp2_change_mtu+0x184/0x264 [mvpp2]
[ 1726.159233] dev_set_mtu_ext+0xdc/0x1b4
[ 1726.163087] do_setlink+0x1d4/0xa90
[ 1726.166593] __rtnl_newlink+0x4a8/0x4f0
[ 1726.170443] rtnl_newlink+0x4c/0x80
[ 1726.173944] rtnetlink_rcv_msg+0x12c/0x37c
[ 1726.178058] netlink_rcv_skb+0x5c/0x130
[ 1726.181910] rtnetlink_rcv+0x18/0x2c
[ 1726.185500] netlink_unicast+0x2c4/0x31c
[ 1726.189438] netlink_sendmsg+0x1bc/0x410
[ 1726.193377] sock_sendmsg+0x54/0x60
[ 1726.196879] ____sys_sendmsg+0x26c/0x290
[ 1726.200817] ___sys_sendmsg+0x7c/0xc0
[ 1726.204494] __sys_sendmsg+0x68/0xd0
[ 1726.208083] __arm64_sys_sendmsg+0x28/0x34
[ 1726.212196] invoke_syscall+0x48/0x114
[ 1726.215962] el0_svc_common.constprop.0+0x44/0xec
[ 1726.220686] do_el0_svc+0x28/0x34
[ 1726.224014] el0_svc+0x2c/0x84
[ 1726.227082] el0t_64_sync_handler+0x11c/0x150
[ 1726.231455] el0t_64_sync+0x18c/0x190
[ 1726.235134] Code: d65f03c0 d65f03c0 d503233f 8b214000 (b9400000)
[ 1726.241253] ---[ end trace 0000000000000000 ]---
[ 1726.245888] note: ip[16253] exited with preempt_count 1
-------- Original Message --------
Subject: [net-next] net: mvpp2: Add TX flow control support for jumbo
frames
Date: 2021-02-15 10:23
From: <stefanc@...vell.com>
To: <netdev@...r.kernel.org>
From: Stefan Chulski <stefanc@...vell.com>
With MTU less than 1500B on all ports, the driver uses per CPU pool
mode.
If one of the ports set to jumbo frame MTU size, all ports move
to shared pools mode.
Here, buffer manager TX Flow Control reconfigured on all ports.
Signed-off-by: Stefan Chulski <stefanc@...vell.com>
---
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c | 26
++++++++++++++++++++
1 file changed, 26 insertions(+)
diff --git a/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
b/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
index 222e9a3..10c17d1 100644
--- a/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
+++ b/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
@@ -924,6 +924,25 @@ static void mvpp2_bm_pool_update_fc(struct
mvpp2_port *port,
spin_unlock_irqrestore(&port->priv->mss_spinlock, flags);
}
+/* disable/enable flow control for BM pool on all ports */
+static void mvpp2_bm_pool_update_priv_fc(struct mvpp2 *priv, bool en)
+{
+ struct mvpp2_port *port;
+ int i;
+
+ for (i = 0; i < priv->port_count; i++) {
+ port = priv->port_list[i];
+ if (port->priv->percpu_pools) {
+ for (i = 0; i < port->nrxqs; i++)
+ mvpp2_bm_pool_update_fc(port, &port->priv->bm_pools[i],
+ port->tx_fc & en);
+ } else {
+ mvpp2_bm_pool_update_fc(port, port->pool_long, port->tx_fc & en);
+ mvpp2_bm_pool_update_fc(port, port->pool_short, port->tx_fc & en);
+ }
+ }
+}
+
static int mvpp2_enable_global_fc(struct mvpp2 *priv)
{
int val, timeout = 0;
@@ -4913,6 +4932,7 @@ static int mvpp2_set_mac_address(struct net_device
*dev, void *p)
*/
static int mvpp2_bm_switch_buffers(struct mvpp2 *priv, bool percpu)
{
+ bool change_percpu = (percpu != priv->percpu_pools);
int numbufs = MVPP2_BM_POOLS_NUM, i;
struct mvpp2_port *port = NULL;
bool status[MVPP2_MAX_PORTS];
@@ -4928,6 +4948,9 @@ static int mvpp2_bm_switch_buffers(struct mvpp2
*priv, bool percpu)
if (priv->percpu_pools)
numbufs = port->nrxqs * 2;
+ if (change_percpu)
+ mvpp2_bm_pool_update_priv_fc(priv, false);
+
for (i = 0; i < numbufs; i++)
mvpp2_bm_pool_destroy(port->dev->dev.parent, priv,
&priv->bm_pools[i]);
@@ -4942,6 +4965,9 @@ static int mvpp2_bm_switch_buffers(struct mvpp2
*priv, bool percpu)
mvpp2_open(port->dev);
}
+ if (change_percpu)
+ mvpp2_bm_pool_update_priv_fc(priv, true);
+
return 0;
}
Powered by blists - more mailing lists