lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <20200227142259.GF26061@nanopsycho>
Date:   Thu, 27 Feb 2020 15:22:59 +0100
From:   Jiri Pirko <jiri@...nulli.us>
To:     Vadym Kochan <vadym.kochan@...ision.eu>
Cc:     "linux-kernel@...r.kernel.org" <linux-kernel@...r.kernel.org>,
        "netdev@...r.kernel.org" <netdev@...r.kernel.org>,
        "David S . Miller" <davem@...emloft.net>,
        Oleksandr Mazur <oleksandr.mazur@...ision.eu>,
        Taras Chornyi <taras.chornyi@...ision.eu>,
        Serhiy Boiko <serhiy.boiko@...ision.eu>,
        Andrii Savka <andrii.savka@...ision.eu>,
        Volodymyr Mytnyk <volodymyr.mytnyk@...ision.eu>,
        Serhiy Pshyk <serhiy.pshyk@...ision.eu>
Subject: Re: [RFC net-next 1/3] net: marvell: prestera: Add Switchdev driver
 for Prestera family ASIC device 98DX325x (AC3x)

Tue, Feb 25, 2020 at 05:30:54PM CET, vadym.kochan@...ision.eu wrote:
>Marvell Prestera 98DX326x integrates up to 24 ports of 1GbE with 8
>ports of 10GbE uplinks or 2 ports of 40Gbps stacking for a largely
>wireless SMB deployment.
>
>This driver implementation includes only L1 & basic L2 support.
>
>The core Prestera switching logic is implemented in prestera.c, there is
>an intermediate hw layer between core logic and firmware. It is
>implemented in prestera_hw.c, the purpose of it is to encapsulate hw
>related logic, in future there is a plan to support more devices with
>different HW related configurations.
>
>The following Switchdev features are supported:
>
>    - VLAN-aware bridge offloading
>    - VLAN-unaware bridge offloading
>    - FDB offloading (learning, ageing)
>    - Switchport configuration
>
>Signed-off-by: Vadym Kochan <vadym.kochan@...ision.eu>
>Signed-off-by: Andrii Savka <andrii.savka@...ision.eu>
>Signed-off-by: Oleksandr Mazur <oleksandr.mazur@...ision.eu>
>Signed-off-by: Serhiy Boiko <serhiy.boiko@...ision.eu>
>Signed-off-by: Serhiy Pshyk <serhiy.pshyk@...ision.eu>
>Signed-off-by: Taras Chornyi <taras.chornyi@...ision.eu>
>Signed-off-by: Volodymyr Mytnyk <volodymyr.mytnyk@...ision.eu>
>---
> drivers/net/ethernet/marvell/Kconfig          |    1 +
> drivers/net/ethernet/marvell/Makefile         |    1 +
> drivers/net/ethernet/marvell/prestera/Kconfig |   13 +
> .../net/ethernet/marvell/prestera/Makefile    |    3 +
> .../net/ethernet/marvell/prestera/prestera.c  | 1502 +++++++++++++++++
> .../net/ethernet/marvell/prestera/prestera.h  |  244 +++
> .../marvell/prestera/prestera_drv_ver.h       |   23 +
> .../ethernet/marvell/prestera/prestera_hw.c   | 1094 ++++++++++++
> .../ethernet/marvell/prestera/prestera_hw.h   |  159 ++
> .../marvell/prestera/prestera_switchdev.c     | 1217 +++++++++++++
> 10 files changed, 4257 insertions(+)
> create mode 100644 drivers/net/ethernet/marvell/prestera/Kconfig
> create mode 100644 drivers/net/ethernet/marvell/prestera/Makefile
> create mode 100644 drivers/net/ethernet/marvell/prestera/prestera.c
> create mode 100644 drivers/net/ethernet/marvell/prestera/prestera.h
> create mode 100644 drivers/net/ethernet/marvell/prestera/prestera_drv_ver.h
> create mode 100644 drivers/net/ethernet/marvell/prestera/prestera_hw.c
> create mode 100644 drivers/net/ethernet/marvell/prestera/prestera_hw.h
> create mode 100644 drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
>
>diff --git a/drivers/net/ethernet/marvell/Kconfig b/drivers/net/ethernet/marvell/Kconfig
>index 3d5caea096fb..74313d9e1fc0 100644
>--- a/drivers/net/ethernet/marvell/Kconfig
>+++ b/drivers/net/ethernet/marvell/Kconfig
>@@ -171,5 +171,6 @@ config SKY2_DEBUG
> 
> 
> source "drivers/net/ethernet/marvell/octeontx2/Kconfig"
>+source "drivers/net/ethernet/marvell/prestera/Kconfig"
> 
> endif # NET_VENDOR_MARVELL
>diff --git a/drivers/net/ethernet/marvell/Makefile b/drivers/net/ethernet/marvell/Makefile
>index 89dea7284d5b..9f88fe822555 100644
>--- a/drivers/net/ethernet/marvell/Makefile
>+++ b/drivers/net/ethernet/marvell/Makefile
>@@ -12,3 +12,4 @@ obj-$(CONFIG_PXA168_ETH) += pxa168_eth.o
> obj-$(CONFIG_SKGE) += skge.o
> obj-$(CONFIG_SKY2) += sky2.o
> obj-y		+= octeontx2/
>+obj-y		+= prestera/
>diff --git a/drivers/net/ethernet/marvell/prestera/Kconfig b/drivers/net/ethernet/marvell/prestera/Kconfig
>new file mode 100644
>index 000000000000..d0b416dcb677
>--- /dev/null
>+++ b/drivers/net/ethernet/marvell/prestera/Kconfig
>@@ -0,0 +1,13 @@
>+# SPDX-License-Identifier: GPL-2.0-only
>+#
>+# Marvell Prestera drivers configuration
>+#
>+
>+config PRESTERA
>+	tristate "Marvell Prestera Switch ASICs support"
>+	depends on NET_SWITCHDEV && VLAN_8021Q
>+	---help---
>+	  This driver supports Marvell Prestera Switch ASICs family.
>+
>+	  To compile this driver as a module, choose M here: the
>+	  module will be called prestera_sw.
>diff --git a/drivers/net/ethernet/marvell/prestera/Makefile b/drivers/net/ethernet/marvell/prestera/Makefile
>new file mode 100644
>index 000000000000..9446298fb7f4
>--- /dev/null
>+++ b/drivers/net/ethernet/marvell/prestera/Makefile
>@@ -0,0 +1,3 @@
>+# SPDX-License-Identifier: GPL-2.0
>+obj-$(CONFIG_PRESTERA)	+= prestera_sw.o
>+prestera_sw-objs	:= prestera.o prestera_hw.o prestera_switchdev.o
>diff --git a/drivers/net/ethernet/marvell/prestera/prestera.c b/drivers/net/ethernet/marvell/prestera/prestera.c
>new file mode 100644
>index 000000000000..12d0eb590bbb
>--- /dev/null
>+++ b/drivers/net/ethernet/marvell/prestera/prestera.c
>@@ -0,0 +1,1502 @@
>+/* SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0


# SPDX-License-Identifier: GPL-2.0-only
# SPDX-License-Identifier: GPL-2.0
/* SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0

You have to make up your mind :)




>+ *
>+ * Copyright (c) 2019-2020 Marvell International Ltd. All rights reserved.
>+ *
>+ */
>+#include <linux/kernel.h>
>+#include <linux/module.h>
>+#include <linux/list.h>
>+#include <linux/netdevice.h>
>+#include <linux/netdev_features.h>
>+#include <linux/etherdevice.h>
>+#include <linux/ethtool.h>
>+#include <linux/jiffies.h>
>+#include <net/switchdev.h>
>+
>+#include "prestera.h"
>+#include "prestera_hw.h"
>+#include "prestera_drv_ver.h"
>+
>+#define MVSW_PR_MTU_DEFAULT 1536
>+
>+#define PORT_STATS_CACHE_TIMEOUT_MS	(msecs_to_jiffies(1000))
>+#define PORT_STATS_CNT	(sizeof(struct mvsw_pr_port_stats) / sizeof(u64))
>+#define PORT_STATS_IDX(name) \
>+	(offsetof(struct mvsw_pr_port_stats, name) / sizeof(u64))
>+#define PORT_STATS_FIELD(name)	\
>+	[PORT_STATS_IDX(name)] = __stringify(name)
>+
>+static struct list_head switches_registered;

Avoid this global list. You don't use it anyway.


>+
>+static const char mvsw_driver_kind[] = "prestera_sw";
>+static const char mvsw_driver_name[] = "mvsw_switchdev";
>+static const char mvsw_driver_version[] = PRESTERA_DRV_VER;
>+
>+#define mvsw_dev(sw)		((sw)->dev->dev)
>+#define mvsw_dev_name(sw)	dev_name((sw)->dev->dev)
>+
>+static struct workqueue_struct *mvsw_pr_wq;
>+
>+struct mvsw_pr_link_mode {
>+	enum ethtool_link_mode_bit_indices eth_mode;
>+	u32 speed;
>+	u64 pr_mask;
>+	u8 duplex;
>+	u8 port_type;
>+};
>+
>+static const struct mvsw_pr_link_mode
>+mvsw_pr_link_modes[MVSW_LINK_MODE_MAX] = {
>+	[MVSW_LINK_MODE_10baseT_Half_BIT] = {
>+		.eth_mode =  ETHTOOL_LINK_MODE_10baseT_Half_BIT,
>+		.speed = 10,
>+		.pr_mask = 1 << MVSW_LINK_MODE_10baseT_Half_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_HALF,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_10baseT_Full_BIT] = {
>+		.eth_mode =  ETHTOOL_LINK_MODE_10baseT_Full_BIT,
>+		.speed = 10,
>+		.pr_mask = 1 << MVSW_LINK_MODE_10baseT_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_100baseT_Half_BIT] = {
>+		.eth_mode =  ETHTOOL_LINK_MODE_100baseT_Half_BIT,
>+		.speed = 100,
>+		.pr_mask = 1 << MVSW_LINK_MODE_100baseT_Half_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_HALF,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_100baseT_Full_BIT] = {
>+		.eth_mode =  ETHTOOL_LINK_MODE_100baseT_Full_BIT,
>+		.speed = 100,
>+		.pr_mask = 1 << MVSW_LINK_MODE_100baseT_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_1000baseT_Half_BIT] = {
>+		.eth_mode =  ETHTOOL_LINK_MODE_1000baseT_Half_BIT,
>+		.speed = 1000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_1000baseT_Half_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_HALF,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_1000baseT_Full_BIT] = {
>+		.eth_mode =  ETHTOOL_LINK_MODE_1000baseT_Full_BIT,
>+		.speed = 1000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_1000baseT_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_1000baseX_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_1000baseX_Full_BIT,
>+		.speed = 1000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_1000baseX_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_FIBRE,
>+	},
>+	[MVSW_LINK_MODE_1000baseKX_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_1000baseKX_Full_BIT,
>+		.speed = 1000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_1000baseKX_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_10GbaseKR_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_10000baseKR_Full_BIT,
>+		.speed = 10000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_10GbaseKR_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_10GbaseSR_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_10000baseSR_Full_BIT,
>+		.speed = 10000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_10GbaseSR_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_FIBRE,
>+	},
>+	[MVSW_LINK_MODE_10GbaseLR_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_10000baseLR_Full_BIT,
>+		.speed = 10000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_10GbaseLR_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_FIBRE,
>+	},
>+	[MVSW_LINK_MODE_20GbaseKR2_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_20000baseKR2_Full_BIT,
>+		.speed = 20000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_20GbaseKR2_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_25GbaseCR_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_25000baseCR_Full_BIT,
>+		.speed = 25000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_25GbaseCR_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_DA,
>+	},
>+	[MVSW_LINK_MODE_25GbaseKR_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_25000baseKR_Full_BIT,
>+		.speed = 25000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_25GbaseKR_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_25GbaseSR_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_25000baseSR_Full_BIT,
>+		.speed = 25000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_25GbaseSR_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_FIBRE,
>+	},
>+	[MVSW_LINK_MODE_40GbaseKR4_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_40000baseKR4_Full_BIT,
>+		.speed = 40000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_40GbaseKR4_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_40GbaseCR4_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_40000baseCR4_Full_BIT,
>+		.speed = 40000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_40GbaseCR4_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_DA,
>+	},
>+	[MVSW_LINK_MODE_40GbaseSR4_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_40000baseSR4_Full_BIT,
>+		.speed = 40000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_40GbaseSR4_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_FIBRE,
>+	},
>+	[MVSW_LINK_MODE_50GbaseCR2_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_50000baseCR2_Full_BIT,
>+		.speed = 50000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_50GbaseCR2_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_DA,
>+	},
>+	[MVSW_LINK_MODE_50GbaseKR2_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_50000baseKR2_Full_BIT,
>+		.speed = 50000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_50GbaseKR2_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_50GbaseSR2_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_50000baseSR2_Full_BIT,
>+		.speed = 50000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_50GbaseSR2_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_FIBRE,
>+	},
>+	[MVSW_LINK_MODE_100GbaseKR4_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_100000baseKR4_Full_BIT,
>+		.speed = 100000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_100GbaseKR4_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_TP,
>+	},
>+	[MVSW_LINK_MODE_100GbaseSR4_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_100000baseSR4_Full_BIT,
>+		.speed = 100000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_100GbaseSR4_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_FIBRE,
>+	},
>+	[MVSW_LINK_MODE_100GbaseCR4_Full_BIT] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_100000baseCR4_Full_BIT,
>+		.speed = 100000,
>+		.pr_mask = 1 << MVSW_LINK_MODE_100GbaseCR4_Full_BIT,
>+		.duplex = MVSW_PORT_DUPLEX_FULL,
>+		.port_type = MVSW_PORT_TYPE_DA,
>+	}
>+};
>+
>+struct mvsw_pr_fec {
>+	u32 eth_fec;
>+	enum ethtool_link_mode_bit_indices eth_mode;
>+	u8 pr_fec;
>+};
>+
>+static const struct mvsw_pr_fec mvsw_pr_fec_caps[MVSW_PORT_FEC_MAX] = {
>+	[MVSW_PORT_FEC_OFF_BIT] = {
>+		.eth_fec = ETHTOOL_FEC_OFF,
>+		.eth_mode = ETHTOOL_LINK_MODE_FEC_NONE_BIT,
>+		.pr_fec = 1 << MVSW_PORT_FEC_OFF_BIT,
>+	},
>+	[MVSW_PORT_FEC_BASER_BIT] = {
>+		.eth_fec = ETHTOOL_FEC_BASER,
>+		.eth_mode = ETHTOOL_LINK_MODE_FEC_BASER_BIT,
>+		.pr_fec = 1 << MVSW_PORT_FEC_BASER_BIT,
>+	},
>+	[MVSW_PORT_FEC_RS_BIT] = {
>+		.eth_fec = ETHTOOL_FEC_RS,
>+		.eth_mode = ETHTOOL_LINK_MODE_FEC_RS_BIT,
>+		.pr_fec = 1 << MVSW_PORT_FEC_RS_BIT,
>+	}
>+};
>+
>+struct mvsw_pr_port_type {
>+	enum ethtool_link_mode_bit_indices eth_mode;
>+	u8 eth_type;
>+};
>+
>+static const struct mvsw_pr_port_type
>+mvsw_pr_port_types[MVSW_PORT_TYPE_MAX] = {
>+	[MVSW_PORT_TYPE_NONE] = {
>+		.eth_mode = __ETHTOOL_LINK_MODE_MASK_NBITS,
>+		.eth_type = PORT_NONE,
>+	},
>+	[MVSW_PORT_TYPE_TP] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_TP_BIT,
>+		.eth_type = PORT_TP,
>+	},
>+	[MVSW_PORT_TYPE_AUI] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_AUI_BIT,
>+		.eth_type = PORT_AUI,
>+	},
>+	[MVSW_PORT_TYPE_MII] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_MII_BIT,
>+		.eth_type = PORT_MII,
>+	},
>+	[MVSW_PORT_TYPE_FIBRE] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_FIBRE_BIT,
>+		.eth_type = PORT_FIBRE,
>+	},
>+	[MVSW_PORT_TYPE_BNC] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_BNC_BIT,
>+		.eth_type = PORT_BNC,
>+	},
>+	[MVSW_PORT_TYPE_DA] = {
>+		.eth_mode = ETHTOOL_LINK_MODE_TP_BIT,
>+		.eth_type = PORT_TP,
>+	},
>+	[MVSW_PORT_TYPE_OTHER] = {
>+		.eth_mode = __ETHTOOL_LINK_MODE_MASK_NBITS,
>+		.eth_type = PORT_OTHER,
>+	}
>+};
>+
>+static const char mvsw_pr_port_cnt_name[PORT_STATS_CNT][ETH_GSTRING_LEN] = {
>+	PORT_STATS_FIELD(good_octets_received),
>+	PORT_STATS_FIELD(bad_octets_received),
>+	PORT_STATS_FIELD(mac_trans_error),
>+	PORT_STATS_FIELD(broadcast_frames_received),
>+	PORT_STATS_FIELD(multicast_frames_received),
>+	PORT_STATS_FIELD(frames_64_octets),
>+	PORT_STATS_FIELD(frames_65_to_127_octets),
>+	PORT_STATS_FIELD(frames_128_to_255_octets),
>+	PORT_STATS_FIELD(frames_256_to_511_octets),
>+	PORT_STATS_FIELD(frames_512_to_1023_octets),
>+	PORT_STATS_FIELD(frames_1024_to_max_octets),
>+	PORT_STATS_FIELD(excessive_collision),
>+	PORT_STATS_FIELD(multicast_frames_sent),
>+	PORT_STATS_FIELD(broadcast_frames_sent),
>+	PORT_STATS_FIELD(fc_sent),
>+	PORT_STATS_FIELD(fc_received),
>+	PORT_STATS_FIELD(buffer_overrun),
>+	PORT_STATS_FIELD(undersize),
>+	PORT_STATS_FIELD(fragments),
>+	PORT_STATS_FIELD(oversize),
>+	PORT_STATS_FIELD(jabber),
>+	PORT_STATS_FIELD(rx_error_frame_received),
>+	PORT_STATS_FIELD(bad_crc),
>+	PORT_STATS_FIELD(collisions),
>+	PORT_STATS_FIELD(late_collision),
>+	PORT_STATS_FIELD(unicast_frames_received),
>+	PORT_STATS_FIELD(unicast_frames_sent),
>+	PORT_STATS_FIELD(sent_multiple),
>+	PORT_STATS_FIELD(sent_deferred),
>+	PORT_STATS_FIELD(frames_1024_to_1518_octets),
>+	PORT_STATS_FIELD(frames_1519_to_max_octets),
>+	PORT_STATS_FIELD(good_octets_sent),
>+};
>+
>+static struct mvsw_pr_port *__find_pr_port(const struct mvsw_pr_switch *sw,
>+					   u32 port_id)
>+{
>+	struct mvsw_pr_port *port;
>+
>+	list_for_each_entry(port, &sw->port_list, list) {
>+		if (port->id == port_id)
>+			return port;
>+	}
>+
>+	return NULL;
>+}
>+
>+static int mvsw_pr_port_state_set(struct net_device *dev, bool is_up)
>+{
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+	int err;
>+
>+	if (!is_up)
>+		netif_stop_queue(dev);
>+
>+	err = mvsw_pr_hw_port_state_set(port, is_up);
>+
>+	if (is_up && !err)
>+		netif_start_queue(dev);
>+
>+	return err;
>+}
>+
>+static int mvsw_pr_port_get_port_parent_id(struct net_device *dev,
>+					   struct netdev_phys_item_id *ppid)
>+{
>+	const struct mvsw_pr_port *port = netdev_priv(dev);
>+
>+	ppid->id_len = sizeof(port->sw->id);
>+
>+	memcpy(&ppid->id, &port->sw->id, ppid->id_len);
>+	return 0;
>+}
>+
>+static int mvsw_pr_port_get_phys_port_name(struct net_device *dev,
>+					   char *buf, size_t len)

Don't implement this please. Just implement basic devlink and devlink
port support, devlink is going to take care of the netdevice names.


>+{
>+	const struct mvsw_pr_port *port = netdev_priv(dev);
>+
>+	snprintf(buf, len, "%u", port->fp_id);
>+	return 0;
>+}
>+
>+static int mvsw_pr_port_open(struct net_device *dev)
>+{
>+	return mvsw_pr_port_state_set(dev, true);
>+}
>+
>+static int mvsw_pr_port_close(struct net_device *dev)
>+{
>+	return mvsw_pr_port_state_set(dev, false);
>+}
>+
>+static netdev_tx_t mvsw_pr_port_xmit(struct sk_buff *skb,
>+				     struct net_device *dev)
>+{

You need to implement this function. In fact, that is the basic
functionality of a netdevice, to transmit and receive traffic.

As Andy suggested, first implement 24 netdevices with slowpath only,
then add offloading and other features.


>+	dev_kfree_skb(skb);
>+	return NETDEV_TX_OK;
>+}
>+
>+static int mvsw_is_valid_mac_addr(struct mvsw_pr_port *port, u8 *addr)
>+{
>+	int err;
>+
>+	if (!is_valid_ether_addr(addr))
>+		return -EADDRNOTAVAIL;
>+
>+	err = memcmp(port->sw->base_mac, addr, ETH_ALEN - 1);

For memcmp, as it does not return 0/-ESOMETHING, you can do:
	if (memcmp...


>+	if (err)
>+		return -EINVAL;
>+
>+	return 0;
>+}
>+
>+static int mvsw_pr_port_set_mac_address(struct net_device *dev, void *p)
>+{
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+	struct sockaddr *addr = p;
>+	int err;
>+
>+	err = mvsw_is_valid_mac_addr(port, addr->sa_data);
>+	if (err)
>+		return err;
>+
>+	err = mvsw_pr_hw_port_mac_set(port, addr->sa_data);
>+	if (!err)

do error path in if:
	if (err)
		return err;


>+		memcpy(dev->dev_addr, addr->sa_data, dev->addr_len);
>+
>+	return err;
>+}
>+
>+static int mvsw_pr_port_change_mtu(struct net_device *dev, int mtu)
>+{
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+	int err;
>+
>+	if (port->sw->mtu_min <= mtu && mtu <= port->sw->mtu_max)
>+		err = mvsw_pr_hw_port_mtu_set(port, mtu);
>+	else
>+		err = -EINVAL;
>+
>+	if (!err)
>+		dev->mtu = mtu;
>+
>+	return err;

How about rather:

	if (mtu < port->sw->mtu_min || mtu > port->sw->mtu_max)
		return -EINVAL;
	err = mvsw_pr_hw_port_mtu_set(port, mtu);
	if (err)
		return err;
	dev->mtu = mtu;
	return 0;


Btw, since you have dev->mtu_min/mtu_max set, you can avoid checking it
here. dev_validate_mtu() will do it for you.



>+}
>+
>+static void mvsw_pr_port_get_stats64(struct net_device *dev,
>+				     struct rtnl_link_stats64 *stats)
>+{
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+	struct mvsw_pr_port_stats *port_stats = &port->cached_hw_stats.stats;
>+
>+	stats->rx_packets =	port_stats->broadcast_frames_received +

Don't use tab after "="


>+				port_stats->multicast_frames_received +
>+				port_stats->unicast_frames_received;
>+
>+	stats->tx_packets =	port_stats->broadcast_frames_sent +
>+				port_stats->multicast_frames_sent +
>+				port_stats->unicast_frames_sent;
>+
>+	stats->rx_bytes = port_stats->good_octets_received;
>+
>+	stats->tx_bytes = port_stats->good_octets_sent;
>+
>+	stats->rx_errors = port_stats->rx_error_frame_received;
>+	stats->tx_errors = port_stats->mac_trans_error;
>+
>+	stats->rx_dropped = port_stats->buffer_overrun;
>+	stats->tx_dropped = 0;
>+
>+	stats->multicast = port_stats->multicast_frames_received;
>+	stats->collisions = port_stats->excessive_collision;
>+
>+	stats->rx_crc_errors = port_stats->bad_crc;
>+}
>+
>+static void mvsw_pr_port_get_hw_stats(struct mvsw_pr_port *port)
>+{
>+	mvsw_pr_hw_port_stats_get(port, &port->cached_hw_stats.stats);
>+}
>+
>+static void update_stats_cache(struct work_struct *work)
>+{
>+	struct mvsw_pr_port *port =
>+		container_of(work, struct mvsw_pr_port,
>+			     cached_hw_stats.caching_dw.work);
>+
>+	mvsw_pr_port_get_hw_stats(port);
>+
>+	queue_delayed_work(mvsw_pr_wq, &port->cached_hw_stats.caching_dw,
>+			   PORT_STATS_CACHE_TIMEOUT_MS);
>+}
>+
>+static void mvsw_pr_port_get_drvinfo(struct net_device *dev,
>+				     struct ethtool_drvinfo *drvinfo)
>+{
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+	struct mvsw_pr_switch *sw = port->sw;
>+
>+	strlcpy(drvinfo->driver, mvsw_driver_kind, sizeof(drvinfo->driver));
>+	strlcpy(drvinfo->bus_info, mvsw_dev_name(sw), sizeof(drvinfo->bus_info));
>+	snprintf(drvinfo->fw_version, sizeof(drvinfo->fw_version),
>+		 "%d.%d.%d",

Unnecessary wrapping.


>+		 sw->dev->fw_rev.maj,
>+		 sw->dev->fw_rev.min,
>+		 sw->dev->fw_rev.sub);
>+}
>+
>+static const struct net_device_ops mvsw_pr_netdev_ops = {
>+	.ndo_open = mvsw_pr_port_open,
>+	.ndo_stop = mvsw_pr_port_close,
>+	.ndo_start_xmit = mvsw_pr_port_xmit,
>+	.ndo_change_mtu = mvsw_pr_port_change_mtu,
>+	.ndo_get_stats64 = mvsw_pr_port_get_stats64,
>+	.ndo_set_mac_address = mvsw_pr_port_set_mac_address,
>+	.ndo_get_phys_port_name = mvsw_pr_port_get_phys_port_name,
>+	.ndo_get_port_parent_id = mvsw_pr_port_get_port_parent_id
>+};
>+
>+bool mvsw_pr_netdev_check(const struct net_device *dev)
>+{
>+	return dev->netdev_ops == &mvsw_pr_netdev_ops;
>+}
>+
>+static int mvsw_pr_lower_dev_walk(struct net_device *lower_dev, void *data)
>+{
>+	struct mvsw_pr_port **pport = data;
>+
>+	if (mvsw_pr_netdev_check(lower_dev)) {
>+		*pport = netdev_priv(lower_dev);
>+		return 1;
>+	}
>+
>+	return 0;
>+}
>+
>+struct mvsw_pr_port *mvsw_pr_port_dev_lower_find(struct net_device *dev)
>+{
>+	struct mvsw_pr_port *port;
>+
>+	if (mvsw_pr_netdev_check(dev))
>+		return netdev_priv(dev);
>+
>+	port = NULL;
>+	netdev_walk_all_lower_dev(dev, mvsw_pr_lower_dev_walk, &port);
>+
>+	return port;
>+}
>+
>+static void mvsw_modes_to_eth(unsigned long *eth_modes, u64 link_modes, u8 fec,
>+			      u8 type)
>+{
>+	u32 mode;
>+
>+	for (mode = 0; mode < MVSW_LINK_MODE_MAX; mode++) {
>+		if ((mvsw_pr_link_modes[mode].pr_mask & link_modes) == 0)
>+			continue;
>+		if (type != MVSW_PORT_TYPE_NONE &&
>+		    mvsw_pr_link_modes[mode].port_type != type)
>+			continue;
>+		__set_bit(mvsw_pr_link_modes[mode].eth_mode, eth_modes);
>+	}
>+
>+	for (mode = 0; mode < MVSW_PORT_FEC_MAX; mode++) {
>+		if ((mvsw_pr_fec_caps[mode].pr_fec & fec) == 0)
>+			continue;
>+		__set_bit(mvsw_pr_fec_caps[mode].eth_mode, eth_modes);
>+	}
>+}
>+
>+static void mvsw_modes_from_eth(const unsigned long *eth_modes, u64 *link_modes,
>+				u8 *fec)
>+{
>+	u32 mode;
>+
>+	for (mode = 0; mode < MVSW_LINK_MODE_MAX; mode++) {
>+		if (!test_bit(mvsw_pr_link_modes[mode].eth_mode, eth_modes))
>+			continue;
>+		*link_modes |= mvsw_pr_link_modes[mode].pr_mask;
>+	}
>+
>+	for (mode = 0; mode < MVSW_PORT_FEC_MAX; mode++) {
>+		if (!test_bit(mvsw_pr_fec_caps[mode].eth_mode, eth_modes))
>+			continue;
>+		*fec |= mvsw_pr_fec_caps[mode].pr_fec;
>+	}
>+}
>+
>+static void mvsw_pr_port_supp_types_get(struct ethtool_link_ksettings *ecmd,
>+					struct mvsw_pr_port *port)
>+{
>+	u32 mode;
>+	u8 ptype;
>+
>+	for (mode = 0; mode < MVSW_LINK_MODE_MAX; mode++) {
>+		if ((mvsw_pr_link_modes[mode].pr_mask &
>+		    port->caps.supp_link_modes) == 0)
>+			continue;
>+		ptype = mvsw_pr_link_modes[mode].port_type;
>+		__set_bit(mvsw_pr_port_types[ptype].eth_mode,
>+			  ecmd->link_modes.supported);
>+	}
>+}
>+
>+static void mvsw_pr_port_speed_get(struct ethtool_link_ksettings *ecmd,
>+				   struct mvsw_pr_port *port)
>+{
>+	u32 speed;
>+	int err;
>+
>+	err = mvsw_pr_hw_port_speed_get(port, &speed);
>+	ecmd->base.speed = !err ? speed : SPEED_UNKNOWN;
>+}
>+
>+static int mvsw_pr_port_link_mode_set(struct mvsw_pr_port *port,
>+				      u32 speed, u8 duplex, u8 type)
>+{
>+	u32 new_mode = MVSW_LINK_MODE_MAX;
>+	u32 mode;
>+
>+	for (mode = 0; mode < MVSW_LINK_MODE_MAX; mode++) {
>+		if (speed != mvsw_pr_link_modes[mode].speed)
>+			continue;
>+		if (duplex != mvsw_pr_link_modes[mode].duplex)
>+			continue;
>+		if (!(mvsw_pr_link_modes[mode].pr_mask &
>+		    port->caps.supp_link_modes))
>+			continue;
>+		if (type != mvsw_pr_link_modes[mode].port_type)
>+			continue;
>+
>+		new_mode = mode;
>+		break;
>+	}
>+
>+	if (new_mode == MVSW_LINK_MODE_MAX) {
>+		netdev_err(port->net_dev, "Unsupported speed/duplex requested");
>+		return -EINVAL;
>+	}
>+
>+	return mvsw_pr_hw_port_link_mode_set(port, new_mode);
>+}
>+
>+static int mvsw_pr_port_speed_duplex_set(const struct ethtool_link_ksettings
>+					 *ecmd, struct mvsw_pr_port *port)
>+{
>+	int err;
>+	u8 duplex;
>+	u32 speed;
>+	u32 curr_mode;

You have to maintain reverse christmas tree ordering for all variables
in functions:

	u32 curr_mode;
	u32 speed;
	u8 duplex;
	int err;


>+
>+	err = mvsw_pr_hw_port_link_mode_get(port, &curr_mode);
>+	if (err || curr_mode >= MVSW_LINK_MODE_MAX)
>+		return -EINVAL;
>+
>+	if (ecmd->base.duplex != DUPLEX_UNKNOWN)
>+		duplex = ecmd->base.duplex == DUPLEX_FULL ?
>+			 MVSW_PORT_DUPLEX_FULL : MVSW_PORT_DUPLEX_HALF;
>+	else
>+		duplex = mvsw_pr_link_modes[curr_mode].duplex;
>+
>+	if (ecmd->base.speed != SPEED_UNKNOWN)
>+		speed = ecmd->base.speed;
>+	else
>+		speed = mvsw_pr_link_modes[curr_mode].speed;
>+
>+	return mvsw_pr_port_link_mode_set(port, speed, duplex, port->caps.type);
>+}
>+
>+static u8 mvsw_pr_port_type_get(struct mvsw_pr_port *port)
>+{
>+	if (port->caps.type < MVSW_PORT_TYPE_MAX)
>+		return mvsw_pr_port_types[port->caps.type].eth_type;
>+	return PORT_OTHER;
>+}
>+
>+static int mvsw_pr_port_type_set(const struct ethtool_link_ksettings *ecmd,
>+				 struct mvsw_pr_port *port)
>+{
>+	int err;
>+	u32 type, mode;
>+	u32 new_mode = MVSW_LINK_MODE_MAX;
>+
>+	for (type = 0; type < MVSW_PORT_TYPE_MAX; type++) {
>+		if (mvsw_pr_port_types[type].eth_type == ecmd->base.port &&
>+		    test_bit(mvsw_pr_port_types[type].eth_mode,
>+			     ecmd->link_modes.supported)) {
>+			break;
>+		}
>+	}
>+
>+	if (type == port->caps.type)
>+		return 0;
>+
>+	if (type == MVSW_PORT_TYPE_MAX) {
>+		pr_err("Unsupported port type requested\n");
>+		return -EINVAL;
>+	}
>+
>+	for (mode = 0; mode < MVSW_LINK_MODE_MAX; mode++) {
>+		if ((mvsw_pr_link_modes[mode].pr_mask &
>+		    port->caps.supp_link_modes) &&
>+		    type == mvsw_pr_link_modes[mode].port_type) {
>+			new_mode = mode;
>+		}
>+	}
>+
>+	if (new_mode < MVSW_LINK_MODE_MAX)
>+		err = mvsw_pr_hw_port_link_mode_set(port, new_mode);
>+	else
>+		err = -EINVAL;
>+
>+	if (!err)
>+		port->caps.type = type;
>+
>+	return err;
>+}
>+
>+static void mvsw_pr_port_remote_cap_get(struct ethtool_link_ksettings *ecmd,
>+					struct mvsw_pr_port *port)
>+{
>+	u64 bitmap;
>+
>+	if (!mvsw_pr_hw_port_remote_cap_get(port, &bitmap)) {
>+		mvsw_modes_to_eth(ecmd->link_modes.lp_advertising,
>+				  bitmap, 0, MVSW_PORT_TYPE_NONE);
>+	}
>+}
>+
>+static void mvsw_pr_port_duplex_get(struct ethtool_link_ksettings *ecmd,
>+				    struct mvsw_pr_port *port)
>+{
>+	u8 duplex;
>+
>+	if (!mvsw_pr_hw_port_duplex_get(port, &duplex)) {
>+		ecmd->base.duplex = duplex == MVSW_PORT_DUPLEX_FULL ?
>+				    DUPLEX_FULL : DUPLEX_HALF;
>+	} else {
>+		ecmd->base.duplex = DUPLEX_UNKNOWN;
>+	}
>+}
>+
>+static int mvsw_pr_port_autoneg_set(struct mvsw_pr_port *port, bool enable,
>+				    u64 link_modes, u8 fec)
>+{
>+	bool refresh = false;
>+	int err = 0;
>+
>+	if (port->caps.type != MVSW_PORT_TYPE_TP)
>+		return enable ? -EINVAL : 0;
>+
>+	if (port->adver_link_modes != link_modes || port->adver_fec != fec) {
>+		port->adver_link_modes = link_modes;
>+		port->adver_fec = fec != 0 ? fec : BIT(MVSW_PORT_FEC_OFF_BIT);
>+		refresh = true;
>+	}
>+
>+	if (port->autoneg == enable && !(port->autoneg && refresh))
>+		return 0;
>+
>+	err = mvsw_pr_hw_port_autoneg_set(port, enable,
>+					  port->adver_link_modes,
>+					  port->adver_fec);
>+	if (err)
>+		return -EINVAL;
>+
>+	port->autoneg = enable;
>+	return 0;
>+}
>+
>+static void mvsw_pr_port_mdix_get(struct ethtool_link_ksettings *ecmd,
>+				  struct mvsw_pr_port *port)
>+{
>+	u8 mode;
>+
>+	if (mvsw_pr_hw_port_mdix_get(port, &mode))

always store the return value in "err" and do "if (err)".


>+		return;
>+
>+	ecmd->base.eth_tp_mdix = mode;
>+}
>+
>+static int mvsw_pr_port_mdix_set(const struct ethtool_link_ksettings *ecmd,
>+				 struct mvsw_pr_port *port)
>+{
>+	if (ecmd->base.eth_tp_mdix_ctrl)
>+		return -EOPNOTSUPP;
>+
>+	return 0;
>+}
>+
>+static int mvsw_pr_port_get_link_ksettings(struct net_device *dev,
>+					   struct ethtool_link_ksettings *ecmd)
>+{
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+
>+	ethtool_link_ksettings_zero_link_mode(ecmd, supported);
>+	ethtool_link_ksettings_zero_link_mode(ecmd, advertising);
>+	ethtool_link_ksettings_zero_link_mode(ecmd, lp_advertising);
>+
>+	ecmd->base.autoneg = port->autoneg ? AUTONEG_ENABLE : AUTONEG_DISABLE;
>+
>+	if (port->caps.type == MVSW_PORT_TYPE_TP) {
>+		ethtool_link_ksettings_add_link_mode(ecmd, supported, Autoneg);
>+		if (netif_running(dev) &&
>+		    (port->autoneg ||
>+		     port->caps.transceiver == MVSW_PORT_TRANSCEIVER_COPPER))
>+			ethtool_link_ksettings_add_link_mode(ecmd, advertising,
>+							     Autoneg);
>+	}
>+
>+	mvsw_modes_to_eth(ecmd->link_modes.supported,
>+			  port->caps.supp_link_modes,
>+			  port->caps.supp_fec,
>+			  port->caps.type);
>+
>+	mvsw_pr_port_supp_types_get(ecmd, port);
>+
>+	if (netif_carrier_ok(dev)) {
>+		mvsw_pr_port_speed_get(ecmd, port);
>+		mvsw_pr_port_duplex_get(ecmd, port);
>+	} else {
>+		ecmd->base.speed = SPEED_UNKNOWN;
>+		ecmd->base.duplex = DUPLEX_UNKNOWN;
>+	}
>+
>+	ecmd->base.port = mvsw_pr_port_type_get(port);
>+
>+	if (port->autoneg) {
>+		if (netif_running(dev))
>+			mvsw_modes_to_eth(ecmd->link_modes.advertising,
>+					  port->adver_link_modes,
>+					  port->adver_fec,
>+					  port->caps.type);
>+
>+		if (netif_carrier_ok(dev) &&
>+		    port->caps.transceiver == MVSW_PORT_TRANSCEIVER_COPPER) {
>+			ethtool_link_ksettings_add_link_mode(ecmd,
>+							     lp_advertising,
>+							     Autoneg);
>+			mvsw_pr_port_remote_cap_get(ecmd, port);
>+		}
>+	}
>+
>+	if (port->caps.type == MVSW_PORT_TYPE_TP &&
>+	    port->caps.transceiver == MVSW_PORT_TRANSCEIVER_COPPER)
>+		mvsw_pr_port_mdix_get(ecmd, port);
>+
>+	return 0;
>+}
>+
>+static bool mvsw_pr_check_supp_modes(const struct mvsw_pr_port_caps *caps,
>+				     u64 adver_modes, u8 adver_fec)
>+{
>+	if ((caps->supp_link_modes & adver_modes) == 0)
>+		return true;
>+	if ((adver_fec & ~caps->supp_fec) != 0)
>+		return true;
>+
>+	return false;
>+}
>+
>+static int mvsw_pr_port_set_link_ksettings(struct net_device *dev,
>+					   const struct ethtool_link_ksettings
>+					   *ecmd)
>+{
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+	bool is_up = netif_running(dev);
>+	u64 adver_modes = 0;
>+	u8 adver_fec = 0;
>+	int err, err1;
>+
>+	if (is_up) {
>+		err = mvsw_pr_port_state_set(dev, false);
>+		if (err)
>+			return err;
>+	}
>+
>+	err = mvsw_pr_port_type_set(ecmd, port);
>+	if (err)
>+		goto fini_link_ksettings;
>+
>+	if (port->caps.transceiver == MVSW_PORT_TRANSCEIVER_COPPER) {
>+		err = mvsw_pr_port_mdix_set(ecmd, port);
>+		if (err)
>+			goto fini_link_ksettings;
>+	}
>+
>+	mvsw_modes_from_eth(ecmd->link_modes.advertising, &adver_modes,
>+			    &adver_fec);
>+
>+	if (ecmd->base.autoneg == AUTONEG_ENABLE &&
>+	    mvsw_pr_check_supp_modes(&port->caps, adver_modes, adver_fec)) {
>+		netdev_err(dev, "Unsupported link mode requested");
>+		err = -EINVAL;
>+		goto fini_link_ksettings;
>+	}
>+
>+	err = mvsw_pr_port_autoneg_set(port,
>+				       ecmd->base.autoneg == AUTONEG_ENABLE,
>+				       adver_modes, adver_fec);
>+	if (err)
>+		goto fini_link_ksettings;
>+
>+	if (ecmd->base.autoneg == AUTONEG_DISABLE) {
>+		err = mvsw_pr_port_speed_duplex_set(ecmd, port);
>+		if (err)
>+			goto fini_link_ksettings;
>+	}
>+
>+fini_link_ksettings:
>+	err1 = mvsw_pr_port_state_set(dev, is_up);
>+	if (err1)
>+		return err1;
>+
>+	return err;
>+}
>+
>+static int mvsw_pr_port_get_fecparam(struct net_device *dev,
>+				     struct ethtool_fecparam *fecparam)
>+{
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+	u32 mode;
>+	u8 active;
>+	int err;
>+
>+	err = mvsw_pr_hw_port_fec_get(port, &active);
>+	if (err)
>+		return err;
>+
>+	fecparam->fec = 0;
>+	for (mode = 0; mode < MVSW_PORT_FEC_MAX; mode++) {
>+		if ((mvsw_pr_fec_caps[mode].pr_fec & port->caps.supp_fec) == 0)
>+			continue;
>+		fecparam->fec |= mvsw_pr_fec_caps[mode].eth_fec;
>+	}
>+
>+	if (active < MVSW_PORT_FEC_MAX)
>+		fecparam->active_fec = mvsw_pr_fec_caps[active].eth_fec;
>+	else
>+		fecparam->active_fec = ETHTOOL_FEC_AUTO;
>+
>+	return 0;
>+}
>+
>+static int mvsw_pr_port_set_fecparam(struct net_device *dev,
>+				     struct ethtool_fecparam *fecparam)
>+{
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+	u8 fec, active;
>+	u32 mode;
>+	int err;
>+
>+	if (port->autoneg) {
>+		netdev_err(dev, "FEC set is not allowed while autoneg is on\n");
>+		return -EINVAL;
>+	}
>+
>+	err = mvsw_pr_hw_port_fec_get(port, &active);
>+	if (err)
>+		return err;
>+
>+	fec = MVSW_PORT_FEC_MAX;
>+	for (mode = 0; mode < MVSW_PORT_FEC_MAX; mode++) {
>+		if ((mvsw_pr_fec_caps[mode].eth_fec & fecparam->fec) &&
>+		    (mvsw_pr_fec_caps[mode].pr_fec & port->caps.supp_fec)) {
>+			fec = mode;
>+			break;
>+		}
>+	}
>+
>+	if (fec == active)
>+		return 0;
>+
>+	if (fec == MVSW_PORT_FEC_MAX) {
>+		netdev_err(dev, "Unsupported FEC requested");
>+		return -EINVAL;
>+	}
>+
>+	return mvsw_pr_hw_port_fec_set(port, fec);
>+}
>+
>+static void mvsw_pr_port_get_ethtool_stats(struct net_device *dev,
>+					   struct ethtool_stats *stats,
>+					   u64 *data)
>+{
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+	struct mvsw_pr_port_stats *port_stats = &port->cached_hw_stats.stats;
>+
>+	memcpy((u8 *)data, port_stats, sizeof(*port_stats));
>+}
>+
>+static void mvsw_pr_port_get_strings(struct net_device *dev,
>+				     u32 stringset, u8 *data)
>+{
>+	if (stringset != ETH_SS_STATS)
>+		return;
>+
>+	memcpy(data, *mvsw_pr_port_cnt_name, sizeof(mvsw_pr_port_cnt_name));
>+}
>+
>+static int mvsw_pr_port_get_sset_count(struct net_device *dev, int sset)
>+{
>+	switch (sset) {
>+	case ETH_SS_STATS:
>+		return PORT_STATS_CNT;
>+	default:
>+		return -EOPNOTSUPP;
>+	}
>+}
>+
>+static const struct ethtool_ops mvsw_pr_ethtool_ops = {
>+	.get_drvinfo = mvsw_pr_port_get_drvinfo,
>+	.get_link_ksettings = mvsw_pr_port_get_link_ksettings,
>+	.set_link_ksettings = mvsw_pr_port_set_link_ksettings,
>+	.get_fecparam = mvsw_pr_port_get_fecparam,
>+	.set_fecparam = mvsw_pr_port_set_fecparam,
>+	.get_sset_count = mvsw_pr_port_get_sset_count,
>+	.get_strings = mvsw_pr_port_get_strings,
>+	.get_ethtool_stats = mvsw_pr_port_get_ethtool_stats,
>+	.get_link = ethtool_op_get_link
>+};

Please put the ethtool code introduction in a separate patch, easier to
review multiple smaller patches than 1 big one.

Please do the same with other blocks, like bridge offload.


>+
>+int mvsw_pr_port_learning_set(struct mvsw_pr_port *port, bool learn)
>+{
>+	return mvsw_pr_hw_port_learning_set(port, learn);
>+}
>+
>+int mvsw_pr_port_flood_set(struct mvsw_pr_port *port, bool flood)
>+{
>+	return mvsw_pr_hw_port_flood_set(port, flood);
>+}
>+
>+int mvsw_pr_port_pvid_set(struct mvsw_pr_port *port, u16 vid)
>+{
>+	int err;
>+
>+	if (!vid) {
>+		err = mvsw_pr_hw_port_accept_frame_type_set
>+		    (port, MVSW_ACCEPT_FRAME_TYPE_TAGGED);
>+		if (err)
>+			return err;
>+	} else {
>+		err = mvsw_pr_hw_vlan_port_vid_set(port, vid);
>+		if (err)
>+			return err;
>+		err = mvsw_pr_hw_port_accept_frame_type_set
>+		    (port, MVSW_ACCEPT_FRAME_TYPE_ALL);
>+		if (err)
>+			goto err_port_allow_untagged_set;
>+	}
>+
>+	port->pvid = vid;
>+	return 0;
>+
>+err_port_allow_untagged_set:
>+	mvsw_pr_hw_vlan_port_vid_set(port, port->pvid);
>+	return err;
>+}
>+
>+struct mvsw_pr_port_vlan*
>+mvsw_pr_port_vlan_find_by_vid(const struct mvsw_pr_port *port, u16 vid)
>+{
>+	struct mvsw_pr_port_vlan *port_vlan;
>+
>+	list_for_each_entry(port_vlan, &port->vlans_list, list) {
>+		if (port_vlan->vid == vid)
>+			return port_vlan;
>+	}
>+
>+	return NULL;
>+}
>+
>+struct mvsw_pr_port_vlan*
>+mvsw_pr_port_vlan_create(struct mvsw_pr_port *port, u16 vid)
>+{
>+	bool untagged = vid == MVSW_PR_DEFAULT_VID;
>+	struct mvsw_pr_port_vlan *port_vlan;
>+	int err;
>+
>+	port_vlan = mvsw_pr_port_vlan_find_by_vid(port, vid);
>+	if (port_vlan)
>+		return ERR_PTR(-EEXIST);
>+
>+	err = mvsw_pr_port_vlan_set(port, vid, true, untagged);
>+	if (err)
>+		return ERR_PTR(err);
>+
>+	port_vlan = kzalloc(sizeof(*port_vlan), GFP_KERNEL);
>+	if (!port_vlan) {
>+		err = -ENOMEM;
>+		goto err_port_vlan_alloc;
>+	}
>+
>+	port_vlan->mvsw_pr_port = port;
>+	port_vlan->vid = vid;
>+
>+	list_add(&port_vlan->list, &port->vlans_list);
>+
>+	return port_vlan;
>+
>+err_port_vlan_alloc:
>+	mvsw_pr_port_vlan_set(port, vid, false, false);
>+	return ERR_PTR(err);
>+}
>+
>+static void
>+mvsw_pr_port_vlan_cleanup(struct mvsw_pr_port_vlan *port_vlan)
>+{
>+	if (port_vlan->bridge_port)
>+		mvsw_pr_port_vlan_bridge_leave(port_vlan);
>+}
>+
>+void mvsw_pr_port_vlan_destroy(struct mvsw_pr_port_vlan *port_vlan)
>+{
>+	struct mvsw_pr_port *port = port_vlan->mvsw_pr_port;
>+	u16 vid = port_vlan->vid;
>+
>+	mvsw_pr_port_vlan_cleanup(port_vlan);
>+	list_del(&port_vlan->list);
>+	kfree(port_vlan);
>+	mvsw_pr_hw_vlan_port_set(port, vid, false, false);
>+}
>+
>+int mvsw_pr_port_vlan_set(struct mvsw_pr_port *port, u16 vid,
>+			  bool is_member, bool untagged)
>+{
>+	return mvsw_pr_hw_vlan_port_set(port, vid, is_member, untagged);
>+}
>+
>+static int mvsw_pr_port_create(struct mvsw_pr_switch *sw, u32 id)
>+{
>+	struct net_device *net_dev;

Be consistent with the rest of the code:
"struct net_device *dev"



>+	struct mvsw_pr_port *port;
>+	char *mac;
>+	int err;
>+
>+	net_dev = alloc_etherdev(sizeof(*port));
>+	if (!net_dev)
>+		return -ENOMEM;
>+
>+	port = netdev_priv(net_dev);
>+
>+	INIT_LIST_HEAD(&port->vlans_list);
>+	port->pvid = MVSW_PR_DEFAULT_VID;
>+	port->net_dev = net_dev;
>+	port->id = id;
>+	port->sw = sw;
>+
>+	err = mvsw_pr_hw_port_info_get(port, &port->fp_id,
>+				       &port->hw_id, &port->dev_id);
>+	if (err) {
>+		dev_err(mvsw_dev(sw), "Failed to get port(%u) info\n", id);
>+		goto err_register_netdev;
>+	}
>+
>+	net_dev->features |= NETIF_F_NETNS_LOCAL | NETIF_F_HW_L2FW_DOFFLOAD;
>+	net_dev->ethtool_ops = &mvsw_pr_ethtool_ops;
>+	net_dev->netdev_ops = &mvsw_pr_netdev_ops;
>+
>+	netif_carrier_off(net_dev);
>+
>+	net_dev->mtu = min_t(unsigned int, sw->mtu_max, MVSW_PR_MTU_DEFAULT);
>+	net_dev->min_mtu = sw->mtu_min;
>+	net_dev->max_mtu = sw->mtu_max;
>+
>+	err = mvsw_pr_hw_port_mtu_set(port, net_dev->mtu);
>+	if (err) {
>+		dev_err(mvsw_dev(sw), "Failed to set port(%u) mtu\n", id);
>+		goto err_register_netdev;
>+	}
>+
>+	/* Only 0xFF mac addrs are supported */
>+	if (port->fp_id >= 0xFF)
>+		goto err_register_netdev;
>+
>+	mac = net_dev->dev_addr;
>+	memcpy(mac, sw->base_mac, net_dev->addr_len - 1);
>+	mac[net_dev->addr_len - 1] = (char)port->fp_id;
>+
>+	err = mvsw_pr_hw_port_mac_set(port, mac);
>+	if (err) {
>+		dev_err(mvsw_dev(sw), "Failed to set port(%u) mac addr\n", id);
>+		goto err_register_netdev;
>+	}
>+
>+	err = mvsw_pr_hw_port_cap_get(port, &port->caps);
>+	if (err) {
>+		dev_err(mvsw_dev(sw), "Failed to get port(%u) caps\n", id);
>+		goto err_register_netdev;
>+	}
>+
>+	port->adver_link_modes = 0;

No need. Mem id zeroed.


>+	port->adver_fec = 1 << MVSW_PORT_FEC_OFF_BIT;
>+	port->autoneg = false;

No need. Mem id zeroed.

	
>+	mvsw_pr_port_autoneg_set(port, true, port->caps.supp_link_modes,
>+				 port->caps.supp_fec);
>+
>+	err = mvsw_pr_hw_port_state_set(port, false);
>+	if (err) {
>+		dev_err(mvsw_dev(sw), "Failed to set port(%u) down\n", id);
>+		goto err_register_netdev;
>+	}
>+
>+	INIT_DELAYED_WORK(&port->cached_hw_stats.caching_dw,
>+			  &update_stats_cache);
>+
>+	err = register_netdev(net_dev);
>+	if (err)
>+		goto err_register_netdev;
>+
>+	list_add(&port->list, &sw->port_list);
>+
>+	return 0;
>+
>+err_register_netdev:
>+	free_netdev(net_dev);
>+	return err;
>+}
>+
>+static void mvsw_pr_port_vlan_flush(struct mvsw_pr_port *port,
>+				    bool flush_default)
>+{
>+	struct mvsw_pr_port_vlan *port_vlan, *tmp;
>+
>+	list_for_each_entry_safe(port_vlan, tmp, &port->vlans_list, list) {
>+		if (!flush_default && port_vlan->vid == MVSW_PR_DEFAULT_VID)
>+			continue;
>+
>+		mvsw_pr_port_vlan_destroy(port_vlan);
>+	}
>+}
>+
>+int mvsw_pr_8021d_bridge_create(struct mvsw_pr_switch *sw, u16 *bridge_id)
>+{
>+	return mvsw_pr_hw_bridge_create(sw, bridge_id);
>+}
>+
>+int mvsw_pr_8021d_bridge_delete(struct mvsw_pr_switch *sw, u16 bridge_id)
>+{
>+	return mvsw_pr_hw_bridge_delete(sw, bridge_id);
>+}
>+
>+int mvsw_pr_8021d_bridge_port_add(struct mvsw_pr_port *port, u16 bridge_id)
>+{
>+	return mvsw_pr_hw_bridge_port_add(port, bridge_id);
>+}
>+
>+int mvsw_pr_8021d_bridge_port_delete(struct mvsw_pr_port *port, u16 bridge_id)
>+{
>+	return mvsw_pr_hw_bridge_port_delete(port, bridge_id);
>+}
>+
>+int mvsw_pr_switch_ageing_set(struct mvsw_pr_switch *sw, u32 ageing_time)
>+{
>+	return mvsw_pr_hw_switch_ageing_set(sw, ageing_time);
>+}
>+
>+int mvsw_pr_fdb_flush_vlan(struct mvsw_pr_switch *sw, u16 vid,
>+			   enum mvsw_pr_fdb_flush_mode mode)
>+{
>+	return mvsw_pr_hw_fdb_flush_vlan(sw, vid, mode);
>+}
>+
>+int mvsw_pr_fdb_flush_port_vlan(struct mvsw_pr_port *port, u16 vid,
>+				enum mvsw_pr_fdb_flush_mode mode)
>+{
>+	return mvsw_pr_hw_fdb_flush_port_vlan(port, vid, mode);
>+}
>+
>+int mvsw_pr_fdb_flush_port(struct mvsw_pr_port *port,
>+			   enum mvsw_pr_fdb_flush_mode mode)
>+{
>+	return mvsw_pr_hw_fdb_flush_port(port, mode);
>+}
>+
>+static int mvsw_pr_clear_ports(struct mvsw_pr_switch *sw)
>+{

Be consistent with the rest of the code:
"struct net_device *dev"



>+	struct net_device *net_dev;
>+	struct list_head *pos, *n;
>+	struct mvsw_pr_port *port;
>+
>+	list_for_each_safe(pos, n, &sw->port_list) {
>+		port = list_entry(pos, typeof(*port), list);
>+		net_dev = port->net_dev;
>+
>+		cancel_delayed_work_sync(&port->cached_hw_stats.caching_dw);
>+		unregister_netdev(net_dev);
>+		mvsw_pr_port_vlan_flush(port, true);
>+		WARN_ON_ONCE(!list_empty(&port->vlans_list));
>+		free_netdev(net_dev);
>+		list_del(pos);
>+	}
>+	return (!list_empty(&sw->port_list));
>+}
>+
>+static void mvsw_pr_port_handle_event(struct mvsw_pr_switch *sw,
>+				      struct mvsw_pr_event *evt)
>+{
>+	struct mvsw_pr_port *port;
>+	struct delayed_work *caching_dw;
>+
>+	port = __find_pr_port(sw, evt->port_evt.port_id);
>+	if (!port)
>+		return;
>+
>+	caching_dw = &port->cached_hw_stats.caching_dw;
>+
>+	switch (evt->id) {
>+	case MVSW_PORT_EVENT_STATE_CHANGED:
>+		if (evt->port_evt.data.oper_state) {
>+			netif_carrier_on(port->net_dev);
>+			if (!delayed_work_pending(caching_dw))
>+				queue_delayed_work(mvsw_pr_wq, caching_dw, 0);
>+		} else {
>+			netif_carrier_off(port->net_dev);
>+			if (delayed_work_pending(caching_dw))
>+				cancel_delayed_work(caching_dw);
>+		}
>+		break;
>+	}
>+}
>+
>+static void mvsw_pr_fdb_handle_event(struct mvsw_pr_switch *sw,
>+				     struct mvsw_pr_event *evt)

Hmm, I think you should register this handler from prestera_switchdev.c


>+{
>+	struct switchdev_notifier_fdb_info info;
>+	struct mvsw_pr_port *port;
>+
>+	port = __find_pr_port(sw, evt->fdb_evt.port_id);
>+	if (!port)
>+		return;
>+
>+	info.addr = evt->fdb_evt.data.mac;
>+	info.vid = evt->fdb_evt.vid;
>+	info.offloaded = true;
>+
>+	rtnl_lock();
>+	switch (evt->id) {
>+	case MVSW_FDB_EVENT_LEARNED:
>+		call_switchdev_notifiers(SWITCHDEV_FDB_ADD_TO_BRIDGE,
>+					 port->net_dev, &info.info, NULL);
>+		break;
>+	case MVSW_FDB_EVENT_AGED:
>+		call_switchdev_notifiers(SWITCHDEV_FDB_DEL_TO_BRIDGE,
>+					 port->net_dev, &info.info, NULL);
>+		break;
>+	}
>+	rtnl_unlock();
>+	return;
>+}
>+
>+int mvsw_pr_fdb_add(struct mvsw_pr_port *port, const unsigned char *mac,
>+		    u16 vid, bool dynamic)
>+{
>+	return mvsw_pr_hw_fdb_add(port, mac, vid, dynamic);
>+}
>+
>+int mvsw_pr_fdb_del(struct mvsw_pr_port *port, const unsigned char *mac,
>+		    u16 vid)
>+{
>+	return mvsw_pr_hw_fdb_del(port, mac, vid);
>+}
>+
>+static void mvsw_pr_fdb_event_handler_unregister(struct mvsw_pr_switch *sw)
>+{
>+	mvsw_pr_hw_event_handler_unregister(sw, MVSW_EVENT_TYPE_FDB,
>+					    mvsw_pr_fdb_handle_event);
>+}
>+
>+static void mvsw_pr_port_event_handler_unregister(struct mvsw_pr_switch *sw)
>+{
>+	mvsw_pr_hw_event_handler_unregister(sw, MVSW_EVENT_TYPE_PORT,
>+					    mvsw_pr_port_handle_event);
>+}
>+
>+static void mvsw_pr_event_handlers_unregister(struct mvsw_pr_switch *sw)
>+{
>+	mvsw_pr_fdb_event_handler_unregister(sw);
>+	mvsw_pr_port_event_handler_unregister(sw);
>+}
>+
>+static int mvsw_pr_fdb_event_handler_register(struct mvsw_pr_switch *sw)
>+{
>+	return mvsw_pr_hw_event_handler_register(sw, MVSW_EVENT_TYPE_FDB,
>+						 mvsw_pr_fdb_handle_event);
>+}
>+
>+static int mvsw_pr_port_event_handler_register(struct mvsw_pr_switch *sw)
>+{
>+	return mvsw_pr_hw_event_handler_register(sw, MVSW_EVENT_TYPE_PORT,
>+						 mvsw_pr_port_handle_event);
>+}
>+
>+static int mvsw_pr_event_handlers_register(struct mvsw_pr_switch *sw)
>+{
>+	int err;
>+
>+	err = mvsw_pr_port_event_handler_register(sw);
>+	if (err)
>+		return err;
>+
>+	err = mvsw_pr_fdb_event_handler_register(sw);
>+	if (err)
>+		goto err_fdb_handler_register;
>+
>+	return 0;
>+
>+err_fdb_handler_register:
>+	mvsw_pr_port_event_handler_unregister(sw);
>+	return err;
>+}
>+
>+static int mvsw_pr_init(struct mvsw_pr_switch *sw)
>+{
>+	u32 port;
>+	int err;
>+
>+	err = mvsw_pr_hw_switch_init(sw);
>+	if (err) {
>+		dev_err(mvsw_dev(sw), "Failed to init Switch device\n");
>+		return err;
>+	}
>+
>+	dev_info(mvsw_dev(sw), "Initialized Switch device\n");

Remove prints like this.


>+
>+	err = mvsw_pr_switchdev_register(sw);
>+	if (err)
>+		return err;
>+
>+	INIT_LIST_HEAD(&sw->port_list);
>+
>+	for (port = 0; port < sw->port_count; port++) {
>+		err = mvsw_pr_port_create(sw, port);
>+		if (err)
>+			goto err_ports_init;
>+	}
>+
>+	err = mvsw_pr_event_handlers_register(sw);
>+	if (err)
>+		goto err_ports_init;
>+
>+	return 0;
>+
>+err_ports_init:
>+	mvsw_pr_clear_ports(sw);
>+	return err;
>+}
>+
>+static void mvsw_pr_fini(struct mvsw_pr_switch *sw)
>+{
>+	mvsw_pr_event_handlers_unregister(sw);
>+

Remove the empty line.


>+	mvsw_pr_switchdev_unregister(sw);
>+	mvsw_pr_clear_ports(sw);
>+}
>+
>+int mvsw_pr_device_register(struct mvsw_pr_device *dev)
>+{
>+	struct mvsw_pr_switch *sw;
>+	int err;
>+
>+	sw = kzalloc(sizeof(*sw), GFP_KERNEL);
>+	if (!sw)
>+		return -ENOMEM;
>+
>+	dev->priv = sw;
>+	sw->dev = dev;
>+
>+	err = mvsw_pr_init(sw);
>+	if (err) {
>+		kfree(sw);
>+		return err;
>+	}
>+
>+	list_add(&sw->list, &switches_registered);
>+
>+	return 0;
>+}
>+EXPORT_SYMBOL(mvsw_pr_device_register);
>+
>+void mvsw_pr_device_unregister(struct mvsw_pr_device *dev)
>+{
>+	struct mvsw_pr_switch *sw = dev->priv;
>+
>+	list_del(&sw->list);
>+	mvsw_pr_fini(sw);
>+	kfree(sw);
>+}
>+EXPORT_SYMBOL(mvsw_pr_device_unregister);
>+
>+static int __init mvsw_pr_module_init(void)
>+{
>+	INIT_LIST_HEAD(&switches_registered);
>+
>+	mvsw_pr_wq = alloc_workqueue(mvsw_driver_name, 0, 0);
>+	if (!mvsw_pr_wq)
>+		return -ENOMEM;
>+
>+	pr_info("Loading Marvell Prestera Switch Driver\n");
>+	return 0;
>+}
>+
>+static void __exit mvsw_pr_module_exit(void)
>+{
>+	destroy_workqueue(mvsw_pr_wq);
>+
>+	pr_info("Unloading Marvell Prestera Switch Driver\n");
>+}
>+
>+module_init(mvsw_pr_module_init);
>+module_exit(mvsw_pr_module_exit);
>+
>+MODULE_AUTHOR("Marvell Semi.");
>+MODULE_LICENSE("GPL");

Inconsistent licences.


>+MODULE_DESCRIPTION("Marvell Prestera switch driver");
>+MODULE_VERSION(PRESTERA_DRV_VER);

Why do you need this? I believe it is better to avoid it.


>diff --git a/drivers/net/ethernet/marvell/prestera/prestera.h b/drivers/net/ethernet/marvell/prestera/prestera.h
>new file mode 100644
>index 000000000000..cbc6b0c78937
>--- /dev/null
>+++ b/drivers/net/ethernet/marvell/prestera/prestera.h
>@@ -0,0 +1,244 @@
>+/* SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0
>+ *
>+ * Copyright (c) 2019-2020 Marvell International Ltd. All rights reserved.
>+ *
>+ */
>+
>+#ifndef _MVSW_PRESTERA_H_
>+#define _MVSW_PRESTERA_H_
>+
>+#include <linux/skbuff.h>
>+#include <linux/notifier.h>
>+#include <uapi/linux/if_ether.h>
>+#include <linux/workqueue.h>
>+
>+#define MVSW_MSG_MAX_SIZE 1500
>+
>+#define MVSW_PR_DEFAULT_VID 1
>+
>+#define MVSW_PR_MIN_AGEING_TIME 10
>+#define MVSW_PR_MAX_AGEING_TIME 1000000
>+#define MVSW_PR_DEFAULT_AGEING_TIME 300
>+
>+struct mvsw_fw_rev {
>+	u16 maj;
>+	u16 min;
>+	u16 sub;
>+};
>+
>+struct mvsw_pr_bridge_port;
>+
>+struct mvsw_pr_port_vlan {
>+	struct list_head list;
>+	struct mvsw_pr_port *mvsw_pr_port;
>+	u16 vid;
>+	struct mvsw_pr_bridge_port *bridge_port;
>+	struct list_head bridge_vlan_node;
>+};
>+
>+struct mvsw_pr_port_stats {
>+	u64 good_octets_received;
>+	u64 bad_octets_received;
>+	u64 mac_trans_error;
>+	u64 broadcast_frames_received;
>+	u64 multicast_frames_received;
>+	u64 frames_64_octets;
>+	u64 frames_65_to_127_octets;
>+	u64 frames_128_to_255_octets;
>+	u64 frames_256_to_511_octets;
>+	u64 frames_512_to_1023_octets;
>+	u64 frames_1024_to_max_octets;
>+	u64 excessive_collision;
>+	u64 multicast_frames_sent;
>+	u64 broadcast_frames_sent;
>+	u64 fc_sent;
>+	u64 fc_received;
>+	u64 buffer_overrun;
>+	u64 undersize;
>+	u64 fragments;
>+	u64 oversize;
>+	u64 jabber;
>+	u64 rx_error_frame_received;
>+	u64 bad_crc;
>+	u64 collisions;
>+	u64 late_collision;
>+	u64 unicast_frames_received;
>+	u64 unicast_frames_sent;
>+	u64 sent_multiple;
>+	u64 sent_deferred;
>+	u64 frames_1024_to_1518_octets;
>+	u64 frames_1519_to_max_octets;
>+	u64 good_octets_sent;
>+};
>+
>+struct mvsw_pr_port_caps {
>+	u64 supp_link_modes;
>+	u8 supp_fec;
>+	u8 type;
>+	u8 transceiver;
>+};
>+
>+struct mvsw_pr_port {
>+	struct net_device *net_dev;

Be consistent with the rest of the code:
"struct net_device *dev"


>+	struct mvsw_pr_switch *sw;
>+	u32 id;
>+	u32 hw_id;
>+	u32 dev_id;
>+	u16 fp_id;
>+	u16 pvid;
>+	bool autoneg;
>+	u64 adver_link_modes;
>+	u8 adver_fec;
>+	struct mvsw_pr_port_caps caps;
>+	struct list_head list;
>+	struct list_head vlans_list;
>+	struct {
>+		struct mvsw_pr_port_stats stats;
>+		struct delayed_work caching_dw;
>+	} cached_hw_stats;
>+};
>+
>+struct mvsw_pr_switchdev {
>+	struct mvsw_pr_switch *sw;
>+	struct notifier_block swdev_n;
>+	struct notifier_block swdev_blocking_n;
>+};
>+
>+struct mvsw_pr_fib {
>+	struct mvsw_pr_switch *sw;
>+	struct notifier_block fib_nb;
>+	struct notifier_block netevent_nb;
>+};
>+
>+struct mvsw_pr_device {
>+	struct device *dev;
>+	struct mvsw_fw_rev fw_rev;
>+	void *priv;
>+
>+	/* called by device driver to pass event up to the higher layer */
>+	int (*recv_msg)(struct mvsw_pr_device *dev, u8 *msg, size_t size);
>+
>+	/* called by higher layer to send request to the firmware */
>+	int (*send_req)(struct mvsw_pr_device *dev, u8 *in_msg,
>+			size_t in_size, u8 *out_msg, size_t out_size,
>+			unsigned int wait);
>+};
>+
>+enum mvsw_pr_event_type {
>+	MVSW_EVENT_TYPE_UNSPEC,
>+	MVSW_EVENT_TYPE_PORT,
>+	MVSW_EVENT_TYPE_FDB,
>+
>+	MVSW_EVENT_TYPE_MAX,
>+};
>+
>+enum mvsw_pr_port_event_id {
>+	MVSW_PORT_EVENT_UNSPEC,
>+	MVSW_PORT_EVENT_STATE_CHANGED,
>+
>+	MVSW_PORT_EVENT_MAX,
>+};
>+
>+enum mvsw_pr_fdb_event_id {
>+	MVSW_FDB_EVENT_UNSPEC,
>+	MVSW_FDB_EVENT_LEARNED,
>+	MVSW_FDB_EVENT_AGED,
>+
>+	MVSW_FDB_EVENT_MAX,
>+};
>+
>+struct mvsw_pr_fdb_event {
>+	u32 port_id;
>+	u32 vid;
>+	union {
>+		u8 mac[ETH_ALEN];
>+	} data;
>+};
>+
>+struct mvsw_pr_port_event {
>+	u32 port_id;
>+	union {
>+		u32 oper_state;
>+	} data;
>+};
>+
>+struct mvsw_pr_event {
>+	u16 id;
>+	union {
>+		struct mvsw_pr_port_event port_evt;
>+		struct mvsw_pr_fdb_event fdb_evt;
>+	};
>+};
>+
>+struct mvsw_pr_bridge;
>+
>+struct mvsw_pr_switch {
>+	struct list_head list;
>+	struct mvsw_pr_device *dev;
>+	struct list_head event_handlers;
>+	char base_mac[ETH_ALEN];
>+	struct list_head port_list;
>+	u32 port_count;
>+	u32 mtu_min;
>+	u32 mtu_max;
>+	u8 id;
>+	struct mvsw_pr_bridge *bridge;
>+	struct mvsw_pr_switchdev *switchdev;
>+	struct mvsw_pr_fib *fib;
>+	struct notifier_block netdevice_nb;
>+};
>+
>+enum mvsw_pr_fdb_flush_mode {
>+	MVSW_PR_FDB_FLUSH_MODE_DYNAMIC = BIT(0),
>+	MVSW_PR_FDB_FLUSH_MODE_STATIC = BIT(1),
>+	MVSW_PR_FDB_FLUSH_MODE_ALL = MVSW_PR_FDB_FLUSH_MODE_DYNAMIC
>+				   | MVSW_PR_FDB_FLUSH_MODE_STATIC,
>+};
>+
>+int mvsw_pr_switch_ageing_set(struct mvsw_pr_switch *sw, u32 ageing_time);
>+
>+int mvsw_pr_port_learning_set(struct mvsw_pr_port *mvsw_pr_port,
>+			      bool learn_enable);
>+int mvsw_pr_port_flood_set(struct mvsw_pr_port *mvsw_pr_port, bool flood);
>+int mvsw_pr_port_pvid_set(struct mvsw_pr_port *mvsw_pr_port, u16 vid);
>+struct mvsw_pr_port_vlan *
>+mvsw_pr_port_vlan_create(struct mvsw_pr_port *mvsw_pr_port, u16 vid);
>+void mvsw_pr_port_vlan_destroy(struct mvsw_pr_port_vlan *mvsw_pr_port_vlan);
>+int mvsw_pr_port_vlan_set(struct mvsw_pr_port *mvsw_pr_port, u16 vid,
>+			  bool is_member, bool untagged);
>+
>+int mvsw_pr_8021d_bridge_create(struct mvsw_pr_switch *sw, u16 *bridge_id);
>+int mvsw_pr_8021d_bridge_delete(struct mvsw_pr_switch *sw, u16 bridge_id);
>+int mvsw_pr_8021d_bridge_port_add(struct mvsw_pr_port *mvsw_pr_port,
>+				  u16 bridge_id);
>+int mvsw_pr_8021d_bridge_port_delete(struct mvsw_pr_port *mvsw_pr_port,
>+				     u16 bridge_id);
>+
>+int mvsw_pr_fdb_add(struct mvsw_pr_port *mvsw_pr_port, const unsigned char *mac,
>+		    u16 vid, bool dynamic);
>+int mvsw_pr_fdb_del(struct mvsw_pr_port *mvsw_pr_port, const unsigned char *mac,
>+		    u16 vid);
>+int mvsw_pr_fdb_flush_vlan(struct mvsw_pr_switch *sw, u16 vid,
>+			   enum mvsw_pr_fdb_flush_mode mode);
>+int mvsw_pr_fdb_flush_port_vlan(struct mvsw_pr_port *port, u16 vid,
>+				enum mvsw_pr_fdb_flush_mode mode);
>+int mvsw_pr_fdb_flush_port(struct mvsw_pr_port *port,
>+			   enum mvsw_pr_fdb_flush_mode mode);
>+
>+struct mvsw_pr_port_vlan *
>+mvsw_pr_port_vlan_find_by_vid(const struct mvsw_pr_port *mvsw_pr_port, u16 vid);
>+void
>+mvsw_pr_port_vlan_bridge_leave(struct mvsw_pr_port_vlan *mvsw_pr_port_vlan);
>+
>+int mvsw_pr_switchdev_register(struct mvsw_pr_switch *sw);
>+void mvsw_pr_switchdev_unregister(struct mvsw_pr_switch *sw);
>+
>+int mvsw_pr_device_register(struct mvsw_pr_device *dev);
>+void mvsw_pr_device_unregister(struct mvsw_pr_device *dev);
>+
>+bool mvsw_pr_netdev_check(const struct net_device *dev);
>+struct mvsw_pr_port *mvsw_pr_port_dev_lower_find(struct net_device *dev);
>+
>+const struct mvsw_pr_port *mvsw_pr_port_find(u32 dev_hw_id, u32 port_hw_id);
>+
>+#endif /* _MVSW_PRESTERA_H_ */
>diff --git a/drivers/net/ethernet/marvell/prestera/prestera_drv_ver.h b/drivers/net/ethernet/marvell/prestera/prestera_drv_ver.h
>new file mode 100644
>index 000000000000..d6617a16d7e1
>--- /dev/null
>+++ b/drivers/net/ethernet/marvell/prestera/prestera_drv_ver.h
>@@ -0,0 +1,23 @@
>+/* SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0
>+ *
>+ * Copyright (c) 2019-2020 Marvell International Ltd. All rights reserved.
>+ *
>+ */
>+#ifndef _PRESTERA_DRV_VER_H_
>+#define _PRESTERA_DRV_VER_H_
>+
>+#include <linux/stringify.h>
>+
>+/* Prestera driver version */
>+#define PRESTERA_DRV_VER_MAJOR	1
>+#define PRESTERA_DRV_VER_MINOR	0
>+#define PRESTERA_DRV_VER_PATCH	0
>+#define PRESTERA_DRV_VER_EXTRA
>+
>+#define PRESTERA_DRV_VER \
>+		__stringify(PRESTERA_DRV_VER_MAJOR)  "." \
>+		__stringify(PRESTERA_DRV_VER_MINOR)  "." \
>+		__stringify(PRESTERA_DRV_VER_PATCH)  \
>+		__stringify(PRESTERA_DRV_VER_EXTRA)
>+
>+#endif  /* _PRESTERA_DRV_VER_H_ */
>diff --git a/drivers/net/ethernet/marvell/prestera/prestera_hw.c b/drivers/net/ethernet/marvell/prestera/prestera_hw.c
>new file mode 100644
>index 000000000000..c97bafdd734e
>--- /dev/null
>+++ b/drivers/net/ethernet/marvell/prestera/prestera_hw.c
>@@ -0,0 +1,1094 @@
>+/* SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0
>+ *
>+ * Copyright (c) 2019-2020 Marvell International Ltd. All rights reserved.
>+ *
>+ */
>+#include <linux/etherdevice.h>
>+#include <linux/ethtool.h>
>+#include <linux/netdevice.h>
>+#include <linux/list.h>
>+
>+#include "prestera.h"
>+#include "prestera_hw.h"
>+
>+#define MVSW_PR_INIT_TIMEOUT 30000000	/* 30sec */
>+#define MVSW_PR_MIN_MTU 64
>+
>+enum mvsw_msg_type {
>+	MVSW_MSG_TYPE_SWITCH_UNSPEC,
>+	MVSW_MSG_TYPE_SWITCH_INIT,
>+
>+	MVSW_MSG_TYPE_AGEING_TIMEOUT_SET,
>+
>+	MVSW_MSG_TYPE_PORT_ATTR_SET,
>+	MVSW_MSG_TYPE_PORT_ATTR_GET,
>+	MVSW_MSG_TYPE_PORT_INFO_GET,
>+
>+	MVSW_MSG_TYPE_VLAN_CREATE,
>+	MVSW_MSG_TYPE_VLAN_DELETE,
>+	MVSW_MSG_TYPE_VLAN_PORT_SET,
>+	MVSW_MSG_TYPE_VLAN_PVID_SET,
>+
>+	MVSW_MSG_TYPE_FDB_ADD,
>+	MVSW_MSG_TYPE_FDB_DELETE,
>+	MVSW_MSG_TYPE_FDB_FLUSH_PORT,
>+	MVSW_MSG_TYPE_FDB_FLUSH_VLAN,
>+	MVSW_MSG_TYPE_FDB_FLUSH_PORT_VLAN,
>+
>+	MVSW_MSG_TYPE_LOG_LEVEL_SET,
>+
>+	MVSW_MSG_TYPE_BRIDGE_CREATE,
>+	MVSW_MSG_TYPE_BRIDGE_DELETE,
>+	MVSW_MSG_TYPE_BRIDGE_PORT_ADD,
>+	MVSW_MSG_TYPE_BRIDGE_PORT_DELETE,
>+
>+	MVSW_MSG_TYPE_ACK,
>+	MVSW_MSG_TYPE_MAX
>+};
>+
>+enum mvsw_msg_port_attr {
>+	MVSW_MSG_PORT_ATTR_ADMIN_STATE,
>+	MVSW_MSG_PORT_ATTR_OPER_STATE,
>+	MVSW_MSG_PORT_ATTR_MTU,
>+	MVSW_MSG_PORT_ATTR_MAC,
>+	MVSW_MSG_PORT_ATTR_SPEED,
>+	MVSW_MSG_PORT_ATTR_ACCEPT_FRAME_TYPE,
>+	MVSW_MSG_PORT_ATTR_LEARNING,
>+	MVSW_MSG_PORT_ATTR_FLOOD,
>+	MVSW_MSG_PORT_ATTR_CAPABILITY,
>+	MVSW_MSG_PORT_ATTR_REMOTE_CAPABILITY,
>+	MVSW_MSG_PORT_ATTR_LINK_MODE,
>+	MVSW_MSG_PORT_ATTR_TYPE,
>+	MVSW_MSG_PORT_ATTR_FEC,
>+	MVSW_MSG_PORT_ATTR_AUTONEG,
>+	MVSW_MSG_PORT_ATTR_DUPLEX,
>+	MVSW_MSG_PORT_ATTR_STATS,
>+	MVSW_MSG_PORT_ATTR_MDIX,
>+	MVSW_MSG_PORT_ATTR_MAX
>+};
>+
>+enum {
>+	MVSW_MSG_ACK_OK,
>+	MVSW_MSG_ACK_FAILED,
>+	MVSW_MSG_ACK_MAX
>+};
>+
>+enum {
>+	MVSW_MODE_FORCED_MDI,
>+	MVSW_MODE_FORCED_MDIX,
>+	MVSW_MODE_AUTO_MDI,
>+	MVSW_MODE_AUTO_MDIX,
>+	MVSW_MODE_AUTO
>+};
>+
>+enum {
>+	MVSW_PORT_GOOD_OCTETS_RCV_CNT,
>+	MVSW_PORT_BAD_OCTETS_RCV_CNT,
>+	MVSW_PORT_MAC_TRANSMIT_ERR_CNT,
>+	MVSW_PORT_BRDC_PKTS_RCV_CNT,
>+	MVSW_PORT_MC_PKTS_RCV_CNT,
>+	MVSW_PORT_PKTS_64_OCTETS_CNT,
>+	MVSW_PORT_PKTS_65TO127_OCTETS_CNT,
>+	MVSW_PORT_PKTS_128TO255_OCTETS_CNT,
>+	MVSW_PORT_PKTS_256TO511_OCTETS_CNT,
>+	MVSW_PORT_PKTS_512TO1023_OCTETS_CNT,
>+	MVSW_PORT_PKTS_1024TOMAX_OCTETS_CNT,
>+	MVSW_PORT_EXCESSIVE_COLLISIONS_CNT,
>+	MVSW_PORT_MC_PKTS_SENT_CNT,
>+	MVSW_PORT_BRDC_PKTS_SENT_CNT,
>+	MVSW_PORT_FC_SENT_CNT,
>+	MVSW_PORT_GOOD_FC_RCV_CNT,
>+	MVSW_PORT_DROP_EVENTS_CNT,
>+	MVSW_PORT_UNDERSIZE_PKTS_CNT,
>+	MVSW_PORT_FRAGMENTS_PKTS_CNT,
>+	MVSW_PORT_OVERSIZE_PKTS_CNT,
>+	MVSW_PORT_JABBER_PKTS_CNT,
>+	MVSW_PORT_MAC_RCV_ERROR_CNT,
>+	MVSW_PORT_BAD_CRC_CNT,
>+	MVSW_PORT_COLLISIONS_CNT,
>+	MVSW_PORT_LATE_COLLISIONS_CNT,
>+	MVSW_PORT_GOOD_UC_PKTS_RCV_CNT,
>+	MVSW_PORT_GOOD_UC_PKTS_SENT_CNT,
>+	MVSW_PORT_MULTIPLE_PKTS_SENT_CNT,
>+	MVSW_PORT_DEFERRED_PKTS_SENT_CNT,
>+	MVSW_PORT_PKTS_1024TO1518_OCTETS_CNT,
>+	MVSW_PORT_PKTS_1519TOMAX_OCTETS_CNT,
>+	MVSW_PORT_GOOD_OCTETS_SENT_CNT,
>+	MVSW_PORT_CNT_MAX,
>+};
>+
>+struct mvsw_msg_cmd {
>+	u32 type;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_ret {
>+	struct mvsw_msg_cmd cmd;
>+	u32 status;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_common_request {
>+	struct mvsw_msg_cmd cmd;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_common_response {
>+	struct mvsw_msg_ret ret;
>+} __packed __aligned(4);
>+
>+union mvsw_msg_switch_param {
>+	u32 ageing_timeout;
>+};
>+
>+struct mvsw_msg_switch_attr_cmd {
>+	struct mvsw_msg_cmd cmd;
>+	union mvsw_msg_switch_param param;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_switch_init_ret {
>+	struct mvsw_msg_ret ret;
>+	u32 port_count;
>+	u32 mtu_max;
>+	u8  switch_id;
>+	u8  mac[ETH_ALEN];
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_port_autoneg_param {
>+	u64 link_mode;
>+	u8  enable;
>+	u8  fec;
>+};
>+
>+struct mvsw_msg_port_cap_param {
>+	u64 link_mode;
>+	u8  type;
>+	u8  fec;
>+	u8  transceiver;
>+};
>+
>+union mvsw_msg_port_param {
>+	u8  admin_state;
>+	u8  oper_state;
>+	u32 mtu;
>+	u8  mac[ETH_ALEN];
>+	u8  accept_frm_type;
>+	u8  learning;
>+	u32 speed;
>+	u8  flood;
>+	u32 link_mode;
>+	u8  type;
>+	u8  duplex;
>+	u8  fec;
>+	u8  mdix;
>+	struct mvsw_msg_port_autoneg_param autoneg;
>+	struct mvsw_msg_port_cap_param cap;
>+};
>+
>+struct mvsw_msg_port_attr_cmd {
>+	struct mvsw_msg_cmd cmd;
>+	u32 attr;
>+	u32 port;
>+	u32 dev;
>+	union mvsw_msg_port_param param;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_port_attr_ret {
>+	struct mvsw_msg_ret ret;
>+	union mvsw_msg_port_param param;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_port_stats_ret {
>+	struct mvsw_msg_ret ret;
>+	u64 stats[MVSW_PORT_CNT_MAX];
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_port_info_cmd {
>+	struct mvsw_msg_cmd cmd;
>+	u32 port;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_port_info_ret {
>+	struct mvsw_msg_ret ret;
>+	u32 hw_id;
>+	u32 dev_id;
>+	u16 fp_id;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_vlan_cmd {
>+	struct mvsw_msg_cmd cmd;
>+	u32 port;
>+	u32 dev;
>+	u16 vid;
>+	u8  is_member;
>+	u8  is_tagged;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_fdb_cmd {
>+	struct mvsw_msg_cmd cmd;
>+	u32 port;
>+	u32 dev;
>+	u8  mac[ETH_ALEN];
>+	u16 vid;
>+	u8  dynamic;
>+	u32 flush_mode;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_event {
>+	u16 type;
>+	u16 id;
>+} __packed __aligned(4);
>+
>+union mvsw_msg_event_fdb_param {
>+	u8 mac[ETH_ALEN];
>+};
>+
>+struct mvsw_msg_event_fdb {
>+	struct mvsw_msg_event id;
>+	u32 port_id;
>+	u32 vid;
>+	union mvsw_msg_event_fdb_param param;
>+} __packed __aligned(4);
>+
>+union mvsw_msg_event_port_param {
>+	u32 oper_state;
>+};
>+
>+struct mvsw_msg_event_port {
>+	struct mvsw_msg_event id;
>+	u32 port_id;
>+	union mvsw_msg_event_port_param param;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_bridge_cmd {
>+	struct mvsw_msg_cmd cmd;
>+	u32 port;
>+	u32 dev;
>+	u16 bridge;
>+} __packed __aligned(4);
>+
>+struct mvsw_msg_bridge_ret {
>+	struct mvsw_msg_ret ret;
>+	u16 bridge;
>+} __packed __aligned(4);
>+
>+#define fw_check_resp(_response)	\
>+({								\
>+	int __er = 0;						\
>+	typeof(_response) __r = (_response);			\
>+	if (__r->ret.cmd.type != MVSW_MSG_TYPE_ACK)		\
>+		__er = -EBADE;					\
>+	else if (__r->ret.status != MVSW_MSG_ACK_OK)		\
>+		__er = -EINVAL;					\
>+	(__er);							\
>+})
>+
>+#define __fw_send_req_resp(_switch, _type, _request, _response, _wait)	\

Please try to avoid doing functions in macros like this one and the
previous one.


>+({								\
>+	int __e;						\
>+	typeof(_switch) __sw = (_switch);			\
>+	typeof(_request) __req = (_request);			\
>+	typeof(_response) __resp = (_response);			\
>+	__req->cmd.type = (_type);				\
>+	__e = __sw->dev->send_req(__sw->dev,			\
>+		(u8 *)__req, sizeof(*__req),			\
>+		(u8 *)__resp, sizeof(*__resp),			\
>+		_wait);						\
>+	if (!__e)						\
>+		__e = fw_check_resp(__resp);			\
>+	(__e);							\
>+})
>+
>+#define fw_send_req_resp(_sw, _t, _req, _resp)	\
>+	__fw_send_req_resp(_sw, _t, _req, _resp, 0)
>+
>+#define fw_send_req_resp_wait(_sw, _t, _req, _resp, _wait)	\
>+	__fw_send_req_resp(_sw, _t, _req, _resp, _wait)
>+
>+#define fw_send_req(_sw, _t, _req)	\

This should be function, not define


>+({							\
>+	struct mvsw_msg_common_response __re;		\
>+	(fw_send_req_resp(_sw, _t, _req, &__re));	\
>+})
>+
>+struct mvsw_fw_event_handler {
>+	struct list_head list;
>+	enum mvsw_pr_event_type type;
>+	void (*func)(struct mvsw_pr_switch *sw, struct mvsw_pr_event *evt);
>+};
>+
>+static int fw_parse_port_evt(u8 *msg, struct mvsw_pr_event *evt)
>+{
>+	struct mvsw_msg_event_port *hw_evt = (struct mvsw_msg_event_port *)msg;
>+
>+	evt->port_evt.port_id = hw_evt->port_id;
>+
>+	if (evt->id == MVSW_PORT_EVENT_STATE_CHANGED)
>+		evt->port_evt.data.oper_state = hw_evt->param.oper_state;
>+	else
>+		return -EINVAL;
>+
>+	return 0;
>+}
>+
>+static int fw_parse_fdb_evt(u8 *msg, struct mvsw_pr_event *evt)
>+{
>+	struct mvsw_msg_event_fdb *hw_evt = (struct mvsw_msg_event_fdb *)msg;
>+
>+	evt->fdb_evt.port_id	= hw_evt->port_id;
>+	evt->fdb_evt.vid	= hw_evt->vid;
>+
>+	memcpy(&evt->fdb_evt.data, &hw_evt->param, sizeof(u8) * ETH_ALEN);
>+
>+	return 0;
>+}
>+
>+struct mvsw_fw_evt_parser {
>+	int (*func)(u8 *msg, struct mvsw_pr_event *evt);
>+};
>+
>+static struct mvsw_fw_evt_parser fw_event_parsers[MVSW_EVENT_TYPE_MAX] = {
>+	[MVSW_EVENT_TYPE_PORT] = {.func = fw_parse_port_evt},
>+	[MVSW_EVENT_TYPE_FDB] = {.func = fw_parse_fdb_evt},
>+};
>+
>+static struct mvsw_fw_event_handler *
>+__find_event_handler(const struct mvsw_pr_switch *sw,

Maintain the function prefix even for helpers like this one:
__prestera_fw_find_event_handler()


>+		     enum mvsw_pr_event_type type)
>+{
>+	struct mvsw_fw_event_handler *eh;
>+
>+	list_for_each_entry_rcu(eh, &sw->event_handlers, list) {
>+		if (eh->type == type)
>+			return eh;
>+	}
>+
>+	return NULL;
>+}
>+
>+static int fw_event_recv(struct mvsw_pr_device *dev, u8 *buf, size_t size)
>+{
>+	void (*cb)(struct mvsw_pr_switch *sw, struct mvsw_pr_event *evt) = NULL;

Please typedef this and use it in struct mvsw_fw_event_handler
definition as well.


>+	struct mvsw_msg_event *msg = (struct mvsw_msg_event *)buf;
>+	struct mvsw_pr_switch *sw = dev->priv;
>+	struct mvsw_fw_event_handler *eh;
>+	struct mvsw_pr_event evt;
>+	int err;
>+
>+	if (msg->type >= MVSW_EVENT_TYPE_MAX)
>+		return -EINVAL;
>+
>+	rcu_read_lock();
>+	eh = __find_event_handler(sw, msg->type);
>+	if (eh)
>+		cb = eh->func;
>+	rcu_read_unlock();
>+
>+	if (!cb || !fw_event_parsers[msg->type].func)
>+		return 0;
>+
>+	evt.id = msg->id;
>+
>+	err = fw_event_parsers[msg->type].func(buf, &evt);
>+	if (!err)
>+		cb(sw, &evt);
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_port_info_get(const struct mvsw_pr_port *port,

Prefix should be rather something like:
"prestera_hw_"


>+			     u16 *fp_id, u32 *hw_id, u32 *dev_id)
>+{
>+	struct mvsw_msg_port_info_ret resp;
>+	struct mvsw_msg_port_info_cmd req = {
>+		.port = port->id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_INFO_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	*hw_id = resp.hw_id;
>+	*dev_id = resp.dev_id;
>+	*fp_id = resp.fp_id;
>+
>+	return 0;
>+}
>+
>+int mvsw_pr_hw_switch_init(struct mvsw_pr_switch *sw)
>+{
>+	struct mvsw_msg_switch_init_ret resp;
>+	struct mvsw_msg_common_request req;
>+	int err = 0;

Pointless init;


>+
>+	INIT_LIST_HEAD(&sw->event_handlers);
>+
>+	err = fw_send_req_resp_wait(sw, MVSW_MSG_TYPE_SWITCH_INIT, &req, &resp,
>+				    MVSW_PR_INIT_TIMEOUT);
>+	if (err)
>+		return err;
>+
>+	sw->id = resp.switch_id;

What is this "switch_id"? u8 does not look like something globally
uniqueue. Rather use base MAC address for example.


>+	sw->port_count = resp.port_count;
>+	sw->mtu_min = MVSW_PR_MIN_MTU;
>+	sw->mtu_max = resp.mtu_max;
>+	sw->dev->recv_msg = fw_event_recv;
>+	memcpy(sw->base_mac, resp.mac, ETH_ALEN);
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_switch_ageing_set(const struct mvsw_pr_switch *sw,
>+				 u32 ageing_time)
>+{
>+	struct mvsw_msg_switch_attr_cmd req = {
>+		.param = {.ageing_timeout = ageing_time}
>+	};
>+
>+	return fw_send_req(sw, MVSW_MSG_TYPE_AGEING_TIMEOUT_SET, &req);
>+}
>+
>+int mvsw_pr_hw_port_state_set(const struct mvsw_pr_port *port,
>+			      bool admin_state)
>+{
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_ADMIN_STATE,
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.param = {.admin_state = admin_state ? 1 : 0}

Just do:
	.admin_state = admin_state


>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_PORT_ATTR_SET, &req);
>+}
>+
>+int mvsw_pr_hw_port_state_get(const struct mvsw_pr_port *port,
>+			      bool *admin_state, bool *oper_state)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	if (admin_state) {
>+		req.attr = MVSW_MSG_PORT_ATTR_ADMIN_STATE;
>+		err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+				       &req, &resp);
>+		if (err)
>+			return err;
>+		*admin_state = resp.param.admin_state != 0;
>+	}
>+
>+	if (oper_state) {
>+		req.attr = MVSW_MSG_PORT_ATTR_OPER_STATE;
>+		err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+				       &req, &resp);
>+		if (err)
>+			return err;
>+		*oper_state = resp.param.oper_state != 0;
>+	}
>+
>+	return 0;
>+}
>+
>+int mvsw_pr_hw_port_mtu_set(const struct mvsw_pr_port *port, u32 mtu)
>+{
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_MTU,
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.param = {.mtu = mtu}
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_PORT_ATTR_SET, &req);
>+}
>+
>+int mvsw_pr_hw_port_mtu_get(const struct mvsw_pr_port *port, u32 *mtu)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_MTU,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	*mtu = resp.param.mtu;
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_port_mac_set(const struct mvsw_pr_port *port, char *mac)
>+{
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_MAC,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	memcpy(&req.param.mac, mac, sizeof(req.param.mac));
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_PORT_ATTR_SET, &req);
>+}
>+
>+int mvsw_pr_hw_port_mac_get(const struct mvsw_pr_port *port, char *mac)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_MAC,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	memcpy(mac, resp.param.mac, sizeof(resp.param.mac));
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_port_accept_frame_type_set(const struct mvsw_pr_port *port,
>+					  enum mvsw_pr_accept_frame_type type)
>+{
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_ACCEPT_FRAME_TYPE,
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.param = {.accept_frm_type = type}
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_PORT_ATTR_SET, &req);
>+}
>+
>+int mvsw_pr_hw_port_learning_set(const struct mvsw_pr_port *port, bool enable)
>+{
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_LEARNING,
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.param = {.learning = enable ? 1 : 0}
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_PORT_ATTR_SET, &req);
>+}
>+
>+int mvsw_pr_hw_event_handler_register(struct mvsw_pr_switch *sw,
>+				      enum mvsw_pr_event_type type,
>+				      void (*cb)(struct mvsw_pr_switch *sw,
>+						 struct mvsw_pr_event *evt))
>+{
>+	struct mvsw_fw_event_handler *eh;
>+
>+	eh = __find_event_handler(sw, type);
>+	if (eh)
>+		return -EEXIST;
>+	eh = kmalloc(sizeof(*eh), GFP_KERNEL);
>+	if (!eh)
>+		return -ENOMEM;
>+
>+	eh->type = type;
>+	eh->func = cb;
>+
>+	INIT_LIST_HEAD(&eh->list);
>+
>+	list_add_rcu(&eh->list, &sw->event_handlers);
>+
>+	return 0;
>+}
>+
>+void mvsw_pr_hw_event_handler_unregister(struct mvsw_pr_switch *sw,
>+					 enum mvsw_pr_event_type type,
>+					 void (*cb)(struct mvsw_pr_switch *sw,
>+						    struct mvsw_pr_event *evt))
>+{
>+	struct mvsw_fw_event_handler *eh;
>+
>+	eh = __find_event_handler(sw, type);
>+	if (!eh)
>+		return;
>+
>+	list_del_rcu(&eh->list);
>+	synchronize_rcu();
>+	kfree(eh);
>+}
>+
>+int mvsw_pr_hw_vlan_create(const struct mvsw_pr_switch *sw, u16 vid)
>+{
>+	struct mvsw_msg_vlan_cmd req = {
>+		.vid = vid,
>+	};
>+
>+	return fw_send_req(sw, MVSW_MSG_TYPE_VLAN_CREATE, &req);
>+}
>+
>+int mvsw_pr_hw_vlan_delete(const struct mvsw_pr_switch *sw, u16 vid)
>+{
>+	struct mvsw_msg_vlan_cmd req = {
>+		.vid = vid,
>+	};
>+
>+	return fw_send_req(sw, MVSW_MSG_TYPE_VLAN_DELETE, &req);
>+}
>+
>+int mvsw_pr_hw_vlan_port_set(const struct mvsw_pr_port *port,
>+			     u16 vid, bool is_member, bool untagged)
>+{
>+	struct mvsw_msg_vlan_cmd req = {
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.vid = vid,
>+		.is_member = is_member ? 1 : 0,
>+		.is_tagged = untagged ? 0 : 1
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_VLAN_PORT_SET, &req);
>+}
>+
>+int mvsw_pr_hw_vlan_port_vid_set(const struct mvsw_pr_port *port, u16 vid)
>+{
>+	struct mvsw_msg_vlan_cmd req = {
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.vid = vid
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_VLAN_PVID_SET, &req);
>+}
>+
>+int mvsw_pr_hw_port_speed_get(const struct mvsw_pr_port *port, u32 *speed)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_SPEED,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	*speed = resp.param.speed;
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_port_flood_set(const struct mvsw_pr_port *port, bool flood)
>+{
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_FLOOD,
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.param = {.flood = flood ? 1 : 0}
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_PORT_ATTR_SET, &req);
>+}
>+
>+int mvsw_pr_hw_fdb_add(const struct mvsw_pr_port *port,
>+		       const unsigned char *mac, u16 vid, bool dynamic)
>+{
>+	struct mvsw_msg_fdb_cmd req = {
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.vid = vid,
>+		.dynamic = dynamic ? 1 : 0
>+	};
>+
>+	memcpy(req.mac, mac, sizeof(req.mac));
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_FDB_ADD, &req);
>+}
>+
>+int mvsw_pr_hw_fdb_del(const struct mvsw_pr_port *port,
>+		       const unsigned char *mac, u16 vid)
>+{
>+	struct mvsw_msg_fdb_cmd req = {
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.vid = vid
>+	};
>+
>+	memcpy(req.mac, mac, sizeof(req.mac));
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_FDB_DELETE, &req);
>+}
>+
>+int mvsw_pr_hw_port_cap_get(const struct mvsw_pr_port *port,
>+			    struct mvsw_pr_port_caps *caps)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_CAPABILITY,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	caps->supp_link_modes = resp.param.cap.link_mode;
>+	caps->supp_fec = resp.param.cap.fec;
>+	caps->type = resp.param.cap.type;
>+	caps->transceiver = resp.param.cap.transceiver;
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_port_remote_cap_get(const struct mvsw_pr_port *port,
>+				   u64 *link_mode_bitmap)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_REMOTE_CAPABILITY,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	*link_mode_bitmap = resp.param.cap.link_mode;
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_port_mdix_get(const struct mvsw_pr_port *port, u8 *mode)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_MDIX,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	switch (resp.param.mdix) {
>+	case MVSW_MODE_FORCED_MDI:
>+	case MVSW_MODE_AUTO_MDI:
>+		*mode = ETH_TP_MDI;
>+		break;
>+
>+	case MVSW_MODE_FORCED_MDIX:
>+	case MVSW_MODE_AUTO_MDIX:
>+		*mode = ETH_TP_MDI_X;
>+		break;
>+
>+	default:
>+		return -EINVAL;
>+	}
>+
>+	return 0;
>+}
>+
>+int mvsw_pr_hw_port_mdix_set(const struct mvsw_pr_port *port, u8 mode)
>+{
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_MDIX,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+
>+	switch (mode) {
>+	case ETH_TP_MDI:
>+		req.param.mdix = MVSW_MODE_FORCED_MDI;
>+		break;
>+
>+	case ETH_TP_MDI_X:
>+		req.param.mdix = MVSW_MODE_FORCED_MDIX;
>+		break;
>+
>+	case ETH_TP_MDI_AUTO:
>+		req.param.mdix = MVSW_MODE_AUTO;
>+		break;
>+
>+	default:
>+		return -EINVAL;
>+	}
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_PORT_ATTR_SET, &req);
>+}
>+
>+int mvsw_pr_hw_port_type_get(const struct mvsw_pr_port *port, u8 *type)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_TYPE,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	*type = resp.param.type;
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_port_fec_get(const struct mvsw_pr_port *port, u8 *fec)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_FEC,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	*fec = resp.param.fec;
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_port_fec_set(const struct mvsw_pr_port *port, u8 fec)
>+{
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_FEC,
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.param = {.fec = fec}
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_PORT_ATTR_SET, &req);
>+}
>+
>+int mvsw_pr_hw_port_autoneg_set(const struct mvsw_pr_port *port,
>+				bool autoneg, u64 link_modes, u8 fec)
>+{
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_AUTONEG,
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.param = {.autoneg = {.link_mode = link_modes,
>+				      .enable = autoneg ? 1 : 0,

You can do just:
	.enable = autoneg;


>+				      .fec = fec}
>+		}
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_PORT_ATTR_SET, &req);
>+}
>+
>+int mvsw_pr_hw_port_duplex_get(const struct mvsw_pr_port *port, u8 *duplex)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_DUPLEX,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	*duplex = resp.param.duplex;
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_port_stats_get(const struct mvsw_pr_port *port,
>+			      struct mvsw_pr_port_stats *stats)
>+{
>+	struct mvsw_msg_port_stats_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_STATS,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	u64 *hw_val = resp.stats;
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	stats->good_octets_received = hw_val[MVSW_PORT_GOOD_OCTETS_RCV_CNT];
>+	stats->bad_octets_received = hw_val[MVSW_PORT_BAD_OCTETS_RCV_CNT];
>+	stats->mac_trans_error = hw_val[MVSW_PORT_MAC_TRANSMIT_ERR_CNT];
>+	stats->broadcast_frames_received = hw_val[MVSW_PORT_BRDC_PKTS_RCV_CNT];
>+	stats->multicast_frames_received = hw_val[MVSW_PORT_MC_PKTS_RCV_CNT];
>+	stats->frames_64_octets = hw_val[MVSW_PORT_PKTS_64_OCTETS_CNT];
>+	stats->frames_65_to_127_octets =
>+		hw_val[MVSW_PORT_PKTS_65TO127_OCTETS_CNT];
>+	stats->frames_128_to_255_octets =
>+		hw_val[MVSW_PORT_PKTS_128TO255_OCTETS_CNT];
>+	stats->frames_256_to_511_octets =
>+		hw_val[MVSW_PORT_PKTS_256TO511_OCTETS_CNT];
>+	stats->frames_512_to_1023_octets =
>+		hw_val[MVSW_PORT_PKTS_512TO1023_OCTETS_CNT];
>+	stats->frames_1024_to_max_octets =
>+		hw_val[MVSW_PORT_PKTS_1024TOMAX_OCTETS_CNT];
>+	stats->excessive_collision = hw_val[MVSW_PORT_EXCESSIVE_COLLISIONS_CNT];
>+	stats->multicast_frames_sent = hw_val[MVSW_PORT_MC_PKTS_SENT_CNT];
>+	stats->broadcast_frames_sent = hw_val[MVSW_PORT_BRDC_PKTS_SENT_CNT];
>+	stats->fc_sent = hw_val[MVSW_PORT_FC_SENT_CNT];
>+	stats->fc_received = hw_val[MVSW_PORT_GOOD_FC_RCV_CNT];
>+	stats->buffer_overrun = hw_val[MVSW_PORT_DROP_EVENTS_CNT];
>+	stats->undersize = hw_val[MVSW_PORT_UNDERSIZE_PKTS_CNT];
>+	stats->fragments = hw_val[MVSW_PORT_FRAGMENTS_PKTS_CNT];
>+	stats->oversize = hw_val[MVSW_PORT_OVERSIZE_PKTS_CNT];
>+	stats->jabber = hw_val[MVSW_PORT_JABBER_PKTS_CNT];
>+	stats->rx_error_frame_received = hw_val[MVSW_PORT_MAC_RCV_ERROR_CNT];
>+	stats->bad_crc = hw_val[MVSW_PORT_BAD_CRC_CNT];
>+	stats->collisions = hw_val[MVSW_PORT_COLLISIONS_CNT];
>+	stats->late_collision = hw_val[MVSW_PORT_LATE_COLLISIONS_CNT];
>+	stats->unicast_frames_received = hw_val[MVSW_PORT_GOOD_UC_PKTS_RCV_CNT];
>+	stats->unicast_frames_sent = hw_val[MVSW_PORT_GOOD_UC_PKTS_SENT_CNT];
>+	stats->sent_multiple = hw_val[MVSW_PORT_MULTIPLE_PKTS_SENT_CNT];
>+	stats->sent_deferred = hw_val[MVSW_PORT_DEFERRED_PKTS_SENT_CNT];
>+	stats->frames_1024_to_1518_octets =
>+		hw_val[MVSW_PORT_PKTS_1024TO1518_OCTETS_CNT];
>+	stats->frames_1519_to_max_octets =
>+		hw_val[MVSW_PORT_PKTS_1519TOMAX_OCTETS_CNT];
>+	stats->good_octets_sent = hw_val[MVSW_PORT_GOOD_OCTETS_SENT_CNT];
>+
>+	return 0;
>+}
>+
>+int mvsw_pr_hw_bridge_create(const struct mvsw_pr_switch *sw, u16 *bridge_id)
>+{
>+	struct mvsw_msg_bridge_cmd req;
>+	struct mvsw_msg_bridge_ret resp;
>+	int err;
>+
>+	err = fw_send_req_resp(sw, MVSW_MSG_TYPE_BRIDGE_CREATE, &req, &resp);
>+	if (err)
>+		return err;
>+
>+	*bridge_id = resp.bridge;
>+	return err;
>+}
>+
>+int mvsw_pr_hw_bridge_delete(const struct mvsw_pr_switch *sw, u16 bridge_id)
>+{
>+	struct mvsw_msg_bridge_cmd req = {
>+		.bridge = bridge_id
>+	};
>+
>+	return fw_send_req(sw, MVSW_MSG_TYPE_BRIDGE_DELETE, &req);
>+}
>+
>+int mvsw_pr_hw_bridge_port_add(const struct mvsw_pr_port *port, u16 bridge_id)
>+{
>+	struct mvsw_msg_bridge_cmd req = {
>+		.bridge = bridge_id,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_BRIDGE_PORT_ADD, &req);
>+}
>+
>+int mvsw_pr_hw_bridge_port_delete(const struct mvsw_pr_port *port,
>+				  u16 bridge_id)
>+{
>+	struct mvsw_msg_bridge_cmd req = {
>+		.bridge = bridge_id,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_BRIDGE_PORT_DELETE, &req);
>+}
>+
>+int mvsw_pr_hw_fdb_flush_port(const struct mvsw_pr_port *port, u32 mode)
>+{
>+	struct mvsw_msg_fdb_cmd req = {
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.flush_mode = mode,
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_FDB_FLUSH_PORT, &req);
>+}
>+
>+int mvsw_pr_hw_fdb_flush_vlan(const struct mvsw_pr_switch *sw, u16 vid,
>+			      u32 mode)
>+{
>+	struct mvsw_msg_fdb_cmd req = {
>+		.vid = vid,
>+		.flush_mode = mode,
>+	};
>+
>+	return fw_send_req(sw, MVSW_MSG_TYPE_FDB_FLUSH_VLAN, &req);
>+}
>+
>+int mvsw_pr_hw_fdb_flush_port_vlan(const struct mvsw_pr_port *port, u16 vid,
>+				   u32 mode)
>+{
>+	struct mvsw_msg_fdb_cmd req = {
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.vid = vid,
>+		.flush_mode = mode,
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_FDB_FLUSH_PORT_VLAN, &req);
>+}
>+
>+int mvsw_pr_hw_port_link_mode_get(const struct mvsw_pr_port *port,
>+				  u32 *mode)
>+{
>+	struct mvsw_msg_port_attr_ret resp;
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_LINK_MODE,
>+		.port = port->hw_id,
>+		.dev = port->dev_id
>+	};
>+	int err;
>+
>+	err = fw_send_req_resp(port->sw, MVSW_MSG_TYPE_PORT_ATTR_GET,
>+			       &req, &resp);
>+	if (err)
>+		return err;
>+
>+	*mode = resp.param.link_mode;
>+
>+	return err;
>+}
>+
>+int mvsw_pr_hw_port_link_mode_set(const struct mvsw_pr_port *port,
>+				  u32 mode)
>+{
>+	struct mvsw_msg_port_attr_cmd req = {
>+		.attr = MVSW_MSG_PORT_ATTR_LINK_MODE,
>+		.port = port->hw_id,
>+		.dev = port->dev_id,
>+		.param = {.link_mode = mode}
>+	};
>+
>+	return fw_send_req(port->sw, MVSW_MSG_TYPE_PORT_ATTR_SET, &req);
>+}
>diff --git a/drivers/net/ethernet/marvell/prestera/prestera_hw.h b/drivers/net/ethernet/marvell/prestera/prestera_hw.h
>new file mode 100644
>index 000000000000..dfae2631160e
>--- /dev/null
>+++ b/drivers/net/ethernet/marvell/prestera/prestera_hw.h
>@@ -0,0 +1,159 @@
>+/* SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0
>+ *
>+ * Copyright (c) 2019-2020 Marvell International Ltd. All rights reserved.
>+ *
>+ */
>+
>+#ifndef _MVSW_PRESTERA_HW_H_
>+#define _MVSW_PRESTERA_HW_H_
>+
>+#include <linux/types.h>
>+
>+enum mvsw_pr_accept_frame_type {
>+	MVSW_ACCEPT_FRAME_TYPE_TAGGED,
>+	MVSW_ACCEPT_FRAME_TYPE_UNTAGGED,
>+	MVSW_ACCEPT_FRAME_TYPE_ALL
>+};
>+
>+enum {
>+	MVSW_LINK_MODE_10baseT_Half_BIT,
>+	MVSW_LINK_MODE_10baseT_Full_BIT,
>+	MVSW_LINK_MODE_100baseT_Half_BIT,
>+	MVSW_LINK_MODE_100baseT_Full_BIT,
>+	MVSW_LINK_MODE_1000baseT_Half_BIT,
>+	MVSW_LINK_MODE_1000baseT_Full_BIT,
>+	MVSW_LINK_MODE_1000baseX_Full_BIT,
>+	MVSW_LINK_MODE_1000baseKX_Full_BIT,
>+	MVSW_LINK_MODE_10GbaseKR_Full_BIT,
>+	MVSW_LINK_MODE_10GbaseSR_Full_BIT,
>+	MVSW_LINK_MODE_10GbaseLR_Full_BIT,
>+	MVSW_LINK_MODE_20GbaseKR2_Full_BIT,
>+	MVSW_LINK_MODE_25GbaseCR_Full_BIT,
>+	MVSW_LINK_MODE_25GbaseKR_Full_BIT,
>+	MVSW_LINK_MODE_25GbaseSR_Full_BIT,
>+	MVSW_LINK_MODE_40GbaseKR4_Full_BIT,
>+	MVSW_LINK_MODE_40GbaseCR4_Full_BIT,
>+	MVSW_LINK_MODE_40GbaseSR4_Full_BIT,
>+	MVSW_LINK_MODE_50GbaseCR2_Full_BIT,
>+	MVSW_LINK_MODE_50GbaseKR2_Full_BIT,
>+	MVSW_LINK_MODE_50GbaseSR2_Full_BIT,
>+	MVSW_LINK_MODE_100GbaseKR4_Full_BIT,
>+	MVSW_LINK_MODE_100GbaseSR4_Full_BIT,
>+	MVSW_LINK_MODE_100GbaseCR4_Full_BIT,
>+	MVSW_LINK_MODE_MAX,
>+};
>+
>+enum {
>+	MVSW_PORT_TYPE_NONE,
>+	MVSW_PORT_TYPE_TP,
>+	MVSW_PORT_TYPE_AUI,
>+	MVSW_PORT_TYPE_MII,
>+	MVSW_PORT_TYPE_FIBRE,
>+	MVSW_PORT_TYPE_BNC,
>+	MVSW_PORT_TYPE_DA,
>+	MVSW_PORT_TYPE_OTHER,
>+	MVSW_PORT_TYPE_MAX,
>+};
>+
>+enum {
>+	MVSW_PORT_TRANSCEIVER_COPPER,
>+	MVSW_PORT_TRANSCEIVER_SFP,
>+	MVSW_PORT_TRANSCEIVER_MAX,
>+};
>+
>+enum {
>+	MVSW_PORT_FEC_OFF_BIT,
>+	MVSW_PORT_FEC_BASER_BIT,
>+	MVSW_PORT_FEC_RS_BIT,
>+	MVSW_PORT_FEC_MAX,
>+};
>+
>+enum {
>+	MVSW_PORT_DUPLEX_HALF,
>+	MVSW_PORT_DUPLEX_FULL
>+};
>+
>+struct mvsw_pr_switch;
>+struct mvsw_pr_port;

For consistently sake, name this rather like this:
struct prestera;
struct prestera_port;

And use it as this for variables and args:

struct prestera *prestera;
struct prestera_port *port;

Or something like that. The point is, avoid "switch" and "sw".


>+struct mvsw_pr_port_stats;
>+struct mvsw_pr_port_caps;
>+
>+enum mvsw_pr_event_type;
>+struct mvsw_pr_event;
>+
>+/* Switch API */
>+int mvsw_pr_hw_switch_init(struct mvsw_pr_switch *sw);
>+int mvsw_pr_hw_switch_ageing_set(const struct mvsw_pr_switch *sw,
>+				 u32 ageing_time);
>+
>+/* Port API */
>+int mvsw_pr_hw_port_info_get(const struct mvsw_pr_port *port,
>+			     u16 *fp_id, u32 *hw_id, u32 *dev_id);
>+int mvsw_pr_hw_port_state_set(const struct mvsw_pr_port *port,
>+			      bool admin_state);
>+int mvsw_pr_hw_port_state_get(const struct mvsw_pr_port *port,
>+			      bool *admin_state, bool *oper_state);
>+int mvsw_pr_hw_port_mtu_set(const struct mvsw_pr_port *port, u32 mtu);
>+int mvsw_pr_hw_port_mtu_get(const struct mvsw_pr_port *port, u32 *mtu);
>+int mvsw_pr_hw_port_mac_set(const struct mvsw_pr_port *port, char *mac);
>+int mvsw_pr_hw_port_mac_get(const struct mvsw_pr_port *port, char *mac);
>+int mvsw_pr_hw_port_accept_frame_type_set(const struct mvsw_pr_port *port,
>+					  enum mvsw_pr_accept_frame_type type);
>+int mvsw_pr_hw_port_learning_set(const struct mvsw_pr_port *port, bool enable);
>+int mvsw_pr_hw_port_speed_get(const struct mvsw_pr_port *port, u32 *speed);
>+int mvsw_pr_hw_port_flood_set(const struct mvsw_pr_port *port, bool flood);
>+int mvsw_pr_hw_port_cap_get(const struct mvsw_pr_port *port,
>+			    struct mvsw_pr_port_caps *caps);
>+int mvsw_pr_hw_port_remote_cap_get(const struct mvsw_pr_port *port,
>+				   u64 *link_mode_bitmap);
>+int mvsw_pr_hw_port_type_get(const struct mvsw_pr_port *port, u8 *type);
>+int mvsw_pr_hw_port_fec_get(const struct mvsw_pr_port *port, u8 *fec);
>+int mvsw_pr_hw_port_fec_set(const struct mvsw_pr_port *port, u8 fec);
>+int mvsw_pr_hw_port_autoneg_set(const struct mvsw_pr_port *port,
>+				bool autoneg, u64 link_modes, u8 fec);
>+int mvsw_pr_hw_port_duplex_get(const struct mvsw_pr_port *port, u8 *duplex);
>+int mvsw_pr_hw_port_stats_get(const struct mvsw_pr_port *port,
>+			      struct mvsw_pr_port_stats *stats);
>+int mvsw_pr_hw_port_link_mode_get(const struct mvsw_pr_port *port,
>+				  u32 *mode);
>+int mvsw_pr_hw_port_link_mode_set(const struct mvsw_pr_port *port,
>+				  u32 mode);
>+int mvsw_pr_hw_port_mdix_get(const struct mvsw_pr_port *port, u8 *mode);
>+int mvsw_pr_hw_port_mdix_set(const struct mvsw_pr_port *port, u8 mode);
>+
>+/* Vlan API */
>+int mvsw_pr_hw_vlan_create(const struct mvsw_pr_switch *sw, u16 vid);
>+int mvsw_pr_hw_vlan_delete(const struct mvsw_pr_switch *sw, u16 vid);
>+int mvsw_pr_hw_vlan_port_set(const struct mvsw_pr_port *port,
>+			     u16 vid, bool is_member, bool untagged);
>+int mvsw_pr_hw_vlan_port_vid_set(const struct mvsw_pr_port *port, u16 vid);
>+
>+/* FDB API */
>+int mvsw_pr_hw_fdb_add(const struct mvsw_pr_port *port,
>+		       const unsigned char *mac, u16 vid, bool dynamic);
>+int mvsw_pr_hw_fdb_del(const struct mvsw_pr_port *port,
>+		       const unsigned char *mac, u16 vid);
>+int mvsw_pr_hw_fdb_flush_port(const struct mvsw_pr_port *port, u32 mode);
>+int mvsw_pr_hw_fdb_flush_vlan(const struct mvsw_pr_switch *sw, u16 vid,
>+			      u32 mode);
>+int mvsw_pr_hw_fdb_flush_port_vlan(const struct mvsw_pr_port *port, u16 vid,
>+				   u32 mode);
>+
>+/* Bridge API */
>+int mvsw_pr_hw_bridge_create(const struct mvsw_pr_switch *sw, u16 *bridge_id);
>+int mvsw_pr_hw_bridge_delete(const struct mvsw_pr_switch *sw, u16 bridge_id);
>+int mvsw_pr_hw_bridge_port_add(const struct mvsw_pr_port *port, u16 bridge_id);
>+int mvsw_pr_hw_bridge_port_delete(const struct mvsw_pr_port *port,
>+				  u16 bridge_id);
>+
>+/* Event handlers */
>+int mvsw_pr_hw_event_handler_register(struct mvsw_pr_switch *sw,
>+				      enum mvsw_pr_event_type type,
>+				      void (*cb)(struct mvsw_pr_switch *sw,
>+						 struct mvsw_pr_event *evt));
>+void mvsw_pr_hw_event_handler_unregister(struct mvsw_pr_switch *sw,
>+					 enum mvsw_pr_event_type type,
>+					 void (*cb)(struct mvsw_pr_switch *sw,
>+						    struct mvsw_pr_event *evt));
>+
>+#endif /* _MVSW_PRESTERA_HW_H_ */
>diff --git a/drivers/net/ethernet/marvell/prestera/prestera_switchdev.c b/drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
>new file mode 100644
>index 000000000000..18fa6bbe5ace
>--- /dev/null
>+++ b/drivers/net/ethernet/marvell/prestera/prestera_switchdev.c
>@@ -0,0 +1,1217 @@
>+/* SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0
>+ *
>+ * Copyright (c) 2019-2020 Marvell International Ltd. All rights reserved.
>+ *
>+ */
>+#include <linux/kernel.h>
>+#include <linux/module.h>
>+#include <linux/if_vlan.h>
>+#include <linux/if_bridge.h>
>+#include <linux/notifier.h>
>+#include <net/switchdev.h>
>+#include <net/netevent.h>
>+#include <net/vxlan.h>
>+
>+#include "prestera.h"
>+
>+struct mvsw_pr_bridge {
>+	struct mvsw_pr_switch *sw;
>+	u32 ageing_time;

Ageing time is global for ASIC? In linux, it is per-bridge.


>+	struct list_head bridge_list;

I

>+	bool bridge_8021q_exists;
>+};
>+
>+struct mvsw_pr_bridge_device {
>+	struct net_device *dev;
>+	struct list_head bridge_node;
>+	struct list_head port_list;
>+	u16 bridge_id;
>+	u8 vlan_enabled:1, multicast_enabled:1, mrouter:1;
>+};
>+
>+struct mvsw_pr_bridge_port {
>+	struct net_device *dev;
>+	struct mvsw_pr_bridge_device *bridge_device;
>+	struct list_head bridge_device_node;
>+	struct list_head vlan_list;
>+	unsigned int ref_count;
>+	u8 stp_state;
>+	unsigned long flags;
>+};
>+
>+struct mvsw_pr_bridge_vlan {
>+	struct list_head bridge_port_node;
>+	struct list_head port_vlan_list;
>+	u16 vid;
>+};
>+
>+struct mvsw_pr_event_work {
>+	struct work_struct work;
>+	struct switchdev_notifier_fdb_info fdb_info;
>+	struct net_device *dev;
>+	unsigned long event;
>+};
>+
>+static struct workqueue_struct *mvsw_owq;
>+
>+static struct mvsw_pr_bridge_port *
>+mvsw_pr_bridge_port_get(struct mvsw_pr_bridge *bridge,
>+			struct net_device *brport_dev);
>+
>+static void mvsw_pr_bridge_port_put(struct mvsw_pr_bridge *bridge,
>+				    struct mvsw_pr_bridge_port *br_port);
>+
>+static struct mvsw_pr_bridge_device *
>+mvsw_pr_bridge_device_find(const struct mvsw_pr_bridge *bridge,
>+			   const struct net_device *br_dev)
>+{
>+	struct mvsw_pr_bridge_device *bridge_device;
>+
>+	list_for_each_entry(bridge_device, &bridge->bridge_list,
>+			    bridge_node)
>+		if (bridge_device->dev == br_dev)
>+			return bridge_device;
>+
>+	return NULL;
>+}
>+
>+static bool
>+mvsw_pr_bridge_device_is_offloaded(const struct mvsw_pr_switch *sw,
>+				   const struct net_device *br_dev)
>+{
>+	return !!mvsw_pr_bridge_device_find(sw->bridge, br_dev);
>+}
>+
>+static struct mvsw_pr_bridge_port *
>+__mvsw_pr_bridge_port_find(const struct mvsw_pr_bridge_device *bridge_device,
>+			   const struct net_device *brport_dev)
>+{
>+	struct mvsw_pr_bridge_port *br_port;
>+
>+	list_for_each_entry(br_port, &bridge_device->port_list,
>+			    bridge_device_node) {
>+		if (br_port->dev == brport_dev)
>+			return br_port;
>+	}

No need for "{}"


>+
>+	return NULL;
>+}
>+
>+static struct mvsw_pr_bridge_port *
>+mvsw_pr_bridge_port_find(struct mvsw_pr_bridge *bridge,
>+			 struct net_device *brport_dev)
>+{
>+	struct net_device *br_dev = netdev_master_upper_dev_get(brport_dev);
>+	struct mvsw_pr_bridge_device *bridge_device;
>+
>+	if (!br_dev)
>+		return NULL;
>+
>+	bridge_device = mvsw_pr_bridge_device_find(bridge, br_dev);
>+	if (!bridge_device)
>+		return NULL;
>+
>+	return __mvsw_pr_bridge_port_find(bridge_device, brport_dev);
>+}
>+
>+static struct mvsw_pr_bridge_vlan *
>+mvsw_pr_bridge_vlan_find(const struct mvsw_pr_bridge_port *br_port, u16 vid)
>+{
>+	struct mvsw_pr_bridge_vlan *br_vlan;
>+
>+	list_for_each_entry(br_vlan, &br_port->vlan_list, bridge_port_node) {
>+		if (br_vlan->vid == vid)
>+			return br_vlan;
>+	}
>+
>+	return NULL;
>+}
>+
>+static struct mvsw_pr_bridge_vlan *
>+mvsw_pr_bridge_vlan_create(struct mvsw_pr_bridge_port *br_port, u16 vid)
>+{
>+	struct mvsw_pr_bridge_vlan *br_vlan;
>+
>+	br_vlan = kzalloc(sizeof(*br_vlan), GFP_KERNEL);
>+	if (!br_vlan)
>+		return NULL;
>+
>+	INIT_LIST_HEAD(&br_vlan->port_vlan_list);
>+	br_vlan->vid = vid;
>+	list_add(&br_vlan->bridge_port_node, &br_port->vlan_list);
>+
>+	return br_vlan;
>+}
>+
>+static void
>+mvsw_pr_bridge_vlan_destroy(struct mvsw_pr_bridge_vlan *br_vlan)
>+{
>+	list_del(&br_vlan->bridge_port_node);
>+	WARN_ON(!list_empty(&br_vlan->port_vlan_list));
>+	kfree(br_vlan);
>+}
>+
>+static struct mvsw_pr_bridge_vlan *
>+mvsw_pr_bridge_vlan_get(struct mvsw_pr_bridge_port *br_port, u16 vid)
>+{
>+	struct mvsw_pr_bridge_vlan *br_vlan;
>+
>+	br_vlan = mvsw_pr_bridge_vlan_find(br_port, vid);
>+	if (br_vlan)
>+		return br_vlan;
>+
>+	return mvsw_pr_bridge_vlan_create(br_port, vid);
>+}
>+
>+static void mvsw_pr_bridge_vlan_put(struct mvsw_pr_bridge_vlan *br_vlan)
>+{
>+	if (list_empty(&br_vlan->port_vlan_list))
>+		mvsw_pr_bridge_vlan_destroy(br_vlan);
>+}
>+
>+static int
>+mvsw_pr_port_vlan_bridge_join(struct mvsw_pr_port_vlan *port_vlan,
>+			      struct mvsw_pr_bridge_port *br_port,
>+			      struct netlink_ext_ack *extack)
>+{
>+	struct mvsw_pr_port *port = port_vlan->mvsw_pr_port;
>+	struct mvsw_pr_bridge_vlan *br_vlan;
>+	u16 vid = port_vlan->vid;
>+	int err;
>+
>+	if (port_vlan->bridge_port)
>+		return 0;
>+
>+	err = mvsw_pr_port_flood_set(port, br_port->flags & BR_FLOOD);
>+	if (err)
>+		return err;
>+
>+	err = mvsw_pr_port_learning_set(port, br_port->flags & BR_LEARNING);
>+	if (err)
>+		goto err_port_learning_set;
>+
>+	br_vlan = mvsw_pr_bridge_vlan_get(br_port, vid);
>+	if (!br_vlan) {
>+		err = -ENOMEM;
>+		goto err_bridge_vlan_get;
>+	}
>+
>+	list_add(&port_vlan->bridge_vlan_node, &br_vlan->port_vlan_list);
>+
>+	mvsw_pr_bridge_port_get(port->sw->bridge, br_port->dev);
>+	port_vlan->bridge_port = br_port;
>+
>+	return 0;
>+
>+err_bridge_vlan_get:
>+	mvsw_pr_port_learning_set(port, false);
>+err_port_learning_set:
>+	return err;
>+}
>+
>+static int
>+mvsw_pr_bridge_vlan_port_count_get(struct mvsw_pr_bridge_device *bridge_device,
>+				   u16 vid)
>+{
>+	int count = 0;
>+	struct mvsw_pr_bridge_port *br_port;
>+	struct mvsw_pr_bridge_vlan *br_vlan;
>+
>+	list_for_each_entry(br_port, &bridge_device->port_list,
>+			    bridge_device_node) {
>+		list_for_each_entry(br_vlan, &br_port->vlan_list,
>+				    bridge_port_node) {
>+			if (br_vlan->vid == vid) {
>+				count += 1;
>+				break;
>+			}
>+		}
>+	}
>+
>+	return count;
>+}
>+
>+void
>+mvsw_pr_port_vlan_bridge_leave(struct mvsw_pr_port_vlan *port_vlan)
>+{
>+	struct mvsw_pr_port *port = port_vlan->mvsw_pr_port;
>+	struct mvsw_pr_bridge_vlan *br_vlan;
>+	struct mvsw_pr_bridge_port *br_port;
>+	int port_count;
>+	u16 vid = port_vlan->vid;
>+	bool last_port, last_vlan;
>+
>+	br_port = port_vlan->bridge_port;
>+	last_vlan = list_is_singular(&br_port->vlan_list);
>+	port_count =
>+	    mvsw_pr_bridge_vlan_port_count_get(br_port->bridge_device, vid);
>+	br_vlan = mvsw_pr_bridge_vlan_find(br_port, vid);
>+	last_port = port_count == 1;
>+	if (last_vlan) {
>+		mvsw_pr_fdb_flush_port(port, MVSW_PR_FDB_FLUSH_MODE_DYNAMIC);
>+	} else if (last_port) {
>+		mvsw_pr_fdb_flush_vlan(port->sw, vid,
>+				       MVSW_PR_FDB_FLUSH_MODE_DYNAMIC);
>+	} else {
>+		mvsw_pr_fdb_flush_port_vlan(port, vid,
>+					    MVSW_PR_FDB_FLUSH_MODE_DYNAMIC);
>+	}
>+
>+	list_del(&port_vlan->bridge_vlan_node);
>+	mvsw_pr_bridge_vlan_put(br_vlan);
>+	mvsw_pr_bridge_port_put(port->sw->bridge, br_port);
>+	port_vlan->bridge_port = NULL;
>+}
>+
>+static int
>+mvsw_pr_bridge_port_vlan_add(struct mvsw_pr_port *port,
>+			     struct mvsw_pr_bridge_port *br_port,
>+			     u16 vid, bool is_untagged, bool is_pvid,
>+			     struct netlink_ext_ack *extack)
>+{
>+	u16 pvid;
>+	struct mvsw_pr_port_vlan *port_vlan;
>+	u16 old_pvid = port->pvid;
>+	int err;
>+
>+	if (is_pvid)
>+		pvid = vid;
>+	else
>+		pvid = port->pvid == vid ? 0 : port->pvid;
>+
>+	port_vlan = mvsw_pr_port_vlan_find_by_vid(port, vid);
>+	if (port_vlan && port_vlan->bridge_port != br_port)
>+		return -EEXIST;
>+
>+	if (!port_vlan) {
>+		port_vlan = mvsw_pr_port_vlan_create(port, vid);
>+		if (IS_ERR(port_vlan))
>+			return PTR_ERR(port_vlan);
>+	}
>+
>+	err = mvsw_pr_port_vlan_set(port, vid, true, is_untagged);
>+	if (err)
>+		goto err_port_vlan_set;
>+
>+	err = mvsw_pr_port_pvid_set(port, pvid);
>+	if (err)
>+		goto err_port_pvid_set;
>+
>+	err = mvsw_pr_port_vlan_bridge_join(port_vlan, br_port, extack);
>+	if (err)
>+		goto err_port_vlan_bridge_join;
>+
>+	return 0;
>+
>+err_port_vlan_bridge_join:
>+	mvsw_pr_port_pvid_set(port, old_pvid);
>+err_port_pvid_set:
>+	mvsw_pr_port_vlan_set(port, vid, false, false);
>+err_port_vlan_set:
>+	mvsw_pr_port_vlan_destroy(port_vlan);
>+
>+	return err;
>+}
>+
>+static int mvsw_pr_port_vlans_add(struct mvsw_pr_port *port,
>+				  const struct switchdev_obj_port_vlan *vlan,
>+				  struct switchdev_trans *trans,
>+				  struct netlink_ext_ack *extack)
>+{
>+	bool flag_untagged = vlan->flags & BRIDGE_VLAN_INFO_UNTAGGED;
>+	bool flag_pvid = vlan->flags & BRIDGE_VLAN_INFO_PVID;
>+	struct net_device *orig_dev = vlan->obj.orig_dev;
>+	struct mvsw_pr_bridge_port *br_port;
>+	struct mvsw_pr_switch *sw = port->sw;
>+	u16 vid;
>+
>+	if (netif_is_bridge_master(orig_dev))
>+		return 0;
>+
>+	if (switchdev_trans_ph_commit(trans))
>+		return 0;
>+
>+	br_port = mvsw_pr_bridge_port_find(sw->bridge, orig_dev);
>+	if (WARN_ON(!br_port))
>+		return -EINVAL;
>+
>+	if (!br_port->bridge_device->vlan_enabled)
>+		return 0;
>+
>+	for (vid = vlan->vid_begin; vid <= vlan->vid_end; vid++) {
>+		int err;
>+
>+		err = mvsw_pr_bridge_port_vlan_add(port, br_port,
>+						   vid, flag_untagged,
>+						   flag_pvid, extack);
>+		if (err)
>+			return err;
>+	}
>+
>+	return 0;
>+}
>+
>+static int mvsw_pr_port_obj_add(struct net_device *dev,
>+				const struct switchdev_obj *obj,
>+				struct switchdev_trans *trans,
>+				struct netlink_ext_ack *extack)
>+{
>+	int err = 0;
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+	const struct switchdev_obj_port_vlan *vlan;
>+
>+	switch (obj->id) {
>+	case SWITCHDEV_OBJ_ID_PORT_VLAN:
>+		vlan = SWITCHDEV_OBJ_PORT_VLAN(obj);
>+		err = mvsw_pr_port_vlans_add(port, vlan, trans, extack);
>+		break;
>+	default:
>+		err = -EOPNOTSUPP;
>+	}
>+
>+	return err;
>+}
>+
>+static void
>+mvsw_pr_bridge_port_vlan_del(struct mvsw_pr_port *port,
>+			     struct mvsw_pr_bridge_port *br_port, u16 vid)
>+{
>+	u16 pvid = port->pvid == vid ? 0 : port->pvid;
>+	struct mvsw_pr_port_vlan *port_vlan;
>+
>+	port_vlan = mvsw_pr_port_vlan_find_by_vid(port, vid);
>+	if (WARN_ON(!port_vlan))
>+		return;
>+
>+	mvsw_pr_port_vlan_bridge_leave(port_vlan);
>+	mvsw_pr_port_pvid_set(port, pvid);
>+	mvsw_pr_port_vlan_destroy(port_vlan);
>+}
>+
>+static int mvsw_pr_port_vlans_del(struct mvsw_pr_port *port,
>+				  const struct switchdev_obj_port_vlan *vlan)
>+{
>+	struct mvsw_pr_switch *sw = port->sw;
>+	struct net_device *orig_dev = vlan->obj.orig_dev;
>+	struct mvsw_pr_bridge_port *br_port;
>+	u16 vid;
>+
>+	if (netif_is_bridge_master(orig_dev))
>+		return -EOPNOTSUPP;
>+
>+	br_port = mvsw_pr_bridge_port_find(sw->bridge, orig_dev);
>+	if (WARN_ON(!br_port))
>+		return -EINVAL;
>+
>+	if (!br_port->bridge_device->vlan_enabled)
>+		return 0;
>+
>+	for (vid = vlan->vid_begin; vid <= vlan->vid_end; vid++)
>+		mvsw_pr_bridge_port_vlan_del(port, br_port, vid);
>+
>+	return 0;
>+}
>+
>+static int mvsw_pr_port_obj_del(struct net_device *dev,
>+				const struct switchdev_obj *obj)
>+{
>+	int err = 0;
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+
>+	switch (obj->id) {
>+	case SWITCHDEV_OBJ_ID_PORT_VLAN:
>+		err = mvsw_pr_port_vlans_del(port,
>+					     SWITCHDEV_OBJ_PORT_VLAN(obj));
>+		break;
>+	default:
>+		err = -EOPNOTSUPP;
>+		break;
>+	}
>+
>+	return err;
>+}
>+
>+static int mvsw_pr_port_attr_br_vlan_set(struct mvsw_pr_port *port,
>+					 struct switchdev_trans *trans,
>+					 struct net_device *orig_dev,
>+					 bool vlan_enabled)
>+{
>+	struct mvsw_pr_switch *sw = port->sw;
>+	struct mvsw_pr_bridge_device *bridge_device;
>+
>+	if (!switchdev_trans_ph_prepare(trans))
>+		return 0;
>+
>+	bridge_device = mvsw_pr_bridge_device_find(sw->bridge, orig_dev);
>+	if (WARN_ON(!bridge_device))
>+		return -EINVAL;
>+
>+	if (bridge_device->vlan_enabled == vlan_enabled)
>+		return 0;
>+
>+	netdev_err(bridge_device->dev,
>+		   "VLAN filtering can't be changed for existing bridge\n");
>+	return -EINVAL;
>+}
>+
>+static int mvsw_pr_port_attr_br_flags_set(struct mvsw_pr_port *port,
>+					  struct switchdev_trans *trans,
>+					  struct net_device *orig_dev,
>+					  unsigned long flags)
>+{
>+	struct mvsw_pr_bridge_port *br_port;
>+	int err;
>+
>+	if (switchdev_trans_ph_prepare(trans))
>+		return 0;
>+
>+	br_port = mvsw_pr_bridge_port_find(port->sw->bridge, orig_dev);
>+	if (!br_port)
>+		return 0;
>+
>+	err = mvsw_pr_port_flood_set(port, flags & BR_FLOOD);
>+	if (err)
>+		return err;
>+
>+	err = mvsw_pr_port_learning_set(port, flags & BR_LEARNING);
>+	if (err)
>+		return err;
>+
>+	memcpy(&br_port->flags, &flags, sizeof(flags));
>+	return 0;
>+}
>+
>+static int mvsw_pr_port_attr_br_ageing_set(struct mvsw_pr_port *port,
>+					   struct switchdev_trans *trans,
>+					   unsigned long ageing_clock_t)
>+{
>+	int err;
>+	struct mvsw_pr_switch *sw = port->sw;
>+	unsigned long ageing_jiffies = clock_t_to_jiffies(ageing_clock_t);
>+	u32 ageing_time = jiffies_to_msecs(ageing_jiffies) / 1000;
>+
>+	if (switchdev_trans_ph_prepare(trans)) {
>+		if (ageing_time < MVSW_PR_MIN_AGEING_TIME ||
>+		    ageing_time > MVSW_PR_MAX_AGEING_TIME)
>+			return -ERANGE;
>+		else
>+			return 0;
>+	}
>+
>+	err = mvsw_pr_switch_ageing_set(sw, ageing_time);
>+	if (!err)
>+		sw->bridge->ageing_time = ageing_time;
>+
>+	return err;
>+}
>+
>+static int mvsw_pr_port_obj_attr_set(struct net_device *dev,
>+				     const struct switchdev_attr *attr,
>+				     struct switchdev_trans *trans)
>+{
>+	int err = 0;
>+	struct mvsw_pr_port *port = netdev_priv(dev);
>+
>+	switch (attr->id) {
>+	case SWITCHDEV_ATTR_ID_PORT_STP_STATE:
>+		err = -EOPNOTSUPP;
>+		break;
>+	case SWITCHDEV_ATTR_ID_PORT_PRE_BRIDGE_FLAGS:
>+		if (attr->u.brport_flags &
>+		    ~(BR_LEARNING | BR_FLOOD | BR_MCAST_FLOOD))
>+			err = -EINVAL;
>+		break;
>+	case SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS:
>+		err = mvsw_pr_port_attr_br_flags_set(port, trans,
>+						     attr->orig_dev,
>+						     attr->u.brport_flags);
>+		break;
>+	case SWITCHDEV_ATTR_ID_BRIDGE_AGEING_TIME:
>+		err = mvsw_pr_port_attr_br_ageing_set(port, trans,
>+						      attr->u.ageing_time);
>+		break;
>+	case SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING:
>+		err = mvsw_pr_port_attr_br_vlan_set(port, trans,
>+						    attr->orig_dev,
>+						    attr->u.vlan_filtering);
>+		break;
>+	default:
>+		err = -EOPNOTSUPP;
>+	}
>+
>+	return err;
>+}
>+
>+static void mvsw_fdb_offload_notify(struct mvsw_pr_port *port,
>+				    struct switchdev_notifier_fdb_info *info)
>+{
>+	struct switchdev_notifier_fdb_info send_info;
>+
>+	send_info.addr = info->addr;
>+	send_info.vid = info->vid;
>+	send_info.offloaded = true;
>+	call_switchdev_notifiers(SWITCHDEV_FDB_OFFLOADED,
>+				 port->net_dev, &send_info.info, NULL);
>+}
>+
>+static int
>+mvsw_pr_port_fdb_set(struct mvsw_pr_port *port,
>+		     struct switchdev_notifier_fdb_info *fdb_info, bool adding)
>+{
>+	struct mvsw_pr_switch *sw = port->sw;
>+	struct mvsw_pr_bridge_port *br_port;
>+	struct mvsw_pr_bridge_device *bridge_device;
>+	struct net_device *orig_dev = fdb_info->info.dev;
>+	int err;
>+	u16 vid;
>+
>+	br_port = mvsw_pr_bridge_port_find(sw->bridge, orig_dev);
>+	if (!br_port)
>+		return -EINVAL;
>+
>+	bridge_device = br_port->bridge_device;
>+
>+	if (bridge_device->vlan_enabled)
>+		vid = fdb_info->vid;
>+	else
>+		vid = bridge_device->bridge_id;
>+
>+	if (adding)
>+		err = mvsw_pr_fdb_add(port, fdb_info->addr, vid, false);
>+	else
>+		err = mvsw_pr_fdb_del(port, fdb_info->addr, vid);
>+
>+	return err;
>+}
>+
>+static void mvsw_pr_bridge_fdb_event_work(struct work_struct *work)

Why do you need to do this in work? Why can't you call down to hw
directly from mvsw_pr_switchdev_event(). I might be missing something
but I don't see that accessing hw in this case can sleep.


>+{
>+	int err = 0;
>+	struct mvsw_pr_event_work *switchdev_work =
>+	    container_of(work, struct mvsw_pr_event_work, work);
>+	struct net_device *dev = switchdev_work->dev;
>+	struct switchdev_notifier_fdb_info *fdb_info;
>+	struct mvsw_pr_port *port;
>+
>+	rtnl_lock();
>+	if (netif_is_vxlan(dev))
>+		goto out;
>+
>+	port = mvsw_pr_port_dev_lower_find(dev);
>+	if (!port)
>+		goto out;
>+
>+	switch (switchdev_work->event) {
>+	case SWITCHDEV_FDB_ADD_TO_DEVICE:
>+		fdb_info = &switchdev_work->fdb_info;
>+		if (!fdb_info->added_by_user)
>+			break;
>+		err = mvsw_pr_port_fdb_set(port, fdb_info, true);
>+		if (err)
>+			break;
>+		mvsw_fdb_offload_notify(port, fdb_info);
>+		break;
>+	case SWITCHDEV_FDB_DEL_TO_DEVICE:
>+		fdb_info = &switchdev_work->fdb_info;
>+		mvsw_pr_port_fdb_set(port, fdb_info, false);
>+		break;
>+	case SWITCHDEV_FDB_ADD_TO_BRIDGE:
>+	case SWITCHDEV_FDB_DEL_TO_BRIDGE:
>+		break;
>+	}
>+
>+out:
>+	rtnl_unlock();
>+	kfree(switchdev_work->fdb_info.addr);
>+	kfree(switchdev_work);
>+	dev_put(dev);
>+}
>+
>+static int mvsw_pr_switchdev_event(struct notifier_block *unused,
>+				   unsigned long event, void *ptr)
>+{
>+	int err = 0;
>+	struct net_device *net_dev = switchdev_notifier_info_to_dev(ptr);
>+	struct mvsw_pr_event_work *switchdev_work;
>+	struct switchdev_notifier_fdb_info *fdb_info;
>+	struct switchdev_notifier_info *info = ptr;
>+	struct net_device *upper_br;
>+
>+	if (event == SWITCHDEV_PORT_ATTR_SET) {
>+		err = switchdev_handle_port_attr_set(net_dev, ptr,
>+						     mvsw_pr_netdev_check,
>+						     mvsw_pr_port_obj_attr_set);
>+		return notifier_from_errno(err);
>+	}
>+
>+	upper_br = netdev_master_upper_dev_get_rcu(net_dev);
>+	if (!upper_br)
>+		return NOTIFY_DONE;
>+
>+	if (!netif_is_bridge_master(upper_br))
>+		return NOTIFY_DONE;
>+
>+	switchdev_work = kzalloc(sizeof(*switchdev_work), GFP_ATOMIC);
>+	if (!switchdev_work)
>+		return NOTIFY_BAD;
>+
>+	switchdev_work->dev = net_dev;
>+	switchdev_work->event = event;
>+
>+	switch (event) {
>+	case SWITCHDEV_FDB_ADD_TO_DEVICE:
>+	case SWITCHDEV_FDB_DEL_TO_DEVICE:
>+	case SWITCHDEV_FDB_ADD_TO_BRIDGE:
>+	case SWITCHDEV_FDB_DEL_TO_BRIDGE:
>+		fdb_info = container_of(info,
>+					struct switchdev_notifier_fdb_info,
>+					info);
>+
>+		INIT_WORK(&switchdev_work->work, mvsw_pr_bridge_fdb_event_work);
>+		memcpy(&switchdev_work->fdb_info, ptr,
>+		       sizeof(switchdev_work->fdb_info));
>+		switchdev_work->fdb_info.addr = kzalloc(ETH_ALEN, GFP_ATOMIC);
>+		if (!switchdev_work->fdb_info.addr)
>+			goto out;
>+		ether_addr_copy((u8 *)switchdev_work->fdb_info.addr,
>+				fdb_info->addr);
>+		dev_hold(net_dev);
>+
>+		break;
>+	case SWITCHDEV_VXLAN_FDB_ADD_TO_DEVICE:
>+	case SWITCHDEV_VXLAN_FDB_DEL_TO_DEVICE:
>+	default:
>+		kfree(switchdev_work);
>+		return NOTIFY_DONE;
>+	}
>+
>+	queue_work(mvsw_owq, &switchdev_work->work);
>+	return NOTIFY_DONE;
>+out:
>+	kfree(switchdev_work);
>+	return NOTIFY_BAD;
>+}
>+
>+static int mvsw_pr_switchdev_blocking_event(struct notifier_block *unused,
>+					    unsigned long event, void *ptr)
>+{
>+	int err = 0;

Pointless init.


>+	struct net_device *net_dev = switchdev_notifier_info_to_dev(ptr);
>+
>+	switch (event) {
>+	case SWITCHDEV_PORT_OBJ_ADD:
>+		if (netif_is_vxlan(net_dev)) {
>+			err = -EOPNOTSUPP;
>+		} else {
>+			err = switchdev_handle_port_obj_add
>+			    (net_dev, ptr, mvsw_pr_netdev_check,
>+			     mvsw_pr_port_obj_add);
>+		}

Remove the "{}"s


>+		break;
>+	case SWITCHDEV_PORT_OBJ_DEL:
>+		if (netif_is_vxlan(net_dev)) {
>+			err = -EOPNOTSUPP;
>+		} else {
>+			err = switchdev_handle_port_obj_del
>+			    (net_dev, ptr, mvsw_pr_netdev_check,
>+			     mvsw_pr_port_obj_del);
>+		}

Remove the "{}"s



>+		break;
>+	case SWITCHDEV_PORT_ATTR_SET:
>+		err = switchdev_handle_port_attr_set
>+		    (net_dev, ptr, mvsw_pr_netdev_check,
>+		    mvsw_pr_port_obj_attr_set);
>+		break;
>+	default:
>+		err = -EOPNOTSUPP;
>+	}
>+
>+	return notifier_from_errno(err);
>+}
>+
>+static struct mvsw_pr_bridge_device *
>+mvsw_pr_bridge_device_create(struct mvsw_pr_bridge *bridge,
>+			     struct net_device *br_dev)
>+{
>+	struct mvsw_pr_bridge_device *bridge_device;
>+	bool vlan_enabled = br_vlan_enabled(br_dev);
>+	u16 bridge_id;
>+	int err;
>+
>+	if (vlan_enabled && bridge->bridge_8021q_exists) {
>+		netdev_err(br_dev, "Only one VLAN-aware bridge is supported\n");

Messages like this one should not be printed to dmesg, but should
be rather pushed to the caller using extack.


>+		return ERR_PTR(-EINVAL);
>+	}
>+
>+	bridge_device = kzalloc(sizeof(*bridge_device), GFP_KERNEL);
>+	if (!bridge_device)
>+		return ERR_PTR(-ENOMEM);
>+
>+	if (vlan_enabled) {
>+		bridge->bridge_8021q_exists = true;
>+	} else {
>+		err = mvsw_pr_8021d_bridge_create(bridge->sw, &bridge_id);
>+		if (err) {
>+			kfree(bridge_device);
>+			return ERR_PTR(err);
>+		}
>+
>+		bridge_device->bridge_id = bridge_id;
>+	}
>+
>+	bridge_device->dev = br_dev;
>+	bridge_device->vlan_enabled = vlan_enabled;
>+	bridge_device->multicast_enabled = br_multicast_enabled(br_dev);
>+	bridge_device->mrouter = br_multicast_router(br_dev);
>+	INIT_LIST_HEAD(&bridge_device->port_list);
>+
>+	list_add(&bridge_device->bridge_node, &bridge->bridge_list);
>+
>+	return bridge_device;
>+}
>+
>+static void
>+mvsw_pr_bridge_device_destroy(struct mvsw_pr_bridge *bridge,
>+			      struct mvsw_pr_bridge_device *bridge_device)
>+{
>+	list_del(&bridge_device->bridge_node);
>+	if (bridge_device->vlan_enabled)
>+		bridge->bridge_8021q_exists = false;
>+	else
>+		mvsw_pr_8021d_bridge_delete(bridge->sw,
>+					    bridge_device->bridge_id);
>+
>+	WARN_ON(!list_empty(&bridge_device->port_list));
>+	kfree(bridge_device);
>+}
>+
>+static struct mvsw_pr_bridge_device *
>+mvsw_pr_bridge_device_get(struct mvsw_pr_bridge *bridge,
>+			  struct net_device *br_dev)
>+{
>+	struct mvsw_pr_bridge_device *bridge_device;
>+
>+	bridge_device = mvsw_pr_bridge_device_find(bridge, br_dev);
>+	if (bridge_device)
>+		return bridge_device;
>+
>+	return mvsw_pr_bridge_device_create(bridge, br_dev);
>+}
>+
>+static void
>+mvsw_pr_bridge_device_put(struct mvsw_pr_bridge *bridge,
>+			  struct mvsw_pr_bridge_device *bridge_device)
>+{
>+	if (list_empty(&bridge_device->port_list))
>+		mvsw_pr_bridge_device_destroy(bridge, bridge_device);
>+}
>+
>+static struct mvsw_pr_bridge_port *
>+mvsw_pr_bridge_port_create(struct mvsw_pr_bridge_device *bridge_device,
>+			   struct net_device *brport_dev)
>+{
>+	struct mvsw_pr_bridge_port *br_port;
>+	struct mvsw_pr_port *port;
>+
>+	br_port = kzalloc(sizeof(*br_port), GFP_KERNEL);
>+	if (!br_port)
>+		return NULL;
>+
>+	port = mvsw_pr_port_dev_lower_find(brport_dev);
>+
>+	br_port->dev = brport_dev;
>+	br_port->bridge_device = bridge_device;
>+	br_port->stp_state = BR_STATE_DISABLED;
>+	br_port->flags = BR_LEARNING | BR_FLOOD | BR_LEARNING_SYNC |
>+				BR_MCAST_FLOOD;
>+	INIT_LIST_HEAD(&br_port->vlan_list);
>+	list_add(&br_port->bridge_device_node, &bridge_device->port_list);
>+	br_port->ref_count = 1;
>+
>+	return br_port;
>+}
>+
>+static void
>+mvsw_pr_bridge_port_destroy(struct mvsw_pr_bridge_port *br_port)
>+{
>+	list_del(&br_port->bridge_device_node);
>+	WARN_ON(!list_empty(&br_port->vlan_list));
>+	kfree(br_port);
>+}
>+
>+static struct mvsw_pr_bridge_port *
>+mvsw_pr_bridge_port_get(struct mvsw_pr_bridge *bridge,
>+			struct net_device *brport_dev)
>+{
>+	struct net_device *br_dev = netdev_master_upper_dev_get(brport_dev);
>+	struct mvsw_pr_bridge_device *bridge_device;
>+	struct mvsw_pr_bridge_port *br_port;
>+	int err;
>+
>+	br_port = mvsw_pr_bridge_port_find(bridge, brport_dev);
>+	if (br_port) {
>+		br_port->ref_count++;

Use refcount_t


>+		return br_port;
>+	}
>+
>+	bridge_device = mvsw_pr_bridge_device_get(bridge, br_dev);
>+	if (IS_ERR(bridge_device))
>+		return ERR_CAST(bridge_device);
>+
>+	br_port = mvsw_pr_bridge_port_create(bridge_device, brport_dev);
>+	if (!br_port) {
>+		err = -ENOMEM;
>+		goto err_brport_create;
>+	}
>+
>+	return br_port;
>+
>+err_brport_create:
>+	mvsw_pr_bridge_device_put(bridge, bridge_device);
>+	return ERR_PTR(err);
>+}
>+
>+static void mvsw_pr_bridge_port_put(struct mvsw_pr_bridge *bridge,
>+				    struct mvsw_pr_bridge_port *br_port)
>+{
>+	struct mvsw_pr_bridge_device *bridge_device;
>+
>+	if (--br_port->ref_count != 0)
>+		return;
>+	bridge_device = br_port->bridge_device;
>+	mvsw_pr_bridge_port_destroy(br_port);
>+	mvsw_pr_bridge_device_put(bridge, bridge_device);
>+}
>+
>+static int
>+mvsw_pr_bridge_8021q_port_join(struct mvsw_pr_bridge_device *bridge_device,
>+			       struct mvsw_pr_bridge_port *br_port,
>+			       struct mvsw_pr_port *port,
>+			       struct netlink_ext_ack *extack)
>+{
>+	if (is_vlan_dev(br_port->dev)) {
>+		NL_SET_ERR_MSG_MOD(extack,
>+				   "Can not enslave a VLAN device to a VLAN-aware bridge");
>+		return -EINVAL;
>+	}
>+
>+	return 0;
>+}
>+
>+static int
>+mvsw_pr_bridge_8021d_port_join(struct mvsw_pr_bridge_device *bridge_device,
>+			       struct mvsw_pr_bridge_port *br_port,
>+			       struct mvsw_pr_port *port,
>+			       struct netlink_ext_ack *extack)
>+{
>+	int err;
>+
>+	if (is_vlan_dev(br_port->dev)) {
>+		NL_SET_ERR_MSG_MOD(extack,
>+				   "Enslaving of a VLAN device is not supported");
>+		return -ENOTSUPP;
>+	}
>+	err = mvsw_pr_8021d_bridge_port_add(port, bridge_device->bridge_id);
>+	if (err)
>+		return err;
>+
>+	err = mvsw_pr_port_flood_set(port, br_port->flags & BR_FLOOD);
>+	if (err)
>+		goto err_port_flood_set;
>+
>+	err = mvsw_pr_port_learning_set(port, br_port->flags & BR_LEARNING);
>+	if (err)
>+		goto err_port_learning_set;
>+
>+	return err;
>+
>+err_port_learning_set:
>+	mvsw_pr_port_flood_set(port, false);
>+err_port_flood_set:
>+	mvsw_pr_8021d_bridge_port_delete(port, bridge_device->bridge_id);
>+	return err;
>+}
>+
>+static int mvsw_pr_port_bridge_join(struct mvsw_pr_port *port,
>+				    struct net_device *brport_dev,
>+				    struct net_device *br_dev,
>+				    struct netlink_ext_ack *extack)
>+{
>+	struct mvsw_pr_bridge_device *bridge_device;
>+	struct mvsw_pr_switch *sw = port->sw;
>+	struct mvsw_pr_bridge_port *br_port;
>+	int err;
>+
>+	br_port = mvsw_pr_bridge_port_get(sw->bridge, brport_dev);
>+	if (IS_ERR(br_port))
>+		return PTR_ERR(br_port);
>+
>+	bridge_device = br_port->bridge_device;
>+
>+	if (bridge_device->vlan_enabled) {
>+		err = mvsw_pr_bridge_8021q_port_join(bridge_device, br_port,
>+						     port, extack);
>+	} else {
>+		err = mvsw_pr_bridge_8021d_port_join(bridge_device, br_port,
>+						     port, extack);
>+	}
>+
>+	if (err)
>+		goto err_port_join;
>+
>+	return 0;
>+
>+err_port_join:
>+	mvsw_pr_bridge_port_put(sw->bridge, br_port);
>+	return err;
>+}
>+
>+static void
>+mvsw_pr_bridge_8021d_port_leave(struct mvsw_pr_bridge_device *bridge_device,
>+				struct mvsw_pr_bridge_port *br_port,
>+				struct mvsw_pr_port *port)
>+{
>+	mvsw_pr_fdb_flush_port(port, MVSW_PR_FDB_FLUSH_MODE_ALL);
>+	mvsw_pr_8021d_bridge_port_delete(port, bridge_device->bridge_id);
>+}
>+
>+static void
>+mvsw_pr_bridge_8021q_port_leave(struct mvsw_pr_bridge_device *bridge_device,
>+				struct mvsw_pr_bridge_port *br_port,
>+				struct mvsw_pr_port *port)
>+{
>+	mvsw_pr_fdb_flush_port(port, MVSW_PR_FDB_FLUSH_MODE_ALL);
>+	mvsw_pr_port_pvid_set(port, MVSW_PR_DEFAULT_VID);
>+}
>+
>+static void mvsw_pr_port_bridge_leave(struct mvsw_pr_port *port,
>+				      struct net_device *brport_dev,
>+				      struct net_device *br_dev)
>+{
>+	struct mvsw_pr_switch *sw = port->sw;
>+	struct mvsw_pr_bridge_device *bridge_device;
>+	struct mvsw_pr_bridge_port *br_port;
>+
>+	bridge_device = mvsw_pr_bridge_device_find(sw->bridge, br_dev);
>+	if (!bridge_device)
>+		return;
>+	br_port = __mvsw_pr_bridge_port_find(bridge_device, brport_dev);
>+	if (!br_port)
>+		return;
>+
>+	if (bridge_device->vlan_enabled)
>+		mvsw_pr_bridge_8021q_port_leave(bridge_device, br_port, port);
>+	else
>+		mvsw_pr_bridge_8021d_port_leave(bridge_device, br_port, port);
>+
>+	mvsw_pr_port_learning_set(port, false);
>+	mvsw_pr_port_flood_set(port, false);
>+	mvsw_pr_bridge_port_put(sw->bridge, br_port);
>+}
>+
>+static int mvsw_pr_netdevice_port_upper_event(struct net_device *lower_dev,
>+					      struct net_device *dev,
>+					      unsigned long event, void *ptr)
>+{
>+	struct netdev_notifier_changeupper_info *info;
>+	struct mvsw_pr_port *port;
>+	struct netlink_ext_ack *extack;
>+	struct net_device *upper_dev;
>+	struct mvsw_pr_switch *sw;
>+	int err = 0;
>+
>+	port = netdev_priv(dev);
>+	sw = port->sw;
>+	info = ptr;
>+	extack = netdev_notifier_info_to_extack(&info->info);
>+
>+	switch (event) {
>+	case NETDEV_PRECHANGEUPPER:
>+		upper_dev = info->upper_dev;
>+		if (!netif_is_bridge_master(upper_dev)) {
>+			NL_SET_ERR_MSG_MOD(extack, "Unknown upper device type");
>+			return -EINVAL;
>+		}
>+		if (!info->linking)
>+			break;
>+		if (netdev_has_any_upper_dev(upper_dev) &&
>+		    (!netif_is_bridge_master(upper_dev) ||
>+		     !mvsw_pr_bridge_device_is_offloaded(sw, upper_dev))) {
>+			NL_SET_ERR_MSG_MOD(extack,
>+					   "Enslaving a port to a device that already has an upper device is not supported");
>+			return -EINVAL;
>+		}
>+		break;
>+	case NETDEV_CHANGEUPPER:
>+		upper_dev = info->upper_dev;
>+		if (netif_is_bridge_master(upper_dev)) {
>+			if (info->linking)
>+				err = mvsw_pr_port_bridge_join(port,
>+							       lower_dev,
>+							       upper_dev,
>+							       extack);
>+			else
>+				mvsw_pr_port_bridge_leave(port,
>+							  lower_dev,
>+							  upper_dev);
>+		}
>+		break;
>+	}
>+
>+	return err;
>+}
>+
>+static int mvsw_pr_netdevice_port_event(struct net_device *lower_dev,
>+					struct net_device *port_dev,
>+					unsigned long event, void *ptr)
>+{
>+	switch (event) {
>+	case NETDEV_PRECHANGEUPPER:
>+	case NETDEV_CHANGEUPPER:
>+		return mvsw_pr_netdevice_port_upper_event(lower_dev, port_dev,
>+							  event, ptr);
>+	}
>+
>+	return 0;
>+}
>+
>+static int mvsw_pr_netdevice_event(struct notifier_block *nb,
>+				   unsigned long event, void *ptr)
>+{
>+	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
>+	struct mvsw_pr_switch *sw;
>+	int err = 0;
>+
>+	sw = container_of(nb, struct mvsw_pr_switch, netdevice_nb);
>+
>+	if (mvsw_pr_netdev_check(dev))
>+		err = mvsw_pr_netdevice_port_event(dev, dev, event, ptr);
>+
>+	return notifier_from_errno(err);
>+}
>+
>+static int mvsw_pr_fdb_init(struct mvsw_pr_switch *sw)
>+{

Just return mvsw_pr_switch_ageing_set(...


>+	int err;
>+
>+	err = mvsw_pr_switch_ageing_set(sw, MVSW_PR_DEFAULT_AGEING_TIME);
>+	if (err)
>+		return err;
>+
>+	return 0;
>+}
>+
>+static int mvsw_pr_switchdev_init(struct mvsw_pr_switch *sw)
>+{
>+	int err = 0;

Pointless init.


>+	struct mvsw_pr_switchdev *swdev;
>+	struct mvsw_pr_bridge *bridge;
>+
>+	if (sw->switchdev)
>+		return -EPERM;
>+
>+	bridge = kzalloc(sizeof(*sw->bridge), GFP_KERNEL);

This is confusing. This struct is not per-bridge but for all bridges.
Perhaps better to squash it together with struct mvsw_pr_switchdev.


>+	if (!bridge)
>+		return -ENOMEM;
>+
>+	swdev = kzalloc(sizeof(*sw->switchdev), GFP_KERNEL);


Why do you need to allocate the memory for bridge and switchdev structs
and not rather have the struct embedded in struct mvsw_pr_switch?



>+	if (!swdev) {
>+		kfree(bridge);
>+		return -ENOMEM;
>+	}
>+
>+	sw->bridge = bridge;
>+	bridge->sw = sw;
>+	sw->switchdev = swdev;
>+	swdev->sw = sw;
>+
>+	INIT_LIST_HEAD(&sw->bridge->bridge_list);
>+
>+	mvsw_owq = alloc_ordered_workqueue("%s_ordered", 0, "prestera_sw");
>+	if (!mvsw_owq) {
>+		err = -ENOMEM;
>+		goto err_alloc_workqueue;
>+	}
>+
>+	swdev->swdev_n.notifier_call = mvsw_pr_switchdev_event;
>+	err = register_switchdev_notifier(&swdev->swdev_n);
>+	if (err)
>+		goto err_register_switchdev_notifier;
>+
>+	swdev->swdev_blocking_n.notifier_call =
>+			mvsw_pr_switchdev_blocking_event;
>+	err = register_switchdev_blocking_notifier(&swdev->swdev_blocking_n);
>+	if (err)
>+		goto err_register_block_switchdev_notifier;
>+
>+	mvsw_pr_fdb_init(sw);
>+
>+	return 0;
>+
>+err_register_block_switchdev_notifier:
>+	unregister_switchdev_notifier(&swdev->swdev_n);
>+err_register_switchdev_notifier:
>+	destroy_workqueue(mvsw_owq);
>+err_alloc_workqueue:
>+	kfree(swdev);
>+	kfree(bridge);
>+	return err;
>+}
>+
>+static void mvsw_pr_switchdev_fini(struct mvsw_pr_switch *sw)
>+{
>+	if (!sw->switchdev)

How this can happen? Remove the check.

>+		return;
>+
>+	unregister_switchdev_notifier(&sw->switchdev->swdev_n);
>+	unregister_switchdev_blocking_notifier
>+	    (&sw->switchdev->swdev_blocking_n);
>+	flush_workqueue(mvsw_owq);
>+	destroy_workqueue(mvsw_owq);
>+	kfree(sw->switchdev);
>+	sw->switchdev = NULL;

Pointless set.


>+	kfree(sw->bridge);
>+}
>+
>+static int mvsw_pr_netdev_init(struct mvsw_pr_switch *sw)

The name is misleading. Looks like it has something to do with netdev,
however it only registers the notifier. Please rename or maybe better
just put the code in mvsw_pr_switchdev_register()

In fact, you are going to need the notifier for things outside switchdev
code too. Perhaps you can register in prestera.c, and call in
prestera_switchdev.c.
See mlxsw_sp_netdevice_event() for inspiration.


>+{
>+	int err = 0;

Pointless init.

>+
>+	if (sw->netdevice_nb.notifier_call)

I don't undestand how this could happen. Remove the check.


>+		return -EPERM;
>+
>+	sw->netdevice_nb.notifier_call = mvsw_pr_netdevice_event;
>+	err = register_netdevice_notifier(&sw->netdevice_nb);

just return.

>+	return err;
>+}
>+
>+static void mvsw_pr_netdev_fini(struct mvsw_pr_switch *sw)
>+{
>+	if (sw->netdevice_nb.notifier_call)

How could it be null? Please remove the check.


>+		unregister_netdevice_notifier(&sw->netdevice_nb);
>+}
>+
>+int mvsw_pr_switchdev_register(struct mvsw_pr_switch *sw)

"register" name is misleading a big. Just do
"prestera_switchdev_init";


>+{
>+	int err;
>+
>+	err = mvsw_pr_switchdev_init(sw);
>+	if (err)
>+		return err;
>+
>+	err = mvsw_pr_netdev_init(sw);
>+	if (err)
>+		goto err_netdevice_notifier;
>+
>+	return 0;
>+
>+err_netdevice_notifier:
>+	mvsw_pr_switchdev_fini(sw);
>+	return err;
>+}
>+
>+void mvsw_pr_switchdev_unregister(struct mvsw_pr_switch *sw)
>+{
>+	mvsw_pr_netdev_fini(sw);
>+	mvsw_pr_switchdev_fini(sw);
>+}
>-- 
>2.17.1
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ