[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <CAPv3WKecZuSfk4LpCehWoijiA6Ea306qn5iyNbg4TucYuOZauw@mail.gmail.com>
Date: Fri, 6 Nov 2015 20:15:31 +0100
From: Marcin Wojtas <mw@...ihalf.com>
To: Gregory CLEMENT <gregory.clement@...e-electrons.com>
Cc: "David S. Miller" <davem@...emloft.net>,
linux-kernel@...r.kernel.org, netdev@...r.kernel.org,
Thomas Petazzoni <thomas.petazzoni@...e-electrons.com>,
Jason Cooper <jason@...edaemon.net>,
Andrew Lunn <andrew@...n.ch>,
Sebastian Hesselbarth <sebastian.hesselbarth@...il.com>,
Ezequiel Garcia <ezequiel.garcia@...e-electrons.com>,
"linux-arm-kernel@...ts.infradead.org"
<linux-arm-kernel@...ts.infradead.org>,
Lior Amsalem <alior@...vell.com>,
Nadav Haklai <nadavh@...vell.com>,
Simon Guinot <simon.guinot@...uanux.org>,
Maxime Ripard <maxime.ripard@...e-electrons.com>,
Boris BREZILLON <boris.brezillon@...e-electrons.com>,
Russell King - ARM Linux <linux@....linux.org.uk>,
Willy Tarreau <w@....eu>
Subject: Re: [RFC PATCH 2/2] net: mvneta: Add naive RSS support
Hi Gregory,
2015-11-06 19:35 GMT+01:00 Gregory CLEMENT <gregory.clement@...e-electrons.com>:
> This patch add the support for the RSS related ethtool
> function. Currently it only use one entry in the indirection table which
> allows associating an mveneta interface to a given CPU.
>
> Signed-off-by: Gregory CLEMENT <gregory.clement@...e-electrons.com>
> ---
> drivers/net/ethernet/marvell/mvneta.c | 114 ++++++++++++++++++++++++++++++++++
> 1 file changed, 114 insertions(+)
>
> diff --git a/drivers/net/ethernet/marvell/mvneta.c b/drivers/net/ethernet/marvell/mvneta.c
> index c38326b848f9..5f810a458443 100644
> --- a/drivers/net/ethernet/marvell/mvneta.c
> +++ b/drivers/net/ethernet/marvell/mvneta.c
> @@ -259,6 +259,11 @@
>
> #define MVNETA_TX_MTU_MAX 0x3ffff
>
> +/* The RSS lookup table actually has 256 entries but we do not use
> + * them yet
> + */
> +#define MVNETA_RSS_LU_TABLE_SIZE 1
> +
> /* TSO header size */
> #define TSO_HEADER_SIZE 128
>
> @@ -380,6 +385,8 @@ struct mvneta_port {
> int use_inband_status:1;
>
> u64 ethtool_stats[ARRAY_SIZE(mvneta_statistics)];
> +
> + u32 indir[MVNETA_RSS_LU_TABLE_SIZE];
> };
>
> /* The mvneta_tx_desc and mvneta_rx_desc structures describe the
> @@ -3173,6 +3180,107 @@ static int mvneta_ethtool_get_sset_count(struct net_device *dev, int sset)
> return -EOPNOTSUPP;
> }
>
> +static u32 mvneta_ethtool_get_rxfh_indir_size(struct net_device *dev)
> +{
> + return MVNETA_RSS_LU_TABLE_SIZE;
> +}
> +
> +static int mvneta_ethtool_get_rxnfc(struct net_device *dev,
> + struct ethtool_rxnfc *info,
> + u32 *rules __always_unused)
> +{
> + switch (info->cmd) {
> + case ETHTOOL_GRXRINGS:
> + info->data = rxq_number;
> + return 0;
> + case ETHTOOL_GRXFH:
> + return -EOPNOTSUPP;
> + default:
> + return -EOPNOTSUPP;
> + }
> +}
> +
> +static int mvneta_config_rss(struct mvneta_port *pp)
> +{
> + int cpu;
> + u32 val;
> +
> + netif_tx_stop_all_queues(pp->dev);
> +
> + /* Mask all ethernet port interrupts */
> + mvreg_write(pp, MVNETA_INTR_NEW_MASK, 0);
Shouldn't the interrupts be masked on each online cpu? There is percpu
unmask function (mvneta_percpu_unmask_interrupt), so maybe ther should
be also mvneta_percpu_mask_interrupt. With this masking should look
like below:
for_each_online_cpu(cpu)
smp_call_function_single(cpu, mvneta_percpu_unmask_interrupt,
pp, true);
> + mvreg_write(pp, MVNETA_INTR_OLD_MASK, 0);
> + mvreg_write(pp, MVNETA_INTR_MISC_MASK, 0);
> +
> + /* We have to synchronise on the napi of each CPU */
> + for_each_online_cpu(cpu) {
> + struct mvneta_pcpu_port *pcpu_port =
> + per_cpu_ptr(pp->ports, cpu);
> +
> + napi_synchronize(&pcpu_port->napi);
> + napi_disable(&pcpu_port->napi);
> + }
> +
> + pp->rxq_def = pp->indir[0];
> +
> + /* update unicast mapping */
> + mvneta_set_rx_mode(pp->dev);
> +
> + /* Update val of portCfg register accordingly with all RxQueue types */
> + val = MVNETA_PORT_CONFIG_DEFL_VALUE(pp->rxq_def);
> + mvreg_write(pp, MVNETA_PORT_CONFIG, val);
> +
> + /* Update the elected CPU matching the new rxq_def */
> + mvneta_percpu_elect(pp);
> +
> + /* We have to synchronise on the napi of each CPU */
> + for_each_online_cpu(cpu) {
> + struct mvneta_pcpu_port *pcpu_port =
> + per_cpu_ptr(pp->ports, cpu);
> +
> + napi_enable(&pcpu_port->napi);
> + }
> +
rxq_def changed, but txq vs CPU mapping remained as in the beginning -
is it intentional?
Best regards,
Marcin
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists