Re: [RFC PATCH 2/2] net: mvneta: Add naive RSS support

From: Marcin Wojtas
Date: Fri Nov 06 2015 - 14:15:37 EST


Hi Gregory,

2015-11-06 19:35 GMT+01:00 Gregory CLEMENT <gregory.clement@xxxxxxxxxxxxxxxxxx>:
> This patch add the support for the RSS related ethtool
> function. Currently it only use one entry in the indirection table which
> allows associating an mveneta interface to a given CPU.
>
> Signed-off-by: Gregory CLEMENT <gregory.clement@xxxxxxxxxxxxxxxxxx>
> ---
> drivers/net/ethernet/marvell/mvneta.c | 114 ++++++++++++++++++++++++++++++++++
> 1 file changed, 114 insertions(+)
>
> diff --git a/drivers/net/ethernet/marvell/mvneta.c b/drivers/net/ethernet/marvell/mvneta.c
> index c38326b848f9..5f810a458443 100644
> --- a/drivers/net/ethernet/marvell/mvneta.c
> +++ b/drivers/net/ethernet/marvell/mvneta.c
> @@ -259,6 +259,11 @@
>
> #define MVNETA_TX_MTU_MAX 0x3ffff
>
> +/* The RSS lookup table actually has 256 entries but we do not use
> + * them yet
> + */
> +#define MVNETA_RSS_LU_TABLE_SIZE 1
> +
> /* TSO header size */
> #define TSO_HEADER_SIZE 128
>
> @@ -380,6 +385,8 @@ struct mvneta_port {
> int use_inband_status:1;
>
> u64 ethtool_stats[ARRAY_SIZE(mvneta_statistics)];
> +
> + u32 indir[MVNETA_RSS_LU_TABLE_SIZE];
> };
>
> /* The mvneta_tx_desc and mvneta_rx_desc structures describe the
> @@ -3173,6 +3180,107 @@ static int mvneta_ethtool_get_sset_count(struct net_device *dev, int sset)
> return -EOPNOTSUPP;
> }
>
> +static u32 mvneta_ethtool_get_rxfh_indir_size(struct net_device *dev)
> +{
> + return MVNETA_RSS_LU_TABLE_SIZE;
> +}
> +
> +static int mvneta_ethtool_get_rxnfc(struct net_device *dev,
> + struct ethtool_rxnfc *info,
> + u32 *rules __always_unused)
> +{
> + switch (info->cmd) {
> + case ETHTOOL_GRXRINGS:
> + info->data = rxq_number;
> + return 0;
> + case ETHTOOL_GRXFH:
> + return -EOPNOTSUPP;
> + default:
> + return -EOPNOTSUPP;
> + }
> +}
> +
> +static int mvneta_config_rss(struct mvneta_port *pp)
> +{
> + int cpu;
> + u32 val;
> +
> + netif_tx_stop_all_queues(pp->dev);
> +
> + /* Mask all ethernet port interrupts */
> + mvreg_write(pp, MVNETA_INTR_NEW_MASK, 0);

Shouldn't the interrupts be masked on each online cpu? There is percpu
unmask function (mvneta_percpu_unmask_interrupt), so maybe ther should
be also mvneta_percpu_mask_interrupt. With this masking should look
like below:

for_each_online_cpu(cpu)
smp_call_function_single(cpu, mvneta_percpu_unmask_interrupt,
pp, true);

> + mvreg_write(pp, MVNETA_INTR_OLD_MASK, 0);
> + mvreg_write(pp, MVNETA_INTR_MISC_MASK, 0);
> +
> + /* We have to synchronise on the napi of each CPU */
> + for_each_online_cpu(cpu) {
> + struct mvneta_pcpu_port *pcpu_port =
> + per_cpu_ptr(pp->ports, cpu);
> +
> + napi_synchronize(&pcpu_port->napi);
> + napi_disable(&pcpu_port->napi);
> + }
> +
> + pp->rxq_def = pp->indir[0];
> +
> + /* update unicast mapping */
> + mvneta_set_rx_mode(pp->dev);
> +
> + /* Update val of portCfg register accordingly with all RxQueue types */
> + val = MVNETA_PORT_CONFIG_DEFL_VALUE(pp->rxq_def);
> + mvreg_write(pp, MVNETA_PORT_CONFIG, val);
> +
> + /* Update the elected CPU matching the new rxq_def */
> + mvneta_percpu_elect(pp);
> +
> + /* We have to synchronise on the napi of each CPU */
> + for_each_online_cpu(cpu) {
> + struct mvneta_pcpu_port *pcpu_port =
> + per_cpu_ptr(pp->ports, cpu);
> +
> + napi_enable(&pcpu_port->napi);
> + }
> +

rxq_def changed, but txq vs CPU mapping remained as in the beginning -
is it intentional?

Best regards,
Marcin
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/