mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-15 15:04:27 +08:00
41d942d56c
Currently all TX/RX rings and completion queues are part of the netdev priv structure and are allocated statically. This patch will change the priv to hold only arrays of pointers and therefore all TX/RX rings and completetion queues will be allocated dynamically. This is in preparation for NUMA aware allocations. Signed-off-by: Yevgeny Petrilin <yevgenyp@mellanox.com> Signed-off-by: Eugenia Emantayev <eugenia@mellanox.com> Reviewed-by: Hadar Hen Zion <hadarh@mellanox.com> Signed-off-by: Amir Vadai <amirv@mellanox.com> Signed-off-by: David S. Miller <davem@davemloft.net>
1234 lines
32 KiB
C
1234 lines
32 KiB
C
/*
|
|
* Copyright (c) 2007 Mellanox Technologies. All rights reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the
|
|
* OpenIB.org BSD license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or
|
|
* without modification, are permitted provided that the following
|
|
* conditions are met:
|
|
*
|
|
* - Redistributions of source code must retain the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer.
|
|
*
|
|
* - Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/ethtool.h>
|
|
#include <linux/netdevice.h>
|
|
#include <linux/mlx4/driver.h>
|
|
#include <linux/in.h>
|
|
#include <net/ip.h>
|
|
|
|
#include "mlx4_en.h"
|
|
#include "en_port.h"
|
|
|
|
#define EN_ETHTOOL_QP_ATTACH (1ull << 63)
|
|
#define EN_ETHTOOL_SHORT_MASK cpu_to_be16(0xffff)
|
|
#define EN_ETHTOOL_WORD_MASK cpu_to_be32(0xffffffff)
|
|
|
|
static int mlx4_en_moderation_update(struct mlx4_en_priv *priv)
|
|
{
|
|
int i;
|
|
int err = 0;
|
|
|
|
for (i = 0; i < priv->tx_ring_num; i++) {
|
|
priv->tx_cq[i]->moder_cnt = priv->tx_frames;
|
|
priv->tx_cq[i]->moder_time = priv->tx_usecs;
|
|
if (priv->port_up) {
|
|
err = mlx4_en_set_cq_moder(priv, priv->tx_cq[i]);
|
|
if (err)
|
|
return err;
|
|
}
|
|
}
|
|
|
|
if (priv->adaptive_rx_coal)
|
|
return 0;
|
|
|
|
for (i = 0; i < priv->rx_ring_num; i++) {
|
|
priv->rx_cq[i]->moder_cnt = priv->rx_frames;
|
|
priv->rx_cq[i]->moder_time = priv->rx_usecs;
|
|
priv->last_moder_time[i] = MLX4_EN_AUTO_CONF;
|
|
if (priv->port_up) {
|
|
err = mlx4_en_set_cq_moder(priv, priv->rx_cq[i]);
|
|
if (err)
|
|
return err;
|
|
}
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void
|
|
mlx4_en_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *drvinfo)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
struct mlx4_en_dev *mdev = priv->mdev;
|
|
|
|
strlcpy(drvinfo->driver, DRV_NAME, sizeof(drvinfo->driver));
|
|
strlcpy(drvinfo->version, DRV_VERSION " (" DRV_RELDATE ")",
|
|
sizeof(drvinfo->version));
|
|
snprintf(drvinfo->fw_version, sizeof(drvinfo->fw_version),
|
|
"%d.%d.%d",
|
|
(u16) (mdev->dev->caps.fw_ver >> 32),
|
|
(u16) ((mdev->dev->caps.fw_ver >> 16) & 0xffff),
|
|
(u16) (mdev->dev->caps.fw_ver & 0xffff));
|
|
strlcpy(drvinfo->bus_info, pci_name(mdev->dev->pdev),
|
|
sizeof(drvinfo->bus_info));
|
|
drvinfo->n_stats = 0;
|
|
drvinfo->regdump_len = 0;
|
|
drvinfo->eedump_len = 0;
|
|
}
|
|
|
|
static const char main_strings[][ETH_GSTRING_LEN] = {
|
|
"rx_packets", "tx_packets", "rx_bytes", "tx_bytes", "rx_errors",
|
|
"tx_errors", "rx_dropped", "tx_dropped", "multicast", "collisions",
|
|
"rx_length_errors", "rx_over_errors", "rx_crc_errors",
|
|
"rx_frame_errors", "rx_fifo_errors", "rx_missed_errors",
|
|
"tx_aborted_errors", "tx_carrier_errors", "tx_fifo_errors",
|
|
"tx_heartbeat_errors", "tx_window_errors",
|
|
|
|
/* port statistics */
|
|
"tso_packets",
|
|
"queue_stopped", "wake_queue", "tx_timeout", "rx_alloc_failed",
|
|
"rx_csum_good", "rx_csum_none", "tx_chksum_offload",
|
|
|
|
/* packet statistics */
|
|
"broadcast", "rx_prio_0", "rx_prio_1", "rx_prio_2", "rx_prio_3",
|
|
"rx_prio_4", "rx_prio_5", "rx_prio_6", "rx_prio_7", "tx_prio_0",
|
|
"tx_prio_1", "tx_prio_2", "tx_prio_3", "tx_prio_4", "tx_prio_5",
|
|
"tx_prio_6", "tx_prio_7",
|
|
};
|
|
#define NUM_MAIN_STATS 21
|
|
#define NUM_ALL_STATS (NUM_MAIN_STATS + NUM_PORT_STATS + NUM_PKT_STATS + NUM_PERF_STATS)
|
|
|
|
static const char mlx4_en_test_names[][ETH_GSTRING_LEN]= {
|
|
"Interrupt Test",
|
|
"Link Test",
|
|
"Speed Test",
|
|
"Register Test",
|
|
"Loopback Test",
|
|
};
|
|
|
|
static u32 mlx4_en_get_msglevel(struct net_device *dev)
|
|
{
|
|
return ((struct mlx4_en_priv *) netdev_priv(dev))->msg_enable;
|
|
}
|
|
|
|
static void mlx4_en_set_msglevel(struct net_device *dev, u32 val)
|
|
{
|
|
((struct mlx4_en_priv *) netdev_priv(dev))->msg_enable = val;
|
|
}
|
|
|
|
static void mlx4_en_get_wol(struct net_device *netdev,
|
|
struct ethtool_wolinfo *wol)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(netdev);
|
|
int err = 0;
|
|
u64 config = 0;
|
|
u64 mask;
|
|
|
|
if ((priv->port < 1) || (priv->port > 2)) {
|
|
en_err(priv, "Failed to get WoL information\n");
|
|
return;
|
|
}
|
|
|
|
mask = (priv->port == 1) ? MLX4_DEV_CAP_FLAG_WOL_PORT1 :
|
|
MLX4_DEV_CAP_FLAG_WOL_PORT2;
|
|
|
|
if (!(priv->mdev->dev->caps.flags & mask)) {
|
|
wol->supported = 0;
|
|
wol->wolopts = 0;
|
|
return;
|
|
}
|
|
|
|
err = mlx4_wol_read(priv->mdev->dev, &config, priv->port);
|
|
if (err) {
|
|
en_err(priv, "Failed to get WoL information\n");
|
|
return;
|
|
}
|
|
|
|
if (config & MLX4_EN_WOL_MAGIC)
|
|
wol->supported = WAKE_MAGIC;
|
|
else
|
|
wol->supported = 0;
|
|
|
|
if (config & MLX4_EN_WOL_ENABLED)
|
|
wol->wolopts = WAKE_MAGIC;
|
|
else
|
|
wol->wolopts = 0;
|
|
}
|
|
|
|
static int mlx4_en_set_wol(struct net_device *netdev,
|
|
struct ethtool_wolinfo *wol)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(netdev);
|
|
u64 config = 0;
|
|
int err = 0;
|
|
u64 mask;
|
|
|
|
if ((priv->port < 1) || (priv->port > 2))
|
|
return -EOPNOTSUPP;
|
|
|
|
mask = (priv->port == 1) ? MLX4_DEV_CAP_FLAG_WOL_PORT1 :
|
|
MLX4_DEV_CAP_FLAG_WOL_PORT2;
|
|
|
|
if (!(priv->mdev->dev->caps.flags & mask))
|
|
return -EOPNOTSUPP;
|
|
|
|
if (wol->supported & ~WAKE_MAGIC)
|
|
return -EINVAL;
|
|
|
|
err = mlx4_wol_read(priv->mdev->dev, &config, priv->port);
|
|
if (err) {
|
|
en_err(priv, "Failed to get WoL info, unable to modify\n");
|
|
return err;
|
|
}
|
|
|
|
if (wol->wolopts & WAKE_MAGIC) {
|
|
config |= MLX4_EN_WOL_DO_MODIFY | MLX4_EN_WOL_ENABLED |
|
|
MLX4_EN_WOL_MAGIC;
|
|
} else {
|
|
config &= ~(MLX4_EN_WOL_ENABLED | MLX4_EN_WOL_MAGIC);
|
|
config |= MLX4_EN_WOL_DO_MODIFY;
|
|
}
|
|
|
|
err = mlx4_wol_write(priv->mdev->dev, config, priv->port);
|
|
if (err)
|
|
en_err(priv, "Failed to set WoL information\n");
|
|
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_en_get_sset_count(struct net_device *dev, int sset)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
int bit_count = hweight64(priv->stats_bitmap);
|
|
|
|
switch (sset) {
|
|
case ETH_SS_STATS:
|
|
return (priv->stats_bitmap ? bit_count : NUM_ALL_STATS) +
|
|
(priv->tx_ring_num * 2) +
|
|
#ifdef CONFIG_NET_RX_BUSY_POLL
|
|
(priv->rx_ring_num * 5);
|
|
#else
|
|
(priv->rx_ring_num * 2);
|
|
#endif
|
|
case ETH_SS_TEST:
|
|
return MLX4_EN_NUM_SELF_TEST - !(priv->mdev->dev->caps.flags
|
|
& MLX4_DEV_CAP_FLAG_UC_LOOPBACK) * 2;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
}
|
|
|
|
static void mlx4_en_get_ethtool_stats(struct net_device *dev,
|
|
struct ethtool_stats *stats, uint64_t *data)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
int index = 0;
|
|
int i, j = 0;
|
|
|
|
spin_lock_bh(&priv->stats_lock);
|
|
|
|
if (!(priv->stats_bitmap)) {
|
|
for (i = 0; i < NUM_MAIN_STATS; i++)
|
|
data[index++] =
|
|
((unsigned long *) &priv->stats)[i];
|
|
for (i = 0; i < NUM_PORT_STATS; i++)
|
|
data[index++] =
|
|
((unsigned long *) &priv->port_stats)[i];
|
|
for (i = 0; i < NUM_PKT_STATS; i++)
|
|
data[index++] =
|
|
((unsigned long *) &priv->pkstats)[i];
|
|
} else {
|
|
for (i = 0; i < NUM_MAIN_STATS; i++) {
|
|
if ((priv->stats_bitmap >> j) & 1)
|
|
data[index++] =
|
|
((unsigned long *) &priv->stats)[i];
|
|
j++;
|
|
}
|
|
for (i = 0; i < NUM_PORT_STATS; i++) {
|
|
if ((priv->stats_bitmap >> j) & 1)
|
|
data[index++] =
|
|
((unsigned long *) &priv->port_stats)[i];
|
|
j++;
|
|
}
|
|
}
|
|
for (i = 0; i < priv->tx_ring_num; i++) {
|
|
data[index++] = priv->tx_ring[i]->packets;
|
|
data[index++] = priv->tx_ring[i]->bytes;
|
|
}
|
|
for (i = 0; i < priv->rx_ring_num; i++) {
|
|
data[index++] = priv->rx_ring[i]->packets;
|
|
data[index++] = priv->rx_ring[i]->bytes;
|
|
#ifdef CONFIG_NET_RX_BUSY_POLL
|
|
data[index++] = priv->rx_ring[i]->yields;
|
|
data[index++] = priv->rx_ring[i]->misses;
|
|
data[index++] = priv->rx_ring[i]->cleaned;
|
|
#endif
|
|
}
|
|
spin_unlock_bh(&priv->stats_lock);
|
|
|
|
}
|
|
|
|
static void mlx4_en_self_test(struct net_device *dev,
|
|
struct ethtool_test *etest, u64 *buf)
|
|
{
|
|
mlx4_en_ex_selftest(dev, &etest->flags, buf);
|
|
}
|
|
|
|
static void mlx4_en_get_strings(struct net_device *dev,
|
|
uint32_t stringset, uint8_t *data)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
int index = 0;
|
|
int i;
|
|
|
|
switch (stringset) {
|
|
case ETH_SS_TEST:
|
|
for (i = 0; i < MLX4_EN_NUM_SELF_TEST - 2; i++)
|
|
strcpy(data + i * ETH_GSTRING_LEN, mlx4_en_test_names[i]);
|
|
if (priv->mdev->dev->caps.flags & MLX4_DEV_CAP_FLAG_UC_LOOPBACK)
|
|
for (; i < MLX4_EN_NUM_SELF_TEST; i++)
|
|
strcpy(data + i * ETH_GSTRING_LEN, mlx4_en_test_names[i]);
|
|
break;
|
|
|
|
case ETH_SS_STATS:
|
|
/* Add main counters */
|
|
if (!priv->stats_bitmap) {
|
|
for (i = 0; i < NUM_MAIN_STATS; i++)
|
|
strcpy(data + (index++) * ETH_GSTRING_LEN,
|
|
main_strings[i]);
|
|
for (i = 0; i < NUM_PORT_STATS; i++)
|
|
strcpy(data + (index++) * ETH_GSTRING_LEN,
|
|
main_strings[i +
|
|
NUM_MAIN_STATS]);
|
|
for (i = 0; i < NUM_PKT_STATS; i++)
|
|
strcpy(data + (index++) * ETH_GSTRING_LEN,
|
|
main_strings[i +
|
|
NUM_MAIN_STATS +
|
|
NUM_PORT_STATS]);
|
|
} else
|
|
for (i = 0; i < NUM_MAIN_STATS + NUM_PORT_STATS; i++) {
|
|
if ((priv->stats_bitmap >> i) & 1) {
|
|
strcpy(data +
|
|
(index++) * ETH_GSTRING_LEN,
|
|
main_strings[i]);
|
|
}
|
|
if (!(priv->stats_bitmap >> i))
|
|
break;
|
|
}
|
|
for (i = 0; i < priv->tx_ring_num; i++) {
|
|
sprintf(data + (index++) * ETH_GSTRING_LEN,
|
|
"tx%d_packets", i);
|
|
sprintf(data + (index++) * ETH_GSTRING_LEN,
|
|
"tx%d_bytes", i);
|
|
}
|
|
for (i = 0; i < priv->rx_ring_num; i++) {
|
|
sprintf(data + (index++) * ETH_GSTRING_LEN,
|
|
"rx%d_packets", i);
|
|
sprintf(data + (index++) * ETH_GSTRING_LEN,
|
|
"rx%d_bytes", i);
|
|
#ifdef CONFIG_NET_RX_BUSY_POLL
|
|
sprintf(data + (index++) * ETH_GSTRING_LEN,
|
|
"rx%d_napi_yield", i);
|
|
sprintf(data + (index++) * ETH_GSTRING_LEN,
|
|
"rx%d_misses", i);
|
|
sprintf(data + (index++) * ETH_GSTRING_LEN,
|
|
"rx%d_cleaned", i);
|
|
#endif
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
static int mlx4_en_get_settings(struct net_device *dev, struct ethtool_cmd *cmd)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
int trans_type;
|
|
|
|
cmd->autoneg = AUTONEG_DISABLE;
|
|
cmd->supported = SUPPORTED_10000baseT_Full;
|
|
cmd->advertising = ADVERTISED_10000baseT_Full;
|
|
|
|
if (mlx4_en_QUERY_PORT(priv->mdev, priv->port))
|
|
return -ENOMEM;
|
|
|
|
trans_type = priv->port_state.transciver;
|
|
if (netif_carrier_ok(dev)) {
|
|
ethtool_cmd_speed_set(cmd, priv->port_state.link_speed);
|
|
cmd->duplex = DUPLEX_FULL;
|
|
} else {
|
|
ethtool_cmd_speed_set(cmd, -1);
|
|
cmd->duplex = -1;
|
|
}
|
|
|
|
if (trans_type > 0 && trans_type <= 0xC) {
|
|
cmd->port = PORT_FIBRE;
|
|
cmd->transceiver = XCVR_EXTERNAL;
|
|
cmd->supported |= SUPPORTED_FIBRE;
|
|
cmd->advertising |= ADVERTISED_FIBRE;
|
|
} else if (trans_type == 0x80 || trans_type == 0) {
|
|
cmd->port = PORT_TP;
|
|
cmd->transceiver = XCVR_INTERNAL;
|
|
cmd->supported |= SUPPORTED_TP;
|
|
cmd->advertising |= ADVERTISED_TP;
|
|
} else {
|
|
cmd->port = -1;
|
|
cmd->transceiver = -1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int mlx4_en_set_settings(struct net_device *dev, struct ethtool_cmd *cmd)
|
|
{
|
|
if ((cmd->autoneg == AUTONEG_ENABLE) ||
|
|
(ethtool_cmd_speed(cmd) != SPEED_10000) ||
|
|
(cmd->duplex != DUPLEX_FULL))
|
|
return -EINVAL;
|
|
|
|
/* Nothing to change */
|
|
return 0;
|
|
}
|
|
|
|
static int mlx4_en_get_coalesce(struct net_device *dev,
|
|
struct ethtool_coalesce *coal)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
|
|
coal->tx_coalesce_usecs = priv->tx_usecs;
|
|
coal->tx_max_coalesced_frames = priv->tx_frames;
|
|
coal->rx_coalesce_usecs = priv->rx_usecs;
|
|
coal->rx_max_coalesced_frames = priv->rx_frames;
|
|
|
|
coal->pkt_rate_low = priv->pkt_rate_low;
|
|
coal->rx_coalesce_usecs_low = priv->rx_usecs_low;
|
|
coal->pkt_rate_high = priv->pkt_rate_high;
|
|
coal->rx_coalesce_usecs_high = priv->rx_usecs_high;
|
|
coal->rate_sample_interval = priv->sample_interval;
|
|
coal->use_adaptive_rx_coalesce = priv->adaptive_rx_coal;
|
|
return 0;
|
|
}
|
|
|
|
static int mlx4_en_set_coalesce(struct net_device *dev,
|
|
struct ethtool_coalesce *coal)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
|
|
priv->rx_frames = (coal->rx_max_coalesced_frames ==
|
|
MLX4_EN_AUTO_CONF) ?
|
|
MLX4_EN_RX_COAL_TARGET :
|
|
coal->rx_max_coalesced_frames;
|
|
priv->rx_usecs = (coal->rx_coalesce_usecs ==
|
|
MLX4_EN_AUTO_CONF) ?
|
|
MLX4_EN_RX_COAL_TIME :
|
|
coal->rx_coalesce_usecs;
|
|
|
|
/* Setting TX coalescing parameters */
|
|
if (coal->tx_coalesce_usecs != priv->tx_usecs ||
|
|
coal->tx_max_coalesced_frames != priv->tx_frames) {
|
|
priv->tx_usecs = coal->tx_coalesce_usecs;
|
|
priv->tx_frames = coal->tx_max_coalesced_frames;
|
|
}
|
|
|
|
/* Set adaptive coalescing params */
|
|
priv->pkt_rate_low = coal->pkt_rate_low;
|
|
priv->rx_usecs_low = coal->rx_coalesce_usecs_low;
|
|
priv->pkt_rate_high = coal->pkt_rate_high;
|
|
priv->rx_usecs_high = coal->rx_coalesce_usecs_high;
|
|
priv->sample_interval = coal->rate_sample_interval;
|
|
priv->adaptive_rx_coal = coal->use_adaptive_rx_coalesce;
|
|
|
|
return mlx4_en_moderation_update(priv);
|
|
}
|
|
|
|
static int mlx4_en_set_pauseparam(struct net_device *dev,
|
|
struct ethtool_pauseparam *pause)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
struct mlx4_en_dev *mdev = priv->mdev;
|
|
int err;
|
|
|
|
priv->prof->tx_pause = pause->tx_pause != 0;
|
|
priv->prof->rx_pause = pause->rx_pause != 0;
|
|
err = mlx4_SET_PORT_general(mdev->dev, priv->port,
|
|
priv->rx_skb_size + ETH_FCS_LEN,
|
|
priv->prof->tx_pause,
|
|
priv->prof->tx_ppp,
|
|
priv->prof->rx_pause,
|
|
priv->prof->rx_ppp);
|
|
if (err)
|
|
en_err(priv, "Failed setting pause params\n");
|
|
|
|
return err;
|
|
}
|
|
|
|
static void mlx4_en_get_pauseparam(struct net_device *dev,
|
|
struct ethtool_pauseparam *pause)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
|
|
pause->tx_pause = priv->prof->tx_pause;
|
|
pause->rx_pause = priv->prof->rx_pause;
|
|
}
|
|
|
|
static int mlx4_en_set_ringparam(struct net_device *dev,
|
|
struct ethtool_ringparam *param)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
struct mlx4_en_dev *mdev = priv->mdev;
|
|
u32 rx_size, tx_size;
|
|
int port_up = 0;
|
|
int err = 0;
|
|
|
|
if (param->rx_jumbo_pending || param->rx_mini_pending)
|
|
return -EINVAL;
|
|
|
|
rx_size = roundup_pow_of_two(param->rx_pending);
|
|
rx_size = max_t(u32, rx_size, MLX4_EN_MIN_RX_SIZE);
|
|
rx_size = min_t(u32, rx_size, MLX4_EN_MAX_RX_SIZE);
|
|
tx_size = roundup_pow_of_two(param->tx_pending);
|
|
tx_size = max_t(u32, tx_size, MLX4_EN_MIN_TX_SIZE);
|
|
tx_size = min_t(u32, tx_size, MLX4_EN_MAX_TX_SIZE);
|
|
|
|
if (rx_size == (priv->port_up ? priv->rx_ring[0]->actual_size :
|
|
priv->rx_ring[0]->size) &&
|
|
tx_size == priv->tx_ring[0]->size)
|
|
return 0;
|
|
|
|
mutex_lock(&mdev->state_lock);
|
|
if (priv->port_up) {
|
|
port_up = 1;
|
|
mlx4_en_stop_port(dev, 1);
|
|
}
|
|
|
|
mlx4_en_free_resources(priv);
|
|
|
|
priv->prof->tx_ring_size = tx_size;
|
|
priv->prof->rx_ring_size = rx_size;
|
|
|
|
err = mlx4_en_alloc_resources(priv);
|
|
if (err) {
|
|
en_err(priv, "Failed reallocating port resources\n");
|
|
goto out;
|
|
}
|
|
if (port_up) {
|
|
err = mlx4_en_start_port(dev);
|
|
if (err)
|
|
en_err(priv, "Failed starting port\n");
|
|
}
|
|
|
|
err = mlx4_en_moderation_update(priv);
|
|
|
|
out:
|
|
mutex_unlock(&mdev->state_lock);
|
|
return err;
|
|
}
|
|
|
|
static void mlx4_en_get_ringparam(struct net_device *dev,
|
|
struct ethtool_ringparam *param)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
|
|
memset(param, 0, sizeof(*param));
|
|
param->rx_max_pending = MLX4_EN_MAX_RX_SIZE;
|
|
param->tx_max_pending = MLX4_EN_MAX_TX_SIZE;
|
|
param->rx_pending = priv->port_up ?
|
|
priv->rx_ring[0]->actual_size : priv->rx_ring[0]->size;
|
|
param->tx_pending = priv->tx_ring[0]->size;
|
|
}
|
|
|
|
static u32 mlx4_en_get_rxfh_indir_size(struct net_device *dev)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
|
|
return priv->rx_ring_num;
|
|
}
|
|
|
|
static int mlx4_en_get_rxfh_indir(struct net_device *dev, u32 *ring_index)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
struct mlx4_en_rss_map *rss_map = &priv->rss_map;
|
|
int rss_rings;
|
|
size_t n = priv->rx_ring_num;
|
|
int err = 0;
|
|
|
|
rss_rings = priv->prof->rss_rings ?: priv->rx_ring_num;
|
|
|
|
while (n--) {
|
|
ring_index[n] = rss_map->qps[n % rss_rings].qpn -
|
|
rss_map->base_qpn;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_en_set_rxfh_indir(struct net_device *dev,
|
|
const u32 *ring_index)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
struct mlx4_en_dev *mdev = priv->mdev;
|
|
int port_up = 0;
|
|
int err = 0;
|
|
int i;
|
|
int rss_rings = 0;
|
|
|
|
/* Calculate RSS table size and make sure flows are spread evenly
|
|
* between rings
|
|
*/
|
|
for (i = 0; i < priv->rx_ring_num; i++) {
|
|
if (i > 0 && !ring_index[i] && !rss_rings)
|
|
rss_rings = i;
|
|
|
|
if (ring_index[i] != (i % (rss_rings ?: priv->rx_ring_num)))
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!rss_rings)
|
|
rss_rings = priv->rx_ring_num;
|
|
|
|
/* RSS table size must be an order of 2 */
|
|
if (!is_power_of_2(rss_rings))
|
|
return -EINVAL;
|
|
|
|
mutex_lock(&mdev->state_lock);
|
|
if (priv->port_up) {
|
|
port_up = 1;
|
|
mlx4_en_stop_port(dev, 1);
|
|
}
|
|
|
|
priv->prof->rss_rings = rss_rings;
|
|
|
|
if (port_up) {
|
|
err = mlx4_en_start_port(dev);
|
|
if (err)
|
|
en_err(priv, "Failed starting port\n");
|
|
}
|
|
|
|
mutex_unlock(&mdev->state_lock);
|
|
return err;
|
|
}
|
|
|
|
#define all_zeros_or_all_ones(field) \
|
|
((field) == 0 || (field) == (__force typeof(field))-1)
|
|
|
|
static int mlx4_en_validate_flow(struct net_device *dev,
|
|
struct ethtool_rxnfc *cmd)
|
|
{
|
|
struct ethtool_usrip4_spec *l3_mask;
|
|
struct ethtool_tcpip4_spec *l4_mask;
|
|
struct ethhdr *eth_mask;
|
|
|
|
if (cmd->fs.location >= MAX_NUM_OF_FS_RULES)
|
|
return -EINVAL;
|
|
|
|
if (cmd->fs.flow_type & FLOW_MAC_EXT) {
|
|
/* dest mac mask must be ff:ff:ff:ff:ff:ff */
|
|
if (!is_broadcast_ether_addr(cmd->fs.m_ext.h_dest))
|
|
return -EINVAL;
|
|
}
|
|
|
|
switch (cmd->fs.flow_type & ~(FLOW_EXT | FLOW_MAC_EXT)) {
|
|
case TCP_V4_FLOW:
|
|
case UDP_V4_FLOW:
|
|
if (cmd->fs.m_u.tcp_ip4_spec.tos)
|
|
return -EINVAL;
|
|
l4_mask = &cmd->fs.m_u.tcp_ip4_spec;
|
|
/* don't allow mask which isn't all 0 or 1 */
|
|
if (!all_zeros_or_all_ones(l4_mask->ip4src) ||
|
|
!all_zeros_or_all_ones(l4_mask->ip4dst) ||
|
|
!all_zeros_or_all_ones(l4_mask->psrc) ||
|
|
!all_zeros_or_all_ones(l4_mask->pdst))
|
|
return -EINVAL;
|
|
break;
|
|
case IP_USER_FLOW:
|
|
l3_mask = &cmd->fs.m_u.usr_ip4_spec;
|
|
if (l3_mask->l4_4_bytes || l3_mask->tos || l3_mask->proto ||
|
|
cmd->fs.h_u.usr_ip4_spec.ip_ver != ETH_RX_NFC_IP4 ||
|
|
(!l3_mask->ip4src && !l3_mask->ip4dst) ||
|
|
!all_zeros_or_all_ones(l3_mask->ip4src) ||
|
|
!all_zeros_or_all_ones(l3_mask->ip4dst))
|
|
return -EINVAL;
|
|
break;
|
|
case ETHER_FLOW:
|
|
eth_mask = &cmd->fs.m_u.ether_spec;
|
|
/* source mac mask must not be set */
|
|
if (!is_zero_ether_addr(eth_mask->h_source))
|
|
return -EINVAL;
|
|
|
|
/* dest mac mask must be ff:ff:ff:ff:ff:ff */
|
|
if (!is_broadcast_ether_addr(eth_mask->h_dest))
|
|
return -EINVAL;
|
|
|
|
if (!all_zeros_or_all_ones(eth_mask->h_proto))
|
|
return -EINVAL;
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
if ((cmd->fs.flow_type & FLOW_EXT)) {
|
|
if (cmd->fs.m_ext.vlan_etype ||
|
|
!((cmd->fs.m_ext.vlan_tci & cpu_to_be16(VLAN_VID_MASK)) ==
|
|
0 ||
|
|
(cmd->fs.m_ext.vlan_tci & cpu_to_be16(VLAN_VID_MASK)) ==
|
|
cpu_to_be16(VLAN_VID_MASK)))
|
|
return -EINVAL;
|
|
|
|
if (cmd->fs.m_ext.vlan_tci) {
|
|
if (be16_to_cpu(cmd->fs.h_ext.vlan_tci) >= VLAN_N_VID)
|
|
return -EINVAL;
|
|
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mlx4_en_ethtool_add_mac_rule(struct ethtool_rxnfc *cmd,
|
|
struct list_head *rule_list_h,
|
|
struct mlx4_spec_list *spec_l2,
|
|
unsigned char *mac)
|
|
{
|
|
int err = 0;
|
|
__be64 mac_msk = cpu_to_be64(MLX4_MAC_MASK << 16);
|
|
|
|
spec_l2->id = MLX4_NET_TRANS_RULE_ID_ETH;
|
|
memcpy(spec_l2->eth.dst_mac_msk, &mac_msk, ETH_ALEN);
|
|
memcpy(spec_l2->eth.dst_mac, mac, ETH_ALEN);
|
|
|
|
if ((cmd->fs.flow_type & FLOW_EXT) &&
|
|
(cmd->fs.m_ext.vlan_tci & cpu_to_be16(VLAN_VID_MASK))) {
|
|
spec_l2->eth.vlan_id = cmd->fs.h_ext.vlan_tci;
|
|
spec_l2->eth.vlan_id_msk = cpu_to_be16(VLAN_VID_MASK);
|
|
}
|
|
|
|
list_add_tail(&spec_l2->list, rule_list_h);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_en_ethtool_add_mac_rule_by_ipv4(struct mlx4_en_priv *priv,
|
|
struct ethtool_rxnfc *cmd,
|
|
struct list_head *rule_list_h,
|
|
struct mlx4_spec_list *spec_l2,
|
|
__be32 ipv4_dst)
|
|
{
|
|
#ifdef CONFIG_INET
|
|
unsigned char mac[ETH_ALEN];
|
|
|
|
if (!ipv4_is_multicast(ipv4_dst)) {
|
|
if (cmd->fs.flow_type & FLOW_MAC_EXT)
|
|
memcpy(&mac, cmd->fs.h_ext.h_dest, ETH_ALEN);
|
|
else
|
|
memcpy(&mac, priv->dev->dev_addr, ETH_ALEN);
|
|
} else {
|
|
ip_eth_mc_map(ipv4_dst, mac);
|
|
}
|
|
|
|
return mlx4_en_ethtool_add_mac_rule(cmd, rule_list_h, spec_l2, &mac[0]);
|
|
#else
|
|
return -EINVAL;
|
|
#endif
|
|
}
|
|
|
|
static int add_ip_rule(struct mlx4_en_priv *priv,
|
|
struct ethtool_rxnfc *cmd,
|
|
struct list_head *list_h)
|
|
{
|
|
int err;
|
|
struct mlx4_spec_list *spec_l2 = NULL;
|
|
struct mlx4_spec_list *spec_l3 = NULL;
|
|
struct ethtool_usrip4_spec *l3_mask = &cmd->fs.m_u.usr_ip4_spec;
|
|
|
|
spec_l3 = kzalloc(sizeof(*spec_l3), GFP_KERNEL);
|
|
spec_l2 = kzalloc(sizeof(*spec_l2), GFP_KERNEL);
|
|
if (!spec_l2 || !spec_l3) {
|
|
err = -ENOMEM;
|
|
goto free_spec;
|
|
}
|
|
|
|
err = mlx4_en_ethtool_add_mac_rule_by_ipv4(priv, cmd, list_h, spec_l2,
|
|
cmd->fs.h_u.
|
|
usr_ip4_spec.ip4dst);
|
|
if (err)
|
|
goto free_spec;
|
|
spec_l3->id = MLX4_NET_TRANS_RULE_ID_IPV4;
|
|
spec_l3->ipv4.src_ip = cmd->fs.h_u.usr_ip4_spec.ip4src;
|
|
if (l3_mask->ip4src)
|
|
spec_l3->ipv4.src_ip_msk = EN_ETHTOOL_WORD_MASK;
|
|
spec_l3->ipv4.dst_ip = cmd->fs.h_u.usr_ip4_spec.ip4dst;
|
|
if (l3_mask->ip4dst)
|
|
spec_l3->ipv4.dst_ip_msk = EN_ETHTOOL_WORD_MASK;
|
|
list_add_tail(&spec_l3->list, list_h);
|
|
|
|
return 0;
|
|
|
|
free_spec:
|
|
kfree(spec_l2);
|
|
kfree(spec_l3);
|
|
return err;
|
|
}
|
|
|
|
static int add_tcp_udp_rule(struct mlx4_en_priv *priv,
|
|
struct ethtool_rxnfc *cmd,
|
|
struct list_head *list_h, int proto)
|
|
{
|
|
int err;
|
|
struct mlx4_spec_list *spec_l2 = NULL;
|
|
struct mlx4_spec_list *spec_l3 = NULL;
|
|
struct mlx4_spec_list *spec_l4 = NULL;
|
|
struct ethtool_tcpip4_spec *l4_mask = &cmd->fs.m_u.tcp_ip4_spec;
|
|
|
|
spec_l2 = kzalloc(sizeof(*spec_l2), GFP_KERNEL);
|
|
spec_l3 = kzalloc(sizeof(*spec_l3), GFP_KERNEL);
|
|
spec_l4 = kzalloc(sizeof(*spec_l4), GFP_KERNEL);
|
|
if (!spec_l2 || !spec_l3 || !spec_l4) {
|
|
err = -ENOMEM;
|
|
goto free_spec;
|
|
}
|
|
|
|
spec_l3->id = MLX4_NET_TRANS_RULE_ID_IPV4;
|
|
|
|
if (proto == TCP_V4_FLOW) {
|
|
err = mlx4_en_ethtool_add_mac_rule_by_ipv4(priv, cmd, list_h,
|
|
spec_l2,
|
|
cmd->fs.h_u.
|
|
tcp_ip4_spec.ip4dst);
|
|
if (err)
|
|
goto free_spec;
|
|
spec_l4->id = MLX4_NET_TRANS_RULE_ID_TCP;
|
|
spec_l3->ipv4.src_ip = cmd->fs.h_u.tcp_ip4_spec.ip4src;
|
|
spec_l3->ipv4.dst_ip = cmd->fs.h_u.tcp_ip4_spec.ip4dst;
|
|
spec_l4->tcp_udp.src_port = cmd->fs.h_u.tcp_ip4_spec.psrc;
|
|
spec_l4->tcp_udp.dst_port = cmd->fs.h_u.tcp_ip4_spec.pdst;
|
|
} else {
|
|
err = mlx4_en_ethtool_add_mac_rule_by_ipv4(priv, cmd, list_h,
|
|
spec_l2,
|
|
cmd->fs.h_u.
|
|
udp_ip4_spec.ip4dst);
|
|
if (err)
|
|
goto free_spec;
|
|
spec_l4->id = MLX4_NET_TRANS_RULE_ID_UDP;
|
|
spec_l3->ipv4.src_ip = cmd->fs.h_u.udp_ip4_spec.ip4src;
|
|
spec_l3->ipv4.dst_ip = cmd->fs.h_u.udp_ip4_spec.ip4dst;
|
|
spec_l4->tcp_udp.src_port = cmd->fs.h_u.udp_ip4_spec.psrc;
|
|
spec_l4->tcp_udp.dst_port = cmd->fs.h_u.udp_ip4_spec.pdst;
|
|
}
|
|
|
|
if (l4_mask->ip4src)
|
|
spec_l3->ipv4.src_ip_msk = EN_ETHTOOL_WORD_MASK;
|
|
if (l4_mask->ip4dst)
|
|
spec_l3->ipv4.dst_ip_msk = EN_ETHTOOL_WORD_MASK;
|
|
|
|
if (l4_mask->psrc)
|
|
spec_l4->tcp_udp.src_port_msk = EN_ETHTOOL_SHORT_MASK;
|
|
if (l4_mask->pdst)
|
|
spec_l4->tcp_udp.dst_port_msk = EN_ETHTOOL_SHORT_MASK;
|
|
|
|
list_add_tail(&spec_l3->list, list_h);
|
|
list_add_tail(&spec_l4->list, list_h);
|
|
|
|
return 0;
|
|
|
|
free_spec:
|
|
kfree(spec_l2);
|
|
kfree(spec_l3);
|
|
kfree(spec_l4);
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_en_ethtool_to_net_trans_rule(struct net_device *dev,
|
|
struct ethtool_rxnfc *cmd,
|
|
struct list_head *rule_list_h)
|
|
{
|
|
int err;
|
|
struct ethhdr *eth_spec;
|
|
struct mlx4_spec_list *spec_l2;
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
|
|
err = mlx4_en_validate_flow(dev, cmd);
|
|
if (err)
|
|
return err;
|
|
|
|
switch (cmd->fs.flow_type & ~(FLOW_EXT | FLOW_MAC_EXT)) {
|
|
case ETHER_FLOW:
|
|
spec_l2 = kzalloc(sizeof(*spec_l2), GFP_KERNEL);
|
|
if (!spec_l2)
|
|
return -ENOMEM;
|
|
|
|
eth_spec = &cmd->fs.h_u.ether_spec;
|
|
mlx4_en_ethtool_add_mac_rule(cmd, rule_list_h, spec_l2,
|
|
ð_spec->h_dest[0]);
|
|
spec_l2->eth.ether_type = eth_spec->h_proto;
|
|
if (eth_spec->h_proto)
|
|
spec_l2->eth.ether_type_enable = 1;
|
|
break;
|
|
case IP_USER_FLOW:
|
|
err = add_ip_rule(priv, cmd, rule_list_h);
|
|
break;
|
|
case TCP_V4_FLOW:
|
|
err = add_tcp_udp_rule(priv, cmd, rule_list_h, TCP_V4_FLOW);
|
|
break;
|
|
case UDP_V4_FLOW:
|
|
err = add_tcp_udp_rule(priv, cmd, rule_list_h, UDP_V4_FLOW);
|
|
break;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_en_flow_replace(struct net_device *dev,
|
|
struct ethtool_rxnfc *cmd)
|
|
{
|
|
int err;
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
struct ethtool_flow_id *loc_rule;
|
|
struct mlx4_spec_list *spec, *tmp_spec;
|
|
u32 qpn;
|
|
u64 reg_id;
|
|
|
|
struct mlx4_net_trans_rule rule = {
|
|
.queue_mode = MLX4_NET_TRANS_Q_FIFO,
|
|
.exclusive = 0,
|
|
.allow_loopback = 1,
|
|
.promisc_mode = MLX4_FS_REGULAR,
|
|
};
|
|
|
|
rule.port = priv->port;
|
|
rule.priority = MLX4_DOMAIN_ETHTOOL | cmd->fs.location;
|
|
INIT_LIST_HEAD(&rule.list);
|
|
|
|
/* Allow direct QP attaches if the EN_ETHTOOL_QP_ATTACH flag is set */
|
|
if (cmd->fs.ring_cookie == RX_CLS_FLOW_DISC)
|
|
qpn = priv->drop_qp.qpn;
|
|
else if (cmd->fs.ring_cookie & EN_ETHTOOL_QP_ATTACH) {
|
|
qpn = cmd->fs.ring_cookie & (EN_ETHTOOL_QP_ATTACH - 1);
|
|
} else {
|
|
if (cmd->fs.ring_cookie >= priv->rx_ring_num) {
|
|
en_warn(priv, "rxnfc: RX ring (%llu) doesn't exist.\n",
|
|
cmd->fs.ring_cookie);
|
|
return -EINVAL;
|
|
}
|
|
qpn = priv->rss_map.qps[cmd->fs.ring_cookie].qpn;
|
|
if (!qpn) {
|
|
en_warn(priv, "rxnfc: RX ring (%llu) is inactive.\n",
|
|
cmd->fs.ring_cookie);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
rule.qpn = qpn;
|
|
err = mlx4_en_ethtool_to_net_trans_rule(dev, cmd, &rule.list);
|
|
if (err)
|
|
goto out_free_list;
|
|
|
|
loc_rule = &priv->ethtool_rules[cmd->fs.location];
|
|
if (loc_rule->id) {
|
|
err = mlx4_flow_detach(priv->mdev->dev, loc_rule->id);
|
|
if (err) {
|
|
en_err(priv, "Fail to detach network rule at location %d. registration id = %llx\n",
|
|
cmd->fs.location, loc_rule->id);
|
|
goto out_free_list;
|
|
}
|
|
loc_rule->id = 0;
|
|
memset(&loc_rule->flow_spec, 0,
|
|
sizeof(struct ethtool_rx_flow_spec));
|
|
list_del(&loc_rule->list);
|
|
}
|
|
err = mlx4_flow_attach(priv->mdev->dev, &rule, ®_id);
|
|
if (err) {
|
|
en_err(priv, "Fail to attach network rule at location %d.\n",
|
|
cmd->fs.location);
|
|
goto out_free_list;
|
|
}
|
|
loc_rule->id = reg_id;
|
|
memcpy(&loc_rule->flow_spec, &cmd->fs,
|
|
sizeof(struct ethtool_rx_flow_spec));
|
|
list_add_tail(&loc_rule->list, &priv->ethtool_list);
|
|
|
|
out_free_list:
|
|
list_for_each_entry_safe(spec, tmp_spec, &rule.list, list) {
|
|
list_del(&spec->list);
|
|
kfree(spec);
|
|
}
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_en_flow_detach(struct net_device *dev,
|
|
struct ethtool_rxnfc *cmd)
|
|
{
|
|
int err = 0;
|
|
struct ethtool_flow_id *rule;
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
|
|
if (cmd->fs.location >= MAX_NUM_OF_FS_RULES)
|
|
return -EINVAL;
|
|
|
|
rule = &priv->ethtool_rules[cmd->fs.location];
|
|
if (!rule->id) {
|
|
err = -ENOENT;
|
|
goto out;
|
|
}
|
|
|
|
err = mlx4_flow_detach(priv->mdev->dev, rule->id);
|
|
if (err) {
|
|
en_err(priv, "Fail to detach network rule at location %d. registration id = 0x%llx\n",
|
|
cmd->fs.location, rule->id);
|
|
goto out;
|
|
}
|
|
rule->id = 0;
|
|
memset(&rule->flow_spec, 0, sizeof(struct ethtool_rx_flow_spec));
|
|
list_del(&rule->list);
|
|
out:
|
|
return err;
|
|
|
|
}
|
|
|
|
static int mlx4_en_get_flow(struct net_device *dev, struct ethtool_rxnfc *cmd,
|
|
int loc)
|
|
{
|
|
int err = 0;
|
|
struct ethtool_flow_id *rule;
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
|
|
if (loc < 0 || loc >= MAX_NUM_OF_FS_RULES)
|
|
return -EINVAL;
|
|
|
|
rule = &priv->ethtool_rules[loc];
|
|
if (rule->id)
|
|
memcpy(&cmd->fs, &rule->flow_spec,
|
|
sizeof(struct ethtool_rx_flow_spec));
|
|
else
|
|
err = -ENOENT;
|
|
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_en_get_num_flows(struct mlx4_en_priv *priv)
|
|
{
|
|
|
|
int i, res = 0;
|
|
for (i = 0; i < MAX_NUM_OF_FS_RULES; i++) {
|
|
if (priv->ethtool_rules[i].id)
|
|
res++;
|
|
}
|
|
return res;
|
|
|
|
}
|
|
|
|
static int mlx4_en_get_rxnfc(struct net_device *dev, struct ethtool_rxnfc *cmd,
|
|
u32 *rule_locs)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
struct mlx4_en_dev *mdev = priv->mdev;
|
|
int err = 0;
|
|
int i = 0, priority = 0;
|
|
|
|
if ((cmd->cmd == ETHTOOL_GRXCLSRLCNT ||
|
|
cmd->cmd == ETHTOOL_GRXCLSRULE ||
|
|
cmd->cmd == ETHTOOL_GRXCLSRLALL) &&
|
|
(mdev->dev->caps.steering_mode !=
|
|
MLX4_STEERING_MODE_DEVICE_MANAGED || !priv->port_up))
|
|
return -EINVAL;
|
|
|
|
switch (cmd->cmd) {
|
|
case ETHTOOL_GRXRINGS:
|
|
cmd->data = priv->rx_ring_num;
|
|
break;
|
|
case ETHTOOL_GRXCLSRLCNT:
|
|
cmd->rule_cnt = mlx4_en_get_num_flows(priv);
|
|
break;
|
|
case ETHTOOL_GRXCLSRULE:
|
|
err = mlx4_en_get_flow(dev, cmd, cmd->fs.location);
|
|
break;
|
|
case ETHTOOL_GRXCLSRLALL:
|
|
while ((!err || err == -ENOENT) && priority < cmd->rule_cnt) {
|
|
err = mlx4_en_get_flow(dev, cmd, i);
|
|
if (!err)
|
|
rule_locs[priority++] = i;
|
|
i++;
|
|
}
|
|
err = 0;
|
|
break;
|
|
default:
|
|
err = -EOPNOTSUPP;
|
|
break;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_en_set_rxnfc(struct net_device *dev, struct ethtool_rxnfc *cmd)
|
|
{
|
|
int err = 0;
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
struct mlx4_en_dev *mdev = priv->mdev;
|
|
|
|
if (mdev->dev->caps.steering_mode !=
|
|
MLX4_STEERING_MODE_DEVICE_MANAGED || !priv->port_up)
|
|
return -EINVAL;
|
|
|
|
switch (cmd->cmd) {
|
|
case ETHTOOL_SRXCLSRLINS:
|
|
err = mlx4_en_flow_replace(dev, cmd);
|
|
break;
|
|
case ETHTOOL_SRXCLSRLDEL:
|
|
err = mlx4_en_flow_detach(dev, cmd);
|
|
break;
|
|
default:
|
|
en_warn(priv, "Unsupported ethtool command. (%d)\n", cmd->cmd);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static void mlx4_en_get_channels(struct net_device *dev,
|
|
struct ethtool_channels *channel)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
|
|
memset(channel, 0, sizeof(*channel));
|
|
|
|
channel->max_rx = MAX_RX_RINGS;
|
|
channel->max_tx = MLX4_EN_MAX_TX_RING_P_UP;
|
|
|
|
channel->rx_count = priv->rx_ring_num;
|
|
channel->tx_count = priv->tx_ring_num / MLX4_EN_NUM_UP;
|
|
}
|
|
|
|
static int mlx4_en_set_channels(struct net_device *dev,
|
|
struct ethtool_channels *channel)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
struct mlx4_en_dev *mdev = priv->mdev;
|
|
int port_up;
|
|
int err = 0;
|
|
|
|
if (channel->other_count || channel->combined_count ||
|
|
channel->tx_count > MLX4_EN_MAX_TX_RING_P_UP ||
|
|
channel->rx_count > MAX_RX_RINGS ||
|
|
!channel->tx_count || !channel->rx_count)
|
|
return -EINVAL;
|
|
|
|
mutex_lock(&mdev->state_lock);
|
|
if (priv->port_up) {
|
|
port_up = 1;
|
|
mlx4_en_stop_port(dev, 1);
|
|
}
|
|
|
|
mlx4_en_free_resources(priv);
|
|
|
|
priv->num_tx_rings_p_up = channel->tx_count;
|
|
priv->tx_ring_num = channel->tx_count * MLX4_EN_NUM_UP;
|
|
priv->rx_ring_num = channel->rx_count;
|
|
|
|
err = mlx4_en_alloc_resources(priv);
|
|
if (err) {
|
|
en_err(priv, "Failed reallocating port resources\n");
|
|
goto out;
|
|
}
|
|
|
|
netif_set_real_num_tx_queues(dev, priv->tx_ring_num);
|
|
netif_set_real_num_rx_queues(dev, priv->rx_ring_num);
|
|
|
|
mlx4_en_setup_tc(dev, MLX4_EN_NUM_UP);
|
|
|
|
en_warn(priv, "Using %d TX rings\n", priv->tx_ring_num);
|
|
en_warn(priv, "Using %d RX rings\n", priv->rx_ring_num);
|
|
|
|
if (port_up) {
|
|
err = mlx4_en_start_port(dev);
|
|
if (err)
|
|
en_err(priv, "Failed starting port\n");
|
|
}
|
|
|
|
err = mlx4_en_moderation_update(priv);
|
|
|
|
out:
|
|
mutex_unlock(&mdev->state_lock);
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_en_get_ts_info(struct net_device *dev,
|
|
struct ethtool_ts_info *info)
|
|
{
|
|
struct mlx4_en_priv *priv = netdev_priv(dev);
|
|
struct mlx4_en_dev *mdev = priv->mdev;
|
|
int ret;
|
|
|
|
ret = ethtool_op_get_ts_info(dev, info);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_TS) {
|
|
info->so_timestamping |=
|
|
SOF_TIMESTAMPING_TX_HARDWARE |
|
|
SOF_TIMESTAMPING_RX_HARDWARE |
|
|
SOF_TIMESTAMPING_RAW_HARDWARE;
|
|
|
|
info->tx_types =
|
|
(1 << HWTSTAMP_TX_OFF) |
|
|
(1 << HWTSTAMP_TX_ON);
|
|
|
|
info->rx_filters =
|
|
(1 << HWTSTAMP_FILTER_NONE) |
|
|
(1 << HWTSTAMP_FILTER_ALL);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
const struct ethtool_ops mlx4_en_ethtool_ops = {
|
|
.get_drvinfo = mlx4_en_get_drvinfo,
|
|
.get_settings = mlx4_en_get_settings,
|
|
.set_settings = mlx4_en_set_settings,
|
|
.get_link = ethtool_op_get_link,
|
|
.get_strings = mlx4_en_get_strings,
|
|
.get_sset_count = mlx4_en_get_sset_count,
|
|
.get_ethtool_stats = mlx4_en_get_ethtool_stats,
|
|
.self_test = mlx4_en_self_test,
|
|
.get_wol = mlx4_en_get_wol,
|
|
.set_wol = mlx4_en_set_wol,
|
|
.get_msglevel = mlx4_en_get_msglevel,
|
|
.set_msglevel = mlx4_en_set_msglevel,
|
|
.get_coalesce = mlx4_en_get_coalesce,
|
|
.set_coalesce = mlx4_en_set_coalesce,
|
|
.get_pauseparam = mlx4_en_get_pauseparam,
|
|
.set_pauseparam = mlx4_en_set_pauseparam,
|
|
.get_ringparam = mlx4_en_get_ringparam,
|
|
.set_ringparam = mlx4_en_set_ringparam,
|
|
.get_rxnfc = mlx4_en_get_rxnfc,
|
|
.set_rxnfc = mlx4_en_set_rxnfc,
|
|
.get_rxfh_indir_size = mlx4_en_get_rxfh_indir_size,
|
|
.get_rxfh_indir = mlx4_en_get_rxfh_indir,
|
|
.set_rxfh_indir = mlx4_en_set_rxfh_indir,
|
|
.get_channels = mlx4_en_get_channels,
|
|
.set_channels = mlx4_en_set_channels,
|
|
.get_ts_info = mlx4_en_get_ts_info,
|
|
};
|
|
|
|
|
|
|
|
|
|
|