mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-12-28 13:34:38 +08:00
0d54d8f7a1
Currently the driver only supports 802.1Q VLAN insertion and stripping. However, once Double VLAN Mode (DVM) is fully supported, then both 802.1Q and 802.1ad VLAN insertion and stripping will be supported. Unfortunately the VSI context parameters only allow for one VLAN ethertype at a time for VLAN offloads so only one or the other VLAN ethertype offload can be supported at once. To support this, multiple changes are needed. Rx path changes: [1] In DVM, the Rx queue context l2tagsel field needs to be cleared so the outermost tag shows up in the l2tag2_2nd field of the Rx flex descriptor. In Single VLAN Mode (SVM), the l2tagsel field should remain 1 to support SVM configurations. [2] Modify the ice_test_staterr() function to take a __le16 instead of the ice_32b_rx_flex_desc union pointer so this function can be used for both rx_desc->wb.status_error0 and rx_desc->wb.status_error1. [3] Add the new inline function ice_get_vlan_tag_from_rx_desc() that checks if there is a VLAN tag in l2tag1 or l2tag2_2nd. [4] In ice_receive_skb(), add a check to see if NETIF_F_HW_VLAN_STAG_RX is enabled in netdev->features. If it is, then this is the VLAN ethertype that needs to be added to the stripping VLAN tag. Since ice_fix_features() prevents CTAG_RX and STAG_RX from being enabled simultaneously, the VLAN ethertype will only ever be 802.1Q or 802.1ad. Tx path changes: [1] In DVM, the VLAN tag needs to be placed in the l2tag2 field of the Tx context descriptor. The new define ICE_TX_FLAGS_HW_OUTER_SINGLE_VLAN was added to the list of tx_flags to handle this case. [2] When the stack requests the VLAN tag to be offloaded on Tx, the driver needs to set either ICE_TX_FLAGS_HW_OUTER_SINGLE_VLAN or ICE_TX_FLAGS_HW_VLAN, so the tag is inserted in l2tag2 or l2tag1 respectively. To determine which location to use, set a bit in the Tx ring flags field during ring allocation that can be used to determine which field to use in the Tx descriptor. In DVM, always use l2tag2, and in SVM, always use l2tag1. Signed-off-by: Brett Creeley <brett.creeley@intel.com> Tested-by: Gurucharan G <gurucharanx.g@intel.com> Signed-off-by: Tony Nguyen <anthony.l.nguyen@intel.com>
1046 lines
27 KiB
C
1046 lines
27 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* Copyright (c) 2019, Intel Corporation. */
|
|
|
|
#include "ice_dcb_lib.h"
|
|
#include "ice_dcb_nl.h"
|
|
|
|
/**
|
|
* ice_dcb_get_ena_tc - return bitmap of enabled TCs
|
|
* @dcbcfg: DCB config to evaluate for enabled TCs
|
|
*/
|
|
static u8 ice_dcb_get_ena_tc(struct ice_dcbx_cfg *dcbcfg)
|
|
{
|
|
u8 i, num_tc, ena_tc = 1;
|
|
|
|
num_tc = ice_dcb_get_num_tc(dcbcfg);
|
|
|
|
for (i = 0; i < num_tc; i++)
|
|
ena_tc |= BIT(i);
|
|
|
|
return ena_tc;
|
|
}
|
|
|
|
/**
|
|
* ice_is_pfc_causing_hung_q
|
|
* @pf: pointer to PF structure
|
|
* @txqueue: Tx queue which is supposedly hung queue
|
|
*
|
|
* find if PFC is causing the hung queue, if yes return true else false
|
|
*/
|
|
bool ice_is_pfc_causing_hung_q(struct ice_pf *pf, unsigned int txqueue)
|
|
{
|
|
u8 num_tcs = 0, i, tc, up_mapped_tc, up_in_tc = 0;
|
|
u64 ref_prio_xoff[ICE_MAX_UP];
|
|
struct ice_vsi *vsi;
|
|
u32 up2tc;
|
|
|
|
vsi = ice_get_main_vsi(pf);
|
|
if (!vsi)
|
|
return false;
|
|
|
|
ice_for_each_traffic_class(i)
|
|
if (vsi->tc_cfg.ena_tc & BIT(i))
|
|
num_tcs++;
|
|
|
|
/* first find out the TC to which the hung queue belongs to */
|
|
for (tc = 0; tc < num_tcs - 1; tc++)
|
|
if (ice_find_q_in_range(vsi->tc_cfg.tc_info[tc].qoffset,
|
|
vsi->tc_cfg.tc_info[tc + 1].qoffset,
|
|
txqueue))
|
|
break;
|
|
|
|
/* Build a bit map of all UPs associated to the suspect hung queue TC,
|
|
* so that we check for its counter increment.
|
|
*/
|
|
up2tc = rd32(&pf->hw, PRTDCB_TUP2TC);
|
|
for (i = 0; i < ICE_MAX_UP; i++) {
|
|
up_mapped_tc = (up2tc >> (i * 3)) & 0x7;
|
|
if (up_mapped_tc == tc)
|
|
up_in_tc |= BIT(i);
|
|
}
|
|
|
|
/* Now that we figured out that hung queue is PFC enabled, still the
|
|
* Tx timeout can be legitimate. So to make sure Tx timeout is
|
|
* absolutely caused by PFC storm, check if the counters are
|
|
* incrementing.
|
|
*/
|
|
for (i = 0; i < ICE_MAX_UP; i++)
|
|
if (up_in_tc & BIT(i))
|
|
ref_prio_xoff[i] = pf->stats.priority_xoff_rx[i];
|
|
|
|
ice_update_dcb_stats(pf);
|
|
|
|
for (i = 0; i < ICE_MAX_UP; i++)
|
|
if (up_in_tc & BIT(i))
|
|
if (pf->stats.priority_xoff_rx[i] > ref_prio_xoff[i])
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_get_mode - gets the DCB mode
|
|
* @port_info: pointer to port info structure
|
|
* @host: if set it's HOST if not it's MANAGED
|
|
*/
|
|
static u8 ice_dcb_get_mode(struct ice_port_info *port_info, bool host)
|
|
{
|
|
u8 mode;
|
|
|
|
if (host)
|
|
mode = DCB_CAP_DCBX_HOST;
|
|
else
|
|
mode = DCB_CAP_DCBX_LLD_MANAGED;
|
|
|
|
if (port_info->qos_cfg.local_dcbx_cfg.dcbx_mode & ICE_DCBX_MODE_CEE)
|
|
return mode | DCB_CAP_DCBX_VER_CEE;
|
|
else
|
|
return mode | DCB_CAP_DCBX_VER_IEEE;
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_get_num_tc - Get the number of TCs from DCBX config
|
|
* @dcbcfg: config to retrieve number of TCs from
|
|
*/
|
|
u8 ice_dcb_get_num_tc(struct ice_dcbx_cfg *dcbcfg)
|
|
{
|
|
bool tc_unused = false;
|
|
u8 num_tc = 0;
|
|
u8 ret = 0;
|
|
int i;
|
|
|
|
/* Scan the ETS Config Priority Table to find traffic classes
|
|
* enabled and create a bitmask of enabled TCs
|
|
*/
|
|
for (i = 0; i < CEE_DCBX_MAX_PRIO; i++)
|
|
num_tc |= BIT(dcbcfg->etscfg.prio_table[i]);
|
|
|
|
/* Scan bitmask for contiguous TCs starting with TC0 */
|
|
for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++) {
|
|
if (num_tc & BIT(i)) {
|
|
if (!tc_unused) {
|
|
ret++;
|
|
} else {
|
|
pr_err("Non-contiguous TCs - Disabling DCB\n");
|
|
return 1;
|
|
}
|
|
} else {
|
|
tc_unused = true;
|
|
}
|
|
}
|
|
|
|
/* There is always at least 1 TC */
|
|
if (!ret)
|
|
ret = 1;
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* ice_get_first_droptc - returns number of first droptc
|
|
* @vsi: used to find the first droptc
|
|
*
|
|
* This function returns the value of first_droptc.
|
|
* When DCB is enabled, first droptc information is derived from enabled_tc
|
|
* and PFC enabled bits. otherwise this function returns 0 as there is one
|
|
* TC without DCB (tc0)
|
|
*/
|
|
static u8 ice_get_first_droptc(struct ice_vsi *vsi)
|
|
{
|
|
struct ice_dcbx_cfg *cfg = &vsi->port_info->qos_cfg.local_dcbx_cfg;
|
|
struct device *dev = ice_pf_to_dev(vsi->back);
|
|
u8 num_tc, ena_tc_map, pfc_ena_map;
|
|
u8 i;
|
|
|
|
num_tc = ice_dcb_get_num_tc(cfg);
|
|
|
|
/* get bitmap of enabled TCs */
|
|
ena_tc_map = ice_dcb_get_ena_tc(cfg);
|
|
|
|
/* get bitmap of PFC enabled TCs */
|
|
pfc_ena_map = cfg->pfc.pfcena;
|
|
|
|
/* get first TC that is not PFC enabled */
|
|
for (i = 0; i < num_tc; i++) {
|
|
if ((ena_tc_map & BIT(i)) && (!(pfc_ena_map & BIT(i)))) {
|
|
dev_dbg(dev, "first drop tc = %d\n", i);
|
|
return i;
|
|
}
|
|
}
|
|
|
|
dev_dbg(dev, "first drop tc = 0\n");
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ice_vsi_set_dcb_tc_cfg - Set VSI's TC based on DCB configuration
|
|
* @vsi: pointer to the VSI instance
|
|
*/
|
|
void ice_vsi_set_dcb_tc_cfg(struct ice_vsi *vsi)
|
|
{
|
|
struct ice_dcbx_cfg *cfg = &vsi->port_info->qos_cfg.local_dcbx_cfg;
|
|
|
|
switch (vsi->type) {
|
|
case ICE_VSI_PF:
|
|
vsi->tc_cfg.ena_tc = ice_dcb_get_ena_tc(cfg);
|
|
vsi->tc_cfg.numtc = ice_dcb_get_num_tc(cfg);
|
|
break;
|
|
case ICE_VSI_CHNL:
|
|
vsi->tc_cfg.ena_tc = BIT(ice_get_first_droptc(vsi));
|
|
vsi->tc_cfg.numtc = 1;
|
|
break;
|
|
case ICE_VSI_CTRL:
|
|
case ICE_VSI_LB:
|
|
default:
|
|
vsi->tc_cfg.ena_tc = ICE_DFLT_TRAFFIC_CLASS;
|
|
vsi->tc_cfg.numtc = 1;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_get_tc - Get the TC associated with the queue
|
|
* @vsi: ptr to the VSI
|
|
* @queue_index: queue number associated with VSI
|
|
*/
|
|
u8 ice_dcb_get_tc(struct ice_vsi *vsi, int queue_index)
|
|
{
|
|
return vsi->tx_rings[queue_index]->dcb_tc;
|
|
}
|
|
|
|
/**
|
|
* ice_vsi_cfg_dcb_rings - Update rings to reflect DCB TC
|
|
* @vsi: VSI owner of rings being updated
|
|
*/
|
|
void ice_vsi_cfg_dcb_rings(struct ice_vsi *vsi)
|
|
{
|
|
struct ice_tx_ring *tx_ring;
|
|
struct ice_rx_ring *rx_ring;
|
|
u16 qoffset, qcount;
|
|
int i, n;
|
|
|
|
if (!test_bit(ICE_FLAG_DCB_ENA, vsi->back->flags)) {
|
|
/* Reset the TC information */
|
|
ice_for_each_txq(vsi, i) {
|
|
tx_ring = vsi->tx_rings[i];
|
|
tx_ring->dcb_tc = 0;
|
|
}
|
|
ice_for_each_rxq(vsi, i) {
|
|
rx_ring = vsi->rx_rings[i];
|
|
rx_ring->dcb_tc = 0;
|
|
}
|
|
return;
|
|
}
|
|
|
|
ice_for_each_traffic_class(n) {
|
|
if (!(vsi->tc_cfg.ena_tc & BIT(n)))
|
|
break;
|
|
|
|
qoffset = vsi->tc_cfg.tc_info[n].qoffset;
|
|
qcount = vsi->tc_cfg.tc_info[n].qcount_tx;
|
|
for (i = qoffset; i < (qoffset + qcount); i++)
|
|
vsi->tx_rings[i]->dcb_tc = n;
|
|
|
|
qcount = vsi->tc_cfg.tc_info[n].qcount_rx;
|
|
for (i = qoffset; i < (qoffset + qcount); i++)
|
|
vsi->rx_rings[i]->dcb_tc = n;
|
|
}
|
|
/* applicable only if "all_enatc" is set, which will be set from
|
|
* setup_tc method as part of configuring channels
|
|
*/
|
|
if (vsi->all_enatc) {
|
|
u8 first_droptc = ice_get_first_droptc(vsi);
|
|
|
|
/* When DCB is configured, TC for ADQ queues (which are really
|
|
* PF queues) should be the first drop TC of the main VSI
|
|
*/
|
|
ice_for_each_chnl_tc(n) {
|
|
if (!(vsi->all_enatc & BIT(n)))
|
|
break;
|
|
|
|
qoffset = vsi->mqprio_qopt.qopt.offset[n];
|
|
qcount = vsi->mqprio_qopt.qopt.count[n];
|
|
for (i = qoffset; i < (qoffset + qcount); i++) {
|
|
vsi->tx_rings[i]->dcb_tc = first_droptc;
|
|
vsi->rx_rings[i]->dcb_tc = first_droptc;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_ena_dis_vsi - disable certain VSIs for DCB config/reconfig
|
|
* @pf: pointer to the PF instance
|
|
* @ena: true to enable VSIs, false to disable
|
|
* @locked: true if caller holds RTNL lock, false otherwise
|
|
*
|
|
* Before a new DCB configuration can be applied, VSIs of type PF, SWITCHDEV
|
|
* and CHNL need to be brought down. Following completion of DCB configuration
|
|
* the VSIs that were downed need to be brought up again. This helper function
|
|
* does both.
|
|
*/
|
|
static void ice_dcb_ena_dis_vsi(struct ice_pf *pf, bool ena, bool locked)
|
|
{
|
|
int i;
|
|
|
|
ice_for_each_vsi(pf, i) {
|
|
struct ice_vsi *vsi = pf->vsi[i];
|
|
|
|
if (!vsi)
|
|
continue;
|
|
|
|
switch (vsi->type) {
|
|
case ICE_VSI_CHNL:
|
|
case ICE_VSI_SWITCHDEV_CTRL:
|
|
case ICE_VSI_PF:
|
|
if (ena)
|
|
ice_ena_vsi(vsi, locked);
|
|
else
|
|
ice_dis_vsi(vsi, locked);
|
|
break;
|
|
default:
|
|
continue;
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_bwchk - check if ETS bandwidth input parameters are correct
|
|
* @pf: pointer to the PF struct
|
|
* @dcbcfg: pointer to DCB config structure
|
|
*/
|
|
int ice_dcb_bwchk(struct ice_pf *pf, struct ice_dcbx_cfg *dcbcfg)
|
|
{
|
|
struct ice_dcb_ets_cfg *etscfg = &dcbcfg->etscfg;
|
|
u8 num_tc, total_bw = 0;
|
|
int i;
|
|
|
|
/* returns number of contigous TCs and 1 TC for non-contigous TCs,
|
|
* since at least 1 TC has to be configured
|
|
*/
|
|
num_tc = ice_dcb_get_num_tc(dcbcfg);
|
|
|
|
/* no bandwidth checks required if there's only one TC, so assign
|
|
* all bandwidth to TC0 and return
|
|
*/
|
|
if (num_tc == 1) {
|
|
etscfg->tcbwtable[0] = ICE_TC_MAX_BW;
|
|
return 0;
|
|
}
|
|
|
|
for (i = 0; i < num_tc; i++)
|
|
total_bw += etscfg->tcbwtable[i];
|
|
|
|
if (!total_bw) {
|
|
etscfg->tcbwtable[0] = ICE_TC_MAX_BW;
|
|
} else if (total_bw != ICE_TC_MAX_BW) {
|
|
dev_err(ice_pf_to_dev(pf), "Invalid config, total bandwidth must equal 100\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ice_pf_dcb_cfg - Apply new DCB configuration
|
|
* @pf: pointer to the PF struct
|
|
* @new_cfg: DCBX config to apply
|
|
* @locked: is the RTNL held
|
|
*/
|
|
int ice_pf_dcb_cfg(struct ice_pf *pf, struct ice_dcbx_cfg *new_cfg, bool locked)
|
|
{
|
|
struct ice_aqc_port_ets_elem buf = { 0 };
|
|
struct ice_dcbx_cfg *old_cfg, *curr_cfg;
|
|
struct device *dev = ice_pf_to_dev(pf);
|
|
int ret = ICE_DCB_NO_HW_CHG;
|
|
struct iidc_event *event;
|
|
struct ice_vsi *pf_vsi;
|
|
|
|
curr_cfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
|
|
|
|
/* FW does not care if change happened */
|
|
if (!pf->hw.port_info->qos_cfg.is_sw_lldp)
|
|
ret = ICE_DCB_HW_CHG_RST;
|
|
|
|
/* Enable DCB tagging only when more than one TC */
|
|
if (ice_dcb_get_num_tc(new_cfg) > 1) {
|
|
dev_dbg(dev, "DCB tagging enabled (num TC > 1)\n");
|
|
set_bit(ICE_FLAG_DCB_ENA, pf->flags);
|
|
} else {
|
|
dev_dbg(dev, "DCB tagging disabled (num TC = 1)\n");
|
|
clear_bit(ICE_FLAG_DCB_ENA, pf->flags);
|
|
}
|
|
|
|
if (!memcmp(new_cfg, curr_cfg, sizeof(*new_cfg))) {
|
|
dev_dbg(dev, "No change in DCB config required\n");
|
|
return ret;
|
|
}
|
|
|
|
if (ice_dcb_bwchk(pf, new_cfg))
|
|
return -EINVAL;
|
|
|
|
/* Store old config in case FW config fails */
|
|
old_cfg = kmemdup(curr_cfg, sizeof(*old_cfg), GFP_KERNEL);
|
|
if (!old_cfg)
|
|
return -ENOMEM;
|
|
|
|
dev_info(dev, "Commit DCB Configuration to the hardware\n");
|
|
pf_vsi = ice_get_main_vsi(pf);
|
|
if (!pf_vsi) {
|
|
dev_dbg(dev, "PF VSI doesn't exist\n");
|
|
ret = -EINVAL;
|
|
goto free_cfg;
|
|
}
|
|
|
|
/* Notify AUX drivers about impending change to TCs */
|
|
event = kzalloc(sizeof(*event), GFP_KERNEL);
|
|
if (!event) {
|
|
ret = -ENOMEM;
|
|
goto free_cfg;
|
|
}
|
|
|
|
set_bit(IIDC_EVENT_BEFORE_TC_CHANGE, event->type);
|
|
ice_send_event_to_aux(pf, event);
|
|
kfree(event);
|
|
|
|
/* avoid race conditions by holding the lock while disabling and
|
|
* re-enabling the VSI
|
|
*/
|
|
if (!locked)
|
|
rtnl_lock();
|
|
|
|
/* disable VSIs affected by DCB changes */
|
|
ice_dcb_ena_dis_vsi(pf, false, true);
|
|
|
|
memcpy(curr_cfg, new_cfg, sizeof(*curr_cfg));
|
|
memcpy(&curr_cfg->etsrec, &curr_cfg->etscfg, sizeof(curr_cfg->etsrec));
|
|
memcpy(&new_cfg->etsrec, &curr_cfg->etscfg, sizeof(curr_cfg->etsrec));
|
|
|
|
/* Only send new config to HW if we are in SW LLDP mode. Otherwise,
|
|
* the new config came from the HW in the first place.
|
|
*/
|
|
if (pf->hw.port_info->qos_cfg.is_sw_lldp) {
|
|
ret = ice_set_dcb_cfg(pf->hw.port_info);
|
|
if (ret) {
|
|
dev_err(dev, "Set DCB Config failed\n");
|
|
/* Restore previous settings to local config */
|
|
memcpy(curr_cfg, old_cfg, sizeof(*curr_cfg));
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
ret = ice_query_port_ets(pf->hw.port_info, &buf, sizeof(buf), NULL);
|
|
if (ret) {
|
|
dev_err(dev, "Query Port ETS failed\n");
|
|
goto out;
|
|
}
|
|
|
|
ice_pf_dcb_recfg(pf);
|
|
|
|
out:
|
|
/* enable previously downed VSIs */
|
|
ice_dcb_ena_dis_vsi(pf, true, true);
|
|
if (!locked)
|
|
rtnl_unlock();
|
|
free_cfg:
|
|
kfree(old_cfg);
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* ice_cfg_etsrec_defaults - Set default ETS recommended DCB config
|
|
* @pi: port information structure
|
|
*/
|
|
static void ice_cfg_etsrec_defaults(struct ice_port_info *pi)
|
|
{
|
|
struct ice_dcbx_cfg *dcbcfg = &pi->qos_cfg.local_dcbx_cfg;
|
|
u8 i;
|
|
|
|
/* Ensure ETS recommended DCB configuration is not already set */
|
|
if (dcbcfg->etsrec.maxtcs)
|
|
return;
|
|
|
|
/* In CEE mode, set the default to 1 TC */
|
|
dcbcfg->etsrec.maxtcs = 1;
|
|
for (i = 0; i < ICE_MAX_TRAFFIC_CLASS; i++) {
|
|
dcbcfg->etsrec.tcbwtable[i] = i ? 0 : 100;
|
|
dcbcfg->etsrec.tsatable[i] = i ? ICE_IEEE_TSA_STRICT :
|
|
ICE_IEEE_TSA_ETS;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_need_recfg - Check if DCB needs reconfig
|
|
* @pf: board private structure
|
|
* @old_cfg: current DCB config
|
|
* @new_cfg: new DCB config
|
|
*/
|
|
static bool
|
|
ice_dcb_need_recfg(struct ice_pf *pf, struct ice_dcbx_cfg *old_cfg,
|
|
struct ice_dcbx_cfg *new_cfg)
|
|
{
|
|
struct device *dev = ice_pf_to_dev(pf);
|
|
bool need_reconfig = false;
|
|
|
|
/* Check if ETS configuration has changed */
|
|
if (memcmp(&new_cfg->etscfg, &old_cfg->etscfg,
|
|
sizeof(new_cfg->etscfg))) {
|
|
/* If Priority Table has changed reconfig is needed */
|
|
if (memcmp(&new_cfg->etscfg.prio_table,
|
|
&old_cfg->etscfg.prio_table,
|
|
sizeof(new_cfg->etscfg.prio_table))) {
|
|
need_reconfig = true;
|
|
dev_dbg(dev, "ETS UP2TC changed.\n");
|
|
}
|
|
|
|
if (memcmp(&new_cfg->etscfg.tcbwtable,
|
|
&old_cfg->etscfg.tcbwtable,
|
|
sizeof(new_cfg->etscfg.tcbwtable)))
|
|
dev_dbg(dev, "ETS TC BW Table changed.\n");
|
|
|
|
if (memcmp(&new_cfg->etscfg.tsatable,
|
|
&old_cfg->etscfg.tsatable,
|
|
sizeof(new_cfg->etscfg.tsatable)))
|
|
dev_dbg(dev, "ETS TSA Table changed.\n");
|
|
}
|
|
|
|
/* Check if PFC configuration has changed */
|
|
if (memcmp(&new_cfg->pfc, &old_cfg->pfc, sizeof(new_cfg->pfc))) {
|
|
need_reconfig = true;
|
|
dev_dbg(dev, "PFC config change detected.\n");
|
|
}
|
|
|
|
/* Check if APP Table has changed */
|
|
if (memcmp(&new_cfg->app, &old_cfg->app, sizeof(new_cfg->app))) {
|
|
need_reconfig = true;
|
|
dev_dbg(dev, "APP Table change detected.\n");
|
|
}
|
|
|
|
dev_dbg(dev, "dcb need_reconfig=%d\n", need_reconfig);
|
|
return need_reconfig;
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_rebuild - rebuild DCB post reset
|
|
* @pf: physical function instance
|
|
*/
|
|
void ice_dcb_rebuild(struct ice_pf *pf)
|
|
{
|
|
struct ice_aqc_port_ets_elem buf = { 0 };
|
|
struct device *dev = ice_pf_to_dev(pf);
|
|
struct ice_dcbx_cfg *err_cfg;
|
|
int ret;
|
|
|
|
ret = ice_query_port_ets(pf->hw.port_info, &buf, sizeof(buf), NULL);
|
|
if (ret) {
|
|
dev_err(dev, "Query Port ETS failed\n");
|
|
goto dcb_error;
|
|
}
|
|
|
|
mutex_lock(&pf->tc_mutex);
|
|
|
|
if (!pf->hw.port_info->qos_cfg.is_sw_lldp)
|
|
ice_cfg_etsrec_defaults(pf->hw.port_info);
|
|
|
|
ret = ice_set_dcb_cfg(pf->hw.port_info);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to set DCB config in rebuild\n");
|
|
goto dcb_error;
|
|
}
|
|
|
|
if (!pf->hw.port_info->qos_cfg.is_sw_lldp) {
|
|
ret = ice_cfg_lldp_mib_change(&pf->hw, true);
|
|
if (ret && !pf->hw.port_info->qos_cfg.is_sw_lldp) {
|
|
dev_err(dev, "Failed to register for MIB changes\n");
|
|
goto dcb_error;
|
|
}
|
|
}
|
|
|
|
dev_info(dev, "DCB info restored\n");
|
|
ret = ice_query_port_ets(pf->hw.port_info, &buf, sizeof(buf), NULL);
|
|
if (ret) {
|
|
dev_err(dev, "Query Port ETS failed\n");
|
|
goto dcb_error;
|
|
}
|
|
|
|
mutex_unlock(&pf->tc_mutex);
|
|
|
|
return;
|
|
|
|
dcb_error:
|
|
dev_err(dev, "Disabling DCB until new settings occur\n");
|
|
err_cfg = kzalloc(sizeof(*err_cfg), GFP_KERNEL);
|
|
if (!err_cfg) {
|
|
mutex_unlock(&pf->tc_mutex);
|
|
return;
|
|
}
|
|
|
|
err_cfg->etscfg.willing = true;
|
|
err_cfg->etscfg.tcbwtable[0] = ICE_TC_MAX_BW;
|
|
err_cfg->etscfg.tsatable[0] = ICE_IEEE_TSA_ETS;
|
|
memcpy(&err_cfg->etsrec, &err_cfg->etscfg, sizeof(err_cfg->etsrec));
|
|
/* Coverity warns the return code of ice_pf_dcb_cfg() is not checked
|
|
* here as is done for other calls to that function. That check is
|
|
* not necessary since this is in this function's error cleanup path.
|
|
* Suppress the Coverity warning with the following comment...
|
|
*/
|
|
/* coverity[check_return] */
|
|
ice_pf_dcb_cfg(pf, err_cfg, false);
|
|
kfree(err_cfg);
|
|
|
|
mutex_unlock(&pf->tc_mutex);
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_init_cfg - set the initial DCB config in SW
|
|
* @pf: PF to apply config to
|
|
* @locked: Is the RTNL held
|
|
*/
|
|
static int ice_dcb_init_cfg(struct ice_pf *pf, bool locked)
|
|
{
|
|
struct ice_dcbx_cfg *newcfg;
|
|
struct ice_port_info *pi;
|
|
int ret = 0;
|
|
|
|
pi = pf->hw.port_info;
|
|
newcfg = kmemdup(&pi->qos_cfg.local_dcbx_cfg, sizeof(*newcfg),
|
|
GFP_KERNEL);
|
|
if (!newcfg)
|
|
return -ENOMEM;
|
|
|
|
memset(&pi->qos_cfg.local_dcbx_cfg, 0, sizeof(*newcfg));
|
|
|
|
dev_info(ice_pf_to_dev(pf), "Configuring initial DCB values\n");
|
|
if (ice_pf_dcb_cfg(pf, newcfg, locked))
|
|
ret = -EINVAL;
|
|
|
|
kfree(newcfg);
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_sw_dflt_cfg - Apply a default DCB config
|
|
* @pf: PF to apply config to
|
|
* @ets_willing: configure ETS willing
|
|
* @locked: was this function called with RTNL held
|
|
*/
|
|
int ice_dcb_sw_dflt_cfg(struct ice_pf *pf, bool ets_willing, bool locked)
|
|
{
|
|
struct ice_aqc_port_ets_elem buf = { 0 };
|
|
struct ice_dcbx_cfg *dcbcfg;
|
|
struct ice_port_info *pi;
|
|
struct ice_hw *hw;
|
|
int ret;
|
|
|
|
hw = &pf->hw;
|
|
pi = hw->port_info;
|
|
dcbcfg = kzalloc(sizeof(*dcbcfg), GFP_KERNEL);
|
|
if (!dcbcfg)
|
|
return -ENOMEM;
|
|
|
|
memset(&pi->qos_cfg.local_dcbx_cfg, 0, sizeof(*dcbcfg));
|
|
|
|
dcbcfg->etscfg.willing = ets_willing ? 1 : 0;
|
|
dcbcfg->etscfg.maxtcs = hw->func_caps.common_cap.maxtc;
|
|
dcbcfg->etscfg.tcbwtable[0] = 100;
|
|
dcbcfg->etscfg.tsatable[0] = ICE_IEEE_TSA_ETS;
|
|
|
|
memcpy(&dcbcfg->etsrec, &dcbcfg->etscfg,
|
|
sizeof(dcbcfg->etsrec));
|
|
dcbcfg->etsrec.willing = 0;
|
|
|
|
dcbcfg->pfc.willing = 1;
|
|
dcbcfg->pfc.pfccap = hw->func_caps.common_cap.maxtc;
|
|
|
|
dcbcfg->numapps = 1;
|
|
dcbcfg->app[0].selector = ICE_APP_SEL_ETHTYPE;
|
|
dcbcfg->app[0].priority = 3;
|
|
dcbcfg->app[0].prot_id = ETH_P_FCOE;
|
|
|
|
ret = ice_pf_dcb_cfg(pf, dcbcfg, locked);
|
|
kfree(dcbcfg);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return ice_query_port_ets(pi, &buf, sizeof(buf), NULL);
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_tc_contig - Check that TCs are contiguous
|
|
* @prio_table: pointer to priority table
|
|
*
|
|
* Check if TCs begin with TC0 and are contiguous
|
|
*/
|
|
static bool ice_dcb_tc_contig(u8 *prio_table)
|
|
{
|
|
bool found_empty = false;
|
|
u8 used_tc = 0;
|
|
int i;
|
|
|
|
/* Create a bitmap of used TCs */
|
|
for (i = 0; i < CEE_DCBX_MAX_PRIO; i++)
|
|
used_tc |= BIT(prio_table[i]);
|
|
|
|
for (i = 0; i < CEE_DCBX_MAX_PRIO; i++) {
|
|
if (used_tc & BIT(i)) {
|
|
if (found_empty)
|
|
return false;
|
|
} else {
|
|
found_empty = true;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_noncontig_cfg - Configure DCB for non-contiguous TCs
|
|
* @pf: pointer to the PF struct
|
|
*
|
|
* If non-contiguous TCs, then configure SW DCB with TC0 and ETS non-willing
|
|
*/
|
|
static int ice_dcb_noncontig_cfg(struct ice_pf *pf)
|
|
{
|
|
struct ice_dcbx_cfg *dcbcfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
|
|
struct device *dev = ice_pf_to_dev(pf);
|
|
int ret;
|
|
|
|
/* Configure SW DCB default with ETS non-willing */
|
|
ret = ice_dcb_sw_dflt_cfg(pf, false, true);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to set local DCB config %d\n", ret);
|
|
return ret;
|
|
}
|
|
|
|
/* Reconfigure with ETS willing so that FW will send LLDP MIB event */
|
|
dcbcfg->etscfg.willing = 1;
|
|
ret = ice_set_dcb_cfg(pf->hw.port_info);
|
|
if (ret)
|
|
dev_err(dev, "Failed to set DCB to unwilling\n");
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* ice_pf_dcb_recfg - Reconfigure all VEBs and VSIs
|
|
* @pf: pointer to the PF struct
|
|
*
|
|
* Assumed caller has already disabled all VSIs before
|
|
* calling this function. Reconfiguring DCB based on
|
|
* local_dcbx_cfg.
|
|
*/
|
|
void ice_pf_dcb_recfg(struct ice_pf *pf)
|
|
{
|
|
struct ice_dcbx_cfg *dcbcfg = &pf->hw.port_info->qos_cfg.local_dcbx_cfg;
|
|
struct iidc_event *event;
|
|
u8 tc_map = 0;
|
|
int v, ret;
|
|
|
|
/* Update each VSI */
|
|
ice_for_each_vsi(pf, v) {
|
|
struct ice_vsi *vsi = pf->vsi[v];
|
|
|
|
if (!vsi)
|
|
continue;
|
|
|
|
if (vsi->type == ICE_VSI_PF) {
|
|
tc_map = ice_dcb_get_ena_tc(dcbcfg);
|
|
|
|
/* If DCBX request non-contiguous TC, then configure
|
|
* default TC
|
|
*/
|
|
if (!ice_dcb_tc_contig(dcbcfg->etscfg.prio_table)) {
|
|
tc_map = ICE_DFLT_TRAFFIC_CLASS;
|
|
ice_dcb_noncontig_cfg(pf);
|
|
}
|
|
} else if (vsi->type == ICE_VSI_CHNL) {
|
|
tc_map = BIT(ice_get_first_droptc(vsi));
|
|
} else {
|
|
tc_map = ICE_DFLT_TRAFFIC_CLASS;
|
|
}
|
|
|
|
ret = ice_vsi_cfg_tc(vsi, tc_map);
|
|
if (ret) {
|
|
dev_err(ice_pf_to_dev(pf), "Failed to config TC for VSI index: %d\n",
|
|
vsi->idx);
|
|
continue;
|
|
}
|
|
/* no need to proceed with remaining cfg if it is CHNL
|
|
* or switchdev VSI
|
|
*/
|
|
if (vsi->type == ICE_VSI_CHNL ||
|
|
vsi->type == ICE_VSI_SWITCHDEV_CTRL)
|
|
continue;
|
|
|
|
ice_vsi_map_rings_to_vectors(vsi);
|
|
if (vsi->type == ICE_VSI_PF)
|
|
ice_dcbnl_set_all(vsi);
|
|
}
|
|
/* Notify the AUX drivers that TC change is finished */
|
|
event = kzalloc(sizeof(*event), GFP_KERNEL);
|
|
if (!event)
|
|
return;
|
|
|
|
set_bit(IIDC_EVENT_AFTER_TC_CHANGE, event->type);
|
|
ice_send_event_to_aux(pf, event);
|
|
kfree(event);
|
|
}
|
|
|
|
/**
|
|
* ice_init_pf_dcb - initialize DCB for a PF
|
|
* @pf: PF to initialize DCB for
|
|
* @locked: Was function called with RTNL held
|
|
*/
|
|
int ice_init_pf_dcb(struct ice_pf *pf, bool locked)
|
|
{
|
|
struct device *dev = ice_pf_to_dev(pf);
|
|
struct ice_port_info *port_info;
|
|
struct ice_hw *hw = &pf->hw;
|
|
int err;
|
|
|
|
port_info = hw->port_info;
|
|
|
|
err = ice_init_dcb(hw, false);
|
|
if (err && !port_info->qos_cfg.is_sw_lldp) {
|
|
dev_err(dev, "Error initializing DCB %d\n", err);
|
|
goto dcb_init_err;
|
|
}
|
|
|
|
dev_info(dev, "DCB is enabled in the hardware, max number of TCs supported on this port are %d\n",
|
|
pf->hw.func_caps.common_cap.maxtc);
|
|
if (err) {
|
|
struct ice_vsi *pf_vsi;
|
|
|
|
/* FW LLDP is disabled, activate SW DCBX/LLDP mode */
|
|
dev_info(dev, "FW LLDP is disabled, DCBx/LLDP in SW mode.\n");
|
|
clear_bit(ICE_FLAG_FW_LLDP_AGENT, pf->flags);
|
|
err = ice_aq_set_pfc_mode(&pf->hw, ICE_AQC_PFC_VLAN_BASED_PFC,
|
|
NULL);
|
|
if (err)
|
|
dev_info(dev, "Failed to set VLAN PFC mode\n");
|
|
|
|
err = ice_dcb_sw_dflt_cfg(pf, true, locked);
|
|
if (err) {
|
|
dev_err(dev, "Failed to set local DCB config %d\n",
|
|
err);
|
|
err = -EIO;
|
|
goto dcb_init_err;
|
|
}
|
|
|
|
/* If the FW DCBX engine is not running then Rx LLDP packets
|
|
* need to be redirected up the stack.
|
|
*/
|
|
pf_vsi = ice_get_main_vsi(pf);
|
|
if (!pf_vsi) {
|
|
dev_err(dev, "Failed to set local DCB config\n");
|
|
err = -EIO;
|
|
goto dcb_init_err;
|
|
}
|
|
|
|
ice_cfg_sw_lldp(pf_vsi, false, true);
|
|
|
|
pf->dcbx_cap = ice_dcb_get_mode(port_info, true);
|
|
return 0;
|
|
}
|
|
|
|
set_bit(ICE_FLAG_FW_LLDP_AGENT, pf->flags);
|
|
|
|
/* DCBX/LLDP enabled in FW, set DCBNL mode advertisement */
|
|
pf->dcbx_cap = ice_dcb_get_mode(port_info, false);
|
|
|
|
err = ice_dcb_init_cfg(pf, locked);
|
|
if (err)
|
|
goto dcb_init_err;
|
|
|
|
return err;
|
|
|
|
dcb_init_err:
|
|
dev_err(dev, "DCB init failed\n");
|
|
return err;
|
|
}
|
|
|
|
/**
|
|
* ice_update_dcb_stats - Update DCB stats counters
|
|
* @pf: PF whose stats needs to be updated
|
|
*/
|
|
void ice_update_dcb_stats(struct ice_pf *pf)
|
|
{
|
|
struct ice_hw_port_stats *prev_ps, *cur_ps;
|
|
struct ice_hw *hw = &pf->hw;
|
|
u8 port;
|
|
int i;
|
|
|
|
port = hw->port_info->lport;
|
|
prev_ps = &pf->stats_prev;
|
|
cur_ps = &pf->stats;
|
|
|
|
for (i = 0; i < 8; i++) {
|
|
ice_stat_update32(hw, GLPRT_PXOFFRXC(port, i),
|
|
pf->stat_prev_loaded,
|
|
&prev_ps->priority_xoff_rx[i],
|
|
&cur_ps->priority_xoff_rx[i]);
|
|
ice_stat_update32(hw, GLPRT_PXONRXC(port, i),
|
|
pf->stat_prev_loaded,
|
|
&prev_ps->priority_xon_rx[i],
|
|
&cur_ps->priority_xon_rx[i]);
|
|
ice_stat_update32(hw, GLPRT_PXONTXC(port, i),
|
|
pf->stat_prev_loaded,
|
|
&prev_ps->priority_xon_tx[i],
|
|
&cur_ps->priority_xon_tx[i]);
|
|
ice_stat_update32(hw, GLPRT_PXOFFTXC(port, i),
|
|
pf->stat_prev_loaded,
|
|
&prev_ps->priority_xoff_tx[i],
|
|
&cur_ps->priority_xoff_tx[i]);
|
|
ice_stat_update32(hw, GLPRT_RXON2OFFCNT(port, i),
|
|
pf->stat_prev_loaded,
|
|
&prev_ps->priority_xon_2_xoff[i],
|
|
&cur_ps->priority_xon_2_xoff[i]);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ice_tx_prepare_vlan_flags_dcb - prepare VLAN tagging for DCB
|
|
* @tx_ring: ring to send buffer on
|
|
* @first: pointer to struct ice_tx_buf
|
|
*
|
|
* This should not be called if the outer VLAN is software offloaded as the VLAN
|
|
* tag will already be configured with the correct ID and priority bits
|
|
*/
|
|
void
|
|
ice_tx_prepare_vlan_flags_dcb(struct ice_tx_ring *tx_ring,
|
|
struct ice_tx_buf *first)
|
|
{
|
|
struct sk_buff *skb = first->skb;
|
|
|
|
if (!test_bit(ICE_FLAG_DCB_ENA, tx_ring->vsi->back->flags))
|
|
return;
|
|
|
|
/* Insert 802.1p priority into VLAN header */
|
|
if ((first->tx_flags & ICE_TX_FLAGS_HW_VLAN ||
|
|
first->tx_flags & ICE_TX_FLAGS_HW_OUTER_SINGLE_VLAN) ||
|
|
skb->priority != TC_PRIO_CONTROL) {
|
|
first->tx_flags &= ~ICE_TX_FLAGS_VLAN_PR_M;
|
|
/* Mask the lower 3 bits to set the 802.1p priority */
|
|
first->tx_flags |= (skb->priority & 0x7) <<
|
|
ICE_TX_FLAGS_VLAN_PR_S;
|
|
/* if this is not already set it means a VLAN 0 + priority needs
|
|
* to be offloaded
|
|
*/
|
|
if (tx_ring->flags & ICE_TX_FLAGS_RING_VLAN_L2TAG2)
|
|
first->tx_flags |= ICE_TX_FLAGS_HW_OUTER_SINGLE_VLAN;
|
|
else
|
|
first->tx_flags |= ICE_TX_FLAGS_HW_VLAN;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ice_dcb_process_lldp_set_mib_change - Process MIB change
|
|
* @pf: ptr to ice_pf
|
|
* @event: pointer to the admin queue receive event
|
|
*/
|
|
void
|
|
ice_dcb_process_lldp_set_mib_change(struct ice_pf *pf,
|
|
struct ice_rq_event_info *event)
|
|
{
|
|
struct ice_aqc_port_ets_elem buf = { 0 };
|
|
struct device *dev = ice_pf_to_dev(pf);
|
|
struct ice_aqc_lldp_get_mib *mib;
|
|
struct ice_dcbx_cfg tmp_dcbx_cfg;
|
|
bool need_reconfig = false;
|
|
struct ice_port_info *pi;
|
|
u8 mib_type;
|
|
int ret;
|
|
|
|
/* Not DCB capable or capability disabled */
|
|
if (!(test_bit(ICE_FLAG_DCB_CAPABLE, pf->flags)))
|
|
return;
|
|
|
|
if (pf->dcbx_cap & DCB_CAP_DCBX_HOST) {
|
|
dev_dbg(dev, "MIB Change Event in HOST mode\n");
|
|
return;
|
|
}
|
|
|
|
pi = pf->hw.port_info;
|
|
mib = (struct ice_aqc_lldp_get_mib *)&event->desc.params.raw;
|
|
/* Ignore if event is not for Nearest Bridge */
|
|
mib_type = ((mib->type >> ICE_AQ_LLDP_BRID_TYPE_S) &
|
|
ICE_AQ_LLDP_BRID_TYPE_M);
|
|
dev_dbg(dev, "LLDP event MIB bridge type 0x%x\n", mib_type);
|
|
if (mib_type != ICE_AQ_LLDP_BRID_TYPE_NEAREST_BRID)
|
|
return;
|
|
|
|
/* Check MIB Type and return if event for Remote MIB update */
|
|
mib_type = mib->type & ICE_AQ_LLDP_MIB_TYPE_M;
|
|
dev_dbg(dev, "LLDP event mib type %s\n", mib_type ? "remote" : "local");
|
|
if (mib_type == ICE_AQ_LLDP_MIB_REMOTE) {
|
|
/* Update the remote cached instance and return */
|
|
ret = ice_aq_get_dcb_cfg(pi->hw, ICE_AQ_LLDP_MIB_REMOTE,
|
|
ICE_AQ_LLDP_BRID_TYPE_NEAREST_BRID,
|
|
&pi->qos_cfg.remote_dcbx_cfg);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to get remote DCB config\n");
|
|
return;
|
|
}
|
|
}
|
|
|
|
mutex_lock(&pf->tc_mutex);
|
|
|
|
/* store the old configuration */
|
|
tmp_dcbx_cfg = pf->hw.port_info->qos_cfg.local_dcbx_cfg;
|
|
|
|
/* Reset the old DCBX configuration data */
|
|
memset(&pi->qos_cfg.local_dcbx_cfg, 0,
|
|
sizeof(pi->qos_cfg.local_dcbx_cfg));
|
|
|
|
/* Get updated DCBX data from firmware */
|
|
ret = ice_get_dcb_cfg(pf->hw.port_info);
|
|
if (ret) {
|
|
dev_err(dev, "Failed to get DCB config\n");
|
|
goto out;
|
|
}
|
|
|
|
/* No change detected in DCBX configs */
|
|
if (!memcmp(&tmp_dcbx_cfg, &pi->qos_cfg.local_dcbx_cfg,
|
|
sizeof(tmp_dcbx_cfg))) {
|
|
dev_dbg(dev, "No change detected in DCBX configuration.\n");
|
|
goto out;
|
|
}
|
|
|
|
pf->dcbx_cap = ice_dcb_get_mode(pi, false);
|
|
|
|
need_reconfig = ice_dcb_need_recfg(pf, &tmp_dcbx_cfg,
|
|
&pi->qos_cfg.local_dcbx_cfg);
|
|
ice_dcbnl_flush_apps(pf, &tmp_dcbx_cfg, &pi->qos_cfg.local_dcbx_cfg);
|
|
if (!need_reconfig)
|
|
goto out;
|
|
|
|
/* Enable DCB tagging only when more than one TC */
|
|
if (ice_dcb_get_num_tc(&pi->qos_cfg.local_dcbx_cfg) > 1) {
|
|
dev_dbg(dev, "DCB tagging enabled (num TC > 1)\n");
|
|
set_bit(ICE_FLAG_DCB_ENA, pf->flags);
|
|
} else {
|
|
dev_dbg(dev, "DCB tagging disabled (num TC = 1)\n");
|
|
clear_bit(ICE_FLAG_DCB_ENA, pf->flags);
|
|
}
|
|
|
|
rtnl_lock();
|
|
/* disable VSIs affected by DCB changes */
|
|
ice_dcb_ena_dis_vsi(pf, false, true);
|
|
|
|
ret = ice_query_port_ets(pf->hw.port_info, &buf, sizeof(buf), NULL);
|
|
if (ret) {
|
|
dev_err(dev, "Query Port ETS failed\n");
|
|
goto unlock_rtnl;
|
|
}
|
|
|
|
/* changes in configuration update VSI */
|
|
ice_pf_dcb_recfg(pf);
|
|
|
|
/* enable previously downed VSIs */
|
|
ice_dcb_ena_dis_vsi(pf, true, true);
|
|
unlock_rtnl:
|
|
rtnl_unlock();
|
|
out:
|
|
mutex_unlock(&pf->tc_mutex);
|
|
}
|