mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-12 22:46:48 +07:00
Merge branch 'master' of git://git.kernel.org/pub/scm/linux/kernel/git/jkirsher/next-queue
Jeff Kirsher says: ==================== Intel Wired LAN Driver Updates 2015-10-19 This series contains updates to i40e and i40evf only. Kiran adds a spinlock around code accessing VSI MAC filter list to ensure that we are synchronizing access to the filter list, otherwise we can end up with multiple accesses at the same time which can cause the VSI MAC filter list to get in an unstable or corrupted state. Jesse fixes overlong BIT defines, where the RSS enabling call were mistakenly missed. Also fixes a bug where the enable function was enabling the interrupt twice while trying to update the two interrupt throttle rate thresholds for Rx and Tx, while refactoring the IRQ enable function to simplify reading the flow. Addressed the high CPU utilization of some small streaming workloads that the driver should reduce CPU in. Anjali fixes two X722 issues with respect to EEPROM checksum verify and reading NVM version info. Fixed where a mask value was accidentally replaced with a bit mask causing Flow Director sideband to be broken. Alex Duyck fixes areas of the drivers which run from hard interrupt context or with interrupts already disabled in netpoll, so use napi_schedule_irqoff() instead of napi_schedule(). Mitch fixes the VF drivers to not easily give up when it is not able to communicate with the PF driver. Carolyn fixes a problem where our tools MAC loopback test, after driver unbind would fail because the hardware was configured for multiqueue and unbind operation did not clear this configuration. Also fixed a issue where the NVMUpdate tool gets bad data from the PHY when using the PHY NVM feature because of contention on the MDIO interface from getting PHY capability calls from the driver during regular operations. Catherine fixed an issue where we were checking if autoneg was allowed to change before checking if autoneg was changing, these checks need to be in the reverse order. Jean Sacren fixes up an function header comment to align the kernel-docs with the actual code. v2: Cleaned up the use of spin_is_locked() in patch 1 based on feedback from David Miller, since it always evaluates to zero on uni-processor builds ==================== Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
commit
c8fdc32491
@ -468,6 +468,8 @@ struct i40e_vsi {
|
||||
#define I40E_VSI_FLAG_VEB_OWNER BIT(1)
|
||||
unsigned long flags;
|
||||
|
||||
/* Per VSI lock to protect elements/list (MAC filter) */
|
||||
spinlock_t mac_filter_list_lock;
|
||||
struct list_head mac_filter_list;
|
||||
|
||||
/* VSI stats */
|
||||
@ -575,6 +577,8 @@ struct i40e_q_vector {
|
||||
struct rcu_head rcu; /* to avoid race with update stats on free */
|
||||
char name[I40E_INT_NAME_STR_LEN];
|
||||
bool arm_wb_state;
|
||||
#define ITR_COUNTDOWN_START 100
|
||||
u8 itr_countdown; /* when 0 should adjust ITR */
|
||||
} ____cacheline_internodealigned_in_smp;
|
||||
|
||||
/* lan device */
|
||||
|
@ -958,6 +958,9 @@ i40e_status i40e_init_shared_code(struct i40e_hw *hw)
|
||||
else
|
||||
hw->pf_id = (u8)(func_rid & 0x7);
|
||||
|
||||
if (hw->mac.type == I40E_MAC_X722)
|
||||
hw->flags |= I40E_HW_FLAG_AQ_SRCTL_ACCESS_ENABLE;
|
||||
|
||||
status = i40e_init_nvm(hw);
|
||||
return status;
|
||||
}
|
||||
@ -2275,13 +2278,15 @@ i40e_status i40e_update_link_info(struct i40e_hw *hw)
|
||||
if (status)
|
||||
return status;
|
||||
|
||||
status = i40e_aq_get_phy_capabilities(hw, false, false, &abilities,
|
||||
NULL);
|
||||
if (hw->phy.link_info.link_info & I40E_AQ_MEDIA_AVAILABLE) {
|
||||
status = i40e_aq_get_phy_capabilities(hw, false, false,
|
||||
&abilities, NULL);
|
||||
if (status)
|
||||
return status;
|
||||
|
||||
memcpy(hw->phy.link_info.module_type, &abilities.module_type,
|
||||
sizeof(hw->phy.link_info.module_type));
|
||||
}
|
||||
|
||||
return status;
|
||||
}
|
||||
|
@ -661,28 +661,31 @@ static int i40e_set_settings(struct net_device *netdev,
|
||||
|
||||
/* Check autoneg */
|
||||
if (autoneg == AUTONEG_ENABLE) {
|
||||
/* If autoneg was not already enabled */
|
||||
if (!(hw->phy.link_info.an_info & I40E_AQ_AN_COMPLETED)) {
|
||||
/* If autoneg is not supported, return error */
|
||||
if (!(safe_ecmd.supported & SUPPORTED_Autoneg)) {
|
||||
netdev_info(netdev, "Autoneg not supported on this phy\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
/* If autoneg was not already enabled */
|
||||
if (!(hw->phy.link_info.an_info & I40E_AQ_AN_COMPLETED)) {
|
||||
/* Autoneg is allowed to change */
|
||||
config.abilities = abilities.abilities |
|
||||
I40E_AQ_PHY_ENABLE_AN;
|
||||
change = true;
|
||||
}
|
||||
} else {
|
||||
/* If autoneg is supported 10GBASE_T is the only phy that
|
||||
* can disable it, so otherwise return error
|
||||
/* If autoneg is currently enabled */
|
||||
if (hw->phy.link_info.an_info & I40E_AQ_AN_COMPLETED) {
|
||||
/* If autoneg is supported 10GBASE_T is the only PHY
|
||||
* that can disable it, so otherwise return error
|
||||
*/
|
||||
if (safe_ecmd.supported & SUPPORTED_Autoneg &&
|
||||
hw->phy.link_info.phy_type != I40E_PHY_TYPE_10GBASE_T) {
|
||||
hw->phy.link_info.phy_type !=
|
||||
I40E_PHY_TYPE_10GBASE_T) {
|
||||
netdev_info(netdev, "Autoneg cannot be disabled on this phy\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
/* If autoneg is currently enabled */
|
||||
if (hw->phy.link_info.an_info & I40E_AQ_AN_COMPLETED) {
|
||||
/* Autoneg is allowed to change */
|
||||
config.abilities = abilities.abilities &
|
||||
~I40E_AQ_PHY_ENABLE_AN;
|
||||
change = true;
|
||||
@ -748,7 +751,7 @@ static int i40e_set_settings(struct net_device *netdev,
|
||||
|
||||
status = i40e_update_link_info(hw);
|
||||
if (status)
|
||||
netdev_info(netdev, "Updating link info failed with err %s aq_err %s\n",
|
||||
netdev_dbg(netdev, "Updating link info failed with err %s aq_err %s\n",
|
||||
i40e_stat_str(hw, status),
|
||||
i40e_aq_str(hw, hw->aq.asq_last_status));
|
||||
|
||||
|
@ -1516,10 +1516,12 @@ void i40e_fcoe_config_netdev(struct net_device *netdev, struct i40e_vsi *vsi)
|
||||
* same PCI function.
|
||||
*/
|
||||
netdev->dev_port = 1;
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
i40e_add_filter(vsi, hw->mac.san_addr, 0, false, false);
|
||||
i40e_add_filter(vsi, (u8[6]) FC_FCOE_FLOGI_MAC, 0, false, false);
|
||||
i40e_add_filter(vsi, FIP_ALL_FCOE_MACS, 0, false, false);
|
||||
i40e_add_filter(vsi, FIP_ALL_ENODE_MACS, 0, false, false);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
/* use san mac */
|
||||
ether_addr_copy(netdev->dev_addr, hw->mac.san_addr);
|
||||
|
@ -39,7 +39,7 @@ static const char i40e_driver_string[] =
|
||||
|
||||
#define DRV_VERSION_MAJOR 1
|
||||
#define DRV_VERSION_MINOR 3
|
||||
#define DRV_VERSION_BUILD 34
|
||||
#define DRV_VERSION_BUILD 38
|
||||
#define DRV_VERSION __stringify(DRV_VERSION_MAJOR) "." \
|
||||
__stringify(DRV_VERSION_MINOR) "." \
|
||||
__stringify(DRV_VERSION_BUILD) DRV_KERN
|
||||
@ -1355,6 +1355,9 @@ static int i40e_rm_default_mac_filter(struct i40e_vsi *vsi, u8 *macaddr)
|
||||
* @is_netdev: make sure its a netdev filter, else doesn't matter
|
||||
*
|
||||
* Returns ptr to the filter object or NULL when no memory available.
|
||||
*
|
||||
* NOTE: This function is expected to be called with mac_filter_list_lock
|
||||
* being held.
|
||||
**/
|
||||
struct i40e_mac_filter *i40e_add_filter(struct i40e_vsi *vsi,
|
||||
u8 *macaddr, s16 vlan,
|
||||
@ -1413,6 +1416,9 @@ struct i40e_mac_filter *i40e_add_filter(struct i40e_vsi *vsi,
|
||||
* @vlan: the vlan
|
||||
* @is_vf: make sure it's a VF filter, else doesn't matter
|
||||
* @is_netdev: make sure it's a netdev filter, else doesn't matter
|
||||
*
|
||||
* NOTE: This function is expected to be called with mac_filter_list_lock
|
||||
* being held.
|
||||
**/
|
||||
void i40e_del_filter(struct i40e_vsi *vsi,
|
||||
u8 *macaddr, s16 vlan,
|
||||
@ -1519,8 +1525,10 @@ static int i40e_set_mac(struct net_device *netdev, void *p)
|
||||
element.flags = I40E_AQC_MACVLAN_DEL_PERFECT_MATCH;
|
||||
i40e_aq_remove_macvlan(&pf->hw, vsi->seid, &element, 1, NULL);
|
||||
} else {
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
i40e_del_filter(vsi, netdev->dev_addr, I40E_VLAN_ANY,
|
||||
false, false);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
}
|
||||
|
||||
if (ether_addr_equal(addr->sa_data, hw->mac.addr)) {
|
||||
@ -1531,10 +1539,12 @@ static int i40e_set_mac(struct net_device *netdev, void *p)
|
||||
element.flags = cpu_to_le16(I40E_AQC_MACVLAN_ADD_PERFECT_MATCH);
|
||||
i40e_aq_add_macvlan(&pf->hw, vsi->seid, &element, 1, NULL);
|
||||
} else {
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
f = i40e_add_filter(vsi, addr->sa_data, I40E_VLAN_ANY,
|
||||
false, false);
|
||||
if (f)
|
||||
f->is_laa = true;
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
}
|
||||
|
||||
i40e_sync_vsi_filters(vsi, false);
|
||||
@ -1707,6 +1717,8 @@ static void i40e_set_rx_mode(struct net_device *netdev)
|
||||
struct netdev_hw_addr *mca;
|
||||
struct netdev_hw_addr *ha;
|
||||
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
/* add addr if not already in the filter list */
|
||||
netdev_for_each_uc_addr(uca, netdev) {
|
||||
if (!i40e_find_mac(vsi, uca->addr, false, true)) {
|
||||
@ -1754,6 +1766,7 @@ static void i40e_set_rx_mode(struct net_device *netdev)
|
||||
bottom_of_search_loop:
|
||||
continue;
|
||||
}
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
/* check for other flag changes */
|
||||
if (vsi->current_netdev_flags != vsi->netdev->flags) {
|
||||
@ -1762,6 +1775,79 @@ static void i40e_set_rx_mode(struct net_device *netdev)
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* i40e_mac_filter_entry_clone - Clones a MAC filter entry
|
||||
* @src: source MAC filter entry to be clones
|
||||
*
|
||||
* Returns the pointer to newly cloned MAC filter entry or NULL
|
||||
* in case of error
|
||||
**/
|
||||
static struct i40e_mac_filter *i40e_mac_filter_entry_clone(
|
||||
struct i40e_mac_filter *src)
|
||||
{
|
||||
struct i40e_mac_filter *f;
|
||||
|
||||
f = kzalloc(sizeof(*f), GFP_ATOMIC);
|
||||
if (!f)
|
||||
return NULL;
|
||||
*f = *src;
|
||||
|
||||
INIT_LIST_HEAD(&f->list);
|
||||
|
||||
return f;
|
||||
}
|
||||
|
||||
/**
|
||||
* i40e_undo_del_filter_entries - Undo the changes made to MAC filter entries
|
||||
* @vsi: pointer to vsi struct
|
||||
* @from: Pointer to list which contains MAC filter entries - changes to
|
||||
* those entries needs to be undone.
|
||||
*
|
||||
* MAC filter entries from list were slated to be removed from device.
|
||||
**/
|
||||
static void i40e_undo_del_filter_entries(struct i40e_vsi *vsi,
|
||||
struct list_head *from)
|
||||
{
|
||||
struct i40e_mac_filter *f, *ftmp;
|
||||
|
||||
list_for_each_entry_safe(f, ftmp, from, list) {
|
||||
f->changed = true;
|
||||
/* Move the element back into MAC filter list*/
|
||||
list_move_tail(&f->list, &vsi->mac_filter_list);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* i40e_undo_add_filter_entries - Undo the changes made to MAC filter entries
|
||||
* @vsi: pointer to vsi struct
|
||||
*
|
||||
* MAC filter entries from list were slated to be added from device.
|
||||
**/
|
||||
static void i40e_undo_add_filter_entries(struct i40e_vsi *vsi)
|
||||
{
|
||||
struct i40e_mac_filter *f, *ftmp;
|
||||
|
||||
list_for_each_entry_safe(f, ftmp, &vsi->mac_filter_list, list) {
|
||||
if (!f->changed && f->counter)
|
||||
f->changed = true;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* i40e_cleanup_add_list - Deletes the element from add list and release
|
||||
* memory
|
||||
* @add_list: Pointer to list which contains MAC filter entries
|
||||
**/
|
||||
static void i40e_cleanup_add_list(struct list_head *add_list)
|
||||
{
|
||||
struct i40e_mac_filter *f, *ftmp;
|
||||
|
||||
list_for_each_entry_safe(f, ftmp, add_list, list) {
|
||||
list_del(&f->list);
|
||||
kfree(f);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* i40e_sync_vsi_filters - Update the VSI filter list to the HW
|
||||
* @vsi: ptr to the VSI
|
||||
@ -1773,11 +1859,13 @@ static void i40e_set_rx_mode(struct net_device *netdev)
|
||||
**/
|
||||
int i40e_sync_vsi_filters(struct i40e_vsi *vsi, bool grab_rtnl)
|
||||
{
|
||||
struct i40e_mac_filter *f, *ftmp;
|
||||
struct list_head tmp_del_list, tmp_add_list;
|
||||
struct i40e_mac_filter *f, *ftmp, *fclone;
|
||||
bool promisc_forced_on = false;
|
||||
bool add_happened = false;
|
||||
int filter_list_len = 0;
|
||||
u32 changed_flags = 0;
|
||||
bool err_cond = false;
|
||||
i40e_status ret = 0;
|
||||
struct i40e_pf *pf;
|
||||
int num_add = 0;
|
||||
@ -1798,17 +1886,13 @@ int i40e_sync_vsi_filters(struct i40e_vsi *vsi, bool grab_rtnl)
|
||||
vsi->current_netdev_flags = vsi->netdev->flags;
|
||||
}
|
||||
|
||||
INIT_LIST_HEAD(&tmp_del_list);
|
||||
INIT_LIST_HEAD(&tmp_add_list);
|
||||
|
||||
if (vsi->flags & I40E_VSI_FLAG_FILTER_CHANGED) {
|
||||
vsi->flags &= ~I40E_VSI_FLAG_FILTER_CHANGED;
|
||||
|
||||
filter_list_len = pf->hw.aq.asq_buf_size /
|
||||
sizeof(struct i40e_aqc_remove_macvlan_element_data);
|
||||
del_list = kcalloc(filter_list_len,
|
||||
sizeof(struct i40e_aqc_remove_macvlan_element_data),
|
||||
GFP_KERNEL);
|
||||
if (!del_list)
|
||||
return -ENOMEM;
|
||||
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
list_for_each_entry_safe(f, ftmp, &vsi->mac_filter_list, list) {
|
||||
if (!f->changed)
|
||||
continue;
|
||||
@ -1816,6 +1900,58 @@ int i40e_sync_vsi_filters(struct i40e_vsi *vsi, bool grab_rtnl)
|
||||
if (f->counter != 0)
|
||||
continue;
|
||||
f->changed = false;
|
||||
|
||||
/* Move the element into temporary del_list */
|
||||
list_move_tail(&f->list, &tmp_del_list);
|
||||
}
|
||||
|
||||
list_for_each_entry_safe(f, ftmp, &vsi->mac_filter_list, list) {
|
||||
if (!f->changed)
|
||||
continue;
|
||||
|
||||
if (f->counter == 0)
|
||||
continue;
|
||||
f->changed = false;
|
||||
|
||||
/* Clone MAC filter entry and add into temporary list */
|
||||
fclone = i40e_mac_filter_entry_clone(f);
|
||||
if (!fclone) {
|
||||
err_cond = true;
|
||||
break;
|
||||
}
|
||||
list_add_tail(&fclone->list, &tmp_add_list);
|
||||
}
|
||||
|
||||
/* if failed to clone MAC filter entry - undo */
|
||||
if (err_cond) {
|
||||
i40e_undo_del_filter_entries(vsi, &tmp_del_list);
|
||||
i40e_undo_add_filter_entries(vsi);
|
||||
}
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
if (err_cond)
|
||||
i40e_cleanup_add_list(&tmp_add_list);
|
||||
}
|
||||
|
||||
/* Now process 'del_list' outside the lock */
|
||||
if (!list_empty(&tmp_del_list)) {
|
||||
filter_list_len = pf->hw.aq.asq_buf_size /
|
||||
sizeof(struct i40e_aqc_remove_macvlan_element_data);
|
||||
del_list = kcalloc(filter_list_len,
|
||||
sizeof(struct i40e_aqc_remove_macvlan_element_data),
|
||||
GFP_KERNEL);
|
||||
if (!del_list) {
|
||||
i40e_cleanup_add_list(&tmp_add_list);
|
||||
|
||||
/* Undo VSI's MAC filter entry element updates */
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
i40e_undo_del_filter_entries(vsi, &tmp_del_list);
|
||||
i40e_undo_add_filter_entries(vsi);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
list_for_each_entry_safe(f, ftmp, &tmp_del_list, list) {
|
||||
cmd_flags = 0;
|
||||
|
||||
/* add to delete list */
|
||||
@ -1828,10 +1964,6 @@ int i40e_sync_vsi_filters(struct i40e_vsi *vsi, bool grab_rtnl)
|
||||
del_list[num_del].flags = cmd_flags;
|
||||
num_del++;
|
||||
|
||||
/* unlink from filter list */
|
||||
list_del(&f->list);
|
||||
kfree(f);
|
||||
|
||||
/* flush a full buffer */
|
||||
if (num_del == filter_list_len) {
|
||||
ret = i40e_aq_remove_macvlan(&pf->hw,
|
||||
@ -1842,12 +1974,18 @@ int i40e_sync_vsi_filters(struct i40e_vsi *vsi, bool grab_rtnl)
|
||||
memset(del_list, 0, sizeof(*del_list));
|
||||
|
||||
if (ret && aq_err != I40E_AQ_RC_ENOENT)
|
||||
dev_info(&pf->pdev->dev,
|
||||
dev_err(&pf->pdev->dev,
|
||||
"ignoring delete macvlan error, err %s, aq_err %s while flushing a full buffer\n",
|
||||
i40e_stat_str(&pf->hw, ret),
|
||||
i40e_aq_str(&pf->hw, aq_err));
|
||||
}
|
||||
/* Release memory for MAC filter entries which were
|
||||
* synced up with HW.
|
||||
*/
|
||||
list_del(&f->list);
|
||||
kfree(f);
|
||||
}
|
||||
|
||||
if (num_del) {
|
||||
ret = i40e_aq_remove_macvlan(&pf->hw, vsi->seid,
|
||||
del_list, num_del, NULL);
|
||||
@ -1863,6 +2001,9 @@ int i40e_sync_vsi_filters(struct i40e_vsi *vsi, bool grab_rtnl)
|
||||
|
||||
kfree(del_list);
|
||||
del_list = NULL;
|
||||
}
|
||||
|
||||
if (!list_empty(&tmp_add_list)) {
|
||||
|
||||
/* do all the adds now */
|
||||
filter_list_len = pf->hw.aq.asq_buf_size /
|
||||
@ -1870,16 +2011,19 @@ int i40e_sync_vsi_filters(struct i40e_vsi *vsi, bool grab_rtnl)
|
||||
add_list = kcalloc(filter_list_len,
|
||||
sizeof(struct i40e_aqc_add_macvlan_element_data),
|
||||
GFP_KERNEL);
|
||||
if (!add_list)
|
||||
if (!add_list) {
|
||||
/* Purge element from temporary lists */
|
||||
i40e_cleanup_add_list(&tmp_add_list);
|
||||
|
||||
/* Undo add filter entries from VSI MAC filter list */
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
i40e_undo_add_filter_entries(vsi);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
list_for_each_entry_safe(f, ftmp, &vsi->mac_filter_list, list) {
|
||||
if (!f->changed)
|
||||
continue;
|
||||
list_for_each_entry_safe(f, ftmp, &tmp_add_list, list) {
|
||||
|
||||
if (f->counter == 0)
|
||||
continue;
|
||||
f->changed = false;
|
||||
add_happened = true;
|
||||
cmd_flags = 0;
|
||||
|
||||
@ -1906,7 +2050,13 @@ int i40e_sync_vsi_filters(struct i40e_vsi *vsi, bool grab_rtnl)
|
||||
break;
|
||||
memset(add_list, 0, sizeof(*add_list));
|
||||
}
|
||||
/* Entries from tmp_add_list were cloned from MAC
|
||||
* filter list, hence clean those cloned entries
|
||||
*/
|
||||
list_del(&f->list);
|
||||
kfree(f);
|
||||
}
|
||||
|
||||
if (num_add) {
|
||||
ret = i40e_aq_add_macvlan(&pf->hw, vsi->seid,
|
||||
add_list, num_add, NULL);
|
||||
@ -2158,6 +2308,9 @@ int i40e_vsi_add_vlan(struct i40e_vsi *vsi, s16 vid)
|
||||
is_vf = (vsi->type == I40E_VSI_SRIOV);
|
||||
is_netdev = !!(vsi->netdev);
|
||||
|
||||
/* Locked once because all functions invoked below iterates list*/
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
if (is_netdev) {
|
||||
add_f = i40e_add_filter(vsi, vsi->netdev->dev_addr, vid,
|
||||
is_vf, is_netdev);
|
||||
@ -2165,6 +2318,7 @@ int i40e_vsi_add_vlan(struct i40e_vsi *vsi, s16 vid)
|
||||
dev_info(&vsi->back->pdev->dev,
|
||||
"Could not add vlan filter %d for %pM\n",
|
||||
vid, vsi->netdev->dev_addr);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
return -ENOMEM;
|
||||
}
|
||||
}
|
||||
@ -2175,6 +2329,7 @@ int i40e_vsi_add_vlan(struct i40e_vsi *vsi, s16 vid)
|
||||
dev_info(&vsi->back->pdev->dev,
|
||||
"Could not add vlan filter %d for %pM\n",
|
||||
vid, f->macaddr);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
return -ENOMEM;
|
||||
}
|
||||
}
|
||||
@ -2196,6 +2351,7 @@ int i40e_vsi_add_vlan(struct i40e_vsi *vsi, s16 vid)
|
||||
dev_info(&vsi->back->pdev->dev,
|
||||
"Could not add filter 0 for %pM\n",
|
||||
vsi->netdev->dev_addr);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
return -ENOMEM;
|
||||
}
|
||||
}
|
||||
@ -2204,8 +2360,9 @@ int i40e_vsi_add_vlan(struct i40e_vsi *vsi, s16 vid)
|
||||
/* Do not assume that I40E_VLAN_ANY should be reset to VLAN 0 */
|
||||
if (vid > 0 && !vsi->info.pvid) {
|
||||
list_for_each_entry(f, &vsi->mac_filter_list, list) {
|
||||
if (i40e_find_filter(vsi, f->macaddr, I40E_VLAN_ANY,
|
||||
is_vf, is_netdev)) {
|
||||
if (!i40e_find_filter(vsi, f->macaddr, I40E_VLAN_ANY,
|
||||
is_vf, is_netdev))
|
||||
continue;
|
||||
i40e_del_filter(vsi, f->macaddr, I40E_VLAN_ANY,
|
||||
is_vf, is_netdev);
|
||||
add_f = i40e_add_filter(vsi, f->macaddr,
|
||||
@ -2214,11 +2371,16 @@ int i40e_vsi_add_vlan(struct i40e_vsi *vsi, s16 vid)
|
||||
dev_info(&vsi->back->pdev->dev,
|
||||
"Could not add filter 0 for %pM\n",
|
||||
f->macaddr);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
return -ENOMEM;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/* Make sure to release before sync_vsi_filter because that
|
||||
* function will lock/unlock as necessary
|
||||
*/
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
if (test_bit(__I40E_DOWN, &vsi->back->state) ||
|
||||
test_bit(__I40E_RESET_RECOVERY_PENDING, &vsi->back->state))
|
||||
@ -2244,6 +2406,9 @@ int i40e_vsi_kill_vlan(struct i40e_vsi *vsi, s16 vid)
|
||||
is_vf = (vsi->type == I40E_VSI_SRIOV);
|
||||
is_netdev = !!(netdev);
|
||||
|
||||
/* Locked once because all functions invoked below iterates list */
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
if (is_netdev)
|
||||
i40e_del_filter(vsi, netdev->dev_addr, vid, is_vf, is_netdev);
|
||||
|
||||
@ -2274,6 +2439,7 @@ int i40e_vsi_kill_vlan(struct i40e_vsi *vsi, s16 vid)
|
||||
dev_info(&vsi->back->pdev->dev,
|
||||
"Could not add filter %d for %pM\n",
|
||||
I40E_VLAN_ANY, netdev->dev_addr);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
return -ENOMEM;
|
||||
}
|
||||
}
|
||||
@ -2287,11 +2453,17 @@ int i40e_vsi_kill_vlan(struct i40e_vsi *vsi, s16 vid)
|
||||
dev_info(&vsi->back->pdev->dev,
|
||||
"Could not add filter %d for %pM\n",
|
||||
I40E_VLAN_ANY, f->macaddr);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
return -ENOMEM;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/* Make sure to release before sync_vsi_filter because that
|
||||
* function with lock/unlock as necessary
|
||||
*/
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
if (test_bit(__I40E_DOWN, &vsi->back->state) ||
|
||||
test_bit(__I40E_RESET_RECOVERY_PENDING, &vsi->back->state))
|
||||
return 0;
|
||||
@ -2915,6 +3087,7 @@ static void i40e_vsi_configure_msix(struct i40e_vsi *vsi)
|
||||
for (i = 0; i < vsi->num_q_vectors; i++, vector++) {
|
||||
struct i40e_q_vector *q_vector = vsi->q_vectors[i];
|
||||
|
||||
q_vector->itr_countdown = ITR_COUNTDOWN_START;
|
||||
q_vector->rx.itr = ITR_TO_REG(vsi->rx_itr_setting);
|
||||
q_vector->rx.latency_range = I40E_LOW_LATENCY;
|
||||
wr32(hw, I40E_PFINT_ITRN(I40E_RX_ITR, vector - 1),
|
||||
@ -3010,6 +3183,7 @@ static void i40e_configure_msi_and_legacy(struct i40e_vsi *vsi)
|
||||
u32 val;
|
||||
|
||||
/* set the ITR configuration */
|
||||
q_vector->itr_countdown = ITR_COUNTDOWN_START;
|
||||
q_vector->rx.itr = ITR_TO_REG(vsi->rx_itr_setting);
|
||||
q_vector->rx.latency_range = I40E_LOW_LATENCY;
|
||||
wr32(hw, I40E_PFINT_ITR0(I40E_RX_ITR), q_vector->rx.itr);
|
||||
@ -3095,7 +3269,7 @@ static irqreturn_t i40e_msix_clean_rings(int irq, void *data)
|
||||
if (!q_vector->tx.ring && !q_vector->rx.ring)
|
||||
return IRQ_HANDLED;
|
||||
|
||||
napi_schedule(&q_vector->napi);
|
||||
napi_schedule_irqoff(&q_vector->napi);
|
||||
|
||||
return IRQ_HANDLED;
|
||||
}
|
||||
@ -3264,6 +3438,8 @@ static irqreturn_t i40e_intr(int irq, void *data)
|
||||
|
||||
/* only q0 is used in MSI/Legacy mode, and none are used in MSIX */
|
||||
if (icr0 & I40E_PFINT_ICR0_QUEUE_0_MASK) {
|
||||
struct i40e_vsi *vsi = pf->vsi[pf->lan_vsi];
|
||||
struct i40e_q_vector *q_vector = vsi->q_vectors[0];
|
||||
|
||||
/* temporarily disable queue cause for NAPI processing */
|
||||
u32 qval = rd32(hw, I40E_QINT_RQCTL(0));
|
||||
@ -3276,7 +3452,7 @@ static irqreturn_t i40e_intr(int irq, void *data)
|
||||
wr32(hw, I40E_QINT_TQCTL(0), qval);
|
||||
|
||||
if (!test_bit(__I40E_DOWN, &pf->state))
|
||||
napi_schedule(&pf->vsi[pf->lan_vsi]->q_vectors[0]->napi);
|
||||
napi_schedule_irqoff(&q_vector->napi);
|
||||
}
|
||||
|
||||
if (icr0 & I40E_PFINT_ICR0_ADMINQ_MASK) {
|
||||
@ -7062,6 +7238,8 @@ static int i40e_vsi_mem_alloc(struct i40e_pf *pf, enum i40e_vsi_type type)
|
||||
/* Setup default MSIX irq handler for VSI */
|
||||
i40e_vsi_setup_irqhandler(vsi, i40e_msix_clean_rings);
|
||||
|
||||
/* Initialize VSI lock */
|
||||
spin_lock_init(&vsi->mac_filter_list_lock);
|
||||
pf->vsi[vsi_idx] = vsi;
|
||||
ret = vsi_idx;
|
||||
goto unlock_pf;
|
||||
@ -8364,7 +8542,7 @@ static int i40e_ndo_bridge_getlink(struct sk_buff *skb, u32 pid, u32 seq,
|
||||
/**
|
||||
* i40e_features_check - Validate encapsulated packet conforms to limits
|
||||
* @skb: skb buff
|
||||
* @netdev: This physical port's netdev
|
||||
* @dev: This physical port's netdev
|
||||
* @features: Offload features that the stack believes apply
|
||||
**/
|
||||
static netdev_features_t i40e_features_check(struct sk_buff *skb,
|
||||
@ -8479,17 +8657,26 @@ static int i40e_config_netdev(struct i40e_vsi *vsi)
|
||||
* default a MAC-VLAN filter that accepts any tagged packet
|
||||
* which must be replaced by a normal filter.
|
||||
*/
|
||||
if (!i40e_rm_default_mac_filter(vsi, mac_addr))
|
||||
if (!i40e_rm_default_mac_filter(vsi, mac_addr)) {
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
i40e_add_filter(vsi, mac_addr,
|
||||
I40E_VLAN_ANY, false, true);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
}
|
||||
} else {
|
||||
/* relate the VSI_VMDQ name to the VSI_MAIN name */
|
||||
snprintf(netdev->name, IFNAMSIZ, "%sv%%d",
|
||||
pf->vsi[pf->lan_vsi]->netdev->name);
|
||||
random_ether_addr(mac_addr);
|
||||
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
i40e_add_filter(vsi, mac_addr, I40E_VLAN_ANY, false, false);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
}
|
||||
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
i40e_add_filter(vsi, brdcast, I40E_VLAN_ANY, false, false);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
ether_addr_copy(netdev->dev_addr, mac_addr);
|
||||
ether_addr_copy(netdev->perm_addr, mac_addr);
|
||||
@ -8563,10 +8750,13 @@ int i40e_is_vsi_uplink_mode_veb(struct i40e_vsi *vsi)
|
||||
static int i40e_add_vsi(struct i40e_vsi *vsi)
|
||||
{
|
||||
int ret = -ENODEV;
|
||||
struct i40e_mac_filter *f, *ftmp;
|
||||
u8 laa_macaddr[ETH_ALEN];
|
||||
bool found_laa_mac_filter = false;
|
||||
struct i40e_pf *pf = vsi->back;
|
||||
struct i40e_hw *hw = &pf->hw;
|
||||
struct i40e_vsi_context ctxt;
|
||||
struct i40e_mac_filter *f, *ftmp;
|
||||
|
||||
u8 enabled_tc = 0x1; /* TC0 enabled */
|
||||
int f_count = 0;
|
||||
|
||||
@ -8738,16 +8928,25 @@ static int i40e_add_vsi(struct i40e_vsi *vsi)
|
||||
vsi->id = ctxt.vsi_number;
|
||||
}
|
||||
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
/* If macvlan filters already exist, force them to get loaded */
|
||||
list_for_each_entry_safe(f, ftmp, &vsi->mac_filter_list, list) {
|
||||
f->changed = true;
|
||||
f_count++;
|
||||
|
||||
/* Expected to have only one MAC filter entry for LAA in list */
|
||||
if (f->is_laa && vsi->type == I40E_VSI_MAIN) {
|
||||
ether_addr_copy(laa_macaddr, f->macaddr);
|
||||
found_laa_mac_filter = true;
|
||||
}
|
||||
}
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
if (found_laa_mac_filter) {
|
||||
struct i40e_aqc_remove_macvlan_element_data element;
|
||||
|
||||
memset(&element, 0, sizeof(element));
|
||||
ether_addr_copy(element.mac_addr, f->macaddr);
|
||||
ether_addr_copy(element.mac_addr, laa_macaddr);
|
||||
element.flags = I40E_AQC_MACVLAN_DEL_PERFECT_MATCH;
|
||||
ret = i40e_aq_remove_macvlan(hw, vsi->seid,
|
||||
&element, 1, NULL);
|
||||
@ -8761,9 +8960,9 @@ static int i40e_add_vsi(struct i40e_vsi *vsi)
|
||||
|
||||
i40e_aq_mac_address_write(hw,
|
||||
I40E_AQC_WRITE_TYPE_LAA_WOL,
|
||||
f->macaddr, NULL);
|
||||
}
|
||||
laa_macaddr, NULL);
|
||||
}
|
||||
|
||||
if (f_count) {
|
||||
vsi->flags |= I40E_VSI_FLAG_FILTER_CHANGED;
|
||||
pf->flags |= I40E_FLAG_FILTER_SYNC;
|
||||
@ -8826,9 +9025,12 @@ int i40e_vsi_release(struct i40e_vsi *vsi)
|
||||
i40e_vsi_disable_irq(vsi);
|
||||
}
|
||||
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
list_for_each_entry_safe(f, ftmp, &vsi->mac_filter_list, list)
|
||||
i40e_del_filter(vsi, f->macaddr, f->vlan,
|
||||
f->is_vf, f->is_netdev);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
i40e_sync_vsi_filters(vsi, false);
|
||||
|
||||
i40e_vsi_delete(vsi);
|
||||
@ -9959,7 +10161,7 @@ static int i40e_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||
static u16 pfs_found;
|
||||
u16 wol_nvm_bits;
|
||||
u16 link_status;
|
||||
int err = 0;
|
||||
int err;
|
||||
u32 len;
|
||||
u32 i;
|
||||
|
||||
@ -10430,6 +10632,7 @@ static int i40e_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||
static void i40e_remove(struct pci_dev *pdev)
|
||||
{
|
||||
struct i40e_pf *pf = pci_get_drvdata(pdev);
|
||||
struct i40e_hw *hw = &pf->hw;
|
||||
i40e_status ret_code;
|
||||
int i;
|
||||
|
||||
@ -10437,6 +10640,10 @@ static void i40e_remove(struct pci_dev *pdev)
|
||||
|
||||
i40e_ptp_stop(pf);
|
||||
|
||||
/* Disable RSS in hw */
|
||||
wr32(hw, I40E_PFQF_HENA(0), 0);
|
||||
wr32(hw, I40E_PFQF_HENA(1), 0);
|
||||
|
||||
/* no more scheduling of any task */
|
||||
set_bit(__I40E_DOWN, &pf->state);
|
||||
del_timer_sync(&pf->service_timer);
|
||||
|
@ -290,9 +290,18 @@ static i40e_status i40e_read_nvm_word_aq(struct i40e_hw *hw, u16 offset,
|
||||
i40e_status i40e_read_nvm_word(struct i40e_hw *hw, u16 offset,
|
||||
u16 *data)
|
||||
{
|
||||
if (hw->mac.type == I40E_MAC_X722)
|
||||
return i40e_read_nvm_word_aq(hw, offset, data);
|
||||
return i40e_read_nvm_word_srctl(hw, offset, data);
|
||||
enum i40e_status_code ret_code = 0;
|
||||
|
||||
if (hw->flags & I40E_HW_FLAG_AQ_SRCTL_ACCESS_ENABLE) {
|
||||
ret_code = i40e_acquire_nvm(hw, I40E_RESOURCE_READ);
|
||||
if (!ret_code) {
|
||||
ret_code = i40e_read_nvm_word_aq(hw, offset, data);
|
||||
i40e_release_nvm(hw);
|
||||
}
|
||||
} else {
|
||||
ret_code = i40e_read_nvm_word_srctl(hw, offset, data);
|
||||
}
|
||||
return ret_code;
|
||||
}
|
||||
|
||||
/**
|
||||
@ -397,9 +406,19 @@ static i40e_status i40e_read_nvm_buffer_aq(struct i40e_hw *hw, u16 offset,
|
||||
i40e_status i40e_read_nvm_buffer(struct i40e_hw *hw, u16 offset,
|
||||
u16 *words, u16 *data)
|
||||
{
|
||||
if (hw->mac.type == I40E_MAC_X722)
|
||||
return i40e_read_nvm_buffer_aq(hw, offset, words, data);
|
||||
return i40e_read_nvm_buffer_srctl(hw, offset, words, data);
|
||||
enum i40e_status_code ret_code = 0;
|
||||
|
||||
if (hw->flags & I40E_HW_FLAG_AQ_SRCTL_ACCESS_ENABLE) {
|
||||
ret_code = i40e_acquire_nvm(hw, I40E_RESOURCE_READ);
|
||||
if (!ret_code) {
|
||||
ret_code = i40e_read_nvm_buffer_aq(hw, offset, words,
|
||||
data);
|
||||
i40e_release_nvm(hw);
|
||||
}
|
||||
} else {
|
||||
ret_code = i40e_read_nvm_buffer_srctl(hw, offset, words, data);
|
||||
}
|
||||
return ret_code;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -815,6 +815,8 @@ void i40e_force_wb(struct i40e_vsi *vsi, struct i40e_q_vector *q_vector)
|
||||
* i40e_set_new_dynamic_itr - Find new ITR level
|
||||
* @rc: structure containing ring performance data
|
||||
*
|
||||
* Returns true if ITR changed, false if not
|
||||
*
|
||||
* Stores a new ITR value based on packets and byte counts during
|
||||
* the last interrupt. The advantage of per interrupt computation
|
||||
* is faster updates and more accurate ITR for the current traffic
|
||||
@ -823,21 +825,32 @@ void i40e_force_wb(struct i40e_vsi *vsi, struct i40e_q_vector *q_vector)
|
||||
* testing data as well as attempting to minimize response time
|
||||
* while increasing bulk throughput.
|
||||
**/
|
||||
static void i40e_set_new_dynamic_itr(struct i40e_ring_container *rc)
|
||||
static bool i40e_set_new_dynamic_itr(struct i40e_ring_container *rc)
|
||||
{
|
||||
enum i40e_latency_range new_latency_range = rc->latency_range;
|
||||
struct i40e_q_vector *qv = rc->ring->q_vector;
|
||||
u32 new_itr = rc->itr;
|
||||
int bytes_per_int;
|
||||
int usecs;
|
||||
|
||||
if (rc->total_packets == 0 || !rc->itr)
|
||||
return;
|
||||
return false;
|
||||
|
||||
/* simple throttlerate management
|
||||
* 0-10MB/s lowest (100000 ints/s)
|
||||
* 0-10MB/s lowest (50000 ints/s)
|
||||
* 10-20MB/s low (20000 ints/s)
|
||||
* 20-1249MB/s bulk (8000 ints/s)
|
||||
* 20-1249MB/s bulk (18000 ints/s)
|
||||
* > 40000 Rx packets per second (8000 ints/s)
|
||||
*
|
||||
* The math works out because the divisor is in 10^(-6) which
|
||||
* turns the bytes/us input value into MB/s values, but
|
||||
* make sure to use usecs, as the register values written
|
||||
* are in 2 usec increments in the ITR registers, and make sure
|
||||
* to use the smoothed values that the countdown timer gives us.
|
||||
*/
|
||||
bytes_per_int = rc->total_bytes / rc->itr;
|
||||
usecs = (rc->itr << 1) * ITR_COUNTDOWN_START;
|
||||
bytes_per_int = rc->total_bytes / usecs;
|
||||
|
||||
switch (new_latency_range) {
|
||||
case I40E_LOWEST_LATENCY:
|
||||
if (bytes_per_int > 10)
|
||||
@ -850,35 +863,52 @@ static void i40e_set_new_dynamic_itr(struct i40e_ring_container *rc)
|
||||
new_latency_range = I40E_LOWEST_LATENCY;
|
||||
break;
|
||||
case I40E_BULK_LATENCY:
|
||||
if (bytes_per_int <= 20)
|
||||
new_latency_range = I40E_LOW_LATENCY;
|
||||
break;
|
||||
case I40E_ULTRA_LATENCY:
|
||||
default:
|
||||
if (bytes_per_int <= 20)
|
||||
new_latency_range = I40E_LOW_LATENCY;
|
||||
break;
|
||||
}
|
||||
|
||||
/* this is to adjust RX more aggressively when streaming small
|
||||
* packets. The value of 40000 was picked as it is just beyond
|
||||
* what the hardware can receive per second if in low latency
|
||||
* mode.
|
||||
*/
|
||||
#define RX_ULTRA_PACKET_RATE 40000
|
||||
|
||||
if ((((rc->total_packets * 1000000) / usecs) > RX_ULTRA_PACKET_RATE) &&
|
||||
(&qv->rx == rc))
|
||||
new_latency_range = I40E_ULTRA_LATENCY;
|
||||
|
||||
rc->latency_range = new_latency_range;
|
||||
|
||||
switch (new_latency_range) {
|
||||
case I40E_LOWEST_LATENCY:
|
||||
new_itr = I40E_ITR_100K;
|
||||
new_itr = I40E_ITR_50K;
|
||||
break;
|
||||
case I40E_LOW_LATENCY:
|
||||
new_itr = I40E_ITR_20K;
|
||||
break;
|
||||
case I40E_BULK_LATENCY:
|
||||
new_itr = I40E_ITR_18K;
|
||||
break;
|
||||
case I40E_ULTRA_LATENCY:
|
||||
new_itr = I40E_ITR_8K;
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
|
||||
if (new_itr != rc->itr)
|
||||
rc->itr = new_itr;
|
||||
|
||||
rc->total_bytes = 0;
|
||||
rc->total_packets = 0;
|
||||
|
||||
if (new_itr != rc->itr) {
|
||||
rc->itr = new_itr;
|
||||
return true;
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
/**
|
||||
@ -1747,6 +1777,21 @@ static int i40e_clean_rx_irq_1buf(struct i40e_ring *rx_ring, int budget)
|
||||
return total_rx_packets;
|
||||
}
|
||||
|
||||
static u32 i40e_buildreg_itr(const int type, const u16 itr)
|
||||
{
|
||||
u32 val;
|
||||
|
||||
val = I40E_PFINT_DYN_CTLN_INTENA_MASK |
|
||||
I40E_PFINT_DYN_CTLN_CLEARPBA_MASK |
|
||||
(type << I40E_PFINT_DYN_CTLN_ITR_INDX_SHIFT) |
|
||||
(itr << I40E_PFINT_DYN_CTLN_INTERVAL_SHIFT);
|
||||
|
||||
return val;
|
||||
}
|
||||
|
||||
/* a small macro to shorten up some long lines */
|
||||
#define INTREG I40E_PFINT_DYN_CTLN
|
||||
|
||||
/**
|
||||
* i40e_update_enable_itr - Update itr and re-enable MSIX interrupt
|
||||
* @vsi: the VSI we care about
|
||||
@ -1757,54 +1802,69 @@ static inline void i40e_update_enable_itr(struct i40e_vsi *vsi,
|
||||
struct i40e_q_vector *q_vector)
|
||||
{
|
||||
struct i40e_hw *hw = &vsi->back->hw;
|
||||
u16 old_itr;
|
||||
bool rx = false, tx = false;
|
||||
u32 rxval, txval;
|
||||
int vector;
|
||||
u32 val;
|
||||
|
||||
vector = (q_vector->v_idx + vsi->base_vector);
|
||||
|
||||
/* avoid dynamic calculation if in countdown mode OR if
|
||||
* all dynamic is disabled
|
||||
*/
|
||||
rxval = txval = i40e_buildreg_itr(I40E_ITR_NONE, 0);
|
||||
|
||||
if (q_vector->itr_countdown > 0 ||
|
||||
(!ITR_IS_DYNAMIC(vsi->rx_itr_setting) &&
|
||||
!ITR_IS_DYNAMIC(vsi->tx_itr_setting))) {
|
||||
goto enable_int;
|
||||
}
|
||||
|
||||
if (ITR_IS_DYNAMIC(vsi->rx_itr_setting)) {
|
||||
old_itr = q_vector->rx.itr;
|
||||
i40e_set_new_dynamic_itr(&q_vector->rx);
|
||||
if (old_itr != q_vector->rx.itr) {
|
||||
val = I40E_PFINT_DYN_CTLN_INTENA_MASK |
|
||||
I40E_PFINT_DYN_CTLN_CLEARPBA_MASK |
|
||||
(I40E_RX_ITR <<
|
||||
I40E_PFINT_DYN_CTLN_ITR_INDX_SHIFT) |
|
||||
(q_vector->rx.itr <<
|
||||
I40E_PFINT_DYN_CTLN_INTERVAL_SHIFT);
|
||||
} else {
|
||||
val = I40E_PFINT_DYN_CTLN_INTENA_MASK |
|
||||
I40E_PFINT_DYN_CTLN_CLEARPBA_MASK |
|
||||
(I40E_ITR_NONE <<
|
||||
I40E_PFINT_DYN_CTLN_ITR_INDX_SHIFT);
|
||||
}
|
||||
if (!test_bit(__I40E_DOWN, &vsi->state))
|
||||
wr32(hw, I40E_PFINT_DYN_CTLN(vector - 1), val);
|
||||
} else {
|
||||
i40e_irq_dynamic_enable(vsi, q_vector->v_idx);
|
||||
rx = i40e_set_new_dynamic_itr(&q_vector->rx);
|
||||
rxval = i40e_buildreg_itr(I40E_RX_ITR, q_vector->rx.itr);
|
||||
}
|
||||
|
||||
if (ITR_IS_DYNAMIC(vsi->tx_itr_setting)) {
|
||||
old_itr = q_vector->tx.itr;
|
||||
i40e_set_new_dynamic_itr(&q_vector->tx);
|
||||
if (old_itr != q_vector->tx.itr) {
|
||||
val = I40E_PFINT_DYN_CTLN_INTENA_MASK |
|
||||
I40E_PFINT_DYN_CTLN_CLEARPBA_MASK |
|
||||
(I40E_TX_ITR <<
|
||||
I40E_PFINT_DYN_CTLN_ITR_INDX_SHIFT) |
|
||||
(q_vector->tx.itr <<
|
||||
I40E_PFINT_DYN_CTLN_INTERVAL_SHIFT);
|
||||
} else {
|
||||
val = I40E_PFINT_DYN_CTLN_INTENA_MASK |
|
||||
I40E_PFINT_DYN_CTLN_CLEARPBA_MASK |
|
||||
(I40E_ITR_NONE <<
|
||||
I40E_PFINT_DYN_CTLN_ITR_INDX_SHIFT);
|
||||
tx = i40e_set_new_dynamic_itr(&q_vector->tx);
|
||||
txval = i40e_buildreg_itr(I40E_TX_ITR, q_vector->tx.itr);
|
||||
}
|
||||
|
||||
if (rx || tx) {
|
||||
/* get the higher of the two ITR adjustments and
|
||||
* use the same value for both ITR registers
|
||||
* when in adaptive mode (Rx and/or Tx)
|
||||
*/
|
||||
u16 itr = max(q_vector->tx.itr, q_vector->rx.itr);
|
||||
|
||||
q_vector->tx.itr = q_vector->rx.itr = itr;
|
||||
txval = i40e_buildreg_itr(I40E_TX_ITR, itr);
|
||||
tx = true;
|
||||
rxval = i40e_buildreg_itr(I40E_RX_ITR, itr);
|
||||
rx = true;
|
||||
}
|
||||
|
||||
/* only need to enable the interrupt once, but need
|
||||
* to possibly update both ITR values
|
||||
*/
|
||||
if (rx) {
|
||||
/* set the INTENA_MSK_MASK so that this first write
|
||||
* won't actually enable the interrupt, instead just
|
||||
* updating the ITR (it's bit 31 PF and VF)
|
||||
*/
|
||||
rxval |= BIT(31);
|
||||
/* don't check _DOWN because interrupt isn't being enabled */
|
||||
wr32(hw, INTREG(vector - 1), rxval);
|
||||
}
|
||||
|
||||
enable_int:
|
||||
if (!test_bit(__I40E_DOWN, &vsi->state))
|
||||
wr32(hw, I40E_PFINT_DYN_CTLN(q_vector->v_idx +
|
||||
vsi->base_vector - 1), val);
|
||||
} else {
|
||||
i40e_irq_dynamic_enable(vsi, q_vector->v_idx);
|
||||
}
|
||||
wr32(hw, INTREG(vector - 1), txval);
|
||||
|
||||
if (q_vector->itr_countdown)
|
||||
q_vector->itr_countdown--;
|
||||
else
|
||||
q_vector->itr_countdown = ITR_COUNTDOWN_START;
|
||||
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -32,12 +32,14 @@
|
||||
#define I40E_MAX_ITR 0x0FF0 /* reg uses 2 usec resolution */
|
||||
#define I40E_MIN_ITR 0x0001 /* reg uses 2 usec resolution */
|
||||
#define I40E_ITR_100K 0x0005
|
||||
#define I40E_ITR_50K 0x000A
|
||||
#define I40E_ITR_20K 0x0019
|
||||
#define I40E_ITR_18K 0x001B
|
||||
#define I40E_ITR_8K 0x003E
|
||||
#define I40E_ITR_4K 0x007A
|
||||
#define I40E_MAX_INTRL 0x3B /* reg uses 4 usec resolution */
|
||||
#define I40E_ITR_RX_DEF I40E_ITR_8K
|
||||
#define I40E_ITR_TX_DEF I40E_ITR_4K
|
||||
#define I40E_ITR_RX_DEF I40E_ITR_20K
|
||||
#define I40E_ITR_TX_DEF I40E_ITR_20K
|
||||
#define I40E_ITR_DYNAMIC 0x8000 /* use top bit as a flag */
|
||||
#define I40E_MIN_INT_RATE 250 /* ~= 1000000 / (I40E_MAX_ITR * 2) */
|
||||
#define I40E_MAX_INT_RATE 500000 /* == 1000000 / (I40E_MIN_ITR * 2) */
|
||||
@ -296,6 +298,7 @@ enum i40e_latency_range {
|
||||
I40E_LOWEST_LATENCY = 0,
|
||||
I40E_LOW_LATENCY = 1,
|
||||
I40E_BULK_LATENCY = 2,
|
||||
I40E_ULTRA_LATENCY = 3,
|
||||
};
|
||||
|
||||
struct i40e_ring_container {
|
||||
|
@ -544,6 +544,9 @@ struct i40e_hw {
|
||||
struct i40e_dcbx_config remote_dcbx_config; /* Peer Cfg */
|
||||
struct i40e_dcbx_config desired_dcbx_config; /* CEE Desired Cfg */
|
||||
|
||||
#define I40E_HW_FLAG_AQ_SRCTL_ACCESS_ENABLE BIT_ULL(0)
|
||||
u64 flags;
|
||||
|
||||
/* debug mask */
|
||||
u32 debug_mask;
|
||||
char err_str[16];
|
||||
@ -1065,8 +1068,8 @@ enum i40e_filter_program_desc_fd_status {
|
||||
};
|
||||
|
||||
#define I40E_TXD_FLTR_QW0_DEST_VSI_SHIFT 23
|
||||
#define I40E_TXD_FLTR_QW0_DEST_VSI_MASK \
|
||||
BIT_ULL(I40E_TXD_FLTR_QW0_DEST_VSI_SHIFT)
|
||||
#define I40E_TXD_FLTR_QW0_DEST_VSI_MASK (0x1FFUL << \
|
||||
I40E_TXD_FLTR_QW0_DEST_VSI_SHIFT)
|
||||
|
||||
#define I40E_TXD_FLTR_QW1_CMD_SHIFT 4
|
||||
#define I40E_TXD_FLTR_QW1_CMD_MASK (0xFFFFULL << \
|
||||
|
@ -547,6 +547,8 @@ static int i40e_alloc_vsi_res(struct i40e_vf *vf, enum i40e_vsi_type type)
|
||||
*/
|
||||
if (vf->port_vlan_id)
|
||||
i40e_vsi_add_pvid(vsi, vf->port_vlan_id);
|
||||
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
f = i40e_add_filter(vsi, vf->default_lan_addr.addr,
|
||||
vf->port_vlan_id ? vf->port_vlan_id : -1,
|
||||
true, false);
|
||||
@ -559,6 +561,7 @@ static int i40e_alloc_vsi_res(struct i40e_vf *vf, enum i40e_vsi_type type)
|
||||
if (!f)
|
||||
dev_info(&pf->pdev->dev,
|
||||
"Could not allocate VF broadcast filter\n");
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
}
|
||||
|
||||
/* program mac filter */
|
||||
@ -1598,6 +1601,11 @@ static int i40e_vc_add_mac_addr_msg(struct i40e_vf *vf, u8 *msg, u16 msglen)
|
||||
}
|
||||
vsi = pf->vsi[vf->lan_vsi_idx];
|
||||
|
||||
/* Lock once, because all function inside for loop accesses VSI's
|
||||
* MAC filter list which needs to be protected using same lock.
|
||||
*/
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
/* add new addresses to the list */
|
||||
for (i = 0; i < al->num_elements; i++) {
|
||||
struct i40e_mac_filter *f;
|
||||
@ -1616,9 +1624,11 @@ static int i40e_vc_add_mac_addr_msg(struct i40e_vf *vf, u8 *msg, u16 msglen)
|
||||
dev_err(&pf->pdev->dev,
|
||||
"Unable to add VF MAC filter\n");
|
||||
ret = I40E_ERR_PARAM;
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
goto error_param;
|
||||
}
|
||||
}
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
/* program the updated filter list */
|
||||
if (i40e_sync_vsi_filters(vsi, false))
|
||||
@ -1666,10 +1676,12 @@ static int i40e_vc_del_mac_addr_msg(struct i40e_vf *vf, u8 *msg, u16 msglen)
|
||||
}
|
||||
vsi = pf->vsi[vf->lan_vsi_idx];
|
||||
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
/* delete addresses from the list */
|
||||
for (i = 0; i < al->num_elements; i++)
|
||||
i40e_del_filter(vsi, al->list[i].addr,
|
||||
I40E_VLAN_ANY, true, false);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
/* program the updated filter list */
|
||||
if (i40e_sync_vsi_filters(vsi, false))
|
||||
@ -2066,6 +2078,11 @@ int i40e_ndo_set_vf_mac(struct net_device *netdev, int vf_id, u8 *mac)
|
||||
goto error_param;
|
||||
}
|
||||
|
||||
/* Lock once because below invoked function add/del_filter requires
|
||||
* mac_filter_list_lock to be held
|
||||
*/
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
/* delete the temporary mac address */
|
||||
i40e_del_filter(vsi, vf->default_lan_addr.addr,
|
||||
vf->port_vlan_id ? vf->port_vlan_id : -1,
|
||||
@ -2077,6 +2094,8 @@ int i40e_ndo_set_vf_mac(struct net_device *netdev, int vf_id, u8 *mac)
|
||||
list_for_each_entry(f, &vsi->mac_filter_list, list)
|
||||
i40e_del_filter(vsi, f->macaddr, f->vlan, true, false);
|
||||
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
dev_info(&pf->pdev->dev, "Setting MAC %pM on VF %d\n", mac, vf_id);
|
||||
/* program mac filter */
|
||||
if (i40e_sync_vsi_filters(vsi, false)) {
|
||||
@ -2109,6 +2128,7 @@ int i40e_ndo_set_vf_port_vlan(struct net_device *netdev,
|
||||
u16 vlanprio = vlan_id | (qos << I40E_VLAN_PRIORITY_SHIFT);
|
||||
struct i40e_netdev_priv *np = netdev_priv(netdev);
|
||||
struct i40e_pf *pf = np->vsi->back;
|
||||
bool is_vsi_in_vlan = false;
|
||||
struct i40e_vsi *vsi;
|
||||
struct i40e_vf *vf;
|
||||
int ret = 0;
|
||||
@ -2138,7 +2158,11 @@ int i40e_ndo_set_vf_port_vlan(struct net_device *netdev,
|
||||
/* duplicate request, so just return success */
|
||||
goto error_pvid;
|
||||
|
||||
if (le16_to_cpu(vsi->info.pvid) == 0 && i40e_is_vsi_in_vlan(vsi)) {
|
||||
spin_lock_bh(&vsi->mac_filter_list_lock);
|
||||
is_vsi_in_vlan = i40e_is_vsi_in_vlan(vsi);
|
||||
spin_unlock_bh(&vsi->mac_filter_list_lock);
|
||||
|
||||
if (le16_to_cpu(vsi->info.pvid) == 0 && is_vsi_in_vlan) {
|
||||
dev_err(&pf->pdev->dev,
|
||||
"VF %d has already configured VLAN filters and the administrator is requesting a port VLAN override.\nPlease unload and reload the VF driver for this change to take effect.\n",
|
||||
vf_id);
|
||||
|
@ -443,9 +443,6 @@ static i40e_status i40e_aq_get_set_rss_lut(struct i40e_hw *hw,
|
||||
I40E_AQC_SET_RSS_LUT_TABLE_TYPE_SHIFT) &
|
||||
I40E_AQC_SET_RSS_LUT_TABLE_TYPE_MASK));
|
||||
|
||||
cmd_resp->addr_high = cpu_to_le32(high_16_bits((u64)lut));
|
||||
cmd_resp->addr_low = cpu_to_le32(lower_32_bits((u64)lut));
|
||||
|
||||
status = i40evf_asq_send_command(hw, &desc, lut, lut_size, NULL);
|
||||
|
||||
return status;
|
||||
@ -520,8 +517,6 @@ static i40e_status i40e_aq_get_set_rss_key(struct i40e_hw *hw,
|
||||
I40E_AQC_SET_RSS_KEY_VSI_ID_SHIFT) &
|
||||
I40E_AQC_SET_RSS_KEY_VSI_ID_MASK));
|
||||
cmd_resp->vsi_id |= cpu_to_le16((u16)I40E_AQC_SET_RSS_KEY_VSI_VALID);
|
||||
cmd_resp->addr_high = cpu_to_le32(high_16_bits((u64)key));
|
||||
cmd_resp->addr_low = cpu_to_le32(lower_32_bits((u64)key));
|
||||
|
||||
status = i40evf_asq_send_command(hw, &desc, key, key_size, NULL);
|
||||
|
||||
|
@ -318,6 +318,8 @@ static void i40evf_force_wb(struct i40e_vsi *vsi, struct i40e_q_vector *q_vector
|
||||
* i40e_set_new_dynamic_itr - Find new ITR level
|
||||
* @rc: structure containing ring performance data
|
||||
*
|
||||
* Returns true if ITR changed, false if not
|
||||
*
|
||||
* Stores a new ITR value based on packets and byte counts during
|
||||
* the last interrupt. The advantage of per interrupt computation
|
||||
* is faster updates and more accurate ITR for the current traffic
|
||||
@ -326,21 +328,32 @@ static void i40evf_force_wb(struct i40e_vsi *vsi, struct i40e_q_vector *q_vector
|
||||
* testing data as well as attempting to minimize response time
|
||||
* while increasing bulk throughput.
|
||||
**/
|
||||
static void i40e_set_new_dynamic_itr(struct i40e_ring_container *rc)
|
||||
static bool i40e_set_new_dynamic_itr(struct i40e_ring_container *rc)
|
||||
{
|
||||
enum i40e_latency_range new_latency_range = rc->latency_range;
|
||||
struct i40e_q_vector *qv = rc->ring->q_vector;
|
||||
u32 new_itr = rc->itr;
|
||||
int bytes_per_int;
|
||||
int usecs;
|
||||
|
||||
if (rc->total_packets == 0 || !rc->itr)
|
||||
return;
|
||||
return false;
|
||||
|
||||
/* simple throttlerate management
|
||||
* 0-10MB/s lowest (100000 ints/s)
|
||||
* 0-10MB/s lowest (50000 ints/s)
|
||||
* 10-20MB/s low (20000 ints/s)
|
||||
* 20-1249MB/s bulk (8000 ints/s)
|
||||
* 20-1249MB/s bulk (18000 ints/s)
|
||||
* > 40000 Rx packets per second (8000 ints/s)
|
||||
*
|
||||
* The math works out because the divisor is in 10^(-6) which
|
||||
* turns the bytes/us input value into MB/s values, but
|
||||
* make sure to use usecs, as the register values written
|
||||
* are in 2 usec increments in the ITR registers, and make sure
|
||||
* to use the smoothed values that the countdown timer gives us.
|
||||
*/
|
||||
bytes_per_int = rc->total_bytes / rc->itr;
|
||||
usecs = (rc->itr << 1) * ITR_COUNTDOWN_START;
|
||||
bytes_per_int = rc->total_bytes / usecs;
|
||||
|
||||
switch (new_latency_range) {
|
||||
case I40E_LOWEST_LATENCY:
|
||||
if (bytes_per_int > 10)
|
||||
@ -353,35 +366,52 @@ static void i40e_set_new_dynamic_itr(struct i40e_ring_container *rc)
|
||||
new_latency_range = I40E_LOWEST_LATENCY;
|
||||
break;
|
||||
case I40E_BULK_LATENCY:
|
||||
if (bytes_per_int <= 20)
|
||||
new_latency_range = I40E_LOW_LATENCY;
|
||||
break;
|
||||
case I40E_ULTRA_LATENCY:
|
||||
default:
|
||||
if (bytes_per_int <= 20)
|
||||
new_latency_range = I40E_LOW_LATENCY;
|
||||
break;
|
||||
}
|
||||
|
||||
/* this is to adjust RX more aggressively when streaming small
|
||||
* packets. The value of 40000 was picked as it is just beyond
|
||||
* what the hardware can receive per second if in low latency
|
||||
* mode.
|
||||
*/
|
||||
#define RX_ULTRA_PACKET_RATE 40000
|
||||
|
||||
if ((((rc->total_packets * 1000000) / usecs) > RX_ULTRA_PACKET_RATE) &&
|
||||
(&qv->rx == rc))
|
||||
new_latency_range = I40E_ULTRA_LATENCY;
|
||||
|
||||
rc->latency_range = new_latency_range;
|
||||
|
||||
switch (new_latency_range) {
|
||||
case I40E_LOWEST_LATENCY:
|
||||
new_itr = I40E_ITR_100K;
|
||||
new_itr = I40E_ITR_50K;
|
||||
break;
|
||||
case I40E_LOW_LATENCY:
|
||||
new_itr = I40E_ITR_20K;
|
||||
break;
|
||||
case I40E_BULK_LATENCY:
|
||||
new_itr = I40E_ITR_18K;
|
||||
break;
|
||||
case I40E_ULTRA_LATENCY:
|
||||
new_itr = I40E_ITR_8K;
|
||||
break;
|
||||
default:
|
||||
break;
|
||||
}
|
||||
|
||||
if (new_itr != rc->itr)
|
||||
rc->itr = new_itr;
|
||||
|
||||
rc->total_bytes = 0;
|
||||
rc->total_packets = 0;
|
||||
|
||||
if (new_itr != rc->itr) {
|
||||
rc->itr = new_itr;
|
||||
return true;
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
/*
|
||||
@ -1187,6 +1217,21 @@ static int i40e_clean_rx_irq_1buf(struct i40e_ring *rx_ring, int budget)
|
||||
return total_rx_packets;
|
||||
}
|
||||
|
||||
static u32 i40e_buildreg_itr(const int type, const u16 itr)
|
||||
{
|
||||
u32 val;
|
||||
|
||||
val = I40E_VFINT_DYN_CTLN1_INTENA_MASK |
|
||||
I40E_VFINT_DYN_CTLN1_CLEARPBA_MASK |
|
||||
(type << I40E_VFINT_DYN_CTLN1_ITR_INDX_SHIFT) |
|
||||
(itr << I40E_VFINT_DYN_CTLN1_INTERVAL_SHIFT);
|
||||
|
||||
return val;
|
||||
}
|
||||
|
||||
/* a small macro to shorten up some long lines */
|
||||
#define INTREG I40E_VFINT_DYN_CTLN1
|
||||
|
||||
/**
|
||||
* i40e_update_enable_itr - Update itr and re-enable MSIX interrupt
|
||||
* @vsi: the VSI we care about
|
||||
@ -1197,55 +1242,67 @@ static inline void i40e_update_enable_itr(struct i40e_vsi *vsi,
|
||||
struct i40e_q_vector *q_vector)
|
||||
{
|
||||
struct i40e_hw *hw = &vsi->back->hw;
|
||||
u16 old_itr;
|
||||
bool rx = false, tx = false;
|
||||
u32 rxval, txval;
|
||||
int vector;
|
||||
u32 val;
|
||||
|
||||
vector = (q_vector->v_idx + vsi->base_vector);
|
||||
if (ITR_IS_DYNAMIC(vsi->rx_itr_setting)) {
|
||||
old_itr = q_vector->rx.itr;
|
||||
i40e_set_new_dynamic_itr(&q_vector->rx);
|
||||
if (old_itr != q_vector->rx.itr) {
|
||||
val = I40E_VFINT_DYN_CTLN1_INTENA_MASK |
|
||||
I40E_VFINT_DYN_CTLN1_CLEARPBA_MASK |
|
||||
(I40E_RX_ITR <<
|
||||
I40E_VFINT_DYN_CTLN1_ITR_INDX_SHIFT) |
|
||||
(q_vector->rx.itr <<
|
||||
I40E_VFINT_DYN_CTLN1_INTERVAL_SHIFT);
|
||||
} else {
|
||||
val = I40E_VFINT_DYN_CTLN1_INTENA_MASK |
|
||||
I40E_VFINT_DYN_CTLN1_CLEARPBA_MASK |
|
||||
(I40E_ITR_NONE <<
|
||||
I40E_VFINT_DYN_CTLN1_ITR_INDX_SHIFT);
|
||||
|
||||
/* avoid dynamic calculation if in countdown mode OR if
|
||||
* all dynamic is disabled
|
||||
*/
|
||||
rxval = txval = i40e_buildreg_itr(I40E_ITR_NONE, 0);
|
||||
|
||||
if (q_vector->itr_countdown > 0 ||
|
||||
(!ITR_IS_DYNAMIC(vsi->rx_itr_setting) &&
|
||||
!ITR_IS_DYNAMIC(vsi->tx_itr_setting))) {
|
||||
goto enable_int;
|
||||
}
|
||||
if (!test_bit(__I40E_DOWN, &vsi->state))
|
||||
wr32(hw, I40E_VFINT_DYN_CTLN1(vector - 1), val);
|
||||
} else {
|
||||
i40evf_irq_enable_queues(vsi->back, 1
|
||||
<< q_vector->v_idx);
|
||||
|
||||
if (ITR_IS_DYNAMIC(vsi->rx_itr_setting)) {
|
||||
rx = i40e_set_new_dynamic_itr(&q_vector->rx);
|
||||
rxval = i40e_buildreg_itr(I40E_RX_ITR, q_vector->rx.itr);
|
||||
}
|
||||
if (ITR_IS_DYNAMIC(vsi->tx_itr_setting)) {
|
||||
old_itr = q_vector->tx.itr;
|
||||
i40e_set_new_dynamic_itr(&q_vector->tx);
|
||||
if (old_itr != q_vector->tx.itr) {
|
||||
val = I40E_VFINT_DYN_CTLN1_INTENA_MASK |
|
||||
I40E_VFINT_DYN_CTLN1_CLEARPBA_MASK |
|
||||
(I40E_TX_ITR <<
|
||||
I40E_VFINT_DYN_CTLN1_ITR_INDX_SHIFT) |
|
||||
(q_vector->tx.itr <<
|
||||
I40E_VFINT_DYN_CTLN1_INTERVAL_SHIFT);
|
||||
tx = i40e_set_new_dynamic_itr(&q_vector->tx);
|
||||
txval = i40e_buildreg_itr(I40E_TX_ITR, q_vector->tx.itr);
|
||||
}
|
||||
if (rx || tx) {
|
||||
/* get the higher of the two ITR adjustments and
|
||||
* use the same value for both ITR registers
|
||||
* when in adaptive mode (Rx and/or Tx)
|
||||
*/
|
||||
u16 itr = max(q_vector->tx.itr, q_vector->rx.itr);
|
||||
|
||||
} else {
|
||||
val = I40E_VFINT_DYN_CTLN1_INTENA_MASK |
|
||||
I40E_VFINT_DYN_CTLN1_CLEARPBA_MASK |
|
||||
(I40E_ITR_NONE <<
|
||||
I40E_VFINT_DYN_CTLN1_ITR_INDX_SHIFT);
|
||||
q_vector->tx.itr = q_vector->rx.itr = itr;
|
||||
txval = i40e_buildreg_itr(I40E_TX_ITR, itr);
|
||||
tx = true;
|
||||
rxval = i40e_buildreg_itr(I40E_RX_ITR, itr);
|
||||
rx = true;
|
||||
}
|
||||
|
||||
/* only need to enable the interrupt once, but need
|
||||
* to possibly update both ITR values
|
||||
*/
|
||||
if (rx) {
|
||||
/* set the INTENA_MSK_MASK so that this first write
|
||||
* won't actually enable the interrupt, instead just
|
||||
* updating the ITR (it's bit 31 PF and VF)
|
||||
*/
|
||||
rxval |= BIT(31);
|
||||
/* don't check _DOWN because interrupt isn't being enabled */
|
||||
wr32(hw, INTREG(vector - 1), rxval);
|
||||
}
|
||||
|
||||
enable_int:
|
||||
if (!test_bit(__I40E_DOWN, &vsi->state))
|
||||
wr32(hw, I40E_VFINT_DYN_CTLN1(vector - 1), val);
|
||||
} else {
|
||||
i40evf_irq_enable_queues(vsi->back, BIT(q_vector->v_idx));
|
||||
}
|
||||
wr32(hw, INTREG(vector - 1), txval);
|
||||
|
||||
if (q_vector->itr_countdown)
|
||||
q_vector->itr_countdown--;
|
||||
else
|
||||
q_vector->itr_countdown = ITR_COUNTDOWN_START;
|
||||
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -32,12 +32,14 @@
|
||||
#define I40E_MAX_ITR 0x0FF0 /* reg uses 2 usec resolution */
|
||||
#define I40E_MIN_ITR 0x0001 /* reg uses 2 usec resolution */
|
||||
#define I40E_ITR_100K 0x0005
|
||||
#define I40E_ITR_50K 0x000A
|
||||
#define I40E_ITR_20K 0x0019
|
||||
#define I40E_ITR_18K 0x001B
|
||||
#define I40E_ITR_8K 0x003E
|
||||
#define I40E_ITR_4K 0x007A
|
||||
#define I40E_MAX_INTRL 0x3B /* reg uses 4 usec resolution */
|
||||
#define I40E_ITR_RX_DEF I40E_ITR_8K
|
||||
#define I40E_ITR_TX_DEF I40E_ITR_4K
|
||||
#define I40E_ITR_RX_DEF I40E_ITR_20K
|
||||
#define I40E_ITR_TX_DEF I40E_ITR_20K
|
||||
#define I40E_ITR_DYNAMIC 0x8000 /* use top bit as a flag */
|
||||
#define I40E_MIN_INT_RATE 250 /* ~= 1000000 / (I40E_MAX_ITR * 2) */
|
||||
#define I40E_MAX_INT_RATE 500000 /* == 1000000 / (I40E_MIN_ITR * 2) */
|
||||
@ -89,12 +91,12 @@ enum i40e_dyn_idx_t {
|
||||
BIT_ULL(I40E_FILTER_PCTYPE_L2_PAYLOAD))
|
||||
|
||||
#define I40E_DEFAULT_RSS_HENA_EXPANDED (I40E_DEFAULT_RSS_HENA | \
|
||||
BIT(I40E_FILTER_PCTYPE_NONF_IPV4_TCP_SYN_NO_ACK) | \
|
||||
BIT(I40E_FILTER_PCTYPE_NONF_UNICAST_IPV4_UDP) | \
|
||||
BIT(I40E_FILTER_PCTYPE_NONF_MULTICAST_IPV4_UDP) | \
|
||||
BIT(I40E_FILTER_PCTYPE_NONF_IPV6_TCP_SYN_NO_ACK) | \
|
||||
BIT(I40E_FILTER_PCTYPE_NONF_UNICAST_IPV6_UDP) | \
|
||||
BIT(I40E_FILTER_PCTYPE_NONF_MULTICAST_IPV6_UDP))
|
||||
BIT_ULL(I40E_FILTER_PCTYPE_NONF_IPV4_TCP_SYN_NO_ACK) | \
|
||||
BIT_ULL(I40E_FILTER_PCTYPE_NONF_UNICAST_IPV4_UDP) | \
|
||||
BIT_ULL(I40E_FILTER_PCTYPE_NONF_MULTICAST_IPV4_UDP) | \
|
||||
BIT_ULL(I40E_FILTER_PCTYPE_NONF_IPV6_TCP_SYN_NO_ACK) | \
|
||||
BIT_ULL(I40E_FILTER_PCTYPE_NONF_UNICAST_IPV6_UDP) | \
|
||||
BIT_ULL(I40E_FILTER_PCTYPE_NONF_MULTICAST_IPV6_UDP))
|
||||
|
||||
#define i40e_pf_get_default_rss_hena(pf) \
|
||||
(((pf)->flags & I40E_FLAG_MULTIPLE_TCP_UDP_RSS_PCTYPE) ? \
|
||||
@ -291,6 +293,7 @@ enum i40e_latency_range {
|
||||
I40E_LOWEST_LATENCY = 0,
|
||||
I40E_LOW_LATENCY = 1,
|
||||
I40E_BULK_LATENCY = 2,
|
||||
I40E_ULTRA_LATENCY = 3,
|
||||
};
|
||||
|
||||
struct i40e_ring_container {
|
||||
|
@ -1055,8 +1055,8 @@ enum i40e_filter_program_desc_fd_status {
|
||||
};
|
||||
|
||||
#define I40E_TXD_FLTR_QW0_DEST_VSI_SHIFT 23
|
||||
#define I40E_TXD_FLTR_QW0_DEST_VSI_MASK \
|
||||
BIT_ULL(I40E_TXD_FLTR_QW0_DEST_VSI_SHIFT)
|
||||
#define I40E_TXD_FLTR_QW0_DEST_VSI_MASK (0x1FFUL << \
|
||||
I40E_TXD_FLTR_QW0_DEST_VSI_SHIFT)
|
||||
|
||||
#define I40E_TXD_FLTR_QW1_CMD_SHIFT 4
|
||||
#define I40E_TXD_FLTR_QW1_CMD_MASK (0xFFFFULL << \
|
||||
|
@ -112,6 +112,8 @@ struct i40e_q_vector {
|
||||
struct i40e_ring_container tx;
|
||||
u32 ring_mask;
|
||||
u8 num_ringpairs; /* total number of ring pairs in vector */
|
||||
#define ITR_COUNTDOWN_START 100
|
||||
u8 itr_countdown; /* when 0 or 1 update ITR */
|
||||
int v_idx; /* vector index in list */
|
||||
char name[IFNAMSIZ + 9];
|
||||
bool arm_wb_state;
|
||||
|
@ -34,7 +34,7 @@ char i40evf_driver_name[] = "i40evf";
|
||||
static const char i40evf_driver_string[] =
|
||||
"Intel(R) XL710/X710 Virtual Function Network Driver";
|
||||
|
||||
#define DRV_VERSION "1.3.21"
|
||||
#define DRV_VERSION "1.3.25"
|
||||
const char i40evf_driver_version[] = DRV_VERSION;
|
||||
static const char i40evf_copyright[] =
|
||||
"Copyright (c) 2013 - 2015 Intel Corporation.";
|
||||
@ -334,7 +334,7 @@ static irqreturn_t i40evf_msix_clean_rings(int irq, void *data)
|
||||
if (!q_vector->tx.ring && !q_vector->rx.ring)
|
||||
return IRQ_HANDLED;
|
||||
|
||||
napi_schedule(&q_vector->napi);
|
||||
napi_schedule_irqoff(&q_vector->napi);
|
||||
|
||||
return IRQ_HANDLED;
|
||||
}
|
||||
@ -357,6 +357,7 @@ i40evf_map_vector_to_rxq(struct i40evf_adapter *adapter, int v_idx, int r_idx)
|
||||
q_vector->rx.ring = rx_ring;
|
||||
q_vector->rx.count++;
|
||||
q_vector->rx.latency_range = I40E_LOW_LATENCY;
|
||||
q_vector->itr_countdown = ITR_COUNTDOWN_START;
|
||||
}
|
||||
|
||||
/**
|
||||
@ -377,6 +378,7 @@ i40evf_map_vector_to_txq(struct i40evf_adapter *adapter, int v_idx, int t_idx)
|
||||
q_vector->tx.ring = tx_ring;
|
||||
q_vector->tx.count++;
|
||||
q_vector->tx.latency_range = I40E_LOW_LATENCY;
|
||||
q_vector->itr_countdown = ITR_COUNTDOWN_START;
|
||||
q_vector->num_ringpairs++;
|
||||
q_vector->ring_mask |= BIT(t_idx);
|
||||
}
|
||||
@ -2357,9 +2359,12 @@ static void i40evf_init_task(struct work_struct *work)
|
||||
err:
|
||||
/* Things went into the weeds, so try again later */
|
||||
if (++adapter->aq_wait_count > I40EVF_AQ_MAX_ERR) {
|
||||
dev_err(&pdev->dev, "Failed to communicate with PF; giving up\n");
|
||||
dev_err(&pdev->dev, "Failed to communicate with PF; waiting before retry\n");
|
||||
adapter->flags |= I40EVF_FLAG_PF_COMMS_FAILED;
|
||||
return; /* do not reschedule */
|
||||
i40evf_shutdown_adminq(hw);
|
||||
adapter->state = __I40EVF_STARTUP;
|
||||
schedule_delayed_work(&adapter->init_task, HZ * 5);
|
||||
return;
|
||||
}
|
||||
schedule_delayed_work(&adapter->init_task, HZ);
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user