mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-25 23:25:37 +07:00
523f9eb1ef
Health buffer address is a 32 bit PCI address offset provided by the FW. This offset is used for reading FW health debug data located on the shared CR space. Cr space is accessible in both driver and FW and allows for different queries and configurations. Health buffer size is always 64B of readable data followed by a lock which is used to block volatile CR space access. Signed-off-by: Alex Vesker <valex@mellanox.com> Signed-off-by: Tariq Toukan <tariqt@mellanox.com> Signed-off-by: Jiri Pirko <jiri@mellanox.com> Signed-off-by: David S. Miller <davem@davemloft.net>
3089 lines
100 KiB
C
3089 lines
100 KiB
C
/*
|
|
* Copyright (c) 2004, 2005 Topspin Communications. All rights reserved.
|
|
* Copyright (c) 2005, 2006, 2007, 2008 Mellanox Technologies. All rights reserved.
|
|
* Copyright (c) 2005, 2006, 2007 Cisco Systems, Inc. All rights reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the
|
|
* OpenIB.org BSD license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or
|
|
* without modification, are permitted provided that the following
|
|
* conditions are met:
|
|
*
|
|
* - Redistributions of source code must retain the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer.
|
|
*
|
|
* - Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*/
|
|
|
|
#include <linux/etherdevice.h>
|
|
#include <linux/mlx4/cmd.h>
|
|
#include <linux/module.h>
|
|
#include <linux/cache.h>
|
|
#include <linux/kernel.h>
|
|
#include <uapi/rdma/mlx4-abi.h>
|
|
|
|
#include "fw.h"
|
|
#include "icm.h"
|
|
|
|
enum {
|
|
MLX4_COMMAND_INTERFACE_MIN_REV = 2,
|
|
MLX4_COMMAND_INTERFACE_MAX_REV = 3,
|
|
MLX4_COMMAND_INTERFACE_NEW_PORT_CMDS = 3,
|
|
};
|
|
|
|
extern void __buggy_use_of_MLX4_GET(void);
|
|
extern void __buggy_use_of_MLX4_PUT(void);
|
|
|
|
static bool enable_qos;
|
|
module_param(enable_qos, bool, 0444);
|
|
MODULE_PARM_DESC(enable_qos, "Enable Enhanced QoS support (default: off)");
|
|
|
|
#define MLX4_GET(dest, source, offset) \
|
|
do { \
|
|
void *__p = (char *) (source) + (offset); \
|
|
__be64 val; \
|
|
switch (sizeof(dest)) { \
|
|
case 1: (dest) = *(u8 *) __p; break; \
|
|
case 2: (dest) = be16_to_cpup(__p); break; \
|
|
case 4: (dest) = be32_to_cpup(__p); break; \
|
|
case 8: val = get_unaligned((__be64 *)__p); \
|
|
(dest) = be64_to_cpu(val); break; \
|
|
default: __buggy_use_of_MLX4_GET(); \
|
|
} \
|
|
} while (0)
|
|
|
|
#define MLX4_PUT(dest, source, offset) \
|
|
do { \
|
|
void *__d = ((char *) (dest) + (offset)); \
|
|
switch (sizeof(source)) { \
|
|
case 1: *(u8 *) __d = (source); break; \
|
|
case 2: *(__be16 *) __d = cpu_to_be16(source); break; \
|
|
case 4: *(__be32 *) __d = cpu_to_be32(source); break; \
|
|
case 8: *(__be64 *) __d = cpu_to_be64(source); break; \
|
|
default: __buggy_use_of_MLX4_PUT(); \
|
|
} \
|
|
} while (0)
|
|
|
|
static void dump_dev_cap_flags(struct mlx4_dev *dev, u64 flags)
|
|
{
|
|
static const char *fname[] = {
|
|
[ 0] = "RC transport",
|
|
[ 1] = "UC transport",
|
|
[ 2] = "UD transport",
|
|
[ 3] = "XRC transport",
|
|
[ 6] = "SRQ support",
|
|
[ 7] = "IPoIB checksum offload",
|
|
[ 8] = "P_Key violation counter",
|
|
[ 9] = "Q_Key violation counter",
|
|
[12] = "Dual Port Different Protocol (DPDP) support",
|
|
[15] = "Big LSO headers",
|
|
[16] = "MW support",
|
|
[17] = "APM support",
|
|
[18] = "Atomic ops support",
|
|
[19] = "Raw multicast support",
|
|
[20] = "Address vector port checking support",
|
|
[21] = "UD multicast support",
|
|
[30] = "IBoE support",
|
|
[32] = "Unicast loopback support",
|
|
[34] = "FCS header control",
|
|
[37] = "Wake On LAN (port1) support",
|
|
[38] = "Wake On LAN (port2) support",
|
|
[40] = "UDP RSS support",
|
|
[41] = "Unicast VEP steering support",
|
|
[42] = "Multicast VEP steering support",
|
|
[48] = "Counters support",
|
|
[52] = "RSS IP fragments support",
|
|
[53] = "Port ETS Scheduler support",
|
|
[55] = "Port link type sensing support",
|
|
[59] = "Port management change event support",
|
|
[61] = "64 byte EQE support",
|
|
[62] = "64 byte CQE support",
|
|
};
|
|
int i;
|
|
|
|
mlx4_dbg(dev, "DEV_CAP flags:\n");
|
|
for (i = 0; i < ARRAY_SIZE(fname); ++i)
|
|
if (fname[i] && (flags & (1LL << i)))
|
|
mlx4_dbg(dev, " %s\n", fname[i]);
|
|
}
|
|
|
|
static void dump_dev_cap_flags2(struct mlx4_dev *dev, u64 flags)
|
|
{
|
|
static const char * const fname[] = {
|
|
[0] = "RSS support",
|
|
[1] = "RSS Toeplitz Hash Function support",
|
|
[2] = "RSS XOR Hash Function support",
|
|
[3] = "Device managed flow steering support",
|
|
[4] = "Automatic MAC reassignment support",
|
|
[5] = "Time stamping support",
|
|
[6] = "VST (control vlan insertion/stripping) support",
|
|
[7] = "FSM (MAC anti-spoofing) support",
|
|
[8] = "Dynamic QP updates support",
|
|
[9] = "Device managed flow steering IPoIB support",
|
|
[10] = "TCP/IP offloads/flow-steering for VXLAN support",
|
|
[11] = "MAD DEMUX (Secure-Host) support",
|
|
[12] = "Large cache line (>64B) CQE stride support",
|
|
[13] = "Large cache line (>64B) EQE stride support",
|
|
[14] = "Ethernet protocol control support",
|
|
[15] = "Ethernet Backplane autoneg support",
|
|
[16] = "CONFIG DEV support",
|
|
[17] = "Asymmetric EQs support",
|
|
[18] = "More than 80 VFs support",
|
|
[19] = "Performance optimized for limited rule configuration flow steering support",
|
|
[20] = "Recoverable error events support",
|
|
[21] = "Port Remap support",
|
|
[22] = "QCN support",
|
|
[23] = "QP rate limiting support",
|
|
[24] = "Ethernet Flow control statistics support",
|
|
[25] = "Granular QoS per VF support",
|
|
[26] = "Port ETS Scheduler support",
|
|
[27] = "Port beacon support",
|
|
[28] = "RX-ALL support",
|
|
[29] = "802.1ad offload support",
|
|
[31] = "Modifying loopback source checks using UPDATE_QP support",
|
|
[32] = "Loopback source checks support",
|
|
[33] = "RoCEv2 support",
|
|
[34] = "DMFS Sniffer support (UC & MC)",
|
|
[35] = "Diag counters per port",
|
|
[36] = "QinQ VST mode support",
|
|
[37] = "sl to vl mapping table change event support",
|
|
[38] = "user MAC support",
|
|
[39] = "Report driver version to FW support",
|
|
};
|
|
int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(fname); ++i)
|
|
if (fname[i] && (flags & (1LL << i)))
|
|
mlx4_dbg(dev, " %s\n", fname[i]);
|
|
}
|
|
|
|
int mlx4_MOD_STAT_CFG(struct mlx4_dev *dev, struct mlx4_mod_stat_cfg *cfg)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 *inbox;
|
|
int err = 0;
|
|
|
|
#define MOD_STAT_CFG_IN_SIZE 0x100
|
|
|
|
#define MOD_STAT_CFG_PG_SZ_M_OFFSET 0x002
|
|
#define MOD_STAT_CFG_PG_SZ_OFFSET 0x003
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
inbox = mailbox->buf;
|
|
|
|
MLX4_PUT(inbox, cfg->log_pg_sz, MOD_STAT_CFG_PG_SZ_OFFSET);
|
|
MLX4_PUT(inbox, cfg->log_pg_sz_m, MOD_STAT_CFG_PG_SZ_M_OFFSET);
|
|
|
|
err = mlx4_cmd(dev, mailbox->dma, 0, 0, MLX4_CMD_MOD_STAT_CFG,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
int mlx4_QUERY_FUNC(struct mlx4_dev *dev, struct mlx4_func *func, int slave)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 *outbox;
|
|
u8 in_modifier;
|
|
u8 field;
|
|
u16 field16;
|
|
int err;
|
|
|
|
#define QUERY_FUNC_BUS_OFFSET 0x00
|
|
#define QUERY_FUNC_DEVICE_OFFSET 0x01
|
|
#define QUERY_FUNC_FUNCTION_OFFSET 0x01
|
|
#define QUERY_FUNC_PHYSICAL_FUNCTION_OFFSET 0x03
|
|
#define QUERY_FUNC_RSVD_EQS_OFFSET 0x04
|
|
#define QUERY_FUNC_MAX_EQ_OFFSET 0x06
|
|
#define QUERY_FUNC_RSVD_UARS_OFFSET 0x0b
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
outbox = mailbox->buf;
|
|
|
|
in_modifier = slave;
|
|
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, in_modifier, 0,
|
|
MLX4_CMD_QUERY_FUNC,
|
|
MLX4_CMD_TIME_CLASS_A,
|
|
MLX4_CMD_NATIVE);
|
|
if (err)
|
|
goto out;
|
|
|
|
MLX4_GET(field, outbox, QUERY_FUNC_BUS_OFFSET);
|
|
func->bus = field & 0xf;
|
|
MLX4_GET(field, outbox, QUERY_FUNC_DEVICE_OFFSET);
|
|
func->device = field & 0xf1;
|
|
MLX4_GET(field, outbox, QUERY_FUNC_FUNCTION_OFFSET);
|
|
func->function = field & 0x7;
|
|
MLX4_GET(field, outbox, QUERY_FUNC_PHYSICAL_FUNCTION_OFFSET);
|
|
func->physical_function = field & 0xf;
|
|
MLX4_GET(field16, outbox, QUERY_FUNC_RSVD_EQS_OFFSET);
|
|
func->rsvd_eqs = field16 & 0xffff;
|
|
MLX4_GET(field16, outbox, QUERY_FUNC_MAX_EQ_OFFSET);
|
|
func->max_eq = field16 & 0xffff;
|
|
MLX4_GET(field, outbox, QUERY_FUNC_RSVD_UARS_OFFSET);
|
|
func->rsvd_uars = field & 0x0f;
|
|
|
|
mlx4_dbg(dev, "Bus: %d, Device: %d, Function: %d, Physical function: %d, Max EQs: %d, Reserved EQs: %d, Reserved UARs: %d\n",
|
|
func->bus, func->device, func->function, func->physical_function,
|
|
func->max_eq, func->rsvd_eqs, func->rsvd_uars);
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_activate_vst_qinq(struct mlx4_priv *priv, int slave, int port)
|
|
{
|
|
struct mlx4_vport_oper_state *vp_oper;
|
|
struct mlx4_vport_state *vp_admin;
|
|
int err;
|
|
|
|
vp_oper = &priv->mfunc.master.vf_oper[slave].vport[port];
|
|
vp_admin = &priv->mfunc.master.vf_admin[slave].vport[port];
|
|
|
|
if (vp_admin->default_vlan != vp_oper->state.default_vlan) {
|
|
err = __mlx4_register_vlan(&priv->dev, port,
|
|
vp_admin->default_vlan,
|
|
&vp_oper->vlan_idx);
|
|
if (err) {
|
|
vp_oper->vlan_idx = NO_INDX;
|
|
mlx4_warn(&priv->dev,
|
|
"No vlan resources slave %d, port %d\n",
|
|
slave, port);
|
|
return err;
|
|
}
|
|
mlx4_dbg(&priv->dev, "alloc vlan %d idx %d slave %d port %d\n",
|
|
(int)(vp_oper->state.default_vlan),
|
|
vp_oper->vlan_idx, slave, port);
|
|
}
|
|
vp_oper->state.vlan_proto = vp_admin->vlan_proto;
|
|
vp_oper->state.default_vlan = vp_admin->default_vlan;
|
|
vp_oper->state.default_qos = vp_admin->default_qos;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mlx4_handle_vst_qinq(struct mlx4_priv *priv, int slave, int port)
|
|
{
|
|
struct mlx4_vport_oper_state *vp_oper;
|
|
struct mlx4_slave_state *slave_state;
|
|
struct mlx4_vport_state *vp_admin;
|
|
int err;
|
|
|
|
vp_oper = &priv->mfunc.master.vf_oper[slave].vport[port];
|
|
vp_admin = &priv->mfunc.master.vf_admin[slave].vport[port];
|
|
slave_state = &priv->mfunc.master.slave_state[slave];
|
|
|
|
if ((vp_admin->vlan_proto != htons(ETH_P_8021AD)) ||
|
|
(!slave_state->active))
|
|
return 0;
|
|
|
|
if (vp_oper->state.vlan_proto == vp_admin->vlan_proto &&
|
|
vp_oper->state.default_vlan == vp_admin->default_vlan &&
|
|
vp_oper->state.default_qos == vp_admin->default_qos)
|
|
return 0;
|
|
|
|
if (!slave_state->vst_qinq_supported) {
|
|
/* Warn and revert the request to set vst QinQ mode */
|
|
vp_admin->vlan_proto = vp_oper->state.vlan_proto;
|
|
vp_admin->default_vlan = vp_oper->state.default_vlan;
|
|
vp_admin->default_qos = vp_oper->state.default_qos;
|
|
|
|
mlx4_warn(&priv->dev,
|
|
"Slave %d does not support VST QinQ mode\n", slave);
|
|
return 0;
|
|
}
|
|
|
|
err = mlx4_activate_vst_qinq(priv, slave, port);
|
|
return err;
|
|
}
|
|
|
|
int mlx4_QUERY_FUNC_CAP_wrapper(struct mlx4_dev *dev, int slave,
|
|
struct mlx4_vhcr *vhcr,
|
|
struct mlx4_cmd_mailbox *inbox,
|
|
struct mlx4_cmd_mailbox *outbox,
|
|
struct mlx4_cmd_info *cmd)
|
|
{
|
|
struct mlx4_priv *priv = mlx4_priv(dev);
|
|
u8 field, port;
|
|
u32 size, proxy_qp, qkey;
|
|
int err = 0;
|
|
struct mlx4_func func;
|
|
|
|
#define QUERY_FUNC_CAP_FLAGS_OFFSET 0x0
|
|
#define QUERY_FUNC_CAP_NUM_PORTS_OFFSET 0x1
|
|
#define QUERY_FUNC_CAP_PF_BHVR_OFFSET 0x4
|
|
#define QUERY_FUNC_CAP_FMR_OFFSET 0x8
|
|
#define QUERY_FUNC_CAP_QP_QUOTA_OFFSET_DEP 0x10
|
|
#define QUERY_FUNC_CAP_CQ_QUOTA_OFFSET_DEP 0x14
|
|
#define QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET_DEP 0x18
|
|
#define QUERY_FUNC_CAP_MPT_QUOTA_OFFSET_DEP 0x20
|
|
#define QUERY_FUNC_CAP_MTT_QUOTA_OFFSET_DEP 0x24
|
|
#define QUERY_FUNC_CAP_MCG_QUOTA_OFFSET_DEP 0x28
|
|
#define QUERY_FUNC_CAP_MAX_EQ_OFFSET 0x2c
|
|
#define QUERY_FUNC_CAP_RESERVED_EQ_OFFSET 0x30
|
|
#define QUERY_FUNC_CAP_QP_RESD_LKEY_OFFSET 0x48
|
|
|
|
#define QUERY_FUNC_CAP_QP_QUOTA_OFFSET 0x50
|
|
#define QUERY_FUNC_CAP_CQ_QUOTA_OFFSET 0x54
|
|
#define QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET 0x58
|
|
#define QUERY_FUNC_CAP_MPT_QUOTA_OFFSET 0x60
|
|
#define QUERY_FUNC_CAP_MTT_QUOTA_OFFSET 0x64
|
|
#define QUERY_FUNC_CAP_MCG_QUOTA_OFFSET 0x68
|
|
|
|
#define QUERY_FUNC_CAP_EXTRA_FLAGS_OFFSET 0x6c
|
|
|
|
#define QUERY_FUNC_CAP_FMR_FLAG 0x80
|
|
#define QUERY_FUNC_CAP_FLAG_RDMA 0x40
|
|
#define QUERY_FUNC_CAP_FLAG_ETH 0x80
|
|
#define QUERY_FUNC_CAP_FLAG_QUOTAS 0x10
|
|
#define QUERY_FUNC_CAP_FLAG_RESD_LKEY 0x08
|
|
#define QUERY_FUNC_CAP_FLAG_VALID_MAILBOX 0x04
|
|
|
|
#define QUERY_FUNC_CAP_EXTRA_FLAGS_BF_QP_ALLOC_FLAG (1UL << 31)
|
|
#define QUERY_FUNC_CAP_EXTRA_FLAGS_A0_QP_ALLOC_FLAG (1UL << 30)
|
|
|
|
/* when opcode modifier = 1 */
|
|
#define QUERY_FUNC_CAP_PHYS_PORT_OFFSET 0x3
|
|
#define QUERY_FUNC_CAP_PRIV_VF_QKEY_OFFSET 0x4
|
|
#define QUERY_FUNC_CAP_FLAGS0_OFFSET 0x8
|
|
#define QUERY_FUNC_CAP_FLAGS1_OFFSET 0xc
|
|
|
|
#define QUERY_FUNC_CAP_QP0_TUNNEL 0x10
|
|
#define QUERY_FUNC_CAP_QP0_PROXY 0x14
|
|
#define QUERY_FUNC_CAP_QP1_TUNNEL 0x18
|
|
#define QUERY_FUNC_CAP_QP1_PROXY 0x1c
|
|
#define QUERY_FUNC_CAP_PHYS_PORT_ID 0x28
|
|
|
|
#define QUERY_FUNC_CAP_FLAGS1_FORCE_MAC 0x40
|
|
#define QUERY_FUNC_CAP_FLAGS1_FORCE_VLAN 0x80
|
|
#define QUERY_FUNC_CAP_FLAGS1_NIC_INFO 0x10
|
|
#define QUERY_FUNC_CAP_VF_ENABLE_QP0 0x08
|
|
|
|
#define QUERY_FUNC_CAP_FLAGS0_FORCE_PHY_WQE_GID 0x80
|
|
#define QUERY_FUNC_CAP_PHV_BIT 0x40
|
|
#define QUERY_FUNC_CAP_VLAN_OFFLOAD_DISABLE 0x20
|
|
|
|
#define QUERY_FUNC_CAP_SUPPORTS_VST_QINQ BIT(30)
|
|
#define QUERY_FUNC_CAP_SUPPORTS_NON_POWER_OF_2_NUM_EQS BIT(31)
|
|
|
|
if (vhcr->op_modifier == 1) {
|
|
struct mlx4_active_ports actv_ports =
|
|
mlx4_get_active_ports(dev, slave);
|
|
int converted_port = mlx4_slave_convert_port(
|
|
dev, slave, vhcr->in_modifier);
|
|
struct mlx4_vport_oper_state *vp_oper;
|
|
|
|
if (converted_port < 0)
|
|
return -EINVAL;
|
|
|
|
vhcr->in_modifier = converted_port;
|
|
/* phys-port = logical-port */
|
|
field = vhcr->in_modifier -
|
|
find_first_bit(actv_ports.ports, dev->caps.num_ports);
|
|
MLX4_PUT(outbox->buf, field, QUERY_FUNC_CAP_PHYS_PORT_OFFSET);
|
|
|
|
port = vhcr->in_modifier;
|
|
proxy_qp = dev->phys_caps.base_proxy_sqpn + 8 * slave + port - 1;
|
|
|
|
/* Set nic_info bit to mark new fields support */
|
|
field = QUERY_FUNC_CAP_FLAGS1_NIC_INFO;
|
|
|
|
if (mlx4_vf_smi_enabled(dev, slave, port) &&
|
|
!mlx4_get_parav_qkey(dev, proxy_qp, &qkey)) {
|
|
field |= QUERY_FUNC_CAP_VF_ENABLE_QP0;
|
|
MLX4_PUT(outbox->buf, qkey,
|
|
QUERY_FUNC_CAP_PRIV_VF_QKEY_OFFSET);
|
|
}
|
|
MLX4_PUT(outbox->buf, field, QUERY_FUNC_CAP_FLAGS1_OFFSET);
|
|
|
|
/* size is now the QP number */
|
|
size = dev->phys_caps.base_tunnel_sqpn + 8 * slave + port - 1;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_QP0_TUNNEL);
|
|
|
|
size += 2;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_QP1_TUNNEL);
|
|
|
|
MLX4_PUT(outbox->buf, proxy_qp, QUERY_FUNC_CAP_QP0_PROXY);
|
|
proxy_qp += 2;
|
|
MLX4_PUT(outbox->buf, proxy_qp, QUERY_FUNC_CAP_QP1_PROXY);
|
|
|
|
MLX4_PUT(outbox->buf, dev->caps.phys_port_id[vhcr->in_modifier],
|
|
QUERY_FUNC_CAP_PHYS_PORT_ID);
|
|
|
|
vp_oper = &priv->mfunc.master.vf_oper[slave].vport[port];
|
|
err = mlx4_handle_vst_qinq(priv, slave, port);
|
|
if (err)
|
|
return err;
|
|
|
|
field = 0;
|
|
if (dev->caps.phv_bit[port])
|
|
field |= QUERY_FUNC_CAP_PHV_BIT;
|
|
if (vp_oper->state.vlan_proto == htons(ETH_P_8021AD))
|
|
field |= QUERY_FUNC_CAP_VLAN_OFFLOAD_DISABLE;
|
|
MLX4_PUT(outbox->buf, field, QUERY_FUNC_CAP_FLAGS0_OFFSET);
|
|
|
|
} else if (vhcr->op_modifier == 0) {
|
|
struct mlx4_active_ports actv_ports =
|
|
mlx4_get_active_ports(dev, slave);
|
|
struct mlx4_slave_state *slave_state =
|
|
&priv->mfunc.master.slave_state[slave];
|
|
|
|
/* enable rdma and ethernet interfaces, new quota locations,
|
|
* and reserved lkey
|
|
*/
|
|
field = (QUERY_FUNC_CAP_FLAG_ETH | QUERY_FUNC_CAP_FLAG_RDMA |
|
|
QUERY_FUNC_CAP_FLAG_QUOTAS | QUERY_FUNC_CAP_FLAG_VALID_MAILBOX |
|
|
QUERY_FUNC_CAP_FLAG_RESD_LKEY);
|
|
MLX4_PUT(outbox->buf, field, QUERY_FUNC_CAP_FLAGS_OFFSET);
|
|
|
|
field = min(
|
|
bitmap_weight(actv_ports.ports, dev->caps.num_ports),
|
|
dev->caps.num_ports);
|
|
MLX4_PUT(outbox->buf, field, QUERY_FUNC_CAP_NUM_PORTS_OFFSET);
|
|
|
|
size = dev->caps.function_caps; /* set PF behaviours */
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_PF_BHVR_OFFSET);
|
|
|
|
field = 0; /* protected FMR support not available as yet */
|
|
MLX4_PUT(outbox->buf, field, QUERY_FUNC_CAP_FMR_OFFSET);
|
|
|
|
size = priv->mfunc.master.res_tracker.res_alloc[RES_QP].quota[slave];
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_QP_QUOTA_OFFSET);
|
|
size = dev->caps.num_qps;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_QP_QUOTA_OFFSET_DEP);
|
|
|
|
size = priv->mfunc.master.res_tracker.res_alloc[RES_SRQ].quota[slave];
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET);
|
|
size = dev->caps.num_srqs;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET_DEP);
|
|
|
|
size = priv->mfunc.master.res_tracker.res_alloc[RES_CQ].quota[slave];
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_CQ_QUOTA_OFFSET);
|
|
size = dev->caps.num_cqs;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_CQ_QUOTA_OFFSET_DEP);
|
|
|
|
if (!(dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_SYS_EQS) ||
|
|
mlx4_QUERY_FUNC(dev, &func, slave)) {
|
|
size = vhcr->in_modifier &
|
|
QUERY_FUNC_CAP_SUPPORTS_NON_POWER_OF_2_NUM_EQS ?
|
|
dev->caps.num_eqs :
|
|
rounddown_pow_of_two(dev->caps.num_eqs);
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MAX_EQ_OFFSET);
|
|
size = dev->caps.reserved_eqs;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_RESERVED_EQ_OFFSET);
|
|
} else {
|
|
size = vhcr->in_modifier &
|
|
QUERY_FUNC_CAP_SUPPORTS_NON_POWER_OF_2_NUM_EQS ?
|
|
func.max_eq :
|
|
rounddown_pow_of_two(func.max_eq);
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MAX_EQ_OFFSET);
|
|
size = func.rsvd_eqs;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_RESERVED_EQ_OFFSET);
|
|
}
|
|
|
|
size = priv->mfunc.master.res_tracker.res_alloc[RES_MPT].quota[slave];
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MPT_QUOTA_OFFSET);
|
|
size = dev->caps.num_mpts;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MPT_QUOTA_OFFSET_DEP);
|
|
|
|
size = priv->mfunc.master.res_tracker.res_alloc[RES_MTT].quota[slave];
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MTT_QUOTA_OFFSET);
|
|
size = dev->caps.num_mtts;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MTT_QUOTA_OFFSET_DEP);
|
|
|
|
size = dev->caps.num_mgms + dev->caps.num_amgms;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MCG_QUOTA_OFFSET);
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MCG_QUOTA_OFFSET_DEP);
|
|
|
|
size = QUERY_FUNC_CAP_EXTRA_FLAGS_BF_QP_ALLOC_FLAG |
|
|
QUERY_FUNC_CAP_EXTRA_FLAGS_A0_QP_ALLOC_FLAG;
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_EXTRA_FLAGS_OFFSET);
|
|
|
|
size = dev->caps.reserved_lkey + ((slave << 8) & 0xFF00);
|
|
MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_QP_RESD_LKEY_OFFSET);
|
|
|
|
if (vhcr->in_modifier & QUERY_FUNC_CAP_SUPPORTS_VST_QINQ)
|
|
slave_state->vst_qinq_supported = true;
|
|
|
|
} else
|
|
err = -EINVAL;
|
|
|
|
return err;
|
|
}
|
|
|
|
int mlx4_QUERY_FUNC_CAP(struct mlx4_dev *dev, u8 gen_or_port,
|
|
struct mlx4_func_cap *func_cap)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 *outbox;
|
|
u8 field, op_modifier;
|
|
u32 size, qkey;
|
|
int err = 0, quotas = 0;
|
|
u32 in_modifier;
|
|
u32 slave_caps;
|
|
|
|
op_modifier = !!gen_or_port; /* 0 = general, 1 = logical port */
|
|
slave_caps = QUERY_FUNC_CAP_SUPPORTS_VST_QINQ |
|
|
QUERY_FUNC_CAP_SUPPORTS_NON_POWER_OF_2_NUM_EQS;
|
|
in_modifier = op_modifier ? gen_or_port : slave_caps;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, in_modifier, op_modifier,
|
|
MLX4_CMD_QUERY_FUNC_CAP,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED);
|
|
if (err)
|
|
goto out;
|
|
|
|
outbox = mailbox->buf;
|
|
|
|
if (!op_modifier) {
|
|
MLX4_GET(field, outbox, QUERY_FUNC_CAP_FLAGS_OFFSET);
|
|
if (!(field & (QUERY_FUNC_CAP_FLAG_ETH | QUERY_FUNC_CAP_FLAG_RDMA))) {
|
|
mlx4_err(dev, "The host supports neither eth nor rdma interfaces\n");
|
|
err = -EPROTONOSUPPORT;
|
|
goto out;
|
|
}
|
|
func_cap->flags = field;
|
|
quotas = !!(func_cap->flags & QUERY_FUNC_CAP_FLAG_QUOTAS);
|
|
|
|
MLX4_GET(field, outbox, QUERY_FUNC_CAP_NUM_PORTS_OFFSET);
|
|
func_cap->num_ports = field;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_PF_BHVR_OFFSET);
|
|
func_cap->pf_context_behaviour = size;
|
|
|
|
if (quotas) {
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_QP_QUOTA_OFFSET);
|
|
func_cap->qp_quota = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET);
|
|
func_cap->srq_quota = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_CQ_QUOTA_OFFSET);
|
|
func_cap->cq_quota = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_MPT_QUOTA_OFFSET);
|
|
func_cap->mpt_quota = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_MTT_QUOTA_OFFSET);
|
|
func_cap->mtt_quota = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_MCG_QUOTA_OFFSET);
|
|
func_cap->mcg_quota = size & 0xFFFFFF;
|
|
|
|
} else {
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_QP_QUOTA_OFFSET_DEP);
|
|
func_cap->qp_quota = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET_DEP);
|
|
func_cap->srq_quota = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_CQ_QUOTA_OFFSET_DEP);
|
|
func_cap->cq_quota = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_MPT_QUOTA_OFFSET_DEP);
|
|
func_cap->mpt_quota = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_MTT_QUOTA_OFFSET_DEP);
|
|
func_cap->mtt_quota = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_MCG_QUOTA_OFFSET_DEP);
|
|
func_cap->mcg_quota = size & 0xFFFFFF;
|
|
}
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_MAX_EQ_OFFSET);
|
|
func_cap->max_eq = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_RESERVED_EQ_OFFSET);
|
|
func_cap->reserved_eq = size & 0xFFFFFF;
|
|
|
|
if (func_cap->flags & QUERY_FUNC_CAP_FLAG_RESD_LKEY) {
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_QP_RESD_LKEY_OFFSET);
|
|
func_cap->reserved_lkey = size;
|
|
} else {
|
|
func_cap->reserved_lkey = 0;
|
|
}
|
|
|
|
func_cap->extra_flags = 0;
|
|
|
|
/* Mailbox data from 0x6c and onward should only be treated if
|
|
* QUERY_FUNC_CAP_FLAG_VALID_MAILBOX is set in func_cap->flags
|
|
*/
|
|
if (func_cap->flags & QUERY_FUNC_CAP_FLAG_VALID_MAILBOX) {
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_EXTRA_FLAGS_OFFSET);
|
|
if (size & QUERY_FUNC_CAP_EXTRA_FLAGS_BF_QP_ALLOC_FLAG)
|
|
func_cap->extra_flags |= MLX4_QUERY_FUNC_FLAGS_BF_RES_QP;
|
|
if (size & QUERY_FUNC_CAP_EXTRA_FLAGS_A0_QP_ALLOC_FLAG)
|
|
func_cap->extra_flags |= MLX4_QUERY_FUNC_FLAGS_A0_RES_QP;
|
|
}
|
|
|
|
goto out;
|
|
}
|
|
|
|
/* logical port query */
|
|
if (gen_or_port > dev->caps.num_ports) {
|
|
err = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
MLX4_GET(func_cap->flags1, outbox, QUERY_FUNC_CAP_FLAGS1_OFFSET);
|
|
if (dev->caps.port_type[gen_or_port] == MLX4_PORT_TYPE_ETH) {
|
|
if (func_cap->flags1 & QUERY_FUNC_CAP_FLAGS1_FORCE_VLAN) {
|
|
mlx4_err(dev, "VLAN is enforced on this port\n");
|
|
err = -EPROTONOSUPPORT;
|
|
goto out;
|
|
}
|
|
|
|
if (func_cap->flags1 & QUERY_FUNC_CAP_FLAGS1_FORCE_MAC) {
|
|
mlx4_err(dev, "Force mac is enabled on this port\n");
|
|
err = -EPROTONOSUPPORT;
|
|
goto out;
|
|
}
|
|
} else if (dev->caps.port_type[gen_or_port] == MLX4_PORT_TYPE_IB) {
|
|
MLX4_GET(field, outbox, QUERY_FUNC_CAP_FLAGS0_OFFSET);
|
|
if (field & QUERY_FUNC_CAP_FLAGS0_FORCE_PHY_WQE_GID) {
|
|
mlx4_err(dev, "phy_wqe_gid is enforced on this ib port\n");
|
|
err = -EPROTONOSUPPORT;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
MLX4_GET(field, outbox, QUERY_FUNC_CAP_PHYS_PORT_OFFSET);
|
|
func_cap->physical_port = field;
|
|
if (func_cap->physical_port != gen_or_port) {
|
|
err = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
if (func_cap->flags1 & QUERY_FUNC_CAP_VF_ENABLE_QP0) {
|
|
MLX4_GET(qkey, outbox, QUERY_FUNC_CAP_PRIV_VF_QKEY_OFFSET);
|
|
func_cap->spec_qps.qp0_qkey = qkey;
|
|
} else {
|
|
func_cap->spec_qps.qp0_qkey = 0;
|
|
}
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_QP0_TUNNEL);
|
|
func_cap->spec_qps.qp0_tunnel = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_QP0_PROXY);
|
|
func_cap->spec_qps.qp0_proxy = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_QP1_TUNNEL);
|
|
func_cap->spec_qps.qp1_tunnel = size & 0xFFFFFF;
|
|
|
|
MLX4_GET(size, outbox, QUERY_FUNC_CAP_QP1_PROXY);
|
|
func_cap->spec_qps.qp1_proxy = size & 0xFFFFFF;
|
|
|
|
if (func_cap->flags1 & QUERY_FUNC_CAP_FLAGS1_NIC_INFO)
|
|
MLX4_GET(func_cap->phys_port_id, outbox,
|
|
QUERY_FUNC_CAP_PHYS_PORT_ID);
|
|
|
|
MLX4_GET(func_cap->flags0, outbox, QUERY_FUNC_CAP_FLAGS0_OFFSET);
|
|
|
|
/* All other resources are allocated by the master, but we still report
|
|
* 'num' and 'reserved' capabilities as follows:
|
|
* - num remains the maximum resource index
|
|
* - 'num - reserved' is the total available objects of a resource, but
|
|
* resource indices may be less than 'reserved'
|
|
* TODO: set per-resource quotas */
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
|
|
return err;
|
|
}
|
|
|
|
static void disable_unsupported_roce_caps(void *buf);
|
|
|
|
int mlx4_QUERY_DEV_CAP(struct mlx4_dev *dev, struct mlx4_dev_cap *dev_cap)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 *outbox;
|
|
u8 field;
|
|
u32 field32, flags, ext_flags;
|
|
u16 size;
|
|
u16 stat_rate;
|
|
int err;
|
|
int i;
|
|
|
|
#define QUERY_DEV_CAP_OUT_SIZE 0x100
|
|
#define QUERY_DEV_CAP_MAX_SRQ_SZ_OFFSET 0x10
|
|
#define QUERY_DEV_CAP_MAX_QP_SZ_OFFSET 0x11
|
|
#define QUERY_DEV_CAP_RSVD_QP_OFFSET 0x12
|
|
#define QUERY_DEV_CAP_MAX_QP_OFFSET 0x13
|
|
#define QUERY_DEV_CAP_RSVD_SRQ_OFFSET 0x14
|
|
#define QUERY_DEV_CAP_MAX_SRQ_OFFSET 0x15
|
|
#define QUERY_DEV_CAP_RSVD_EEC_OFFSET 0x16
|
|
#define QUERY_DEV_CAP_MAX_EEC_OFFSET 0x17
|
|
#define QUERY_DEV_CAP_MAX_CQ_SZ_OFFSET 0x19
|
|
#define QUERY_DEV_CAP_RSVD_CQ_OFFSET 0x1a
|
|
#define QUERY_DEV_CAP_MAX_CQ_OFFSET 0x1b
|
|
#define QUERY_DEV_CAP_MAX_MPT_OFFSET 0x1d
|
|
#define QUERY_DEV_CAP_RSVD_EQ_OFFSET 0x1e
|
|
#define QUERY_DEV_CAP_MAX_EQ_OFFSET 0x1f
|
|
#define QUERY_DEV_CAP_RSVD_MTT_OFFSET 0x20
|
|
#define QUERY_DEV_CAP_MAX_MRW_SZ_OFFSET 0x21
|
|
#define QUERY_DEV_CAP_RSVD_MRW_OFFSET 0x22
|
|
#define QUERY_DEV_CAP_MAX_MTT_SEG_OFFSET 0x23
|
|
#define QUERY_DEV_CAP_NUM_SYS_EQ_OFFSET 0x26
|
|
#define QUERY_DEV_CAP_MAX_AV_OFFSET 0x27
|
|
#define QUERY_DEV_CAP_MAX_REQ_QP_OFFSET 0x29
|
|
#define QUERY_DEV_CAP_MAX_RES_QP_OFFSET 0x2b
|
|
#define QUERY_DEV_CAP_MAX_GSO_OFFSET 0x2d
|
|
#define QUERY_DEV_CAP_RSS_OFFSET 0x2e
|
|
#define QUERY_DEV_CAP_MAX_RDMA_OFFSET 0x2f
|
|
#define QUERY_DEV_CAP_RSZ_SRQ_OFFSET 0x33
|
|
#define QUERY_DEV_CAP_PORT_BEACON_OFFSET 0x34
|
|
#define QUERY_DEV_CAP_ACK_DELAY_OFFSET 0x35
|
|
#define QUERY_DEV_CAP_MTU_WIDTH_OFFSET 0x36
|
|
#define QUERY_DEV_CAP_VL_PORT_OFFSET 0x37
|
|
#define QUERY_DEV_CAP_MAX_MSG_SZ_OFFSET 0x38
|
|
#define QUERY_DEV_CAP_MAX_GID_OFFSET 0x3b
|
|
#define QUERY_DEV_CAP_RATE_SUPPORT_OFFSET 0x3c
|
|
#define QUERY_DEV_CAP_CQ_TS_SUPPORT_OFFSET 0x3e
|
|
#define QUERY_DEV_CAP_MAX_PKEY_OFFSET 0x3f
|
|
#define QUERY_DEV_CAP_EXT_FLAGS_OFFSET 0x40
|
|
#define QUERY_DEV_CAP_WOL_OFFSET 0x43
|
|
#define QUERY_DEV_CAP_FLAGS_OFFSET 0x44
|
|
#define QUERY_DEV_CAP_RSVD_UAR_OFFSET 0x48
|
|
#define QUERY_DEV_CAP_UAR_SZ_OFFSET 0x49
|
|
#define QUERY_DEV_CAP_PAGE_SZ_OFFSET 0x4b
|
|
#define QUERY_DEV_CAP_BF_OFFSET 0x4c
|
|
#define QUERY_DEV_CAP_LOG_BF_REG_SZ_OFFSET 0x4d
|
|
#define QUERY_DEV_CAP_LOG_MAX_BF_REGS_PER_PAGE_OFFSET 0x4e
|
|
#define QUERY_DEV_CAP_LOG_MAX_BF_PAGES_OFFSET 0x4f
|
|
#define QUERY_DEV_CAP_MAX_SG_SQ_OFFSET 0x51
|
|
#define QUERY_DEV_CAP_MAX_DESC_SZ_SQ_OFFSET 0x52
|
|
#define QUERY_DEV_CAP_MAX_SG_RQ_OFFSET 0x55
|
|
#define QUERY_DEV_CAP_MAX_DESC_SZ_RQ_OFFSET 0x56
|
|
#define QUERY_DEV_CAP_USER_MAC_EN_OFFSET 0x5C
|
|
#define QUERY_DEV_CAP_SVLAN_BY_QP_OFFSET 0x5D
|
|
#define QUERY_DEV_CAP_MAX_QP_MCG_OFFSET 0x61
|
|
#define QUERY_DEV_CAP_RSVD_MCG_OFFSET 0x62
|
|
#define QUERY_DEV_CAP_MAX_MCG_OFFSET 0x63
|
|
#define QUERY_DEV_CAP_RSVD_PD_OFFSET 0x64
|
|
#define QUERY_DEV_CAP_MAX_PD_OFFSET 0x65
|
|
#define QUERY_DEV_CAP_RSVD_XRC_OFFSET 0x66
|
|
#define QUERY_DEV_CAP_MAX_XRC_OFFSET 0x67
|
|
#define QUERY_DEV_CAP_MAX_COUNTERS_OFFSET 0x68
|
|
#define QUERY_DEV_CAP_PORT_FLOWSTATS_COUNTERS_OFFSET 0x70
|
|
#define QUERY_DEV_CAP_EXT_2_FLAGS_OFFSET 0x70
|
|
#define QUERY_DEV_CAP_FLOW_STEERING_IPOIB_OFFSET 0x74
|
|
#define QUERY_DEV_CAP_FLOW_STEERING_RANGE_EN_OFFSET 0x76
|
|
#define QUERY_DEV_CAP_FLOW_STEERING_MAX_QP_OFFSET 0x77
|
|
#define QUERY_DEV_CAP_SL2VL_EVENT_OFFSET 0x78
|
|
#define QUERY_DEV_CAP_CQ_EQ_CACHE_LINE_STRIDE 0x7a
|
|
#define QUERY_DEV_CAP_ECN_QCN_VER_OFFSET 0x7b
|
|
#define QUERY_DEV_CAP_RDMARC_ENTRY_SZ_OFFSET 0x80
|
|
#define QUERY_DEV_CAP_QPC_ENTRY_SZ_OFFSET 0x82
|
|
#define QUERY_DEV_CAP_AUX_ENTRY_SZ_OFFSET 0x84
|
|
#define QUERY_DEV_CAP_ALTC_ENTRY_SZ_OFFSET 0x86
|
|
#define QUERY_DEV_CAP_EQC_ENTRY_SZ_OFFSET 0x88
|
|
#define QUERY_DEV_CAP_CQC_ENTRY_SZ_OFFSET 0x8a
|
|
#define QUERY_DEV_CAP_SRQ_ENTRY_SZ_OFFSET 0x8c
|
|
#define QUERY_DEV_CAP_C_MPT_ENTRY_SZ_OFFSET 0x8e
|
|
#define QUERY_DEV_CAP_MTT_ENTRY_SZ_OFFSET 0x90
|
|
#define QUERY_DEV_CAP_D_MPT_ENTRY_SZ_OFFSET 0x92
|
|
#define QUERY_DEV_CAP_BMME_FLAGS_OFFSET 0x94
|
|
#define QUERY_DEV_CAP_CONFIG_DEV_OFFSET 0x94
|
|
#define QUERY_DEV_CAP_PHV_EN_OFFSET 0x96
|
|
#define QUERY_DEV_CAP_RSVD_LKEY_OFFSET 0x98
|
|
#define QUERY_DEV_CAP_MAX_ICM_SZ_OFFSET 0xa0
|
|
#define QUERY_DEV_CAP_ETH_BACKPL_OFFSET 0x9c
|
|
#define QUERY_DEV_CAP_DIAG_RPRT_PER_PORT 0x9c
|
|
#define QUERY_DEV_CAP_FW_REASSIGN_MAC 0x9d
|
|
#define QUERY_DEV_CAP_VXLAN 0x9e
|
|
#define QUERY_DEV_CAP_MAD_DEMUX_OFFSET 0xb0
|
|
#define QUERY_DEV_CAP_DMFS_HIGH_RATE_QPN_BASE_OFFSET 0xa8
|
|
#define QUERY_DEV_CAP_DMFS_HIGH_RATE_QPN_RANGE_OFFSET 0xac
|
|
#define QUERY_DEV_CAP_QP_RATE_LIMIT_NUM_OFFSET 0xcc
|
|
#define QUERY_DEV_CAP_QP_RATE_LIMIT_MAX_OFFSET 0xd0
|
|
#define QUERY_DEV_CAP_QP_RATE_LIMIT_MIN_OFFSET 0xd2
|
|
#define QUERY_DEV_CAP_HEALTH_BUFFER_ADDRESS_OFFSET 0xe4
|
|
|
|
dev_cap->flags2 = 0;
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
outbox = mailbox->buf;
|
|
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, 0, 0, MLX4_CMD_QUERY_DEV_CAP,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
goto out;
|
|
|
|
if (mlx4_is_mfunc(dev))
|
|
disable_unsupported_roce_caps(outbox);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSVD_QP_OFFSET);
|
|
dev_cap->reserved_qps = 1 << (field & 0xf);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_QP_OFFSET);
|
|
dev_cap->max_qps = 1 << (field & 0x1f);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSVD_SRQ_OFFSET);
|
|
dev_cap->reserved_srqs = 1 << (field >> 4);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_SRQ_OFFSET);
|
|
dev_cap->max_srqs = 1 << (field & 0x1f);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_CQ_SZ_OFFSET);
|
|
dev_cap->max_cq_sz = 1 << field;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSVD_CQ_OFFSET);
|
|
dev_cap->reserved_cqs = 1 << (field & 0xf);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_CQ_OFFSET);
|
|
dev_cap->max_cqs = 1 << (field & 0x1f);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_MPT_OFFSET);
|
|
dev_cap->max_mpts = 1 << (field & 0x3f);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSVD_EQ_OFFSET);
|
|
dev_cap->reserved_eqs = 1 << (field & 0xf);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_EQ_OFFSET);
|
|
dev_cap->max_eqs = 1 << (field & 0xf);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSVD_MTT_OFFSET);
|
|
dev_cap->reserved_mtts = 1 << (field >> 4);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSVD_MRW_OFFSET);
|
|
dev_cap->reserved_mrws = 1 << (field & 0xf);
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_NUM_SYS_EQ_OFFSET);
|
|
dev_cap->num_sys_eqs = size & 0xfff;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_REQ_QP_OFFSET);
|
|
dev_cap->max_requester_per_qp = 1 << (field & 0x3f);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_RES_QP_OFFSET);
|
|
dev_cap->max_responder_per_qp = 1 << (field & 0x3f);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_GSO_OFFSET);
|
|
field &= 0x1f;
|
|
if (!field)
|
|
dev_cap->max_gso_sz = 0;
|
|
else
|
|
dev_cap->max_gso_sz = 1 << field;
|
|
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSS_OFFSET);
|
|
if (field & 0x20)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_RSS_XOR;
|
|
if (field & 0x10)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_RSS_TOP;
|
|
field &= 0xf;
|
|
if (field) {
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_RSS;
|
|
dev_cap->max_rss_tbl_sz = 1 << field;
|
|
} else
|
|
dev_cap->max_rss_tbl_sz = 0;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_RDMA_OFFSET);
|
|
dev_cap->max_rdma_global = 1 << (field & 0x3f);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_ACK_DELAY_OFFSET);
|
|
dev_cap->local_ca_ack_delay = field & 0x1f;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_VL_PORT_OFFSET);
|
|
dev_cap->num_ports = field & 0xf;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_MSG_SZ_OFFSET);
|
|
dev_cap->max_msg_sz = 1 << (field & 0x1f);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_PORT_FLOWSTATS_COUNTERS_OFFSET);
|
|
if (field & 0x10)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_FLOWSTATS_EN;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_FLOW_STEERING_RANGE_EN_OFFSET);
|
|
if (field & 0x80)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_FS_EN;
|
|
dev_cap->fs_log_max_ucast_qp_range_size = field & 0x1f;
|
|
if (field & 0x20)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_DMFS_UC_MC_SNIFFER;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_PORT_BEACON_OFFSET);
|
|
if (field & 0x80)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_PORT_BEACON;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_FLOW_STEERING_IPOIB_OFFSET);
|
|
if (field & 0x80)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_DMFS_IPOIB;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_FLOW_STEERING_MAX_QP_OFFSET);
|
|
dev_cap->fs_max_num_qp_per_entry = field;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_SL2VL_EVENT_OFFSET);
|
|
if (field & (1 << 5))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_SL_TO_VL_CHANGE_EVENT;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_ECN_QCN_VER_OFFSET);
|
|
if (field & 0x1)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_QCN;
|
|
MLX4_GET(stat_rate, outbox, QUERY_DEV_CAP_RATE_SUPPORT_OFFSET);
|
|
dev_cap->stat_rate_support = stat_rate;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_CQ_TS_SUPPORT_OFFSET);
|
|
if (field & 0x80)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_TS;
|
|
MLX4_GET(ext_flags, outbox, QUERY_DEV_CAP_EXT_FLAGS_OFFSET);
|
|
MLX4_GET(flags, outbox, QUERY_DEV_CAP_FLAGS_OFFSET);
|
|
dev_cap->flags = flags | (u64)ext_flags << 32;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_WOL_OFFSET);
|
|
dev_cap->wol_port[1] = !!(field & 0x20);
|
|
dev_cap->wol_port[2] = !!(field & 0x40);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSVD_UAR_OFFSET);
|
|
dev_cap->reserved_uars = field >> 4;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_UAR_SZ_OFFSET);
|
|
dev_cap->uar_size = 1 << ((field & 0x3f) + 20);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_PAGE_SZ_OFFSET);
|
|
dev_cap->min_page_sz = 1 << field;
|
|
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_BF_OFFSET);
|
|
if (field & 0x80) {
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_LOG_BF_REG_SZ_OFFSET);
|
|
dev_cap->bf_reg_size = 1 << (field & 0x1f);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_LOG_MAX_BF_REGS_PER_PAGE_OFFSET);
|
|
if ((1 << (field & 0x3f)) > (PAGE_SIZE / dev_cap->bf_reg_size))
|
|
field = 3;
|
|
dev_cap->bf_regs_per_page = 1 << (field & 0x3f);
|
|
} else {
|
|
dev_cap->bf_reg_size = 0;
|
|
}
|
|
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_SG_SQ_OFFSET);
|
|
dev_cap->max_sq_sg = field;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_MAX_DESC_SZ_SQ_OFFSET);
|
|
dev_cap->max_sq_desc_sz = size;
|
|
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_USER_MAC_EN_OFFSET);
|
|
if (field & (1 << 2))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_USER_MAC_EN;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_SVLAN_BY_QP_OFFSET);
|
|
if (field & 0x1)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_SVLAN_BY_QP;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_QP_MCG_OFFSET);
|
|
dev_cap->max_qp_per_mcg = 1 << field;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSVD_MCG_OFFSET);
|
|
dev_cap->reserved_mgms = field & 0xf;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_MCG_OFFSET);
|
|
dev_cap->max_mcgs = 1 << field;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSVD_PD_OFFSET);
|
|
dev_cap->reserved_pds = field >> 4;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_PD_OFFSET);
|
|
dev_cap->max_pds = 1 << (field & 0x3f);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSVD_XRC_OFFSET);
|
|
dev_cap->reserved_xrcds = field >> 4;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_XRC_OFFSET);
|
|
dev_cap->max_xrcds = 1 << (field & 0x1f);
|
|
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_RDMARC_ENTRY_SZ_OFFSET);
|
|
dev_cap->rdmarc_entry_sz = size;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_QPC_ENTRY_SZ_OFFSET);
|
|
dev_cap->qpc_entry_sz = size;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_AUX_ENTRY_SZ_OFFSET);
|
|
dev_cap->aux_entry_sz = size;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_ALTC_ENTRY_SZ_OFFSET);
|
|
dev_cap->altc_entry_sz = size;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_EQC_ENTRY_SZ_OFFSET);
|
|
dev_cap->eqc_entry_sz = size;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_CQC_ENTRY_SZ_OFFSET);
|
|
dev_cap->cqc_entry_sz = size;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_SRQ_ENTRY_SZ_OFFSET);
|
|
dev_cap->srq_entry_sz = size;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_C_MPT_ENTRY_SZ_OFFSET);
|
|
dev_cap->cmpt_entry_sz = size;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_MTT_ENTRY_SZ_OFFSET);
|
|
dev_cap->mtt_entry_sz = size;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_D_MPT_ENTRY_SZ_OFFSET);
|
|
dev_cap->dmpt_entry_sz = size;
|
|
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_SRQ_SZ_OFFSET);
|
|
dev_cap->max_srq_sz = 1 << field;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_QP_SZ_OFFSET);
|
|
dev_cap->max_qp_sz = 1 << field;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_RSZ_SRQ_OFFSET);
|
|
dev_cap->resize_srq = field & 1;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_SG_RQ_OFFSET);
|
|
dev_cap->max_rq_sg = field;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_MAX_DESC_SZ_RQ_OFFSET);
|
|
dev_cap->max_rq_desc_sz = size;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_CQ_EQ_CACHE_LINE_STRIDE);
|
|
if (field & (1 << 4))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_QOS_VPP;
|
|
if (field & (1 << 5))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_ETH_PROT_CTRL;
|
|
if (field & (1 << 6))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_CQE_STRIDE;
|
|
if (field & (1 << 7))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_EQE_STRIDE;
|
|
MLX4_GET(dev_cap->bmme_flags, outbox,
|
|
QUERY_DEV_CAP_BMME_FLAGS_OFFSET);
|
|
if (dev_cap->bmme_flags & MLX4_FLAG_ROCE_V1_V2)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_ROCE_V1_V2;
|
|
if (dev_cap->bmme_flags & MLX4_FLAG_PORT_REMAP)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_PORT_REMAP;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_CONFIG_DEV_OFFSET);
|
|
if (field & 0x20)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_CONFIG_DEV;
|
|
if (field & (1 << 2))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_IGNORE_FCS;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_PHV_EN_OFFSET);
|
|
if (field & 0x80)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_PHV_EN;
|
|
if (field & 0x40)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_SKIP_OUTER_VLAN;
|
|
|
|
MLX4_GET(dev_cap->reserved_lkey, outbox,
|
|
QUERY_DEV_CAP_RSVD_LKEY_OFFSET);
|
|
MLX4_GET(field32, outbox, QUERY_DEV_CAP_ETH_BACKPL_OFFSET);
|
|
if (field32 & (1 << 0))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_ETH_BACKPL_AN_REP;
|
|
if (field32 & (1 << 7))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_RECOVERABLE_ERROR_EVENT;
|
|
if (field32 & (1 << 8))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_DRIVER_VERSION_TO_FW;
|
|
MLX4_GET(field32, outbox, QUERY_DEV_CAP_DIAG_RPRT_PER_PORT);
|
|
if (field32 & (1 << 17))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_DIAG_PER_PORT;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_FW_REASSIGN_MAC);
|
|
if (field & 1<<6)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_REASSIGN_MAC_EN;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_VXLAN);
|
|
if (field & 1<<3)
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_VXLAN_OFFLOADS;
|
|
if (field & (1 << 5))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_ETS_CFG;
|
|
MLX4_GET(dev_cap->max_icm_sz, outbox,
|
|
QUERY_DEV_CAP_MAX_ICM_SZ_OFFSET);
|
|
if (dev_cap->flags & MLX4_DEV_CAP_FLAG_COUNTERS)
|
|
MLX4_GET(dev_cap->max_counters, outbox,
|
|
QUERY_DEV_CAP_MAX_COUNTERS_OFFSET);
|
|
|
|
MLX4_GET(field32, outbox,
|
|
QUERY_DEV_CAP_MAD_DEMUX_OFFSET);
|
|
if (field32 & (1 << 0))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_MAD_DEMUX;
|
|
|
|
MLX4_GET(dev_cap->dmfs_high_rate_qpn_base, outbox,
|
|
QUERY_DEV_CAP_DMFS_HIGH_RATE_QPN_BASE_OFFSET);
|
|
dev_cap->dmfs_high_rate_qpn_base &= MGM_QPN_MASK;
|
|
MLX4_GET(dev_cap->dmfs_high_rate_qpn_range, outbox,
|
|
QUERY_DEV_CAP_DMFS_HIGH_RATE_QPN_RANGE_OFFSET);
|
|
dev_cap->dmfs_high_rate_qpn_range &= MGM_QPN_MASK;
|
|
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_QP_RATE_LIMIT_NUM_OFFSET);
|
|
dev_cap->rl_caps.num_rates = size;
|
|
if (dev_cap->rl_caps.num_rates) {
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_QP_RATE_LIMIT;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_QP_RATE_LIMIT_MAX_OFFSET);
|
|
dev_cap->rl_caps.max_val = size & 0xfff;
|
|
dev_cap->rl_caps.max_unit = size >> 14;
|
|
MLX4_GET(size, outbox, QUERY_DEV_CAP_QP_RATE_LIMIT_MIN_OFFSET);
|
|
dev_cap->rl_caps.min_val = size & 0xfff;
|
|
dev_cap->rl_caps.min_unit = size >> 14;
|
|
}
|
|
|
|
MLX4_GET(dev_cap->health_buffer_addrs, outbox,
|
|
QUERY_DEV_CAP_HEALTH_BUFFER_ADDRESS_OFFSET);
|
|
|
|
MLX4_GET(field32, outbox, QUERY_DEV_CAP_EXT_2_FLAGS_OFFSET);
|
|
if (field32 & (1 << 16))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_UPDATE_QP;
|
|
if (field32 & (1 << 18))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_UPDATE_QP_SRC_CHECK_LB;
|
|
if (field32 & (1 << 19))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_LB_SRC_CHK;
|
|
if (field32 & (1 << 26))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_VLAN_CONTROL;
|
|
if (field32 & (1 << 20))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_FSM;
|
|
if (field32 & (1 << 21))
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_80_VFS;
|
|
|
|
for (i = 1; i <= dev_cap->num_ports; i++) {
|
|
err = mlx4_QUERY_PORT(dev, i, dev_cap->port_cap + i);
|
|
if (err)
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* Each UAR has 4 EQ doorbells; so if a UAR is reserved, then
|
|
* we can't use any EQs whose doorbell falls on that page,
|
|
* even if the EQ itself isn't reserved.
|
|
*/
|
|
if (dev_cap->num_sys_eqs == 0)
|
|
dev_cap->reserved_eqs = max(dev_cap->reserved_uars * 4,
|
|
dev_cap->reserved_eqs);
|
|
else
|
|
dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_SYS_EQS;
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
void mlx4_dev_cap_dump(struct mlx4_dev *dev, struct mlx4_dev_cap *dev_cap)
|
|
{
|
|
if (dev_cap->bf_reg_size > 0)
|
|
mlx4_dbg(dev, "BlueFlame available (reg size %d, regs/page %d)\n",
|
|
dev_cap->bf_reg_size, dev_cap->bf_regs_per_page);
|
|
else
|
|
mlx4_dbg(dev, "BlueFlame not available\n");
|
|
|
|
mlx4_dbg(dev, "Base MM extensions: flags %08x, rsvd L_Key %08x\n",
|
|
dev_cap->bmme_flags, dev_cap->reserved_lkey);
|
|
mlx4_dbg(dev, "Max ICM size %lld MB\n",
|
|
(unsigned long long) dev_cap->max_icm_sz >> 20);
|
|
mlx4_dbg(dev, "Max QPs: %d, reserved QPs: %d, entry size: %d\n",
|
|
dev_cap->max_qps, dev_cap->reserved_qps, dev_cap->qpc_entry_sz);
|
|
mlx4_dbg(dev, "Max SRQs: %d, reserved SRQs: %d, entry size: %d\n",
|
|
dev_cap->max_srqs, dev_cap->reserved_srqs, dev_cap->srq_entry_sz);
|
|
mlx4_dbg(dev, "Max CQs: %d, reserved CQs: %d, entry size: %d\n",
|
|
dev_cap->max_cqs, dev_cap->reserved_cqs, dev_cap->cqc_entry_sz);
|
|
mlx4_dbg(dev, "Num sys EQs: %d, max EQs: %d, reserved EQs: %d, entry size: %d\n",
|
|
dev_cap->num_sys_eqs, dev_cap->max_eqs, dev_cap->reserved_eqs,
|
|
dev_cap->eqc_entry_sz);
|
|
mlx4_dbg(dev, "reserved MPTs: %d, reserved MTTs: %d\n",
|
|
dev_cap->reserved_mrws, dev_cap->reserved_mtts);
|
|
mlx4_dbg(dev, "Max PDs: %d, reserved PDs: %d, reserved UARs: %d\n",
|
|
dev_cap->max_pds, dev_cap->reserved_pds, dev_cap->reserved_uars);
|
|
mlx4_dbg(dev, "Max QP/MCG: %d, reserved MGMs: %d\n",
|
|
dev_cap->max_pds, dev_cap->reserved_mgms);
|
|
mlx4_dbg(dev, "Max CQEs: %d, max WQEs: %d, max SRQ WQEs: %d\n",
|
|
dev_cap->max_cq_sz, dev_cap->max_qp_sz, dev_cap->max_srq_sz);
|
|
mlx4_dbg(dev, "Local CA ACK delay: %d, max MTU: %d, port width cap: %d\n",
|
|
dev_cap->local_ca_ack_delay, 128 << dev_cap->port_cap[1].ib_mtu,
|
|
dev_cap->port_cap[1].max_port_width);
|
|
mlx4_dbg(dev, "Max SQ desc size: %d, max SQ S/G: %d\n",
|
|
dev_cap->max_sq_desc_sz, dev_cap->max_sq_sg);
|
|
mlx4_dbg(dev, "Max RQ desc size: %d, max RQ S/G: %d\n",
|
|
dev_cap->max_rq_desc_sz, dev_cap->max_rq_sg);
|
|
mlx4_dbg(dev, "Max GSO size: %d\n", dev_cap->max_gso_sz);
|
|
mlx4_dbg(dev, "Max counters: %d\n", dev_cap->max_counters);
|
|
mlx4_dbg(dev, "Max RSS Table size: %d\n", dev_cap->max_rss_tbl_sz);
|
|
mlx4_dbg(dev, "DMFS high rate steer QPn base: %d\n",
|
|
dev_cap->dmfs_high_rate_qpn_base);
|
|
mlx4_dbg(dev, "DMFS high rate steer QPn range: %d\n",
|
|
dev_cap->dmfs_high_rate_qpn_range);
|
|
|
|
if (dev_cap->flags2 & MLX4_DEV_CAP_FLAG2_QP_RATE_LIMIT) {
|
|
struct mlx4_rate_limit_caps *rl_caps = &dev_cap->rl_caps;
|
|
|
|
mlx4_dbg(dev, "QP Rate-Limit: #rates %d, unit/val max %d/%d, min %d/%d\n",
|
|
rl_caps->num_rates, rl_caps->max_unit, rl_caps->max_val,
|
|
rl_caps->min_unit, rl_caps->min_val);
|
|
}
|
|
|
|
dump_dev_cap_flags(dev, dev_cap->flags);
|
|
dump_dev_cap_flags2(dev, dev_cap->flags2);
|
|
}
|
|
|
|
int mlx4_QUERY_PORT(struct mlx4_dev *dev, int port, struct mlx4_port_cap *port_cap)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 *outbox;
|
|
u8 field;
|
|
u32 field32;
|
|
int err;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
outbox = mailbox->buf;
|
|
|
|
if (dev->flags & MLX4_FLAG_OLD_PORT_CMDS) {
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, 0, 0, MLX4_CMD_QUERY_DEV_CAP,
|
|
MLX4_CMD_TIME_CLASS_A,
|
|
MLX4_CMD_NATIVE);
|
|
|
|
if (err)
|
|
goto out;
|
|
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_VL_PORT_OFFSET);
|
|
port_cap->max_vl = field >> 4;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MTU_WIDTH_OFFSET);
|
|
port_cap->ib_mtu = field >> 4;
|
|
port_cap->max_port_width = field & 0xf;
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_GID_OFFSET);
|
|
port_cap->max_gids = 1 << (field & 0xf);
|
|
MLX4_GET(field, outbox, QUERY_DEV_CAP_MAX_PKEY_OFFSET);
|
|
port_cap->max_pkeys = 1 << (field & 0xf);
|
|
} else {
|
|
#define QUERY_PORT_SUPPORTED_TYPE_OFFSET 0x00
|
|
#define QUERY_PORT_MTU_OFFSET 0x01
|
|
#define QUERY_PORT_ETH_MTU_OFFSET 0x02
|
|
#define QUERY_PORT_WIDTH_OFFSET 0x06
|
|
#define QUERY_PORT_MAX_GID_PKEY_OFFSET 0x07
|
|
#define QUERY_PORT_MAX_MACVLAN_OFFSET 0x0a
|
|
#define QUERY_PORT_MAX_VL_OFFSET 0x0b
|
|
#define QUERY_PORT_MAC_OFFSET 0x10
|
|
#define QUERY_PORT_TRANS_VENDOR_OFFSET 0x18
|
|
#define QUERY_PORT_WAVELENGTH_OFFSET 0x1c
|
|
#define QUERY_PORT_TRANS_CODE_OFFSET 0x20
|
|
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, port, 0, MLX4_CMD_QUERY_PORT,
|
|
MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
goto out;
|
|
|
|
MLX4_GET(field, outbox, QUERY_PORT_SUPPORTED_TYPE_OFFSET);
|
|
port_cap->link_state = (field & 0x80) >> 7;
|
|
port_cap->supported_port_types = field & 3;
|
|
port_cap->suggested_type = (field >> 3) & 1;
|
|
port_cap->default_sense = (field >> 4) & 1;
|
|
port_cap->dmfs_optimized_state = (field >> 5) & 1;
|
|
MLX4_GET(field, outbox, QUERY_PORT_MTU_OFFSET);
|
|
port_cap->ib_mtu = field & 0xf;
|
|
MLX4_GET(field, outbox, QUERY_PORT_WIDTH_OFFSET);
|
|
port_cap->max_port_width = field & 0xf;
|
|
MLX4_GET(field, outbox, QUERY_PORT_MAX_GID_PKEY_OFFSET);
|
|
port_cap->max_gids = 1 << (field >> 4);
|
|
port_cap->max_pkeys = 1 << (field & 0xf);
|
|
MLX4_GET(field, outbox, QUERY_PORT_MAX_VL_OFFSET);
|
|
port_cap->max_vl = field & 0xf;
|
|
port_cap->max_tc_eth = field >> 4;
|
|
MLX4_GET(field, outbox, QUERY_PORT_MAX_MACVLAN_OFFSET);
|
|
port_cap->log_max_macs = field & 0xf;
|
|
port_cap->log_max_vlans = field >> 4;
|
|
MLX4_GET(port_cap->eth_mtu, outbox, QUERY_PORT_ETH_MTU_OFFSET);
|
|
MLX4_GET(port_cap->def_mac, outbox, QUERY_PORT_MAC_OFFSET);
|
|
MLX4_GET(field32, outbox, QUERY_PORT_TRANS_VENDOR_OFFSET);
|
|
port_cap->trans_type = field32 >> 24;
|
|
port_cap->vendor_oui = field32 & 0xffffff;
|
|
MLX4_GET(port_cap->wavelength, outbox, QUERY_PORT_WAVELENGTH_OFFSET);
|
|
MLX4_GET(port_cap->trans_code, outbox, QUERY_PORT_TRANS_CODE_OFFSET);
|
|
}
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
#define DEV_CAP_EXT_2_FLAG_PFC_COUNTERS (1 << 28)
|
|
#define DEV_CAP_EXT_2_FLAG_VLAN_CONTROL (1 << 26)
|
|
#define DEV_CAP_EXT_2_FLAG_80_VFS (1 << 21)
|
|
#define DEV_CAP_EXT_2_FLAG_FSM (1 << 20)
|
|
|
|
int mlx4_QUERY_DEV_CAP_wrapper(struct mlx4_dev *dev, int slave,
|
|
struct mlx4_vhcr *vhcr,
|
|
struct mlx4_cmd_mailbox *inbox,
|
|
struct mlx4_cmd_mailbox *outbox,
|
|
struct mlx4_cmd_info *cmd)
|
|
{
|
|
u64 flags;
|
|
int err = 0;
|
|
u8 field;
|
|
u16 field16;
|
|
u32 bmme_flags, field32;
|
|
int real_port;
|
|
int slave_port;
|
|
int first_port;
|
|
struct mlx4_active_ports actv_ports;
|
|
|
|
err = mlx4_cmd_box(dev, 0, outbox->dma, 0, 0, MLX4_CMD_QUERY_DEV_CAP,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
return err;
|
|
|
|
disable_unsupported_roce_caps(outbox->buf);
|
|
/* add port mng change event capability and disable mw type 1
|
|
* unconditionally to slaves
|
|
*/
|
|
MLX4_GET(flags, outbox->buf, QUERY_DEV_CAP_EXT_FLAGS_OFFSET);
|
|
flags |= MLX4_DEV_CAP_FLAG_PORT_MNG_CHG_EV;
|
|
flags &= ~MLX4_DEV_CAP_FLAG_MEM_WINDOW;
|
|
actv_ports = mlx4_get_active_ports(dev, slave);
|
|
first_port = find_first_bit(actv_ports.ports, dev->caps.num_ports);
|
|
for (slave_port = 0, real_port = first_port;
|
|
real_port < first_port +
|
|
bitmap_weight(actv_ports.ports, dev->caps.num_ports);
|
|
++real_port, ++slave_port) {
|
|
if (flags & (MLX4_DEV_CAP_FLAG_WOL_PORT1 << real_port))
|
|
flags |= MLX4_DEV_CAP_FLAG_WOL_PORT1 << slave_port;
|
|
else
|
|
flags &= ~(MLX4_DEV_CAP_FLAG_WOL_PORT1 << slave_port);
|
|
}
|
|
for (; slave_port < dev->caps.num_ports; ++slave_port)
|
|
flags &= ~(MLX4_DEV_CAP_FLAG_WOL_PORT1 << slave_port);
|
|
|
|
/* Not exposing RSS IP fragments to guests */
|
|
flags &= ~MLX4_DEV_CAP_FLAG_RSS_IP_FRAG;
|
|
MLX4_PUT(outbox->buf, flags, QUERY_DEV_CAP_EXT_FLAGS_OFFSET);
|
|
|
|
MLX4_GET(field, outbox->buf, QUERY_DEV_CAP_VL_PORT_OFFSET);
|
|
field &= ~0x0F;
|
|
field |= bitmap_weight(actv_ports.ports, dev->caps.num_ports) & 0x0F;
|
|
MLX4_PUT(outbox->buf, field, QUERY_DEV_CAP_VL_PORT_OFFSET);
|
|
|
|
/* For guests, disable timestamp */
|
|
MLX4_GET(field, outbox->buf, QUERY_DEV_CAP_CQ_TS_SUPPORT_OFFSET);
|
|
field &= 0x7f;
|
|
MLX4_PUT(outbox->buf, field, QUERY_DEV_CAP_CQ_TS_SUPPORT_OFFSET);
|
|
|
|
/* For guests, disable vxlan tunneling and QoS support */
|
|
MLX4_GET(field, outbox->buf, QUERY_DEV_CAP_VXLAN);
|
|
field &= 0xd7;
|
|
MLX4_PUT(outbox->buf, field, QUERY_DEV_CAP_VXLAN);
|
|
|
|
/* For guests, disable port BEACON */
|
|
MLX4_GET(field, outbox->buf, QUERY_DEV_CAP_PORT_BEACON_OFFSET);
|
|
field &= 0x7f;
|
|
MLX4_PUT(outbox->buf, field, QUERY_DEV_CAP_PORT_BEACON_OFFSET);
|
|
|
|
/* For guests, report Blueflame disabled */
|
|
MLX4_GET(field, outbox->buf, QUERY_DEV_CAP_BF_OFFSET);
|
|
field &= 0x7f;
|
|
MLX4_PUT(outbox->buf, field, QUERY_DEV_CAP_BF_OFFSET);
|
|
|
|
/* For guests, disable mw type 2 and port remap*/
|
|
MLX4_GET(bmme_flags, outbox->buf, QUERY_DEV_CAP_BMME_FLAGS_OFFSET);
|
|
bmme_flags &= ~MLX4_BMME_FLAG_TYPE_2_WIN;
|
|
bmme_flags &= ~MLX4_FLAG_PORT_REMAP;
|
|
MLX4_PUT(outbox->buf, bmme_flags, QUERY_DEV_CAP_BMME_FLAGS_OFFSET);
|
|
|
|
/* turn off device-managed steering capability if not enabled */
|
|
if (dev->caps.steering_mode != MLX4_STEERING_MODE_DEVICE_MANAGED) {
|
|
MLX4_GET(field, outbox->buf,
|
|
QUERY_DEV_CAP_FLOW_STEERING_RANGE_EN_OFFSET);
|
|
field &= 0x7f;
|
|
MLX4_PUT(outbox->buf, field,
|
|
QUERY_DEV_CAP_FLOW_STEERING_RANGE_EN_OFFSET);
|
|
}
|
|
|
|
/* turn off ipoib managed steering for guests */
|
|
MLX4_GET(field, outbox->buf, QUERY_DEV_CAP_FLOW_STEERING_IPOIB_OFFSET);
|
|
field &= ~0x80;
|
|
MLX4_PUT(outbox->buf, field, QUERY_DEV_CAP_FLOW_STEERING_IPOIB_OFFSET);
|
|
|
|
/* turn off host side virt features (VST, FSM, etc) for guests */
|
|
MLX4_GET(field32, outbox->buf, QUERY_DEV_CAP_EXT_2_FLAGS_OFFSET);
|
|
field32 &= ~(DEV_CAP_EXT_2_FLAG_VLAN_CONTROL | DEV_CAP_EXT_2_FLAG_80_VFS |
|
|
DEV_CAP_EXT_2_FLAG_FSM | DEV_CAP_EXT_2_FLAG_PFC_COUNTERS);
|
|
MLX4_PUT(outbox->buf, field32, QUERY_DEV_CAP_EXT_2_FLAGS_OFFSET);
|
|
|
|
/* turn off QCN for guests */
|
|
MLX4_GET(field, outbox->buf, QUERY_DEV_CAP_ECN_QCN_VER_OFFSET);
|
|
field &= 0xfe;
|
|
MLX4_PUT(outbox->buf, field, QUERY_DEV_CAP_ECN_QCN_VER_OFFSET);
|
|
|
|
/* turn off QP max-rate limiting for guests */
|
|
field16 = 0;
|
|
MLX4_PUT(outbox->buf, field16, QUERY_DEV_CAP_QP_RATE_LIMIT_NUM_OFFSET);
|
|
|
|
/* turn off QoS per VF support for guests */
|
|
MLX4_GET(field, outbox->buf, QUERY_DEV_CAP_CQ_EQ_CACHE_LINE_STRIDE);
|
|
field &= 0xef;
|
|
MLX4_PUT(outbox->buf, field, QUERY_DEV_CAP_CQ_EQ_CACHE_LINE_STRIDE);
|
|
|
|
/* turn off ignore FCS feature for guests */
|
|
MLX4_GET(field, outbox->buf, QUERY_DEV_CAP_CONFIG_DEV_OFFSET);
|
|
field &= 0xfb;
|
|
MLX4_PUT(outbox->buf, field, QUERY_DEV_CAP_CONFIG_DEV_OFFSET);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void disable_unsupported_roce_caps(void *buf)
|
|
{
|
|
u32 flags;
|
|
|
|
MLX4_GET(flags, buf, QUERY_DEV_CAP_EXT_FLAGS_OFFSET);
|
|
flags &= ~(1UL << 31);
|
|
MLX4_PUT(buf, flags, QUERY_DEV_CAP_EXT_FLAGS_OFFSET);
|
|
MLX4_GET(flags, buf, QUERY_DEV_CAP_EXT_2_FLAGS_OFFSET);
|
|
flags &= ~(1UL << 24);
|
|
MLX4_PUT(buf, flags, QUERY_DEV_CAP_EXT_2_FLAGS_OFFSET);
|
|
MLX4_GET(flags, buf, QUERY_DEV_CAP_BMME_FLAGS_OFFSET);
|
|
flags &= ~(MLX4_FLAG_ROCE_V1_V2);
|
|
MLX4_PUT(buf, flags, QUERY_DEV_CAP_BMME_FLAGS_OFFSET);
|
|
}
|
|
|
|
int mlx4_QUERY_PORT_wrapper(struct mlx4_dev *dev, int slave,
|
|
struct mlx4_vhcr *vhcr,
|
|
struct mlx4_cmd_mailbox *inbox,
|
|
struct mlx4_cmd_mailbox *outbox,
|
|
struct mlx4_cmd_info *cmd)
|
|
{
|
|
struct mlx4_priv *priv = mlx4_priv(dev);
|
|
u64 def_mac;
|
|
u8 port_type;
|
|
u16 short_field;
|
|
int err;
|
|
int admin_link_state;
|
|
int port = mlx4_slave_convert_port(dev, slave,
|
|
vhcr->in_modifier & 0xFF);
|
|
|
|
#define MLX4_VF_PORT_NO_LINK_SENSE_MASK 0xE0
|
|
#define MLX4_PORT_LINK_UP_MASK 0x80
|
|
#define QUERY_PORT_CUR_MAX_PKEY_OFFSET 0x0c
|
|
#define QUERY_PORT_CUR_MAX_GID_OFFSET 0x0e
|
|
|
|
if (port < 0)
|
|
return -EINVAL;
|
|
|
|
/* Protect against untrusted guests: enforce that this is the
|
|
* QUERY_PORT general query.
|
|
*/
|
|
if (vhcr->op_modifier || vhcr->in_modifier & ~0xFF)
|
|
return -EINVAL;
|
|
|
|
vhcr->in_modifier = port;
|
|
|
|
err = mlx4_cmd_box(dev, 0, outbox->dma, vhcr->in_modifier, 0,
|
|
MLX4_CMD_QUERY_PORT, MLX4_CMD_TIME_CLASS_B,
|
|
MLX4_CMD_NATIVE);
|
|
|
|
if (!err && dev->caps.function != slave) {
|
|
def_mac = priv->mfunc.master.vf_oper[slave].vport[vhcr->in_modifier].state.mac;
|
|
MLX4_PUT(outbox->buf, def_mac, QUERY_PORT_MAC_OFFSET);
|
|
|
|
/* get port type - currently only eth is enabled */
|
|
MLX4_GET(port_type, outbox->buf,
|
|
QUERY_PORT_SUPPORTED_TYPE_OFFSET);
|
|
|
|
/* No link sensing allowed */
|
|
port_type &= MLX4_VF_PORT_NO_LINK_SENSE_MASK;
|
|
/* set port type to currently operating port type */
|
|
port_type |= (dev->caps.port_type[vhcr->in_modifier] & 0x3);
|
|
|
|
admin_link_state = priv->mfunc.master.vf_oper[slave].vport[vhcr->in_modifier].state.link_state;
|
|
if (IFLA_VF_LINK_STATE_ENABLE == admin_link_state)
|
|
port_type |= MLX4_PORT_LINK_UP_MASK;
|
|
else if (IFLA_VF_LINK_STATE_DISABLE == admin_link_state)
|
|
port_type &= ~MLX4_PORT_LINK_UP_MASK;
|
|
else if (IFLA_VF_LINK_STATE_AUTO == admin_link_state && mlx4_is_bonded(dev)) {
|
|
int other_port = (port == 1) ? 2 : 1;
|
|
struct mlx4_port_cap port_cap;
|
|
|
|
err = mlx4_QUERY_PORT(dev, other_port, &port_cap);
|
|
if (err)
|
|
goto out;
|
|
port_type |= (port_cap.link_state << 7);
|
|
}
|
|
|
|
MLX4_PUT(outbox->buf, port_type,
|
|
QUERY_PORT_SUPPORTED_TYPE_OFFSET);
|
|
|
|
if (dev->caps.port_type[vhcr->in_modifier] == MLX4_PORT_TYPE_ETH)
|
|
short_field = mlx4_get_slave_num_gids(dev, slave, port);
|
|
else
|
|
short_field = 1; /* slave max gids */
|
|
MLX4_PUT(outbox->buf, short_field,
|
|
QUERY_PORT_CUR_MAX_GID_OFFSET);
|
|
|
|
short_field = dev->caps.pkey_table_len[vhcr->in_modifier];
|
|
MLX4_PUT(outbox->buf, short_field,
|
|
QUERY_PORT_CUR_MAX_PKEY_OFFSET);
|
|
}
|
|
out:
|
|
return err;
|
|
}
|
|
|
|
int mlx4_get_slave_pkey_gid_tbl_len(struct mlx4_dev *dev, u8 port,
|
|
int *gid_tbl_len, int *pkey_tbl_len)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 *outbox;
|
|
u16 field;
|
|
int err;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, port, 0,
|
|
MLX4_CMD_QUERY_PORT, MLX4_CMD_TIME_CLASS_B,
|
|
MLX4_CMD_WRAPPED);
|
|
if (err)
|
|
goto out;
|
|
|
|
outbox = mailbox->buf;
|
|
|
|
MLX4_GET(field, outbox, QUERY_PORT_CUR_MAX_GID_OFFSET);
|
|
*gid_tbl_len = field;
|
|
|
|
MLX4_GET(field, outbox, QUERY_PORT_CUR_MAX_PKEY_OFFSET);
|
|
*pkey_tbl_len = field;
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
EXPORT_SYMBOL(mlx4_get_slave_pkey_gid_tbl_len);
|
|
|
|
int mlx4_map_cmd(struct mlx4_dev *dev, u16 op, struct mlx4_icm *icm, u64 virt)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
struct mlx4_icm_iter iter;
|
|
__be64 *pages;
|
|
int lg;
|
|
int nent = 0;
|
|
int i;
|
|
int err = 0;
|
|
int ts = 0, tc = 0;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
pages = mailbox->buf;
|
|
|
|
for (mlx4_icm_first(icm, &iter);
|
|
!mlx4_icm_last(&iter);
|
|
mlx4_icm_next(&iter)) {
|
|
/*
|
|
* We have to pass pages that are aligned to their
|
|
* size, so find the least significant 1 in the
|
|
* address or size and use that as our log2 size.
|
|
*/
|
|
lg = ffs(mlx4_icm_addr(&iter) | mlx4_icm_size(&iter)) - 1;
|
|
if (lg < MLX4_ICM_PAGE_SHIFT) {
|
|
mlx4_warn(dev, "Got FW area not aligned to %d (%llx/%lx)\n",
|
|
MLX4_ICM_PAGE_SIZE,
|
|
(unsigned long long) mlx4_icm_addr(&iter),
|
|
mlx4_icm_size(&iter));
|
|
err = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
for (i = 0; i < mlx4_icm_size(&iter) >> lg; ++i) {
|
|
if (virt != -1) {
|
|
pages[nent * 2] = cpu_to_be64(virt);
|
|
virt += 1ULL << lg;
|
|
}
|
|
|
|
pages[nent * 2 + 1] =
|
|
cpu_to_be64((mlx4_icm_addr(&iter) + (i << lg)) |
|
|
(lg - MLX4_ICM_PAGE_SHIFT));
|
|
ts += 1 << (lg - 10);
|
|
++tc;
|
|
|
|
if (++nent == MLX4_MAILBOX_SIZE / 16) {
|
|
err = mlx4_cmd(dev, mailbox->dma, nent, 0, op,
|
|
MLX4_CMD_TIME_CLASS_B,
|
|
MLX4_CMD_NATIVE);
|
|
if (err)
|
|
goto out;
|
|
nent = 0;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (nent)
|
|
err = mlx4_cmd(dev, mailbox->dma, nent, 0, op,
|
|
MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
goto out;
|
|
|
|
switch (op) {
|
|
case MLX4_CMD_MAP_FA:
|
|
mlx4_dbg(dev, "Mapped %d chunks/%d KB for FW\n", tc, ts);
|
|
break;
|
|
case MLX4_CMD_MAP_ICM_AUX:
|
|
mlx4_dbg(dev, "Mapped %d chunks/%d KB for ICM aux\n", tc, ts);
|
|
break;
|
|
case MLX4_CMD_MAP_ICM:
|
|
mlx4_dbg(dev, "Mapped %d chunks/%d KB at %llx for ICM\n",
|
|
tc, ts, (unsigned long long) virt - (ts << 10));
|
|
break;
|
|
}
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
int mlx4_MAP_FA(struct mlx4_dev *dev, struct mlx4_icm *icm)
|
|
{
|
|
return mlx4_map_cmd(dev, MLX4_CMD_MAP_FA, icm, -1);
|
|
}
|
|
|
|
int mlx4_UNMAP_FA(struct mlx4_dev *dev)
|
|
{
|
|
return mlx4_cmd(dev, 0, 0, 0, MLX4_CMD_UNMAP_FA,
|
|
MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE);
|
|
}
|
|
|
|
|
|
int mlx4_RUN_FW(struct mlx4_dev *dev)
|
|
{
|
|
return mlx4_cmd(dev, 0, 0, 0, MLX4_CMD_RUN_FW,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
}
|
|
|
|
int mlx4_QUERY_FW(struct mlx4_dev *dev)
|
|
{
|
|
struct mlx4_fw *fw = &mlx4_priv(dev)->fw;
|
|
struct mlx4_cmd *cmd = &mlx4_priv(dev)->cmd;
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 *outbox;
|
|
int err = 0;
|
|
u64 fw_ver;
|
|
u16 cmd_if_rev;
|
|
u8 lg;
|
|
|
|
#define QUERY_FW_OUT_SIZE 0x100
|
|
#define QUERY_FW_VER_OFFSET 0x00
|
|
#define QUERY_FW_PPF_ID 0x09
|
|
#define QUERY_FW_CMD_IF_REV_OFFSET 0x0a
|
|
#define QUERY_FW_MAX_CMD_OFFSET 0x0f
|
|
#define QUERY_FW_ERR_START_OFFSET 0x30
|
|
#define QUERY_FW_ERR_SIZE_OFFSET 0x38
|
|
#define QUERY_FW_ERR_BAR_OFFSET 0x3c
|
|
|
|
#define QUERY_FW_SIZE_OFFSET 0x00
|
|
#define QUERY_FW_CLR_INT_BASE_OFFSET 0x20
|
|
#define QUERY_FW_CLR_INT_BAR_OFFSET 0x28
|
|
|
|
#define QUERY_FW_COMM_BASE_OFFSET 0x40
|
|
#define QUERY_FW_COMM_BAR_OFFSET 0x48
|
|
|
|
#define QUERY_FW_CLOCK_OFFSET 0x50
|
|
#define QUERY_FW_CLOCK_BAR 0x58
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
outbox = mailbox->buf;
|
|
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, 0, 0, MLX4_CMD_QUERY_FW,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
goto out;
|
|
|
|
MLX4_GET(fw_ver, outbox, QUERY_FW_VER_OFFSET);
|
|
/*
|
|
* FW subminor version is at more significant bits than minor
|
|
* version, so swap here.
|
|
*/
|
|
dev->caps.fw_ver = (fw_ver & 0xffff00000000ull) |
|
|
((fw_ver & 0xffff0000ull) >> 16) |
|
|
((fw_ver & 0x0000ffffull) << 16);
|
|
|
|
MLX4_GET(lg, outbox, QUERY_FW_PPF_ID);
|
|
dev->caps.function = lg;
|
|
|
|
if (mlx4_is_slave(dev))
|
|
goto out;
|
|
|
|
|
|
MLX4_GET(cmd_if_rev, outbox, QUERY_FW_CMD_IF_REV_OFFSET);
|
|
if (cmd_if_rev < MLX4_COMMAND_INTERFACE_MIN_REV ||
|
|
cmd_if_rev > MLX4_COMMAND_INTERFACE_MAX_REV) {
|
|
mlx4_err(dev, "Installed FW has unsupported command interface revision %d\n",
|
|
cmd_if_rev);
|
|
mlx4_err(dev, "(Installed FW version is %d.%d.%03d)\n",
|
|
(int) (dev->caps.fw_ver >> 32),
|
|
(int) (dev->caps.fw_ver >> 16) & 0xffff,
|
|
(int) dev->caps.fw_ver & 0xffff);
|
|
mlx4_err(dev, "This driver version supports only revisions %d to %d\n",
|
|
MLX4_COMMAND_INTERFACE_MIN_REV, MLX4_COMMAND_INTERFACE_MAX_REV);
|
|
err = -ENODEV;
|
|
goto out;
|
|
}
|
|
|
|
if (cmd_if_rev < MLX4_COMMAND_INTERFACE_NEW_PORT_CMDS)
|
|
dev->flags |= MLX4_FLAG_OLD_PORT_CMDS;
|
|
|
|
MLX4_GET(lg, outbox, QUERY_FW_MAX_CMD_OFFSET);
|
|
cmd->max_cmds = 1 << lg;
|
|
|
|
mlx4_dbg(dev, "FW version %d.%d.%03d (cmd intf rev %d), max commands %d\n",
|
|
(int) (dev->caps.fw_ver >> 32),
|
|
(int) (dev->caps.fw_ver >> 16) & 0xffff,
|
|
(int) dev->caps.fw_ver & 0xffff,
|
|
cmd_if_rev, cmd->max_cmds);
|
|
|
|
MLX4_GET(fw->catas_offset, outbox, QUERY_FW_ERR_START_OFFSET);
|
|
MLX4_GET(fw->catas_size, outbox, QUERY_FW_ERR_SIZE_OFFSET);
|
|
MLX4_GET(fw->catas_bar, outbox, QUERY_FW_ERR_BAR_OFFSET);
|
|
fw->catas_bar = (fw->catas_bar >> 6) * 2;
|
|
|
|
mlx4_dbg(dev, "Catastrophic error buffer at 0x%llx, size 0x%x, BAR %d\n",
|
|
(unsigned long long) fw->catas_offset, fw->catas_size, fw->catas_bar);
|
|
|
|
MLX4_GET(fw->fw_pages, outbox, QUERY_FW_SIZE_OFFSET);
|
|
MLX4_GET(fw->clr_int_base, outbox, QUERY_FW_CLR_INT_BASE_OFFSET);
|
|
MLX4_GET(fw->clr_int_bar, outbox, QUERY_FW_CLR_INT_BAR_OFFSET);
|
|
fw->clr_int_bar = (fw->clr_int_bar >> 6) * 2;
|
|
|
|
MLX4_GET(fw->comm_base, outbox, QUERY_FW_COMM_BASE_OFFSET);
|
|
MLX4_GET(fw->comm_bar, outbox, QUERY_FW_COMM_BAR_OFFSET);
|
|
fw->comm_bar = (fw->comm_bar >> 6) * 2;
|
|
mlx4_dbg(dev, "Communication vector bar:%d offset:0x%llx\n",
|
|
fw->comm_bar, fw->comm_base);
|
|
mlx4_dbg(dev, "FW size %d KB\n", fw->fw_pages >> 2);
|
|
|
|
MLX4_GET(fw->clock_offset, outbox, QUERY_FW_CLOCK_OFFSET);
|
|
MLX4_GET(fw->clock_bar, outbox, QUERY_FW_CLOCK_BAR);
|
|
fw->clock_bar = (fw->clock_bar >> 6) * 2;
|
|
mlx4_dbg(dev, "Internal clock bar:%d offset:0x%llx\n",
|
|
fw->clock_bar, fw->clock_offset);
|
|
|
|
/*
|
|
* Round up number of system pages needed in case
|
|
* MLX4_ICM_PAGE_SIZE < PAGE_SIZE.
|
|
*/
|
|
fw->fw_pages =
|
|
ALIGN(fw->fw_pages, PAGE_SIZE / MLX4_ICM_PAGE_SIZE) >>
|
|
(PAGE_SHIFT - MLX4_ICM_PAGE_SHIFT);
|
|
|
|
mlx4_dbg(dev, "Clear int @ %llx, BAR %d\n",
|
|
(unsigned long long) fw->clr_int_base, fw->clr_int_bar);
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
int mlx4_QUERY_FW_wrapper(struct mlx4_dev *dev, int slave,
|
|
struct mlx4_vhcr *vhcr,
|
|
struct mlx4_cmd_mailbox *inbox,
|
|
struct mlx4_cmd_mailbox *outbox,
|
|
struct mlx4_cmd_info *cmd)
|
|
{
|
|
u8 *outbuf;
|
|
int err;
|
|
|
|
outbuf = outbox->buf;
|
|
err = mlx4_cmd_box(dev, 0, outbox->dma, 0, 0, MLX4_CMD_QUERY_FW,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
return err;
|
|
|
|
/* for slaves, set pci PPF ID to invalid and zero out everything
|
|
* else except FW version */
|
|
outbuf[0] = outbuf[1] = 0;
|
|
memset(&outbuf[8], 0, QUERY_FW_OUT_SIZE - 8);
|
|
outbuf[QUERY_FW_PPF_ID] = MLX4_INVALID_SLAVE_ID;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void get_board_id(void *vsd, char *board_id)
|
|
{
|
|
int i;
|
|
|
|
#define VSD_OFFSET_SIG1 0x00
|
|
#define VSD_OFFSET_SIG2 0xde
|
|
#define VSD_OFFSET_MLX_BOARD_ID 0xd0
|
|
#define VSD_OFFSET_TS_BOARD_ID 0x20
|
|
|
|
#define VSD_SIGNATURE_TOPSPIN 0x5ad
|
|
|
|
memset(board_id, 0, MLX4_BOARD_ID_LEN);
|
|
|
|
if (be16_to_cpup(vsd + VSD_OFFSET_SIG1) == VSD_SIGNATURE_TOPSPIN &&
|
|
be16_to_cpup(vsd + VSD_OFFSET_SIG2) == VSD_SIGNATURE_TOPSPIN) {
|
|
strlcpy(board_id, vsd + VSD_OFFSET_TS_BOARD_ID, MLX4_BOARD_ID_LEN);
|
|
} else {
|
|
/*
|
|
* The board ID is a string but the firmware byte
|
|
* swaps each 4-byte word before passing it back to
|
|
* us. Therefore we need to swab it before printing.
|
|
*/
|
|
u32 *bid_u32 = (u32 *)board_id;
|
|
|
|
for (i = 0; i < 4; ++i) {
|
|
u32 *addr;
|
|
u32 val;
|
|
|
|
addr = (u32 *) (vsd + VSD_OFFSET_MLX_BOARD_ID + i * 4);
|
|
val = get_unaligned(addr);
|
|
val = swab32(val);
|
|
put_unaligned(val, &bid_u32[i]);
|
|
}
|
|
}
|
|
}
|
|
|
|
int mlx4_QUERY_ADAPTER(struct mlx4_dev *dev, struct mlx4_adapter *adapter)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 *outbox;
|
|
int err;
|
|
|
|
#define QUERY_ADAPTER_OUT_SIZE 0x100
|
|
#define QUERY_ADAPTER_INTA_PIN_OFFSET 0x10
|
|
#define QUERY_ADAPTER_VSD_OFFSET 0x20
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
outbox = mailbox->buf;
|
|
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, 0, 0, MLX4_CMD_QUERY_ADAPTER,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
goto out;
|
|
|
|
MLX4_GET(adapter->inta_pin, outbox, QUERY_ADAPTER_INTA_PIN_OFFSET);
|
|
|
|
get_board_id(outbox + QUERY_ADAPTER_VSD_OFFSET / 4,
|
|
adapter->board_id);
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
int mlx4_INIT_HCA(struct mlx4_dev *dev, struct mlx4_init_hca_param *param)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
__be32 *inbox;
|
|
int err;
|
|
static const u8 a0_dmfs_hw_steering[] = {
|
|
[MLX4_STEERING_DMFS_A0_DEFAULT] = 0,
|
|
[MLX4_STEERING_DMFS_A0_DYNAMIC] = 1,
|
|
[MLX4_STEERING_DMFS_A0_STATIC] = 2,
|
|
[MLX4_STEERING_DMFS_A0_DISABLE] = 3
|
|
};
|
|
|
|
#define INIT_HCA_IN_SIZE 0x200
|
|
#define INIT_HCA_VERSION_OFFSET 0x000
|
|
#define INIT_HCA_VERSION 2
|
|
#define INIT_HCA_VXLAN_OFFSET 0x0c
|
|
#define INIT_HCA_CACHELINE_SZ_OFFSET 0x0e
|
|
#define INIT_HCA_FLAGS_OFFSET 0x014
|
|
#define INIT_HCA_RECOVERABLE_ERROR_EVENT_OFFSET 0x018
|
|
#define INIT_HCA_QPC_OFFSET 0x020
|
|
#define INIT_HCA_QPC_BASE_OFFSET (INIT_HCA_QPC_OFFSET + 0x10)
|
|
#define INIT_HCA_LOG_QP_OFFSET (INIT_HCA_QPC_OFFSET + 0x17)
|
|
#define INIT_HCA_SRQC_BASE_OFFSET (INIT_HCA_QPC_OFFSET + 0x28)
|
|
#define INIT_HCA_LOG_SRQ_OFFSET (INIT_HCA_QPC_OFFSET + 0x2f)
|
|
#define INIT_HCA_CQC_BASE_OFFSET (INIT_HCA_QPC_OFFSET + 0x30)
|
|
#define INIT_HCA_LOG_CQ_OFFSET (INIT_HCA_QPC_OFFSET + 0x37)
|
|
#define INIT_HCA_EQE_CQE_OFFSETS (INIT_HCA_QPC_OFFSET + 0x38)
|
|
#define INIT_HCA_EQE_CQE_STRIDE_OFFSET (INIT_HCA_QPC_OFFSET + 0x3b)
|
|
#define INIT_HCA_ALTC_BASE_OFFSET (INIT_HCA_QPC_OFFSET + 0x40)
|
|
#define INIT_HCA_AUXC_BASE_OFFSET (INIT_HCA_QPC_OFFSET + 0x50)
|
|
#define INIT_HCA_EQC_BASE_OFFSET (INIT_HCA_QPC_OFFSET + 0x60)
|
|
#define INIT_HCA_LOG_EQ_OFFSET (INIT_HCA_QPC_OFFSET + 0x67)
|
|
#define INIT_HCA_NUM_SYS_EQS_OFFSET (INIT_HCA_QPC_OFFSET + 0x6a)
|
|
#define INIT_HCA_RDMARC_BASE_OFFSET (INIT_HCA_QPC_OFFSET + 0x70)
|
|
#define INIT_HCA_LOG_RD_OFFSET (INIT_HCA_QPC_OFFSET + 0x77)
|
|
#define INIT_HCA_MCAST_OFFSET 0x0c0
|
|
#define INIT_HCA_MC_BASE_OFFSET (INIT_HCA_MCAST_OFFSET + 0x00)
|
|
#define INIT_HCA_LOG_MC_ENTRY_SZ_OFFSET (INIT_HCA_MCAST_OFFSET + 0x12)
|
|
#define INIT_HCA_LOG_MC_HASH_SZ_OFFSET (INIT_HCA_MCAST_OFFSET + 0x16)
|
|
#define INIT_HCA_UC_STEERING_OFFSET (INIT_HCA_MCAST_OFFSET + 0x18)
|
|
#define INIT_HCA_LOG_MC_TABLE_SZ_OFFSET (INIT_HCA_MCAST_OFFSET + 0x1b)
|
|
#define INIT_HCA_DEVICE_MANAGED_FLOW_STEERING_EN 0x6
|
|
#define INIT_HCA_DRIVER_VERSION_OFFSET 0x140
|
|
#define INIT_HCA_DRIVER_VERSION_SZ 0x40
|
|
#define INIT_HCA_FS_PARAM_OFFSET 0x1d0
|
|
#define INIT_HCA_FS_BASE_OFFSET (INIT_HCA_FS_PARAM_OFFSET + 0x00)
|
|
#define INIT_HCA_FS_LOG_ENTRY_SZ_OFFSET (INIT_HCA_FS_PARAM_OFFSET + 0x12)
|
|
#define INIT_HCA_FS_A0_OFFSET (INIT_HCA_FS_PARAM_OFFSET + 0x18)
|
|
#define INIT_HCA_FS_LOG_TABLE_SZ_OFFSET (INIT_HCA_FS_PARAM_OFFSET + 0x1b)
|
|
#define INIT_HCA_FS_ETH_BITS_OFFSET (INIT_HCA_FS_PARAM_OFFSET + 0x21)
|
|
#define INIT_HCA_FS_ETH_NUM_ADDRS_OFFSET (INIT_HCA_FS_PARAM_OFFSET + 0x22)
|
|
#define INIT_HCA_FS_IB_BITS_OFFSET (INIT_HCA_FS_PARAM_OFFSET + 0x25)
|
|
#define INIT_HCA_FS_IB_NUM_ADDRS_OFFSET (INIT_HCA_FS_PARAM_OFFSET + 0x26)
|
|
#define INIT_HCA_TPT_OFFSET 0x0f0
|
|
#define INIT_HCA_DMPT_BASE_OFFSET (INIT_HCA_TPT_OFFSET + 0x00)
|
|
#define INIT_HCA_TPT_MW_OFFSET (INIT_HCA_TPT_OFFSET + 0x08)
|
|
#define INIT_HCA_LOG_MPT_SZ_OFFSET (INIT_HCA_TPT_OFFSET + 0x0b)
|
|
#define INIT_HCA_MTT_BASE_OFFSET (INIT_HCA_TPT_OFFSET + 0x10)
|
|
#define INIT_HCA_CMPT_BASE_OFFSET (INIT_HCA_TPT_OFFSET + 0x18)
|
|
#define INIT_HCA_UAR_OFFSET 0x120
|
|
#define INIT_HCA_LOG_UAR_SZ_OFFSET (INIT_HCA_UAR_OFFSET + 0x0a)
|
|
#define INIT_HCA_UAR_PAGE_SZ_OFFSET (INIT_HCA_UAR_OFFSET + 0x0b)
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
inbox = mailbox->buf;
|
|
|
|
*((u8 *) mailbox->buf + INIT_HCA_VERSION_OFFSET) = INIT_HCA_VERSION;
|
|
|
|
*((u8 *) mailbox->buf + INIT_HCA_CACHELINE_SZ_OFFSET) =
|
|
((ilog2(cache_line_size()) - 4) << 5) | (1 << 4);
|
|
|
|
#if defined(__LITTLE_ENDIAN)
|
|
*(inbox + INIT_HCA_FLAGS_OFFSET / 4) &= ~cpu_to_be32(1 << 1);
|
|
#elif defined(__BIG_ENDIAN)
|
|
*(inbox + INIT_HCA_FLAGS_OFFSET / 4) |= cpu_to_be32(1 << 1);
|
|
#else
|
|
#error Host endianness not defined
|
|
#endif
|
|
/* Check port for UD address vector: */
|
|
*(inbox + INIT_HCA_FLAGS_OFFSET / 4) |= cpu_to_be32(1);
|
|
|
|
/* Enable IPoIB checksumming if we can: */
|
|
if (dev->caps.flags & MLX4_DEV_CAP_FLAG_IPOIB_CSUM)
|
|
*(inbox + INIT_HCA_FLAGS_OFFSET / 4) |= cpu_to_be32(1 << 3);
|
|
|
|
/* Enable QoS support if module parameter set */
|
|
if (dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_ETS_CFG && enable_qos)
|
|
*(inbox + INIT_HCA_FLAGS_OFFSET / 4) |= cpu_to_be32(1 << 2);
|
|
|
|
/* enable counters */
|
|
if (dev->caps.flags & MLX4_DEV_CAP_FLAG_COUNTERS)
|
|
*(inbox + INIT_HCA_FLAGS_OFFSET / 4) |= cpu_to_be32(1 << 4);
|
|
|
|
/* Enable RSS spread to fragmented IP packets when supported */
|
|
if (dev->caps.flags & MLX4_DEV_CAP_FLAG_RSS_IP_FRAG)
|
|
*(inbox + INIT_HCA_FLAGS_OFFSET / 4) |= cpu_to_be32(1 << 13);
|
|
|
|
/* CX3 is capable of extending CQEs/EQEs from 32 to 64 bytes */
|
|
if (dev->caps.flags & MLX4_DEV_CAP_FLAG_64B_EQE) {
|
|
*(inbox + INIT_HCA_EQE_CQE_OFFSETS / 4) |= cpu_to_be32(1 << 29);
|
|
dev->caps.eqe_size = 64;
|
|
dev->caps.eqe_factor = 1;
|
|
} else {
|
|
dev->caps.eqe_size = 32;
|
|
dev->caps.eqe_factor = 0;
|
|
}
|
|
|
|
if (dev->caps.flags & MLX4_DEV_CAP_FLAG_64B_CQE) {
|
|
*(inbox + INIT_HCA_EQE_CQE_OFFSETS / 4) |= cpu_to_be32(1 << 30);
|
|
dev->caps.cqe_size = 64;
|
|
dev->caps.userspace_caps |= MLX4_USER_DEV_CAP_LARGE_CQE;
|
|
} else {
|
|
dev->caps.cqe_size = 32;
|
|
}
|
|
|
|
/* CX3 is capable of extending CQEs\EQEs to strides larger than 64B */
|
|
if ((dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_EQE_STRIDE) &&
|
|
(dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_CQE_STRIDE)) {
|
|
dev->caps.eqe_size = cache_line_size();
|
|
dev->caps.cqe_size = cache_line_size();
|
|
dev->caps.eqe_factor = 0;
|
|
MLX4_PUT(inbox, (u8)((ilog2(dev->caps.eqe_size) - 5) << 4 |
|
|
(ilog2(dev->caps.eqe_size) - 5)),
|
|
INIT_HCA_EQE_CQE_STRIDE_OFFSET);
|
|
|
|
/* User still need to know to support CQE > 32B */
|
|
dev->caps.userspace_caps |= MLX4_USER_DEV_CAP_LARGE_CQE;
|
|
}
|
|
|
|
if (dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_RECOVERABLE_ERROR_EVENT)
|
|
*(inbox + INIT_HCA_RECOVERABLE_ERROR_EVENT_OFFSET / 4) |= cpu_to_be32(1 << 31);
|
|
|
|
if (dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_DRIVER_VERSION_TO_FW) {
|
|
u8 *dst = (u8 *)(inbox + INIT_HCA_DRIVER_VERSION_OFFSET / 4);
|
|
|
|
strncpy(dst, DRV_NAME_FOR_FW, INIT_HCA_DRIVER_VERSION_SZ - 1);
|
|
mlx4_dbg(dev, "Reporting Driver Version to FW: %s\n", dst);
|
|
}
|
|
|
|
/* QPC/EEC/CQC/EQC/RDMARC attributes */
|
|
|
|
MLX4_PUT(inbox, param->qpc_base, INIT_HCA_QPC_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->log_num_qps, INIT_HCA_LOG_QP_OFFSET);
|
|
MLX4_PUT(inbox, param->srqc_base, INIT_HCA_SRQC_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->log_num_srqs, INIT_HCA_LOG_SRQ_OFFSET);
|
|
MLX4_PUT(inbox, param->cqc_base, INIT_HCA_CQC_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->log_num_cqs, INIT_HCA_LOG_CQ_OFFSET);
|
|
MLX4_PUT(inbox, param->altc_base, INIT_HCA_ALTC_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->auxc_base, INIT_HCA_AUXC_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->eqc_base, INIT_HCA_EQC_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->log_num_eqs, INIT_HCA_LOG_EQ_OFFSET);
|
|
MLX4_PUT(inbox, param->num_sys_eqs, INIT_HCA_NUM_SYS_EQS_OFFSET);
|
|
MLX4_PUT(inbox, param->rdmarc_base, INIT_HCA_RDMARC_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->log_rd_per_qp, INIT_HCA_LOG_RD_OFFSET);
|
|
|
|
/* steering attributes */
|
|
if (dev->caps.steering_mode ==
|
|
MLX4_STEERING_MODE_DEVICE_MANAGED) {
|
|
*(inbox + INIT_HCA_FLAGS_OFFSET / 4) |=
|
|
cpu_to_be32(1 <<
|
|
INIT_HCA_DEVICE_MANAGED_FLOW_STEERING_EN);
|
|
|
|
MLX4_PUT(inbox, param->mc_base, INIT_HCA_FS_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->log_mc_entry_sz,
|
|
INIT_HCA_FS_LOG_ENTRY_SZ_OFFSET);
|
|
MLX4_PUT(inbox, param->log_mc_table_sz,
|
|
INIT_HCA_FS_LOG_TABLE_SZ_OFFSET);
|
|
/* Enable Ethernet flow steering
|
|
* with udp unicast and tcp unicast
|
|
*/
|
|
if (dev->caps.dmfs_high_steer_mode !=
|
|
MLX4_STEERING_DMFS_A0_STATIC)
|
|
MLX4_PUT(inbox,
|
|
(u8)(MLX4_FS_UDP_UC_EN | MLX4_FS_TCP_UC_EN),
|
|
INIT_HCA_FS_ETH_BITS_OFFSET);
|
|
MLX4_PUT(inbox, (u16) MLX4_FS_NUM_OF_L2_ADDR,
|
|
INIT_HCA_FS_ETH_NUM_ADDRS_OFFSET);
|
|
/* Enable IPoIB flow steering
|
|
* with udp unicast and tcp unicast
|
|
*/
|
|
MLX4_PUT(inbox, (u8) (MLX4_FS_UDP_UC_EN | MLX4_FS_TCP_UC_EN),
|
|
INIT_HCA_FS_IB_BITS_OFFSET);
|
|
MLX4_PUT(inbox, (u16) MLX4_FS_NUM_OF_L2_ADDR,
|
|
INIT_HCA_FS_IB_NUM_ADDRS_OFFSET);
|
|
|
|
if (dev->caps.dmfs_high_steer_mode !=
|
|
MLX4_STEERING_DMFS_A0_NOT_SUPPORTED)
|
|
MLX4_PUT(inbox,
|
|
((u8)(a0_dmfs_hw_steering[dev->caps.dmfs_high_steer_mode]
|
|
<< 6)),
|
|
INIT_HCA_FS_A0_OFFSET);
|
|
} else {
|
|
MLX4_PUT(inbox, param->mc_base, INIT_HCA_MC_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->log_mc_entry_sz,
|
|
INIT_HCA_LOG_MC_ENTRY_SZ_OFFSET);
|
|
MLX4_PUT(inbox, param->log_mc_hash_sz,
|
|
INIT_HCA_LOG_MC_HASH_SZ_OFFSET);
|
|
MLX4_PUT(inbox, param->log_mc_table_sz,
|
|
INIT_HCA_LOG_MC_TABLE_SZ_OFFSET);
|
|
if (dev->caps.steering_mode == MLX4_STEERING_MODE_B0)
|
|
MLX4_PUT(inbox, (u8) (1 << 3),
|
|
INIT_HCA_UC_STEERING_OFFSET);
|
|
}
|
|
|
|
/* TPT attributes */
|
|
|
|
MLX4_PUT(inbox, param->dmpt_base, INIT_HCA_DMPT_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->mw_enabled, INIT_HCA_TPT_MW_OFFSET);
|
|
MLX4_PUT(inbox, param->log_mpt_sz, INIT_HCA_LOG_MPT_SZ_OFFSET);
|
|
MLX4_PUT(inbox, param->mtt_base, INIT_HCA_MTT_BASE_OFFSET);
|
|
MLX4_PUT(inbox, param->cmpt_base, INIT_HCA_CMPT_BASE_OFFSET);
|
|
|
|
/* UAR attributes */
|
|
|
|
MLX4_PUT(inbox, param->uar_page_sz, INIT_HCA_UAR_PAGE_SZ_OFFSET);
|
|
MLX4_PUT(inbox, param->log_uar_sz, INIT_HCA_LOG_UAR_SZ_OFFSET);
|
|
|
|
/* set parser VXLAN attributes */
|
|
if (dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_VXLAN_OFFLOADS) {
|
|
u8 parser_params = 0;
|
|
MLX4_PUT(inbox, parser_params, INIT_HCA_VXLAN_OFFSET);
|
|
}
|
|
|
|
err = mlx4_cmd(dev, mailbox->dma, 0, 0, MLX4_CMD_INIT_HCA,
|
|
MLX4_CMD_TIME_CLASS_C, MLX4_CMD_NATIVE);
|
|
|
|
if (err)
|
|
mlx4_err(dev, "INIT_HCA returns %d\n", err);
|
|
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
int mlx4_QUERY_HCA(struct mlx4_dev *dev,
|
|
struct mlx4_init_hca_param *param)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
__be32 *outbox;
|
|
u32 dword_field;
|
|
int err;
|
|
u8 byte_field;
|
|
static const u8 a0_dmfs_query_hw_steering[] = {
|
|
[0] = MLX4_STEERING_DMFS_A0_DEFAULT,
|
|
[1] = MLX4_STEERING_DMFS_A0_DYNAMIC,
|
|
[2] = MLX4_STEERING_DMFS_A0_STATIC,
|
|
[3] = MLX4_STEERING_DMFS_A0_DISABLE
|
|
};
|
|
|
|
#define QUERY_HCA_GLOBAL_CAPS_OFFSET 0x04
|
|
#define QUERY_HCA_CORE_CLOCK_OFFSET 0x0c
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
outbox = mailbox->buf;
|
|
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, 0, 0,
|
|
MLX4_CMD_QUERY_HCA,
|
|
MLX4_CMD_TIME_CLASS_B,
|
|
!mlx4_is_slave(dev));
|
|
if (err)
|
|
goto out;
|
|
|
|
MLX4_GET(param->global_caps, outbox, QUERY_HCA_GLOBAL_CAPS_OFFSET);
|
|
MLX4_GET(param->hca_core_clock, outbox, QUERY_HCA_CORE_CLOCK_OFFSET);
|
|
|
|
/* QPC/EEC/CQC/EQC/RDMARC attributes */
|
|
|
|
MLX4_GET(param->qpc_base, outbox, INIT_HCA_QPC_BASE_OFFSET);
|
|
MLX4_GET(param->log_num_qps, outbox, INIT_HCA_LOG_QP_OFFSET);
|
|
MLX4_GET(param->srqc_base, outbox, INIT_HCA_SRQC_BASE_OFFSET);
|
|
MLX4_GET(param->log_num_srqs, outbox, INIT_HCA_LOG_SRQ_OFFSET);
|
|
MLX4_GET(param->cqc_base, outbox, INIT_HCA_CQC_BASE_OFFSET);
|
|
MLX4_GET(param->log_num_cqs, outbox, INIT_HCA_LOG_CQ_OFFSET);
|
|
MLX4_GET(param->altc_base, outbox, INIT_HCA_ALTC_BASE_OFFSET);
|
|
MLX4_GET(param->auxc_base, outbox, INIT_HCA_AUXC_BASE_OFFSET);
|
|
MLX4_GET(param->eqc_base, outbox, INIT_HCA_EQC_BASE_OFFSET);
|
|
MLX4_GET(param->log_num_eqs, outbox, INIT_HCA_LOG_EQ_OFFSET);
|
|
MLX4_GET(param->num_sys_eqs, outbox, INIT_HCA_NUM_SYS_EQS_OFFSET);
|
|
MLX4_GET(param->rdmarc_base, outbox, INIT_HCA_RDMARC_BASE_OFFSET);
|
|
MLX4_GET(param->log_rd_per_qp, outbox, INIT_HCA_LOG_RD_OFFSET);
|
|
|
|
MLX4_GET(dword_field, outbox, INIT_HCA_FLAGS_OFFSET);
|
|
if (dword_field & (1 << INIT_HCA_DEVICE_MANAGED_FLOW_STEERING_EN)) {
|
|
param->steering_mode = MLX4_STEERING_MODE_DEVICE_MANAGED;
|
|
} else {
|
|
MLX4_GET(byte_field, outbox, INIT_HCA_UC_STEERING_OFFSET);
|
|
if (byte_field & 0x8)
|
|
param->steering_mode = MLX4_STEERING_MODE_B0;
|
|
else
|
|
param->steering_mode = MLX4_STEERING_MODE_A0;
|
|
}
|
|
|
|
if (dword_field & (1 << 13))
|
|
param->rss_ip_frags = 1;
|
|
|
|
/* steering attributes */
|
|
if (param->steering_mode == MLX4_STEERING_MODE_DEVICE_MANAGED) {
|
|
MLX4_GET(param->mc_base, outbox, INIT_HCA_FS_BASE_OFFSET);
|
|
MLX4_GET(param->log_mc_entry_sz, outbox,
|
|
INIT_HCA_FS_LOG_ENTRY_SZ_OFFSET);
|
|
MLX4_GET(param->log_mc_table_sz, outbox,
|
|
INIT_HCA_FS_LOG_TABLE_SZ_OFFSET);
|
|
MLX4_GET(byte_field, outbox,
|
|
INIT_HCA_FS_A0_OFFSET);
|
|
param->dmfs_high_steer_mode =
|
|
a0_dmfs_query_hw_steering[(byte_field >> 6) & 3];
|
|
} else {
|
|
MLX4_GET(param->mc_base, outbox, INIT_HCA_MC_BASE_OFFSET);
|
|
MLX4_GET(param->log_mc_entry_sz, outbox,
|
|
INIT_HCA_LOG_MC_ENTRY_SZ_OFFSET);
|
|
MLX4_GET(param->log_mc_hash_sz, outbox,
|
|
INIT_HCA_LOG_MC_HASH_SZ_OFFSET);
|
|
MLX4_GET(param->log_mc_table_sz, outbox,
|
|
INIT_HCA_LOG_MC_TABLE_SZ_OFFSET);
|
|
}
|
|
|
|
/* CX3 is capable of extending CQEs/EQEs from 32 to 64 bytes */
|
|
MLX4_GET(byte_field, outbox, INIT_HCA_EQE_CQE_OFFSETS);
|
|
if (byte_field & 0x20) /* 64-bytes eqe enabled */
|
|
param->dev_cap_enabled |= MLX4_DEV_CAP_64B_EQE_ENABLED;
|
|
if (byte_field & 0x40) /* 64-bytes cqe enabled */
|
|
param->dev_cap_enabled |= MLX4_DEV_CAP_64B_CQE_ENABLED;
|
|
|
|
/* CX3 is capable of extending CQEs\EQEs to strides larger than 64B */
|
|
MLX4_GET(byte_field, outbox, INIT_HCA_EQE_CQE_STRIDE_OFFSET);
|
|
if (byte_field) {
|
|
param->dev_cap_enabled |= MLX4_DEV_CAP_EQE_STRIDE_ENABLED;
|
|
param->dev_cap_enabled |= MLX4_DEV_CAP_CQE_STRIDE_ENABLED;
|
|
param->cqe_size = 1 << ((byte_field &
|
|
MLX4_CQE_SIZE_MASK_STRIDE) + 5);
|
|
param->eqe_size = 1 << (((byte_field &
|
|
MLX4_EQE_SIZE_MASK_STRIDE) >> 4) + 5);
|
|
}
|
|
|
|
/* TPT attributes */
|
|
|
|
MLX4_GET(param->dmpt_base, outbox, INIT_HCA_DMPT_BASE_OFFSET);
|
|
MLX4_GET(param->mw_enabled, outbox, INIT_HCA_TPT_MW_OFFSET);
|
|
MLX4_GET(param->log_mpt_sz, outbox, INIT_HCA_LOG_MPT_SZ_OFFSET);
|
|
MLX4_GET(param->mtt_base, outbox, INIT_HCA_MTT_BASE_OFFSET);
|
|
MLX4_GET(param->cmpt_base, outbox, INIT_HCA_CMPT_BASE_OFFSET);
|
|
|
|
/* UAR attributes */
|
|
|
|
MLX4_GET(param->uar_page_sz, outbox, INIT_HCA_UAR_PAGE_SZ_OFFSET);
|
|
MLX4_GET(param->log_uar_sz, outbox, INIT_HCA_LOG_UAR_SZ_OFFSET);
|
|
|
|
/* phv_check enable */
|
|
MLX4_GET(byte_field, outbox, INIT_HCA_CACHELINE_SZ_OFFSET);
|
|
if (byte_field & 0x2)
|
|
param->phv_check_en = 1;
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_hca_core_clock_update(struct mlx4_dev *dev)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
__be32 *outbox;
|
|
int err;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox)) {
|
|
mlx4_warn(dev, "hca_core_clock mailbox allocation failed\n");
|
|
return PTR_ERR(mailbox);
|
|
}
|
|
outbox = mailbox->buf;
|
|
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, 0, 0,
|
|
MLX4_CMD_QUERY_HCA,
|
|
MLX4_CMD_TIME_CLASS_B,
|
|
!mlx4_is_slave(dev));
|
|
if (err) {
|
|
mlx4_warn(dev, "hca_core_clock update failed\n");
|
|
goto out;
|
|
}
|
|
|
|
MLX4_GET(dev->caps.hca_core_clock, outbox, QUERY_HCA_CORE_CLOCK_OFFSET);
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
|
|
return err;
|
|
}
|
|
|
|
/* for IB-type ports only in SRIOV mode. Checks that both proxy QP0
|
|
* and real QP0 are active, so that the paravirtualized QP0 is ready
|
|
* to operate */
|
|
static int check_qp0_state(struct mlx4_dev *dev, int function, int port)
|
|
{
|
|
struct mlx4_priv *priv = mlx4_priv(dev);
|
|
/* irrelevant if not infiniband */
|
|
if (priv->mfunc.master.qp0_state[port].proxy_qp0_active &&
|
|
priv->mfunc.master.qp0_state[port].qp0_active)
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
int mlx4_INIT_PORT_wrapper(struct mlx4_dev *dev, int slave,
|
|
struct mlx4_vhcr *vhcr,
|
|
struct mlx4_cmd_mailbox *inbox,
|
|
struct mlx4_cmd_mailbox *outbox,
|
|
struct mlx4_cmd_info *cmd)
|
|
{
|
|
struct mlx4_priv *priv = mlx4_priv(dev);
|
|
int port = mlx4_slave_convert_port(dev, slave, vhcr->in_modifier);
|
|
int err;
|
|
|
|
if (port < 0)
|
|
return -EINVAL;
|
|
|
|
if (priv->mfunc.master.slave_state[slave].init_port_mask & (1 << port))
|
|
return 0;
|
|
|
|
if (dev->caps.port_mask[port] != MLX4_PORT_TYPE_IB) {
|
|
/* Enable port only if it was previously disabled */
|
|
if (!priv->mfunc.master.init_port_ref[port]) {
|
|
err = mlx4_cmd(dev, 0, port, 0, MLX4_CMD_INIT_PORT,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
return err;
|
|
}
|
|
priv->mfunc.master.slave_state[slave].init_port_mask |= (1 << port);
|
|
} else {
|
|
if (slave == mlx4_master_func_num(dev)) {
|
|
if (check_qp0_state(dev, slave, port) &&
|
|
!priv->mfunc.master.qp0_state[port].port_active) {
|
|
err = mlx4_cmd(dev, 0, port, 0, MLX4_CMD_INIT_PORT,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
return err;
|
|
priv->mfunc.master.qp0_state[port].port_active = 1;
|
|
priv->mfunc.master.slave_state[slave].init_port_mask |= (1 << port);
|
|
}
|
|
} else
|
|
priv->mfunc.master.slave_state[slave].init_port_mask |= (1 << port);
|
|
}
|
|
++priv->mfunc.master.init_port_ref[port];
|
|
return 0;
|
|
}
|
|
|
|
int mlx4_INIT_PORT(struct mlx4_dev *dev, int port)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 *inbox;
|
|
int err;
|
|
u32 flags;
|
|
u16 field;
|
|
|
|
if (dev->flags & MLX4_FLAG_OLD_PORT_CMDS) {
|
|
#define INIT_PORT_IN_SIZE 256
|
|
#define INIT_PORT_FLAGS_OFFSET 0x00
|
|
#define INIT_PORT_FLAG_SIG (1 << 18)
|
|
#define INIT_PORT_FLAG_NG (1 << 17)
|
|
#define INIT_PORT_FLAG_G0 (1 << 16)
|
|
#define INIT_PORT_VL_SHIFT 4
|
|
#define INIT_PORT_PORT_WIDTH_SHIFT 8
|
|
#define INIT_PORT_MTU_OFFSET 0x04
|
|
#define INIT_PORT_MAX_GID_OFFSET 0x06
|
|
#define INIT_PORT_MAX_PKEY_OFFSET 0x0a
|
|
#define INIT_PORT_GUID0_OFFSET 0x10
|
|
#define INIT_PORT_NODE_GUID_OFFSET 0x18
|
|
#define INIT_PORT_SI_GUID_OFFSET 0x20
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
inbox = mailbox->buf;
|
|
|
|
flags = 0;
|
|
flags |= (dev->caps.vl_cap[port] & 0xf) << INIT_PORT_VL_SHIFT;
|
|
flags |= (dev->caps.port_width_cap[port] & 0xf) << INIT_PORT_PORT_WIDTH_SHIFT;
|
|
MLX4_PUT(inbox, flags, INIT_PORT_FLAGS_OFFSET);
|
|
|
|
field = 128 << dev->caps.ib_mtu_cap[port];
|
|
MLX4_PUT(inbox, field, INIT_PORT_MTU_OFFSET);
|
|
field = dev->caps.gid_table_len[port];
|
|
MLX4_PUT(inbox, field, INIT_PORT_MAX_GID_OFFSET);
|
|
field = dev->caps.pkey_table_len[port];
|
|
MLX4_PUT(inbox, field, INIT_PORT_MAX_PKEY_OFFSET);
|
|
|
|
err = mlx4_cmd(dev, mailbox->dma, port, 0, MLX4_CMD_INIT_PORT,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
} else
|
|
err = mlx4_cmd(dev, 0, port, 0, MLX4_CMD_INIT_PORT,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED);
|
|
|
|
if (!err)
|
|
mlx4_hca_core_clock_update(dev);
|
|
|
|
return err;
|
|
}
|
|
EXPORT_SYMBOL_GPL(mlx4_INIT_PORT);
|
|
|
|
int mlx4_CLOSE_PORT_wrapper(struct mlx4_dev *dev, int slave,
|
|
struct mlx4_vhcr *vhcr,
|
|
struct mlx4_cmd_mailbox *inbox,
|
|
struct mlx4_cmd_mailbox *outbox,
|
|
struct mlx4_cmd_info *cmd)
|
|
{
|
|
struct mlx4_priv *priv = mlx4_priv(dev);
|
|
int port = mlx4_slave_convert_port(dev, slave, vhcr->in_modifier);
|
|
int err;
|
|
|
|
if (port < 0)
|
|
return -EINVAL;
|
|
|
|
if (!(priv->mfunc.master.slave_state[slave].init_port_mask &
|
|
(1 << port)))
|
|
return 0;
|
|
|
|
if (dev->caps.port_mask[port] != MLX4_PORT_TYPE_IB) {
|
|
if (priv->mfunc.master.init_port_ref[port] == 1) {
|
|
err = mlx4_cmd(dev, 0, port, 0, MLX4_CMD_CLOSE_PORT,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
return err;
|
|
}
|
|
priv->mfunc.master.slave_state[slave].init_port_mask &= ~(1 << port);
|
|
} else {
|
|
/* infiniband port */
|
|
if (slave == mlx4_master_func_num(dev)) {
|
|
if (!priv->mfunc.master.qp0_state[port].qp0_active &&
|
|
priv->mfunc.master.qp0_state[port].port_active) {
|
|
err = mlx4_cmd(dev, 0, port, 0, MLX4_CMD_CLOSE_PORT,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
if (err)
|
|
return err;
|
|
priv->mfunc.master.slave_state[slave].init_port_mask &= ~(1 << port);
|
|
priv->mfunc.master.qp0_state[port].port_active = 0;
|
|
}
|
|
} else
|
|
priv->mfunc.master.slave_state[slave].init_port_mask &= ~(1 << port);
|
|
}
|
|
--priv->mfunc.master.init_port_ref[port];
|
|
return 0;
|
|
}
|
|
|
|
int mlx4_CLOSE_PORT(struct mlx4_dev *dev, int port)
|
|
{
|
|
return mlx4_cmd(dev, 0, port, 0, MLX4_CMD_CLOSE_PORT,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mlx4_CLOSE_PORT);
|
|
|
|
int mlx4_CLOSE_HCA(struct mlx4_dev *dev, int panic)
|
|
{
|
|
return mlx4_cmd(dev, 0, 0, panic, MLX4_CMD_CLOSE_HCA,
|
|
MLX4_CMD_TIME_CLASS_C, MLX4_CMD_NATIVE);
|
|
}
|
|
|
|
struct mlx4_config_dev {
|
|
__be32 update_flags;
|
|
__be32 rsvd1[3];
|
|
__be16 vxlan_udp_dport;
|
|
__be16 rsvd2;
|
|
__be16 roce_v2_entropy;
|
|
__be16 roce_v2_udp_dport;
|
|
__be32 roce_flags;
|
|
__be32 rsvd4[25];
|
|
__be16 rsvd5;
|
|
u8 rsvd6;
|
|
u8 rx_checksum_val;
|
|
};
|
|
|
|
#define MLX4_VXLAN_UDP_DPORT (1 << 0)
|
|
#define MLX4_ROCE_V2_UDP_DPORT BIT(3)
|
|
#define MLX4_DISABLE_RX_PORT BIT(18)
|
|
|
|
static int mlx4_CONFIG_DEV_set(struct mlx4_dev *dev, struct mlx4_config_dev *config_dev)
|
|
{
|
|
int err;
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
|
|
memcpy(mailbox->buf, config_dev, sizeof(*config_dev));
|
|
|
|
err = mlx4_cmd(dev, mailbox->dma, 0, 0, MLX4_CMD_CONFIG_DEV,
|
|
MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE);
|
|
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
static int mlx4_CONFIG_DEV_get(struct mlx4_dev *dev, struct mlx4_config_dev *config_dev)
|
|
{
|
|
int err;
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, 0, 1, MLX4_CMD_CONFIG_DEV,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
|
|
if (!err)
|
|
memcpy(config_dev, mailbox->buf, sizeof(*config_dev));
|
|
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
/* Conversion between the HW values and the actual functionality.
|
|
* The value represented by the array index,
|
|
* and the functionality determined by the flags.
|
|
*/
|
|
static const u8 config_dev_csum_flags[] = {
|
|
[0] = 0,
|
|
[1] = MLX4_RX_CSUM_MODE_VAL_NON_TCP_UDP,
|
|
[2] = MLX4_RX_CSUM_MODE_VAL_NON_TCP_UDP |
|
|
MLX4_RX_CSUM_MODE_L4,
|
|
[3] = MLX4_RX_CSUM_MODE_L4 |
|
|
MLX4_RX_CSUM_MODE_IP_OK_IP_NON_TCP_UDP |
|
|
MLX4_RX_CSUM_MODE_MULTI_VLAN
|
|
};
|
|
|
|
int mlx4_config_dev_retrieval(struct mlx4_dev *dev,
|
|
struct mlx4_config_dev_params *params)
|
|
{
|
|
struct mlx4_config_dev config_dev = {0};
|
|
int err;
|
|
u8 csum_mask;
|
|
|
|
#define CONFIG_DEV_RX_CSUM_MODE_MASK 0x7
|
|
#define CONFIG_DEV_RX_CSUM_MODE_PORT1_BIT_OFFSET 0
|
|
#define CONFIG_DEV_RX_CSUM_MODE_PORT2_BIT_OFFSET 4
|
|
|
|
if (!(dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_CONFIG_DEV))
|
|
return -EOPNOTSUPP;
|
|
|
|
err = mlx4_CONFIG_DEV_get(dev, &config_dev);
|
|
if (err)
|
|
return err;
|
|
|
|
csum_mask = (config_dev.rx_checksum_val >> CONFIG_DEV_RX_CSUM_MODE_PORT1_BIT_OFFSET) &
|
|
CONFIG_DEV_RX_CSUM_MODE_MASK;
|
|
|
|
if (csum_mask >= ARRAY_SIZE(config_dev_csum_flags))
|
|
return -EINVAL;
|
|
params->rx_csum_flags_port_1 = config_dev_csum_flags[csum_mask];
|
|
|
|
csum_mask = (config_dev.rx_checksum_val >> CONFIG_DEV_RX_CSUM_MODE_PORT2_BIT_OFFSET) &
|
|
CONFIG_DEV_RX_CSUM_MODE_MASK;
|
|
|
|
if (csum_mask >= ARRAY_SIZE(config_dev_csum_flags))
|
|
return -EINVAL;
|
|
params->rx_csum_flags_port_2 = config_dev_csum_flags[csum_mask];
|
|
|
|
params->vxlan_udp_dport = be16_to_cpu(config_dev.vxlan_udp_dport);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL_GPL(mlx4_config_dev_retrieval);
|
|
|
|
int mlx4_config_vxlan_port(struct mlx4_dev *dev, __be16 udp_port)
|
|
{
|
|
struct mlx4_config_dev config_dev;
|
|
|
|
memset(&config_dev, 0, sizeof(config_dev));
|
|
config_dev.update_flags = cpu_to_be32(MLX4_VXLAN_UDP_DPORT);
|
|
config_dev.vxlan_udp_dport = udp_port;
|
|
|
|
return mlx4_CONFIG_DEV_set(dev, &config_dev);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mlx4_config_vxlan_port);
|
|
|
|
#define CONFIG_DISABLE_RX_PORT BIT(15)
|
|
int mlx4_disable_rx_port_check(struct mlx4_dev *dev, bool dis)
|
|
{
|
|
struct mlx4_config_dev config_dev;
|
|
|
|
memset(&config_dev, 0, sizeof(config_dev));
|
|
config_dev.update_flags = cpu_to_be32(MLX4_DISABLE_RX_PORT);
|
|
if (dis)
|
|
config_dev.roce_flags =
|
|
cpu_to_be32(CONFIG_DISABLE_RX_PORT);
|
|
|
|
return mlx4_CONFIG_DEV_set(dev, &config_dev);
|
|
}
|
|
|
|
int mlx4_config_roce_v2_port(struct mlx4_dev *dev, u16 udp_port)
|
|
{
|
|
struct mlx4_config_dev config_dev;
|
|
|
|
memset(&config_dev, 0, sizeof(config_dev));
|
|
config_dev.update_flags = cpu_to_be32(MLX4_ROCE_V2_UDP_DPORT);
|
|
config_dev.roce_v2_udp_dport = cpu_to_be16(udp_port);
|
|
|
|
return mlx4_CONFIG_DEV_set(dev, &config_dev);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mlx4_config_roce_v2_port);
|
|
|
|
int mlx4_virt2phy_port_map(struct mlx4_dev *dev, u32 port1, u32 port2)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
struct {
|
|
__be32 v_port1;
|
|
__be32 v_port2;
|
|
} *v2p;
|
|
int err;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return -ENOMEM;
|
|
|
|
v2p = mailbox->buf;
|
|
v2p->v_port1 = cpu_to_be32(port1);
|
|
v2p->v_port2 = cpu_to_be32(port2);
|
|
|
|
err = mlx4_cmd(dev, mailbox->dma, 0,
|
|
MLX4_SET_PORT_VIRT2PHY, MLX4_CMD_VIRT_PORT_MAP,
|
|
MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE);
|
|
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
|
|
int mlx4_SET_ICM_SIZE(struct mlx4_dev *dev, u64 icm_size, u64 *aux_pages)
|
|
{
|
|
int ret = mlx4_cmd_imm(dev, icm_size, aux_pages, 0, 0,
|
|
MLX4_CMD_SET_ICM_SIZE,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/*
|
|
* Round up number of system pages needed in case
|
|
* MLX4_ICM_PAGE_SIZE < PAGE_SIZE.
|
|
*/
|
|
*aux_pages = ALIGN(*aux_pages, PAGE_SIZE / MLX4_ICM_PAGE_SIZE) >>
|
|
(PAGE_SHIFT - MLX4_ICM_PAGE_SHIFT);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int mlx4_NOP(struct mlx4_dev *dev)
|
|
{
|
|
/* Input modifier of 0x1f means "finish as soon as possible." */
|
|
return mlx4_cmd(dev, 0, 0x1f, 0, MLX4_CMD_NOP, MLX4_CMD_TIME_CLASS_A,
|
|
MLX4_CMD_NATIVE);
|
|
}
|
|
|
|
int mlx4_query_diag_counters(struct mlx4_dev *dev, u8 op_modifier,
|
|
const u32 offset[],
|
|
u32 value[], size_t array_len, u8 port)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 *outbox;
|
|
size_t i;
|
|
int ret;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
|
|
outbox = mailbox->buf;
|
|
|
|
ret = mlx4_cmd_box(dev, 0, mailbox->dma, port, op_modifier,
|
|
MLX4_CMD_DIAG_RPRT, MLX4_CMD_TIME_CLASS_A,
|
|
MLX4_CMD_NATIVE);
|
|
if (ret)
|
|
goto out;
|
|
|
|
for (i = 0; i < array_len; i++) {
|
|
if (offset[i] > MLX4_MAILBOX_SIZE) {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
MLX4_GET(value[i], outbox, offset[i]);
|
|
}
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(mlx4_query_diag_counters);
|
|
|
|
int mlx4_get_phys_port_id(struct mlx4_dev *dev)
|
|
{
|
|
u8 port;
|
|
u32 *outbox;
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
u32 in_mod;
|
|
u32 guid_hi, guid_lo;
|
|
int err, ret = 0;
|
|
#define MOD_STAT_CFG_PORT_OFFSET 8
|
|
#define MOD_STAT_CFG_GUID_H 0X14
|
|
#define MOD_STAT_CFG_GUID_L 0X1c
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
outbox = mailbox->buf;
|
|
|
|
for (port = 1; port <= dev->caps.num_ports; port++) {
|
|
in_mod = port << MOD_STAT_CFG_PORT_OFFSET;
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, in_mod, 0x2,
|
|
MLX4_CMD_MOD_STAT_CFG, MLX4_CMD_TIME_CLASS_A,
|
|
MLX4_CMD_NATIVE);
|
|
if (err) {
|
|
mlx4_err(dev, "Fail to get port %d uplink guid\n",
|
|
port);
|
|
ret = err;
|
|
} else {
|
|
MLX4_GET(guid_hi, outbox, MOD_STAT_CFG_GUID_H);
|
|
MLX4_GET(guid_lo, outbox, MOD_STAT_CFG_GUID_L);
|
|
dev->caps.phys_port_id[port] = (u64)guid_lo |
|
|
(u64)guid_hi << 32;
|
|
}
|
|
}
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return ret;
|
|
}
|
|
|
|
#define MLX4_WOL_SETUP_MODE (5 << 28)
|
|
int mlx4_wol_read(struct mlx4_dev *dev, u64 *config, int port)
|
|
{
|
|
u32 in_mod = MLX4_WOL_SETUP_MODE | port << 8;
|
|
|
|
return mlx4_cmd_imm(dev, 0, config, in_mod, 0x3,
|
|
MLX4_CMD_MOD_STAT_CFG, MLX4_CMD_TIME_CLASS_A,
|
|
MLX4_CMD_NATIVE);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mlx4_wol_read);
|
|
|
|
int mlx4_wol_write(struct mlx4_dev *dev, u64 config, int port)
|
|
{
|
|
u32 in_mod = MLX4_WOL_SETUP_MODE | port << 8;
|
|
|
|
return mlx4_cmd(dev, config, in_mod, 0x1, MLX4_CMD_MOD_STAT_CFG,
|
|
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_NATIVE);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mlx4_wol_write);
|
|
|
|
enum {
|
|
ADD_TO_MCG = 0x26,
|
|
};
|
|
|
|
|
|
void mlx4_opreq_action(struct work_struct *work)
|
|
{
|
|
struct mlx4_priv *priv = container_of(work, struct mlx4_priv,
|
|
opreq_task);
|
|
struct mlx4_dev *dev = &priv->dev;
|
|
int num_tasks = atomic_read(&priv->opreq_count);
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
struct mlx4_mgm *mgm;
|
|
u32 *outbox;
|
|
u32 modifier;
|
|
u16 token;
|
|
u16 type;
|
|
int err;
|
|
u32 num_qps;
|
|
struct mlx4_qp qp;
|
|
int i;
|
|
u8 rem_mcg;
|
|
u8 prot;
|
|
|
|
#define GET_OP_REQ_MODIFIER_OFFSET 0x08
|
|
#define GET_OP_REQ_TOKEN_OFFSET 0x14
|
|
#define GET_OP_REQ_TYPE_OFFSET 0x1a
|
|
#define GET_OP_REQ_DATA_OFFSET 0x20
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox)) {
|
|
mlx4_err(dev, "Failed to allocate mailbox for GET_OP_REQ\n");
|
|
return;
|
|
}
|
|
outbox = mailbox->buf;
|
|
|
|
while (num_tasks) {
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, 0, 0,
|
|
MLX4_CMD_GET_OP_REQ, MLX4_CMD_TIME_CLASS_A,
|
|
MLX4_CMD_NATIVE);
|
|
if (err) {
|
|
mlx4_err(dev, "Failed to retrieve required operation: %d\n",
|
|
err);
|
|
return;
|
|
}
|
|
MLX4_GET(modifier, outbox, GET_OP_REQ_MODIFIER_OFFSET);
|
|
MLX4_GET(token, outbox, GET_OP_REQ_TOKEN_OFFSET);
|
|
MLX4_GET(type, outbox, GET_OP_REQ_TYPE_OFFSET);
|
|
type &= 0xfff;
|
|
|
|
switch (type) {
|
|
case ADD_TO_MCG:
|
|
if (dev->caps.steering_mode ==
|
|
MLX4_STEERING_MODE_DEVICE_MANAGED) {
|
|
mlx4_warn(dev, "ADD MCG operation is not supported in DEVICE_MANAGED steering mode\n");
|
|
err = EPERM;
|
|
break;
|
|
}
|
|
mgm = (struct mlx4_mgm *)((u8 *)(outbox) +
|
|
GET_OP_REQ_DATA_OFFSET);
|
|
num_qps = be32_to_cpu(mgm->members_count) &
|
|
MGM_QPN_MASK;
|
|
rem_mcg = ((u8 *)(&mgm->members_count))[0] & 1;
|
|
prot = ((u8 *)(&mgm->members_count))[0] >> 6;
|
|
|
|
for (i = 0; i < num_qps; i++) {
|
|
qp.qpn = be32_to_cpu(mgm->qp[i]);
|
|
if (rem_mcg)
|
|
err = mlx4_multicast_detach(dev, &qp,
|
|
mgm->gid,
|
|
prot, 0);
|
|
else
|
|
err = mlx4_multicast_attach(dev, &qp,
|
|
mgm->gid,
|
|
mgm->gid[5]
|
|
, 0, prot,
|
|
NULL);
|
|
if (err)
|
|
break;
|
|
}
|
|
break;
|
|
default:
|
|
mlx4_warn(dev, "Bad type for required operation\n");
|
|
err = EINVAL;
|
|
break;
|
|
}
|
|
err = mlx4_cmd(dev, 0, ((u32) err |
|
|
(__force u32)cpu_to_be32(token) << 16),
|
|
1, MLX4_CMD_GET_OP_REQ, MLX4_CMD_TIME_CLASS_A,
|
|
MLX4_CMD_NATIVE);
|
|
if (err) {
|
|
mlx4_err(dev, "Failed to acknowledge required request: %d\n",
|
|
err);
|
|
goto out;
|
|
}
|
|
memset(outbox, 0, 0xffc);
|
|
num_tasks = atomic_dec_return(&priv->opreq_count);
|
|
}
|
|
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
}
|
|
|
|
static int mlx4_check_smp_firewall_active(struct mlx4_dev *dev,
|
|
struct mlx4_cmd_mailbox *mailbox)
|
|
{
|
|
#define MLX4_CMD_MAD_DEMUX_SET_ATTR_OFFSET 0x10
|
|
#define MLX4_CMD_MAD_DEMUX_GETRESP_ATTR_OFFSET 0x20
|
|
#define MLX4_CMD_MAD_DEMUX_TRAP_ATTR_OFFSET 0x40
|
|
#define MLX4_CMD_MAD_DEMUX_TRAP_REPRESS_ATTR_OFFSET 0x70
|
|
|
|
u32 set_attr_mask, getresp_attr_mask;
|
|
u32 trap_attr_mask, traprepress_attr_mask;
|
|
|
|
MLX4_GET(set_attr_mask, mailbox->buf,
|
|
MLX4_CMD_MAD_DEMUX_SET_ATTR_OFFSET);
|
|
mlx4_dbg(dev, "SMP firewall set_attribute_mask = 0x%x\n",
|
|
set_attr_mask);
|
|
|
|
MLX4_GET(getresp_attr_mask, mailbox->buf,
|
|
MLX4_CMD_MAD_DEMUX_GETRESP_ATTR_OFFSET);
|
|
mlx4_dbg(dev, "SMP firewall getresp_attribute_mask = 0x%x\n",
|
|
getresp_attr_mask);
|
|
|
|
MLX4_GET(trap_attr_mask, mailbox->buf,
|
|
MLX4_CMD_MAD_DEMUX_TRAP_ATTR_OFFSET);
|
|
mlx4_dbg(dev, "SMP firewall trap_attribute_mask = 0x%x\n",
|
|
trap_attr_mask);
|
|
|
|
MLX4_GET(traprepress_attr_mask, mailbox->buf,
|
|
MLX4_CMD_MAD_DEMUX_TRAP_REPRESS_ATTR_OFFSET);
|
|
mlx4_dbg(dev, "SMP firewall traprepress_attribute_mask = 0x%x\n",
|
|
traprepress_attr_mask);
|
|
|
|
if (set_attr_mask && getresp_attr_mask && trap_attr_mask &&
|
|
traprepress_attr_mask)
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int mlx4_config_mad_demux(struct mlx4_dev *dev)
|
|
{
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
int err;
|
|
|
|
/* Check if mad_demux is supported */
|
|
if (!(dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_MAD_DEMUX))
|
|
return 0;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox)) {
|
|
mlx4_warn(dev, "Failed to allocate mailbox for cmd MAD_DEMUX");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Query mad_demux to find out which MADs are handled by internal sma */
|
|
err = mlx4_cmd_box(dev, 0, mailbox->dma, 0x01 /* subn mgmt class */,
|
|
MLX4_CMD_MAD_DEMUX_QUERY_RESTR, MLX4_CMD_MAD_DEMUX,
|
|
MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE);
|
|
if (err) {
|
|
mlx4_warn(dev, "MLX4_CMD_MAD_DEMUX: query restrictions failed (%d)\n",
|
|
err);
|
|
goto out;
|
|
}
|
|
|
|
if (mlx4_check_smp_firewall_active(dev, mailbox))
|
|
dev->flags |= MLX4_FLAG_SECURE_HOST;
|
|
|
|
/* Config mad_demux to handle all MADs returned by the query above */
|
|
err = mlx4_cmd(dev, mailbox->dma, 0x01 /* subn mgmt class */,
|
|
MLX4_CMD_MAD_DEMUX_CONFIG, MLX4_CMD_MAD_DEMUX,
|
|
MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE);
|
|
if (err) {
|
|
mlx4_warn(dev, "MLX4_CMD_MAD_DEMUX: configure failed (%d)\n", err);
|
|
goto out;
|
|
}
|
|
|
|
if (dev->flags & MLX4_FLAG_SECURE_HOST)
|
|
mlx4_warn(dev, "HCA operating in secure-host mode. SMP firewall activated.\n");
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
/* Access Reg commands */
|
|
enum mlx4_access_reg_masks {
|
|
MLX4_ACCESS_REG_STATUS_MASK = 0x7f,
|
|
MLX4_ACCESS_REG_METHOD_MASK = 0x7f,
|
|
MLX4_ACCESS_REG_LEN_MASK = 0x7ff
|
|
};
|
|
|
|
struct mlx4_access_reg {
|
|
__be16 constant1;
|
|
u8 status;
|
|
u8 resrvd1;
|
|
__be16 reg_id;
|
|
u8 method;
|
|
u8 constant2;
|
|
__be32 resrvd2[2];
|
|
__be16 len_const;
|
|
__be16 resrvd3;
|
|
#define MLX4_ACCESS_REG_HEADER_SIZE (20)
|
|
u8 reg_data[MLX4_MAILBOX_SIZE-MLX4_ACCESS_REG_HEADER_SIZE];
|
|
} __attribute__((__packed__));
|
|
|
|
/**
|
|
* mlx4_ACCESS_REG - Generic access reg command.
|
|
* @dev: mlx4_dev.
|
|
* @reg_id: register ID to access.
|
|
* @method: Access method Read/Write.
|
|
* @reg_len: register length to Read/Write in bytes.
|
|
* @reg_data: reg_data pointer to Read/Write From/To.
|
|
*
|
|
* Access ConnectX registers FW command.
|
|
* Returns 0 on success and copies outbox mlx4_access_reg data
|
|
* field into reg_data or a negative error code.
|
|
*/
|
|
static int mlx4_ACCESS_REG(struct mlx4_dev *dev, u16 reg_id,
|
|
enum mlx4_access_reg_method method,
|
|
u16 reg_len, void *reg_data)
|
|
{
|
|
struct mlx4_cmd_mailbox *inbox, *outbox;
|
|
struct mlx4_access_reg *inbuf, *outbuf;
|
|
int err;
|
|
|
|
inbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(inbox))
|
|
return PTR_ERR(inbox);
|
|
|
|
outbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(outbox)) {
|
|
mlx4_free_cmd_mailbox(dev, inbox);
|
|
return PTR_ERR(outbox);
|
|
}
|
|
|
|
inbuf = inbox->buf;
|
|
outbuf = outbox->buf;
|
|
|
|
inbuf->constant1 = cpu_to_be16(0x1<<11 | 0x4);
|
|
inbuf->constant2 = 0x1;
|
|
inbuf->reg_id = cpu_to_be16(reg_id);
|
|
inbuf->method = method & MLX4_ACCESS_REG_METHOD_MASK;
|
|
|
|
reg_len = min(reg_len, (u16)(sizeof(inbuf->reg_data)));
|
|
inbuf->len_const =
|
|
cpu_to_be16(((reg_len/4 + 1) & MLX4_ACCESS_REG_LEN_MASK) |
|
|
((0x3) << 12));
|
|
|
|
memcpy(inbuf->reg_data, reg_data, reg_len);
|
|
err = mlx4_cmd_box(dev, inbox->dma, outbox->dma, 0, 0,
|
|
MLX4_CMD_ACCESS_REG, MLX4_CMD_TIME_CLASS_C,
|
|
MLX4_CMD_WRAPPED);
|
|
if (err)
|
|
goto out;
|
|
|
|
if (outbuf->status & MLX4_ACCESS_REG_STATUS_MASK) {
|
|
err = outbuf->status & MLX4_ACCESS_REG_STATUS_MASK;
|
|
mlx4_err(dev,
|
|
"MLX4_CMD_ACCESS_REG(%x) returned REG status (%x)\n",
|
|
reg_id, err);
|
|
goto out;
|
|
}
|
|
|
|
memcpy(reg_data, outbuf->reg_data, reg_len);
|
|
out:
|
|
mlx4_free_cmd_mailbox(dev, inbox);
|
|
mlx4_free_cmd_mailbox(dev, outbox);
|
|
return err;
|
|
}
|
|
|
|
/* ConnectX registers IDs */
|
|
enum mlx4_reg_id {
|
|
MLX4_REG_ID_PTYS = 0x5004,
|
|
};
|
|
|
|
/**
|
|
* mlx4_ACCESS_PTYS_REG - Access PTYs (Port Type and Speed)
|
|
* register
|
|
* @dev: mlx4_dev.
|
|
* @method: Access method Read/Write.
|
|
* @ptys_reg: PTYS register data pointer.
|
|
*
|
|
* Access ConnectX PTYS register, to Read/Write Port Type/Speed
|
|
* configuration
|
|
* Returns 0 on success or a negative error code.
|
|
*/
|
|
int mlx4_ACCESS_PTYS_REG(struct mlx4_dev *dev,
|
|
enum mlx4_access_reg_method method,
|
|
struct mlx4_ptys_reg *ptys_reg)
|
|
{
|
|
return mlx4_ACCESS_REG(dev, MLX4_REG_ID_PTYS,
|
|
method, sizeof(*ptys_reg), ptys_reg);
|
|
}
|
|
EXPORT_SYMBOL_GPL(mlx4_ACCESS_PTYS_REG);
|
|
|
|
int mlx4_ACCESS_REG_wrapper(struct mlx4_dev *dev, int slave,
|
|
struct mlx4_vhcr *vhcr,
|
|
struct mlx4_cmd_mailbox *inbox,
|
|
struct mlx4_cmd_mailbox *outbox,
|
|
struct mlx4_cmd_info *cmd)
|
|
{
|
|
struct mlx4_access_reg *inbuf = inbox->buf;
|
|
u8 method = inbuf->method & MLX4_ACCESS_REG_METHOD_MASK;
|
|
u16 reg_id = be16_to_cpu(inbuf->reg_id);
|
|
|
|
if (slave != mlx4_master_func_num(dev) &&
|
|
method == MLX4_ACCESS_REG_WRITE)
|
|
return -EPERM;
|
|
|
|
if (reg_id == MLX4_REG_ID_PTYS) {
|
|
struct mlx4_ptys_reg *ptys_reg =
|
|
(struct mlx4_ptys_reg *)inbuf->reg_data;
|
|
|
|
ptys_reg->local_port =
|
|
mlx4_slave_convert_port(dev, slave,
|
|
ptys_reg->local_port);
|
|
}
|
|
|
|
return mlx4_cmd_box(dev, inbox->dma, outbox->dma, vhcr->in_modifier,
|
|
0, MLX4_CMD_ACCESS_REG, MLX4_CMD_TIME_CLASS_C,
|
|
MLX4_CMD_NATIVE);
|
|
}
|
|
|
|
static int mlx4_SET_PORT_phv_bit(struct mlx4_dev *dev, u8 port, u8 phv_bit)
|
|
{
|
|
#define SET_PORT_GEN_PHV_VALID 0x10
|
|
#define SET_PORT_GEN_PHV_EN 0x80
|
|
|
|
struct mlx4_cmd_mailbox *mailbox;
|
|
struct mlx4_set_port_general_context *context;
|
|
u32 in_mod;
|
|
int err;
|
|
|
|
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
|
if (IS_ERR(mailbox))
|
|
return PTR_ERR(mailbox);
|
|
context = mailbox->buf;
|
|
|
|
context->flags2 |= SET_PORT_GEN_PHV_VALID;
|
|
if (phv_bit)
|
|
context->phv_en |= SET_PORT_GEN_PHV_EN;
|
|
|
|
in_mod = MLX4_SET_PORT_GENERAL << 8 | port;
|
|
err = mlx4_cmd(dev, mailbox->dma, in_mod, MLX4_SET_PORT_ETH_OPCODE,
|
|
MLX4_CMD_SET_PORT, MLX4_CMD_TIME_CLASS_B,
|
|
MLX4_CMD_NATIVE);
|
|
|
|
mlx4_free_cmd_mailbox(dev, mailbox);
|
|
return err;
|
|
}
|
|
|
|
int get_phv_bit(struct mlx4_dev *dev, u8 port, int *phv)
|
|
{
|
|
int err;
|
|
struct mlx4_func_cap func_cap;
|
|
|
|
memset(&func_cap, 0, sizeof(func_cap));
|
|
err = mlx4_QUERY_FUNC_CAP(dev, port, &func_cap);
|
|
if (!err)
|
|
*phv = func_cap.flags0 & QUERY_FUNC_CAP_PHV_BIT;
|
|
return err;
|
|
}
|
|
EXPORT_SYMBOL(get_phv_bit);
|
|
|
|
int set_phv_bit(struct mlx4_dev *dev, u8 port, int new_val)
|
|
{
|
|
int ret;
|
|
|
|
if (mlx4_is_slave(dev))
|
|
return -EPERM;
|
|
|
|
if (dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_PHV_EN &&
|
|
!(dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_SKIP_OUTER_VLAN)) {
|
|
ret = mlx4_SET_PORT_phv_bit(dev, port, new_val);
|
|
if (!ret)
|
|
dev->caps.phv_bit[port] = new_val;
|
|
return ret;
|
|
}
|
|
|
|
return -EOPNOTSUPP;
|
|
}
|
|
EXPORT_SYMBOL(set_phv_bit);
|
|
|
|
int mlx4_get_is_vlan_offload_disabled(struct mlx4_dev *dev, u8 port,
|
|
bool *vlan_offload_disabled)
|
|
{
|
|
struct mlx4_func_cap func_cap;
|
|
int err;
|
|
|
|
memset(&func_cap, 0, sizeof(func_cap));
|
|
err = mlx4_QUERY_FUNC_CAP(dev, port, &func_cap);
|
|
if (!err)
|
|
*vlan_offload_disabled =
|
|
!!(func_cap.flags0 &
|
|
QUERY_FUNC_CAP_VLAN_OFFLOAD_DISABLE);
|
|
return err;
|
|
}
|
|
EXPORT_SYMBOL(mlx4_get_is_vlan_offload_disabled);
|
|
|
|
void mlx4_replace_zero_macs(struct mlx4_dev *dev)
|
|
{
|
|
int i;
|
|
u8 mac_addr[ETH_ALEN];
|
|
|
|
dev->port_random_macs = 0;
|
|
for (i = 1; i <= dev->caps.num_ports; ++i)
|
|
if (!dev->caps.def_mac[i] &&
|
|
dev->caps.port_type[i] == MLX4_PORT_TYPE_ETH) {
|
|
eth_random_addr(mac_addr);
|
|
dev->port_random_macs |= 1 << i;
|
|
dev->caps.def_mac[i] = mlx4_mac_to_u64(mac_addr);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL_GPL(mlx4_replace_zero_macs);
|