mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-26 03:15:37 +07:00
23705adb2c
The following set of NPC registers allow the driver to configure NPC to generate different key value schemes to compare against packet payload in MCAM search. NPC_AF_INTF(0..1)_KEX_CFG NPC_AF_KEX_LDATA(0..1)_FLAGS_CFG NPC_AF_INTF(0..1)_LID(0..7)_LT(0..15)_LD(0..1)_CFG NPC_AF_INTF(0..1)_LDATA(0..1)_FLAGS(0..15)_CFG Currently, the AF driver populates these registers to configure the default values to address the most common use cases such as key generation for channel number + DMAC. The secure firmware stores different configuration value of these registers to enable different NPC use case along with the name for the lookup. Patch loads profile binary from secure firmware over the exiting CGX mailbox interface and apply the profile. AF driver shall fall back to the default configuration in case of any errors. The AF consumer driver can know the selected profile on response to NPC_GET_KEX_CFG mailbox by introducing mkex_pfl_name in the struct npc_get_kex_cfg_rsp. Signed-off-by: Vamsi Attunuru <vamsi.attunuru@marvell.com> Signed-off-by: Jerin Jacob <jerinj@marvell.com> Signed-off-by: David S. Miller <davem@davemloft.net>
870 lines
21 KiB
C
870 lines
21 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* Marvell OcteonTx2 CGX driver
|
|
*
|
|
* Copyright (C) 2018 Marvell International Ltd.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*/
|
|
|
|
#include <linux/acpi.h>
|
|
#include <linux/module.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/netdevice.h>
|
|
#include <linux/etherdevice.h>
|
|
#include <linux/phy.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_mdio.h>
|
|
#include <linux/of_net.h>
|
|
|
|
#include "cgx.h"
|
|
|
|
#define DRV_NAME "octeontx2-cgx"
|
|
#define DRV_STRING "Marvell OcteonTX2 CGX/MAC Driver"
|
|
|
|
/**
|
|
* struct lmac
|
|
* @wq_cmd_cmplt: waitq to keep the process blocked until cmd completion
|
|
* @cmd_lock: Lock to serialize the command interface
|
|
* @resp: command response
|
|
* @link_info: link related information
|
|
* @event_cb: callback for linkchange events
|
|
* @event_cb_lock: lock for serializing callback with unregister
|
|
* @cmd_pend: flag set before new command is started
|
|
* flag cleared after command response is received
|
|
* @cgx: parent cgx port
|
|
* @lmac_id: lmac port id
|
|
* @name: lmac port name
|
|
*/
|
|
struct lmac {
|
|
wait_queue_head_t wq_cmd_cmplt;
|
|
struct mutex cmd_lock;
|
|
u64 resp;
|
|
struct cgx_link_user_info link_info;
|
|
struct cgx_event_cb event_cb;
|
|
spinlock_t event_cb_lock;
|
|
bool cmd_pend;
|
|
struct cgx *cgx;
|
|
u8 lmac_id;
|
|
char *name;
|
|
};
|
|
|
|
struct cgx {
|
|
void __iomem *reg_base;
|
|
struct pci_dev *pdev;
|
|
u8 cgx_id;
|
|
u8 lmac_count;
|
|
struct lmac *lmac_idmap[MAX_LMAC_PER_CGX];
|
|
struct work_struct cgx_cmd_work;
|
|
struct workqueue_struct *cgx_cmd_workq;
|
|
struct list_head cgx_list;
|
|
};
|
|
|
|
static LIST_HEAD(cgx_list);
|
|
|
|
/* Convert firmware speed encoding to user format(Mbps) */
|
|
static u32 cgx_speed_mbps[CGX_LINK_SPEED_MAX];
|
|
|
|
/* Convert firmware lmac type encoding to string */
|
|
static char *cgx_lmactype_string[LMAC_MODE_MAX];
|
|
|
|
/* CGX PHY management internal APIs */
|
|
static int cgx_fwi_link_change(struct cgx *cgx, int lmac_id, bool en);
|
|
|
|
/* Supported devices */
|
|
static const struct pci_device_id cgx_id_table[] = {
|
|
{ PCI_DEVICE(PCI_VENDOR_ID_CAVIUM, PCI_DEVID_OCTEONTX2_CGX) },
|
|
{ 0, } /* end of table */
|
|
};
|
|
|
|
MODULE_DEVICE_TABLE(pci, cgx_id_table);
|
|
|
|
static void cgx_write(struct cgx *cgx, u64 lmac, u64 offset, u64 val)
|
|
{
|
|
writeq(val, cgx->reg_base + (lmac << 18) + offset);
|
|
}
|
|
|
|
static u64 cgx_read(struct cgx *cgx, u64 lmac, u64 offset)
|
|
{
|
|
return readq(cgx->reg_base + (lmac << 18) + offset);
|
|
}
|
|
|
|
static inline struct lmac *lmac_pdata(u8 lmac_id, struct cgx *cgx)
|
|
{
|
|
if (!cgx || lmac_id >= MAX_LMAC_PER_CGX)
|
|
return NULL;
|
|
|
|
return cgx->lmac_idmap[lmac_id];
|
|
}
|
|
|
|
int cgx_get_cgxcnt_max(void)
|
|
{
|
|
struct cgx *cgx_dev;
|
|
int idmax = -ENODEV;
|
|
|
|
list_for_each_entry(cgx_dev, &cgx_list, cgx_list)
|
|
if (cgx_dev->cgx_id > idmax)
|
|
idmax = cgx_dev->cgx_id;
|
|
|
|
if (idmax < 0)
|
|
return 0;
|
|
|
|
return idmax + 1;
|
|
}
|
|
EXPORT_SYMBOL(cgx_get_cgxcnt_max);
|
|
|
|
int cgx_get_lmac_cnt(void *cgxd)
|
|
{
|
|
struct cgx *cgx = cgxd;
|
|
|
|
if (!cgx)
|
|
return -ENODEV;
|
|
|
|
return cgx->lmac_count;
|
|
}
|
|
EXPORT_SYMBOL(cgx_get_lmac_cnt);
|
|
|
|
void *cgx_get_pdata(int cgx_id)
|
|
{
|
|
struct cgx *cgx_dev;
|
|
|
|
list_for_each_entry(cgx_dev, &cgx_list, cgx_list) {
|
|
if (cgx_dev->cgx_id == cgx_id)
|
|
return cgx_dev;
|
|
}
|
|
return NULL;
|
|
}
|
|
EXPORT_SYMBOL(cgx_get_pdata);
|
|
|
|
/* Ensure the required lock for event queue(where asynchronous events are
|
|
* posted) is acquired before calling this API. Else an asynchronous event(with
|
|
* latest link status) can reach the destination before this function returns
|
|
* and could make the link status appear wrong.
|
|
*/
|
|
int cgx_get_link_info(void *cgxd, int lmac_id,
|
|
struct cgx_link_user_info *linfo)
|
|
{
|
|
struct lmac *lmac = lmac_pdata(lmac_id, cgxd);
|
|
|
|
if (!lmac)
|
|
return -ENODEV;
|
|
|
|
*linfo = lmac->link_info;
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_get_link_info);
|
|
|
|
static u64 mac2u64 (u8 *mac_addr)
|
|
{
|
|
u64 mac = 0;
|
|
int index;
|
|
|
|
for (index = ETH_ALEN - 1; index >= 0; index--)
|
|
mac |= ((u64)*mac_addr++) << (8 * index);
|
|
return mac;
|
|
}
|
|
|
|
int cgx_lmac_addr_set(u8 cgx_id, u8 lmac_id, u8 *mac_addr)
|
|
{
|
|
struct cgx *cgx_dev = cgx_get_pdata(cgx_id);
|
|
u64 cfg;
|
|
|
|
/* copy 6bytes from macaddr */
|
|
/* memcpy(&cfg, mac_addr, 6); */
|
|
|
|
cfg = mac2u64 (mac_addr);
|
|
|
|
cgx_write(cgx_dev, 0, (CGXX_CMRX_RX_DMAC_CAM0 + (lmac_id * 0x8)),
|
|
cfg | CGX_DMAC_CAM_ADDR_ENABLE | ((u64)lmac_id << 49));
|
|
|
|
cfg = cgx_read(cgx_dev, lmac_id, CGXX_CMRX_RX_DMAC_CTL0);
|
|
cfg |= CGX_DMAC_CTL0_CAM_ENABLE;
|
|
cgx_write(cgx_dev, lmac_id, CGXX_CMRX_RX_DMAC_CTL0, cfg);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_lmac_addr_set);
|
|
|
|
u64 cgx_lmac_addr_get(u8 cgx_id, u8 lmac_id)
|
|
{
|
|
struct cgx *cgx_dev = cgx_get_pdata(cgx_id);
|
|
u64 cfg;
|
|
|
|
cfg = cgx_read(cgx_dev, 0, CGXX_CMRX_RX_DMAC_CAM0 + lmac_id * 0x8);
|
|
return cfg & CGX_RX_DMAC_ADR_MASK;
|
|
}
|
|
EXPORT_SYMBOL(cgx_lmac_addr_get);
|
|
|
|
int cgx_set_pkind(void *cgxd, u8 lmac_id, int pkind)
|
|
{
|
|
struct cgx *cgx = cgxd;
|
|
|
|
if (!cgx || lmac_id >= cgx->lmac_count)
|
|
return -ENODEV;
|
|
|
|
cgx_write(cgx, lmac_id, CGXX_CMRX_RX_ID_MAP, (pkind & 0x3F));
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_set_pkind);
|
|
|
|
static inline u8 cgx_get_lmac_type(struct cgx *cgx, int lmac_id)
|
|
{
|
|
u64 cfg;
|
|
|
|
cfg = cgx_read(cgx, lmac_id, CGXX_CMRX_CFG);
|
|
return (cfg >> CGX_LMAC_TYPE_SHIFT) & CGX_LMAC_TYPE_MASK;
|
|
}
|
|
|
|
/* Configure CGX LMAC in internal loopback mode */
|
|
int cgx_lmac_internal_loopback(void *cgxd, int lmac_id, bool enable)
|
|
{
|
|
struct cgx *cgx = cgxd;
|
|
u8 lmac_type;
|
|
u64 cfg;
|
|
|
|
if (!cgx || lmac_id >= cgx->lmac_count)
|
|
return -ENODEV;
|
|
|
|
lmac_type = cgx_get_lmac_type(cgx, lmac_id);
|
|
if (lmac_type == LMAC_MODE_SGMII || lmac_type == LMAC_MODE_QSGMII) {
|
|
cfg = cgx_read(cgx, lmac_id, CGXX_GMP_PCS_MRX_CTL);
|
|
if (enable)
|
|
cfg |= CGXX_GMP_PCS_MRX_CTL_LBK;
|
|
else
|
|
cfg &= ~CGXX_GMP_PCS_MRX_CTL_LBK;
|
|
cgx_write(cgx, lmac_id, CGXX_GMP_PCS_MRX_CTL, cfg);
|
|
} else {
|
|
cfg = cgx_read(cgx, lmac_id, CGXX_SPUX_CONTROL1);
|
|
if (enable)
|
|
cfg |= CGXX_SPUX_CONTROL1_LBK;
|
|
else
|
|
cfg &= ~CGXX_SPUX_CONTROL1_LBK;
|
|
cgx_write(cgx, lmac_id, CGXX_SPUX_CONTROL1, cfg);
|
|
}
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_lmac_internal_loopback);
|
|
|
|
void cgx_lmac_promisc_config(int cgx_id, int lmac_id, bool enable)
|
|
{
|
|
struct cgx *cgx = cgx_get_pdata(cgx_id);
|
|
u64 cfg = 0;
|
|
|
|
if (!cgx)
|
|
return;
|
|
|
|
if (enable) {
|
|
/* Enable promiscuous mode on LMAC */
|
|
cfg = cgx_read(cgx, lmac_id, CGXX_CMRX_RX_DMAC_CTL0);
|
|
cfg &= ~(CGX_DMAC_CAM_ACCEPT | CGX_DMAC_MCAST_MODE);
|
|
cfg |= CGX_DMAC_BCAST_MODE;
|
|
cgx_write(cgx, lmac_id, CGXX_CMRX_RX_DMAC_CTL0, cfg);
|
|
|
|
cfg = cgx_read(cgx, 0,
|
|
(CGXX_CMRX_RX_DMAC_CAM0 + lmac_id * 0x8));
|
|
cfg &= ~CGX_DMAC_CAM_ADDR_ENABLE;
|
|
cgx_write(cgx, 0,
|
|
(CGXX_CMRX_RX_DMAC_CAM0 + lmac_id * 0x8), cfg);
|
|
} else {
|
|
/* Disable promiscuous mode */
|
|
cfg = cgx_read(cgx, lmac_id, CGXX_CMRX_RX_DMAC_CTL0);
|
|
cfg |= CGX_DMAC_CAM_ACCEPT | CGX_DMAC_MCAST_MODE;
|
|
cgx_write(cgx, lmac_id, CGXX_CMRX_RX_DMAC_CTL0, cfg);
|
|
cfg = cgx_read(cgx, 0,
|
|
(CGXX_CMRX_RX_DMAC_CAM0 + lmac_id * 0x8));
|
|
cfg |= CGX_DMAC_CAM_ADDR_ENABLE;
|
|
cgx_write(cgx, 0,
|
|
(CGXX_CMRX_RX_DMAC_CAM0 + lmac_id * 0x8), cfg);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(cgx_lmac_promisc_config);
|
|
|
|
int cgx_get_rx_stats(void *cgxd, int lmac_id, int idx, u64 *rx_stat)
|
|
{
|
|
struct cgx *cgx = cgxd;
|
|
|
|
if (!cgx || lmac_id >= cgx->lmac_count)
|
|
return -ENODEV;
|
|
*rx_stat = cgx_read(cgx, lmac_id, CGXX_CMRX_RX_STAT0 + (idx * 8));
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_get_rx_stats);
|
|
|
|
int cgx_get_tx_stats(void *cgxd, int lmac_id, int idx, u64 *tx_stat)
|
|
{
|
|
struct cgx *cgx = cgxd;
|
|
|
|
if (!cgx || lmac_id >= cgx->lmac_count)
|
|
return -ENODEV;
|
|
*tx_stat = cgx_read(cgx, lmac_id, CGXX_CMRX_TX_STAT0 + (idx * 8));
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_get_tx_stats);
|
|
|
|
int cgx_lmac_rx_tx_enable(void *cgxd, int lmac_id, bool enable)
|
|
{
|
|
struct cgx *cgx = cgxd;
|
|
u64 cfg;
|
|
|
|
if (!cgx || lmac_id >= cgx->lmac_count)
|
|
return -ENODEV;
|
|
|
|
cfg = cgx_read(cgx, lmac_id, CGXX_CMRX_CFG);
|
|
if (enable)
|
|
cfg |= CMR_EN | DATA_PKT_RX_EN | DATA_PKT_TX_EN;
|
|
else
|
|
cfg &= ~(CMR_EN | DATA_PKT_RX_EN | DATA_PKT_TX_EN);
|
|
cgx_write(cgx, lmac_id, CGXX_CMRX_CFG, cfg);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_lmac_rx_tx_enable);
|
|
|
|
/* CGX Firmware interface low level support */
|
|
static int cgx_fwi_cmd_send(u64 req, u64 *resp, struct lmac *lmac)
|
|
{
|
|
struct cgx *cgx = lmac->cgx;
|
|
struct device *dev;
|
|
int err = 0;
|
|
u64 cmd;
|
|
|
|
/* Ensure no other command is in progress */
|
|
err = mutex_lock_interruptible(&lmac->cmd_lock);
|
|
if (err)
|
|
return err;
|
|
|
|
/* Ensure command register is free */
|
|
cmd = cgx_read(cgx, lmac->lmac_id, CGX_COMMAND_REG);
|
|
if (FIELD_GET(CMDREG_OWN, cmd) != CGX_CMD_OWN_NS) {
|
|
err = -EBUSY;
|
|
goto unlock;
|
|
}
|
|
|
|
/* Update ownership in command request */
|
|
req = FIELD_SET(CMDREG_OWN, CGX_CMD_OWN_FIRMWARE, req);
|
|
|
|
/* Mark this lmac as pending, before we start */
|
|
lmac->cmd_pend = true;
|
|
|
|
/* Start command in hardware */
|
|
cgx_write(cgx, lmac->lmac_id, CGX_COMMAND_REG, req);
|
|
|
|
/* Ensure command is completed without errors */
|
|
if (!wait_event_timeout(lmac->wq_cmd_cmplt, !lmac->cmd_pend,
|
|
msecs_to_jiffies(CGX_CMD_TIMEOUT))) {
|
|
dev = &cgx->pdev->dev;
|
|
dev_err(dev, "cgx port %d:%d cmd timeout\n",
|
|
cgx->cgx_id, lmac->lmac_id);
|
|
err = -EIO;
|
|
goto unlock;
|
|
}
|
|
|
|
/* we have a valid command response */
|
|
smp_rmb(); /* Ensure the latest updates are visible */
|
|
*resp = lmac->resp;
|
|
|
|
unlock:
|
|
mutex_unlock(&lmac->cmd_lock);
|
|
|
|
return err;
|
|
}
|
|
|
|
static inline int cgx_fwi_cmd_generic(u64 req, u64 *resp,
|
|
struct cgx *cgx, int lmac_id)
|
|
{
|
|
struct lmac *lmac;
|
|
int err;
|
|
|
|
lmac = lmac_pdata(lmac_id, cgx);
|
|
if (!lmac)
|
|
return -ENODEV;
|
|
|
|
err = cgx_fwi_cmd_send(req, resp, lmac);
|
|
|
|
/* Check for valid response */
|
|
if (!err) {
|
|
if (FIELD_GET(EVTREG_STAT, *resp) == CGX_STAT_FAIL)
|
|
return -EIO;
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
return err;
|
|
}
|
|
|
|
static inline void cgx_link_usertable_init(void)
|
|
{
|
|
cgx_speed_mbps[CGX_LINK_NONE] = 0;
|
|
cgx_speed_mbps[CGX_LINK_10M] = 10;
|
|
cgx_speed_mbps[CGX_LINK_100M] = 100;
|
|
cgx_speed_mbps[CGX_LINK_1G] = 1000;
|
|
cgx_speed_mbps[CGX_LINK_2HG] = 2500;
|
|
cgx_speed_mbps[CGX_LINK_5G] = 5000;
|
|
cgx_speed_mbps[CGX_LINK_10G] = 10000;
|
|
cgx_speed_mbps[CGX_LINK_20G] = 20000;
|
|
cgx_speed_mbps[CGX_LINK_25G] = 25000;
|
|
cgx_speed_mbps[CGX_LINK_40G] = 40000;
|
|
cgx_speed_mbps[CGX_LINK_50G] = 50000;
|
|
cgx_speed_mbps[CGX_LINK_100G] = 100000;
|
|
|
|
cgx_lmactype_string[LMAC_MODE_SGMII] = "SGMII";
|
|
cgx_lmactype_string[LMAC_MODE_XAUI] = "XAUI";
|
|
cgx_lmactype_string[LMAC_MODE_RXAUI] = "RXAUI";
|
|
cgx_lmactype_string[LMAC_MODE_10G_R] = "10G_R";
|
|
cgx_lmactype_string[LMAC_MODE_40G_R] = "40G_R";
|
|
cgx_lmactype_string[LMAC_MODE_QSGMII] = "QSGMII";
|
|
cgx_lmactype_string[LMAC_MODE_25G_R] = "25G_R";
|
|
cgx_lmactype_string[LMAC_MODE_50G_R] = "50G_R";
|
|
cgx_lmactype_string[LMAC_MODE_100G_R] = "100G_R";
|
|
cgx_lmactype_string[LMAC_MODE_USXGMII] = "USXGMII";
|
|
}
|
|
|
|
static inline void link_status_user_format(u64 lstat,
|
|
struct cgx_link_user_info *linfo,
|
|
struct cgx *cgx, u8 lmac_id)
|
|
{
|
|
char *lmac_string;
|
|
|
|
linfo->link_up = FIELD_GET(RESP_LINKSTAT_UP, lstat);
|
|
linfo->full_duplex = FIELD_GET(RESP_LINKSTAT_FDUPLEX, lstat);
|
|
linfo->speed = cgx_speed_mbps[FIELD_GET(RESP_LINKSTAT_SPEED, lstat)];
|
|
linfo->lmac_type_id = cgx_get_lmac_type(cgx, lmac_id);
|
|
lmac_string = cgx_lmactype_string[linfo->lmac_type_id];
|
|
strncpy(linfo->lmac_type, lmac_string, LMACTYPE_STR_LEN - 1);
|
|
}
|
|
|
|
/* Hardware event handlers */
|
|
static inline void cgx_link_change_handler(u64 lstat,
|
|
struct lmac *lmac)
|
|
{
|
|
struct cgx_link_user_info *linfo;
|
|
struct cgx *cgx = lmac->cgx;
|
|
struct cgx_link_event event;
|
|
struct device *dev;
|
|
int err_type;
|
|
|
|
dev = &cgx->pdev->dev;
|
|
|
|
link_status_user_format(lstat, &event.link_uinfo, cgx, lmac->lmac_id);
|
|
err_type = FIELD_GET(RESP_LINKSTAT_ERRTYPE, lstat);
|
|
|
|
event.cgx_id = cgx->cgx_id;
|
|
event.lmac_id = lmac->lmac_id;
|
|
|
|
/* update the local copy of link status */
|
|
lmac->link_info = event.link_uinfo;
|
|
linfo = &lmac->link_info;
|
|
|
|
/* Ensure callback doesn't get unregistered until we finish it */
|
|
spin_lock(&lmac->event_cb_lock);
|
|
|
|
if (!lmac->event_cb.notify_link_chg) {
|
|
dev_dbg(dev, "cgx port %d:%d Link change handler null",
|
|
cgx->cgx_id, lmac->lmac_id);
|
|
if (err_type != CGX_ERR_NONE) {
|
|
dev_err(dev, "cgx port %d:%d Link error %d\n",
|
|
cgx->cgx_id, lmac->lmac_id, err_type);
|
|
}
|
|
dev_info(dev, "cgx port %d:%d Link is %s %d Mbps\n",
|
|
cgx->cgx_id, lmac->lmac_id,
|
|
linfo->link_up ? "UP" : "DOWN", linfo->speed);
|
|
goto err;
|
|
}
|
|
|
|
if (lmac->event_cb.notify_link_chg(&event, lmac->event_cb.data))
|
|
dev_err(dev, "event notification failure\n");
|
|
err:
|
|
spin_unlock(&lmac->event_cb_lock);
|
|
}
|
|
|
|
static inline bool cgx_cmdresp_is_linkevent(u64 event)
|
|
{
|
|
u8 id;
|
|
|
|
id = FIELD_GET(EVTREG_ID, event);
|
|
if (id == CGX_CMD_LINK_BRING_UP ||
|
|
id == CGX_CMD_LINK_BRING_DOWN)
|
|
return true;
|
|
else
|
|
return false;
|
|
}
|
|
|
|
static inline bool cgx_event_is_linkevent(u64 event)
|
|
{
|
|
if (FIELD_GET(EVTREG_ID, event) == CGX_EVT_LINK_CHANGE)
|
|
return true;
|
|
else
|
|
return false;
|
|
}
|
|
|
|
static inline int cgx_fwi_get_mkex_prfl_sz(u64 *prfl_sz,
|
|
struct cgx *cgx)
|
|
{
|
|
u64 req = 0;
|
|
u64 resp;
|
|
int err;
|
|
|
|
req = FIELD_SET(CMDREG_ID, CGX_CMD_GET_MKEX_PRFL_SIZE, req);
|
|
err = cgx_fwi_cmd_generic(req, &resp, cgx, 0);
|
|
if (!err)
|
|
*prfl_sz = FIELD_GET(RESP_MKEX_PRFL_SIZE, resp);
|
|
|
|
return err;
|
|
}
|
|
|
|
static inline int cgx_fwi_get_mkex_prfl_addr(u64 *prfl_addr,
|
|
struct cgx *cgx)
|
|
{
|
|
u64 req = 0;
|
|
u64 resp;
|
|
int err;
|
|
|
|
req = FIELD_SET(CMDREG_ID, CGX_CMD_GET_MKEX_PRFL_ADDR, req);
|
|
err = cgx_fwi_cmd_generic(req, &resp, cgx, 0);
|
|
if (!err)
|
|
*prfl_addr = FIELD_GET(RESP_MKEX_PRFL_ADDR, resp);
|
|
|
|
return err;
|
|
}
|
|
|
|
int cgx_get_mkex_prfl_info(u64 *addr, u64 *size)
|
|
{
|
|
struct cgx *cgx_dev;
|
|
int err;
|
|
|
|
if (!addr || !size)
|
|
return -EINVAL;
|
|
|
|
cgx_dev = list_first_entry(&cgx_list, struct cgx, cgx_list);
|
|
if (!cgx_dev)
|
|
return -ENXIO;
|
|
|
|
err = cgx_fwi_get_mkex_prfl_sz(size, cgx_dev);
|
|
if (err)
|
|
return -EIO;
|
|
|
|
err = cgx_fwi_get_mkex_prfl_addr(addr, cgx_dev);
|
|
if (err)
|
|
return -EIO;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_get_mkex_prfl_info);
|
|
|
|
static irqreturn_t cgx_fwi_event_handler(int irq, void *data)
|
|
{
|
|
struct lmac *lmac = data;
|
|
struct cgx *cgx;
|
|
u64 event;
|
|
|
|
cgx = lmac->cgx;
|
|
|
|
event = cgx_read(cgx, lmac->lmac_id, CGX_EVENT_REG);
|
|
|
|
if (!FIELD_GET(EVTREG_ACK, event))
|
|
return IRQ_NONE;
|
|
|
|
switch (FIELD_GET(EVTREG_EVT_TYPE, event)) {
|
|
case CGX_EVT_CMD_RESP:
|
|
/* Copy the response. Since only one command is active at a
|
|
* time, there is no way a response can get overwritten
|
|
*/
|
|
lmac->resp = event;
|
|
/* Ensure response is updated before thread context starts */
|
|
smp_wmb();
|
|
|
|
/* There wont be separate events for link change initiated from
|
|
* software; Hence report the command responses as events
|
|
*/
|
|
if (cgx_cmdresp_is_linkevent(event))
|
|
cgx_link_change_handler(event, lmac);
|
|
|
|
/* Release thread waiting for completion */
|
|
lmac->cmd_pend = false;
|
|
wake_up_interruptible(&lmac->wq_cmd_cmplt);
|
|
break;
|
|
case CGX_EVT_ASYNC:
|
|
if (cgx_event_is_linkevent(event))
|
|
cgx_link_change_handler(event, lmac);
|
|
break;
|
|
}
|
|
|
|
/* Any new event or command response will be posted by firmware
|
|
* only after the current status is acked.
|
|
* Ack the interrupt register as well.
|
|
*/
|
|
cgx_write(lmac->cgx, lmac->lmac_id, CGX_EVENT_REG, 0);
|
|
cgx_write(lmac->cgx, lmac->lmac_id, CGXX_CMRX_INT, FW_CGX_INT);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/* APIs for PHY management using CGX firmware interface */
|
|
|
|
/* callback registration for hardware events like link change */
|
|
int cgx_lmac_evh_register(struct cgx_event_cb *cb, void *cgxd, int lmac_id)
|
|
{
|
|
struct cgx *cgx = cgxd;
|
|
struct lmac *lmac;
|
|
|
|
lmac = lmac_pdata(lmac_id, cgx);
|
|
if (!lmac)
|
|
return -ENODEV;
|
|
|
|
lmac->event_cb = *cb;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_lmac_evh_register);
|
|
|
|
int cgx_lmac_evh_unregister(void *cgxd, int lmac_id)
|
|
{
|
|
struct lmac *lmac;
|
|
unsigned long flags;
|
|
struct cgx *cgx = cgxd;
|
|
|
|
lmac = lmac_pdata(lmac_id, cgx);
|
|
if (!lmac)
|
|
return -ENODEV;
|
|
|
|
spin_lock_irqsave(&lmac->event_cb_lock, flags);
|
|
lmac->event_cb.notify_link_chg = NULL;
|
|
lmac->event_cb.data = NULL;
|
|
spin_unlock_irqrestore(&lmac->event_cb_lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_lmac_evh_unregister);
|
|
|
|
static int cgx_fwi_link_change(struct cgx *cgx, int lmac_id, bool enable)
|
|
{
|
|
u64 req = 0;
|
|
u64 resp;
|
|
|
|
if (enable)
|
|
req = FIELD_SET(CMDREG_ID, CGX_CMD_LINK_BRING_UP, req);
|
|
else
|
|
req = FIELD_SET(CMDREG_ID, CGX_CMD_LINK_BRING_DOWN, req);
|
|
|
|
return cgx_fwi_cmd_generic(req, &resp, cgx, lmac_id);
|
|
}
|
|
|
|
static inline int cgx_fwi_read_version(u64 *resp, struct cgx *cgx)
|
|
{
|
|
u64 req = 0;
|
|
|
|
req = FIELD_SET(CMDREG_ID, CGX_CMD_GET_FW_VER, req);
|
|
return cgx_fwi_cmd_generic(req, resp, cgx, 0);
|
|
}
|
|
|
|
static int cgx_lmac_verify_fwi_version(struct cgx *cgx)
|
|
{
|
|
struct device *dev = &cgx->pdev->dev;
|
|
int major_ver, minor_ver;
|
|
u64 resp;
|
|
int err;
|
|
|
|
if (!cgx->lmac_count)
|
|
return 0;
|
|
|
|
err = cgx_fwi_read_version(&resp, cgx);
|
|
if (err)
|
|
return err;
|
|
|
|
major_ver = FIELD_GET(RESP_MAJOR_VER, resp);
|
|
minor_ver = FIELD_GET(RESP_MINOR_VER, resp);
|
|
dev_dbg(dev, "Firmware command interface version = %d.%d\n",
|
|
major_ver, minor_ver);
|
|
if (major_ver != CGX_FIRMWARE_MAJOR_VER ||
|
|
minor_ver != CGX_FIRMWARE_MINOR_VER)
|
|
return -EIO;
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
static void cgx_lmac_linkup_work(struct work_struct *work)
|
|
{
|
|
struct cgx *cgx = container_of(work, struct cgx, cgx_cmd_work);
|
|
struct device *dev = &cgx->pdev->dev;
|
|
int i, err;
|
|
|
|
/* Do Link up for all the lmacs */
|
|
for (i = 0; i < cgx->lmac_count; i++) {
|
|
err = cgx_fwi_link_change(cgx, i, true);
|
|
if (err)
|
|
dev_info(dev, "cgx port %d:%d Link up command failed\n",
|
|
cgx->cgx_id, i);
|
|
}
|
|
}
|
|
|
|
int cgx_lmac_linkup_start(void *cgxd)
|
|
{
|
|
struct cgx *cgx = cgxd;
|
|
|
|
if (!cgx)
|
|
return -ENODEV;
|
|
|
|
queue_work(cgx->cgx_cmd_workq, &cgx->cgx_cmd_work);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(cgx_lmac_linkup_start);
|
|
|
|
static int cgx_lmac_init(struct cgx *cgx)
|
|
{
|
|
struct lmac *lmac;
|
|
int i, err;
|
|
|
|
cgx->lmac_count = cgx_read(cgx, 0, CGXX_CMRX_RX_LMACS) & 0x7;
|
|
if (cgx->lmac_count > MAX_LMAC_PER_CGX)
|
|
cgx->lmac_count = MAX_LMAC_PER_CGX;
|
|
|
|
for (i = 0; i < cgx->lmac_count; i++) {
|
|
lmac = kcalloc(1, sizeof(struct lmac), GFP_KERNEL);
|
|
if (!lmac)
|
|
return -ENOMEM;
|
|
lmac->name = kcalloc(1, sizeof("cgx_fwi_xxx_yyy"), GFP_KERNEL);
|
|
if (!lmac->name)
|
|
return -ENOMEM;
|
|
sprintf(lmac->name, "cgx_fwi_%d_%d", cgx->cgx_id, i);
|
|
lmac->lmac_id = i;
|
|
lmac->cgx = cgx;
|
|
init_waitqueue_head(&lmac->wq_cmd_cmplt);
|
|
mutex_init(&lmac->cmd_lock);
|
|
spin_lock_init(&lmac->event_cb_lock);
|
|
err = request_irq(pci_irq_vector(cgx->pdev,
|
|
CGX_LMAC_FWI + i * 9),
|
|
cgx_fwi_event_handler, 0, lmac->name, lmac);
|
|
if (err)
|
|
return err;
|
|
|
|
/* Enable interrupt */
|
|
cgx_write(cgx, lmac->lmac_id, CGXX_CMRX_INT_ENA_W1S,
|
|
FW_CGX_INT);
|
|
|
|
/* Add reference */
|
|
cgx->lmac_idmap[i] = lmac;
|
|
}
|
|
|
|
return cgx_lmac_verify_fwi_version(cgx);
|
|
}
|
|
|
|
static int cgx_lmac_exit(struct cgx *cgx)
|
|
{
|
|
struct lmac *lmac;
|
|
int i;
|
|
|
|
if (cgx->cgx_cmd_workq) {
|
|
flush_workqueue(cgx->cgx_cmd_workq);
|
|
destroy_workqueue(cgx->cgx_cmd_workq);
|
|
cgx->cgx_cmd_workq = NULL;
|
|
}
|
|
|
|
/* Free all lmac related resources */
|
|
for (i = 0; i < cgx->lmac_count; i++) {
|
|
lmac = cgx->lmac_idmap[i];
|
|
if (!lmac)
|
|
continue;
|
|
free_irq(pci_irq_vector(cgx->pdev, CGX_LMAC_FWI + i * 9), lmac);
|
|
kfree(lmac->name);
|
|
kfree(lmac);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int cgx_probe(struct pci_dev *pdev, const struct pci_device_id *id)
|
|
{
|
|
struct device *dev = &pdev->dev;
|
|
struct cgx *cgx;
|
|
int err, nvec;
|
|
|
|
cgx = devm_kzalloc(dev, sizeof(*cgx), GFP_KERNEL);
|
|
if (!cgx)
|
|
return -ENOMEM;
|
|
cgx->pdev = pdev;
|
|
|
|
pci_set_drvdata(pdev, cgx);
|
|
|
|
err = pci_enable_device(pdev);
|
|
if (err) {
|
|
dev_err(dev, "Failed to enable PCI device\n");
|
|
pci_set_drvdata(pdev, NULL);
|
|
return err;
|
|
}
|
|
|
|
err = pci_request_regions(pdev, DRV_NAME);
|
|
if (err) {
|
|
dev_err(dev, "PCI request regions failed 0x%x\n", err);
|
|
goto err_disable_device;
|
|
}
|
|
|
|
/* MAP configuration registers */
|
|
cgx->reg_base = pcim_iomap(pdev, PCI_CFG_REG_BAR_NUM, 0);
|
|
if (!cgx->reg_base) {
|
|
dev_err(dev, "CGX: Cannot map CSR memory space, aborting\n");
|
|
err = -ENOMEM;
|
|
goto err_release_regions;
|
|
}
|
|
|
|
nvec = CGX_NVEC;
|
|
err = pci_alloc_irq_vectors(pdev, nvec, nvec, PCI_IRQ_MSIX);
|
|
if (err < 0 || err != nvec) {
|
|
dev_err(dev, "Request for %d msix vectors failed, err %d\n",
|
|
nvec, err);
|
|
goto err_release_regions;
|
|
}
|
|
|
|
cgx->cgx_id = (pci_resource_start(pdev, PCI_CFG_REG_BAR_NUM) >> 24)
|
|
& CGX_ID_MASK;
|
|
|
|
/* init wq for processing linkup requests */
|
|
INIT_WORK(&cgx->cgx_cmd_work, cgx_lmac_linkup_work);
|
|
cgx->cgx_cmd_workq = alloc_workqueue("cgx_cmd_workq", 0, 0);
|
|
if (!cgx->cgx_cmd_workq) {
|
|
dev_err(dev, "alloc workqueue failed for cgx cmd");
|
|
err = -ENOMEM;
|
|
goto err_release_regions;
|
|
}
|
|
|
|
list_add(&cgx->cgx_list, &cgx_list);
|
|
|
|
cgx_link_usertable_init();
|
|
|
|
err = cgx_lmac_init(cgx);
|
|
if (err)
|
|
goto err_release_lmac;
|
|
|
|
return 0;
|
|
|
|
err_release_lmac:
|
|
cgx_lmac_exit(cgx);
|
|
list_del(&cgx->cgx_list);
|
|
err_release_regions:
|
|
pci_release_regions(pdev);
|
|
err_disable_device:
|
|
pci_disable_device(pdev);
|
|
pci_set_drvdata(pdev, NULL);
|
|
return err;
|
|
}
|
|
|
|
static void cgx_remove(struct pci_dev *pdev)
|
|
{
|
|
struct cgx *cgx = pci_get_drvdata(pdev);
|
|
|
|
cgx_lmac_exit(cgx);
|
|
list_del(&cgx->cgx_list);
|
|
pci_free_irq_vectors(pdev);
|
|
pci_release_regions(pdev);
|
|
pci_disable_device(pdev);
|
|
pci_set_drvdata(pdev, NULL);
|
|
}
|
|
|
|
struct pci_driver cgx_driver = {
|
|
.name = DRV_NAME,
|
|
.id_table = cgx_id_table,
|
|
.probe = cgx_probe,
|
|
.remove = cgx_remove,
|
|
};
|