mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-26 02:05:29 +07:00
cf718eaa8f
Enabled the PF->VF Mailbox support. Mailbox message are interpreted as {type, opcode, data}. Supported message types are REQ, ACK and NACK. Signed-off-by: Srikanth Jampala <Jampala.Srikanth@cavium.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
454 lines
11 KiB
C
454 lines
11 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include <linux/pci.h>
|
|
#include <linux/printk.h>
|
|
#include <linux/slab.h>
|
|
|
|
#include "nitrox_dev.h"
|
|
#include "nitrox_csr.h"
|
|
#include "nitrox_common.h"
|
|
#include "nitrox_hal.h"
|
|
#include "nitrox_mbx.h"
|
|
|
|
/**
|
|
* One vector for each type of ring
|
|
* - NPS packet ring, AQMQ ring and ZQMQ ring
|
|
*/
|
|
#define NR_RING_VECTORS 3
|
|
#define NR_NON_RING_VECTORS 1
|
|
/* base entry for packet ring/port */
|
|
#define PKT_RING_MSIX_BASE 0
|
|
#define NON_RING_MSIX_BASE 192
|
|
|
|
/**
|
|
* nps_pkt_slc_isr - IRQ handler for NPS solicit port
|
|
* @irq: irq number
|
|
* @data: argument
|
|
*/
|
|
static irqreturn_t nps_pkt_slc_isr(int irq, void *data)
|
|
{
|
|
struct nitrox_q_vector *qvec = data;
|
|
union nps_pkt_slc_cnts slc_cnts;
|
|
struct nitrox_cmdq *cmdq = qvec->cmdq;
|
|
|
|
slc_cnts.value = readq(cmdq->compl_cnt_csr_addr);
|
|
/* New packet on SLC output port */
|
|
if (slc_cnts.s.slc_int)
|
|
tasklet_hi_schedule(&qvec->resp_tasklet);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static void clear_nps_core_err_intr(struct nitrox_device *ndev)
|
|
{
|
|
u64 value;
|
|
|
|
/* Write 1 to clear */
|
|
value = nitrox_read_csr(ndev, NPS_CORE_INT);
|
|
nitrox_write_csr(ndev, NPS_CORE_INT, value);
|
|
|
|
dev_err_ratelimited(DEV(ndev), "NSP_CORE_INT 0x%016llx\n", value);
|
|
}
|
|
|
|
static void clear_nps_pkt_err_intr(struct nitrox_device *ndev)
|
|
{
|
|
union nps_pkt_int pkt_int;
|
|
unsigned long value, offset;
|
|
int i;
|
|
|
|
pkt_int.value = nitrox_read_csr(ndev, NPS_PKT_INT);
|
|
dev_err_ratelimited(DEV(ndev), "NPS_PKT_INT 0x%016llx\n",
|
|
pkt_int.value);
|
|
|
|
if (pkt_int.s.slc_err) {
|
|
offset = NPS_PKT_SLC_ERR_TYPE;
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
dev_err_ratelimited(DEV(ndev),
|
|
"NPS_PKT_SLC_ERR_TYPE 0x%016lx\n", value);
|
|
|
|
offset = NPS_PKT_SLC_RERR_LO;
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
/* enable the solicit ports */
|
|
for_each_set_bit(i, &value, BITS_PER_LONG)
|
|
enable_pkt_solicit_port(ndev, i);
|
|
|
|
dev_err_ratelimited(DEV(ndev),
|
|
"NPS_PKT_SLC_RERR_LO 0x%016lx\n", value);
|
|
|
|
offset = NPS_PKT_SLC_RERR_HI;
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
dev_err_ratelimited(DEV(ndev),
|
|
"NPS_PKT_SLC_RERR_HI 0x%016lx\n", value);
|
|
}
|
|
|
|
if (pkt_int.s.in_err) {
|
|
offset = NPS_PKT_IN_ERR_TYPE;
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
dev_err_ratelimited(DEV(ndev),
|
|
"NPS_PKT_IN_ERR_TYPE 0x%016lx\n", value);
|
|
offset = NPS_PKT_IN_RERR_LO;
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
/* enable the input ring */
|
|
for_each_set_bit(i, &value, BITS_PER_LONG)
|
|
enable_pkt_input_ring(ndev, i);
|
|
|
|
dev_err_ratelimited(DEV(ndev),
|
|
"NPS_PKT_IN_RERR_LO 0x%016lx\n", value);
|
|
|
|
offset = NPS_PKT_IN_RERR_HI;
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
dev_err_ratelimited(DEV(ndev),
|
|
"NPS_PKT_IN_RERR_HI 0x%016lx\n", value);
|
|
}
|
|
}
|
|
|
|
static void clear_pom_err_intr(struct nitrox_device *ndev)
|
|
{
|
|
u64 value;
|
|
|
|
value = nitrox_read_csr(ndev, POM_INT);
|
|
nitrox_write_csr(ndev, POM_INT, value);
|
|
dev_err_ratelimited(DEV(ndev), "POM_INT 0x%016llx\n", value);
|
|
}
|
|
|
|
static void clear_pem_err_intr(struct nitrox_device *ndev)
|
|
{
|
|
u64 value;
|
|
|
|
value = nitrox_read_csr(ndev, PEM0_INT);
|
|
nitrox_write_csr(ndev, PEM0_INT, value);
|
|
dev_err_ratelimited(DEV(ndev), "PEM(0)_INT 0x%016llx\n", value);
|
|
}
|
|
|
|
static void clear_lbc_err_intr(struct nitrox_device *ndev)
|
|
{
|
|
union lbc_int lbc_int;
|
|
u64 value, offset;
|
|
int i;
|
|
|
|
lbc_int.value = nitrox_read_csr(ndev, LBC_INT);
|
|
dev_err_ratelimited(DEV(ndev), "LBC_INT 0x%016llx\n", lbc_int.value);
|
|
|
|
if (lbc_int.s.dma_rd_err) {
|
|
for (i = 0; i < NR_CLUSTERS; i++) {
|
|
offset = EFL_CORE_VF_ERR_INT0X(i);
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
offset = EFL_CORE_VF_ERR_INT1X(i);
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
}
|
|
}
|
|
|
|
if (lbc_int.s.cam_soft_err) {
|
|
dev_err_ratelimited(DEV(ndev), "CAM_SOFT_ERR, invalidating LBC\n");
|
|
invalidate_lbc(ndev);
|
|
}
|
|
|
|
if (lbc_int.s.pref_dat_len_mismatch_err) {
|
|
offset = LBC_PLM_VF1_64_INT;
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
offset = LBC_PLM_VF65_128_INT;
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
}
|
|
|
|
if (lbc_int.s.rd_dat_len_mismatch_err) {
|
|
offset = LBC_ELM_VF1_64_INT;
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
offset = LBC_ELM_VF65_128_INT;
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
}
|
|
nitrox_write_csr(ndev, LBC_INT, lbc_int.value);
|
|
}
|
|
|
|
static void clear_efl_err_intr(struct nitrox_device *ndev)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < NR_CLUSTERS; i++) {
|
|
union efl_core_int core_int;
|
|
u64 value, offset;
|
|
|
|
offset = EFL_CORE_INTX(i);
|
|
core_int.value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, core_int.value);
|
|
dev_err_ratelimited(DEV(ndev), "ELF_CORE(%d)_INT 0x%016llx\n",
|
|
i, core_int.value);
|
|
if (core_int.s.se_err) {
|
|
offset = EFL_CORE_SE_ERR_INTX(i);
|
|
value = nitrox_read_csr(ndev, offset);
|
|
nitrox_write_csr(ndev, offset, value);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void clear_bmi_err_intr(struct nitrox_device *ndev)
|
|
{
|
|
u64 value;
|
|
|
|
value = nitrox_read_csr(ndev, BMI_INT);
|
|
nitrox_write_csr(ndev, BMI_INT, value);
|
|
dev_err_ratelimited(DEV(ndev), "BMI_INT 0x%016llx\n", value);
|
|
}
|
|
|
|
static void nps_core_int_tasklet(unsigned long data)
|
|
{
|
|
struct nitrox_q_vector *qvec = (void *)(uintptr_t)(data);
|
|
struct nitrox_device *ndev = qvec->ndev;
|
|
|
|
/* if pf mode do queue recovery */
|
|
if (ndev->mode == __NDEV_MODE_PF) {
|
|
} else {
|
|
/**
|
|
* if VF(s) enabled communicate the error information
|
|
* to VF(s)
|
|
*/
|
|
}
|
|
}
|
|
|
|
/**
|
|
* nps_core_int_isr - interrupt handler for NITROX errors and
|
|
* mailbox communication
|
|
*/
|
|
static irqreturn_t nps_core_int_isr(int irq, void *data)
|
|
{
|
|
struct nitrox_q_vector *qvec = data;
|
|
struct nitrox_device *ndev = qvec->ndev;
|
|
union nps_core_int_active core_int;
|
|
|
|
core_int.value = nitrox_read_csr(ndev, NPS_CORE_INT_ACTIVE);
|
|
|
|
if (core_int.s.nps_core)
|
|
clear_nps_core_err_intr(ndev);
|
|
|
|
if (core_int.s.nps_pkt)
|
|
clear_nps_pkt_err_intr(ndev);
|
|
|
|
if (core_int.s.pom)
|
|
clear_pom_err_intr(ndev);
|
|
|
|
if (core_int.s.pem)
|
|
clear_pem_err_intr(ndev);
|
|
|
|
if (core_int.s.lbc)
|
|
clear_lbc_err_intr(ndev);
|
|
|
|
if (core_int.s.efl)
|
|
clear_efl_err_intr(ndev);
|
|
|
|
if (core_int.s.bmi)
|
|
clear_bmi_err_intr(ndev);
|
|
|
|
/* Mailbox interrupt */
|
|
if (core_int.s.mbox)
|
|
nitrox_pf2vf_mbox_handler(ndev);
|
|
|
|
/* If more work callback the ISR, set resend */
|
|
core_int.s.resend = 1;
|
|
nitrox_write_csr(ndev, NPS_CORE_INT_ACTIVE, core_int.value);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
void nitrox_unregister_interrupts(struct nitrox_device *ndev)
|
|
{
|
|
struct pci_dev *pdev = ndev->pdev;
|
|
int i;
|
|
|
|
for (i = 0; i < ndev->num_vecs; i++) {
|
|
struct nitrox_q_vector *qvec;
|
|
int vec;
|
|
|
|
qvec = ndev->qvec + i;
|
|
if (!qvec->valid)
|
|
continue;
|
|
|
|
/* get the vector number */
|
|
vec = pci_irq_vector(pdev, i);
|
|
irq_set_affinity_hint(vec, NULL);
|
|
free_irq(vec, qvec);
|
|
|
|
tasklet_disable(&qvec->resp_tasklet);
|
|
tasklet_kill(&qvec->resp_tasklet);
|
|
qvec->valid = false;
|
|
}
|
|
kfree(ndev->qvec);
|
|
ndev->qvec = NULL;
|
|
pci_free_irq_vectors(pdev);
|
|
}
|
|
|
|
int nitrox_register_interrupts(struct nitrox_device *ndev)
|
|
{
|
|
struct pci_dev *pdev = ndev->pdev;
|
|
struct nitrox_q_vector *qvec;
|
|
int nr_vecs, vec, cpu;
|
|
int ret, i;
|
|
|
|
/*
|
|
* PF MSI-X vectors
|
|
*
|
|
* Entry 0: NPS PKT ring 0
|
|
* Entry 1: AQMQ ring 0
|
|
* Entry 2: ZQM ring 0
|
|
* Entry 3: NPS PKT ring 1
|
|
* Entry 4: AQMQ ring 1
|
|
* Entry 5: ZQM ring 1
|
|
* ....
|
|
* Entry 192: NPS_CORE_INT_ACTIVE
|
|
*/
|
|
nr_vecs = pci_msix_vec_count(pdev);
|
|
|
|
/* Enable MSI-X */
|
|
ret = pci_alloc_irq_vectors(pdev, nr_vecs, nr_vecs, PCI_IRQ_MSIX);
|
|
if (ret < 0) {
|
|
dev_err(DEV(ndev), "msix vectors %d alloc failed\n", nr_vecs);
|
|
return ret;
|
|
}
|
|
ndev->num_vecs = nr_vecs;
|
|
|
|
ndev->qvec = kcalloc(nr_vecs, sizeof(*qvec), GFP_KERNEL);
|
|
if (!ndev->qvec) {
|
|
pci_free_irq_vectors(pdev);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* request irqs for packet rings/ports */
|
|
for (i = PKT_RING_MSIX_BASE; i < (nr_vecs - 1); i += NR_RING_VECTORS) {
|
|
qvec = &ndev->qvec[i];
|
|
|
|
qvec->ring = i / NR_RING_VECTORS;
|
|
if (qvec->ring >= ndev->nr_queues)
|
|
break;
|
|
|
|
qvec->cmdq = &ndev->pkt_inq[qvec->ring];
|
|
snprintf(qvec->name, IRQ_NAMESZ, "nitrox-pkt%d", qvec->ring);
|
|
/* get the vector number */
|
|
vec = pci_irq_vector(pdev, i);
|
|
ret = request_irq(vec, nps_pkt_slc_isr, 0, qvec->name, qvec);
|
|
if (ret) {
|
|
dev_err(DEV(ndev), "irq failed for pkt ring/port%d\n",
|
|
qvec->ring);
|
|
goto irq_fail;
|
|
}
|
|
cpu = qvec->ring % num_online_cpus();
|
|
irq_set_affinity_hint(vec, get_cpu_mask(cpu));
|
|
|
|
tasklet_init(&qvec->resp_tasklet, pkt_slc_resp_tasklet,
|
|
(unsigned long)qvec);
|
|
qvec->valid = true;
|
|
}
|
|
|
|
/* request irqs for non ring vectors */
|
|
i = NON_RING_MSIX_BASE;
|
|
qvec = &ndev->qvec[i];
|
|
qvec->ndev = ndev;
|
|
|
|
snprintf(qvec->name, IRQ_NAMESZ, "nitrox-core-int%d", i);
|
|
/* get the vector number */
|
|
vec = pci_irq_vector(pdev, i);
|
|
ret = request_irq(vec, nps_core_int_isr, 0, qvec->name, qvec);
|
|
if (ret) {
|
|
dev_err(DEV(ndev), "irq failed for nitrox-core-int%d\n", i);
|
|
goto irq_fail;
|
|
}
|
|
cpu = num_online_cpus();
|
|
irq_set_affinity_hint(vec, get_cpu_mask(cpu));
|
|
|
|
tasklet_init(&qvec->resp_tasklet, nps_core_int_tasklet,
|
|
(unsigned long)qvec);
|
|
qvec->valid = true;
|
|
|
|
return 0;
|
|
|
|
irq_fail:
|
|
nitrox_unregister_interrupts(ndev);
|
|
return ret;
|
|
}
|
|
|
|
void nitrox_sriov_unregister_interrupts(struct nitrox_device *ndev)
|
|
{
|
|
struct pci_dev *pdev = ndev->pdev;
|
|
int i;
|
|
|
|
for (i = 0; i < ndev->num_vecs; i++) {
|
|
struct nitrox_q_vector *qvec;
|
|
int vec;
|
|
|
|
qvec = ndev->qvec + i;
|
|
if (!qvec->valid)
|
|
continue;
|
|
|
|
vec = ndev->iov.msix.vector;
|
|
irq_set_affinity_hint(vec, NULL);
|
|
free_irq(vec, qvec);
|
|
|
|
tasklet_disable(&qvec->resp_tasklet);
|
|
tasklet_kill(&qvec->resp_tasklet);
|
|
qvec->valid = false;
|
|
}
|
|
kfree(ndev->qvec);
|
|
ndev->qvec = NULL;
|
|
pci_disable_msix(pdev);
|
|
}
|
|
|
|
int nitrox_sriov_register_interupts(struct nitrox_device *ndev)
|
|
{
|
|
struct pci_dev *pdev = ndev->pdev;
|
|
struct nitrox_q_vector *qvec;
|
|
int vec, cpu;
|
|
int ret;
|
|
|
|
/**
|
|
* only non ring vectors i.e Entry 192 is available
|
|
* for PF in SR-IOV mode.
|
|
*/
|
|
ndev->iov.msix.entry = NON_RING_MSIX_BASE;
|
|
ret = pci_enable_msix_exact(pdev, &ndev->iov.msix, NR_NON_RING_VECTORS);
|
|
if (ret) {
|
|
dev_err(DEV(ndev), "failed to allocate nps-core-int%d\n",
|
|
NON_RING_MSIX_BASE);
|
|
return ret;
|
|
}
|
|
|
|
qvec = kcalloc(NR_NON_RING_VECTORS, sizeof(*qvec), GFP_KERNEL);
|
|
if (!qvec) {
|
|
pci_disable_msix(pdev);
|
|
return -ENOMEM;
|
|
}
|
|
qvec->ndev = ndev;
|
|
|
|
ndev->qvec = qvec;
|
|
ndev->num_vecs = NR_NON_RING_VECTORS;
|
|
snprintf(qvec->name, IRQ_NAMESZ, "nitrox-core-int%d",
|
|
NON_RING_MSIX_BASE);
|
|
|
|
vec = ndev->iov.msix.vector;
|
|
ret = request_irq(vec, nps_core_int_isr, 0, qvec->name, qvec);
|
|
if (ret) {
|
|
dev_err(DEV(ndev), "irq failed for nitrox-core-int%d\n",
|
|
NON_RING_MSIX_BASE);
|
|
goto iov_irq_fail;
|
|
}
|
|
cpu = num_online_cpus();
|
|
irq_set_affinity_hint(vec, get_cpu_mask(cpu));
|
|
|
|
tasklet_init(&qvec->resp_tasklet, nps_core_int_tasklet,
|
|
(unsigned long)qvec);
|
|
qvec->valid = true;
|
|
|
|
return 0;
|
|
|
|
iov_irq_fail:
|
|
nitrox_sriov_unregister_interrupts(ndev);
|
|
return ret;
|
|
}
|