mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-30 01:26:44 +07:00
ed21cfe207
On XLR/XLS, the cpu cores communicate with fast on-chip devices (e.g. network accelerator, security engine etc.) using the Fast Messaging Network(FMN). The FMN queues and credits needs to be configured and intialized before it can be used. The co-processor 2 on XLR/XLS CPU cores has registers for FMN access, and the XLR/XLS has custom instructions for sending and loading messages. The FMN can deliver also per-cpu interrupts when messages are available at the CPU. This patch adds FMN initialization, adds interrupt setup and handling, and also provides support for sending and receiving FMN messages. Signed-off-by: Ganesan Ramalingam <ganesanr@broadcom.com> Signed-off-by: Jayachandran C <jchandra@broadcom.com> Patchwork: http://patchwork.linux-mips.org/patch/4468 Signed-off-by: John Crispin <blogic@openwrt.org>
283 lines
7.1 KiB
C
283 lines
7.1 KiB
C
/*
|
|
* Copyright 2003-2011 NetLogic Microsystems, Inc. (NetLogic). All rights
|
|
* reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the NetLogic
|
|
* license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in
|
|
* the documentation and/or other materials provided with the
|
|
* distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY NETLOGIC ``AS IS'' AND ANY EXPRESS OR
|
|
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
|
|
* WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL NETLOGIC OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
|
|
* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
|
|
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE
|
|
* OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN
|
|
* IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/init.h>
|
|
#include <linux/linkage.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/irq.h>
|
|
|
|
#include <asm/errno.h>
|
|
#include <asm/signal.h>
|
|
#include <asm/ptrace.h>
|
|
#include <asm/mipsregs.h>
|
|
#include <asm/thread_info.h>
|
|
|
|
#include <asm/netlogic/mips-extns.h>
|
|
#include <asm/netlogic/interrupt.h>
|
|
#include <asm/netlogic/haldefs.h>
|
|
#include <asm/netlogic/common.h>
|
|
|
|
#if defined(CONFIG_CPU_XLP)
|
|
#include <asm/netlogic/xlp-hal/iomap.h>
|
|
#include <asm/netlogic/xlp-hal/xlp.h>
|
|
#include <asm/netlogic/xlp-hal/pic.h>
|
|
#elif defined(CONFIG_CPU_XLR)
|
|
#include <asm/netlogic/xlr/iomap.h>
|
|
#include <asm/netlogic/xlr/pic.h>
|
|
#include <asm/netlogic/xlr/fmn.h>
|
|
#else
|
|
#error "Unknown CPU"
|
|
#endif
|
|
|
|
#ifdef CONFIG_SMP
|
|
#define SMP_IRQ_MASK ((1ULL << IRQ_IPI_SMP_FUNCTION) | \
|
|
(1ULL << IRQ_IPI_SMP_RESCHEDULE))
|
|
#else
|
|
#define SMP_IRQ_MASK 0
|
|
#endif
|
|
#define PERCPU_IRQ_MASK (SMP_IRQ_MASK | (1ull << IRQ_TIMER) | \
|
|
(1ull << IRQ_FMN))
|
|
|
|
struct nlm_pic_irq {
|
|
void (*extra_ack)(struct irq_data *);
|
|
struct nlm_soc_info *node;
|
|
int picirq;
|
|
int irt;
|
|
int flags;
|
|
};
|
|
|
|
static void xlp_pic_enable(struct irq_data *d)
|
|
{
|
|
unsigned long flags;
|
|
struct nlm_pic_irq *pd = irq_data_get_irq_handler_data(d);
|
|
|
|
BUG_ON(!pd);
|
|
spin_lock_irqsave(&pd->node->piclock, flags);
|
|
nlm_pic_enable_irt(pd->node->picbase, pd->irt);
|
|
spin_unlock_irqrestore(&pd->node->piclock, flags);
|
|
}
|
|
|
|
static void xlp_pic_disable(struct irq_data *d)
|
|
{
|
|
struct nlm_pic_irq *pd = irq_data_get_irq_handler_data(d);
|
|
unsigned long flags;
|
|
|
|
BUG_ON(!pd);
|
|
spin_lock_irqsave(&pd->node->piclock, flags);
|
|
nlm_pic_disable_irt(pd->node->picbase, pd->irt);
|
|
spin_unlock_irqrestore(&pd->node->piclock, flags);
|
|
}
|
|
|
|
static void xlp_pic_mask_ack(struct irq_data *d)
|
|
{
|
|
struct nlm_pic_irq *pd = irq_data_get_irq_handler_data(d);
|
|
uint64_t mask = 1ull << pd->picirq;
|
|
|
|
write_c0_eirr(mask); /* ack by writing EIRR */
|
|
}
|
|
|
|
static void xlp_pic_unmask(struct irq_data *d)
|
|
{
|
|
struct nlm_pic_irq *pd = irq_data_get_irq_handler_data(d);
|
|
|
|
if (!pd)
|
|
return;
|
|
|
|
if (pd->extra_ack)
|
|
pd->extra_ack(d);
|
|
|
|
/* Ack is a single write, no need to lock */
|
|
nlm_pic_ack(pd->node->picbase, pd->irt);
|
|
}
|
|
|
|
static struct irq_chip xlp_pic = {
|
|
.name = "XLP-PIC",
|
|
.irq_enable = xlp_pic_enable,
|
|
.irq_disable = xlp_pic_disable,
|
|
.irq_mask_ack = xlp_pic_mask_ack,
|
|
.irq_unmask = xlp_pic_unmask,
|
|
};
|
|
|
|
static void cpuintr_disable(struct irq_data *d)
|
|
{
|
|
uint64_t eimr;
|
|
uint64_t mask = 1ull << d->irq;
|
|
|
|
eimr = read_c0_eimr();
|
|
write_c0_eimr(eimr & ~mask);
|
|
}
|
|
|
|
static void cpuintr_enable(struct irq_data *d)
|
|
{
|
|
uint64_t eimr;
|
|
uint64_t mask = 1ull << d->irq;
|
|
|
|
eimr = read_c0_eimr();
|
|
write_c0_eimr(eimr | mask);
|
|
}
|
|
|
|
static void cpuintr_ack(struct irq_data *d)
|
|
{
|
|
uint64_t mask = 1ull << d->irq;
|
|
|
|
write_c0_eirr(mask);
|
|
}
|
|
|
|
static void cpuintr_nop(struct irq_data *d)
|
|
{
|
|
WARN(d->irq >= PIC_IRQ_BASE, "Bad irq %d", d->irq);
|
|
}
|
|
|
|
/*
|
|
* Chip definition for CPU originated interrupts(timer, msg) and
|
|
* IPIs
|
|
*/
|
|
struct irq_chip nlm_cpu_intr = {
|
|
.name = "XLP-CPU-INTR",
|
|
.irq_enable = cpuintr_enable,
|
|
.irq_disable = cpuintr_disable,
|
|
.irq_mask = cpuintr_nop,
|
|
.irq_ack = cpuintr_nop,
|
|
.irq_eoi = cpuintr_ack,
|
|
};
|
|
|
|
static void __init nlm_init_percpu_irqs(void)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < PIC_IRT_FIRST_IRQ; i++)
|
|
irq_set_chip_and_handler(i, &nlm_cpu_intr, handle_percpu_irq);
|
|
#ifdef CONFIG_SMP
|
|
irq_set_chip_and_handler(IRQ_IPI_SMP_FUNCTION, &nlm_cpu_intr,
|
|
nlm_smp_function_ipi_handler);
|
|
irq_set_chip_and_handler(IRQ_IPI_SMP_RESCHEDULE, &nlm_cpu_intr,
|
|
nlm_smp_resched_ipi_handler);
|
|
#endif
|
|
}
|
|
|
|
void nlm_setup_pic_irq(int node, int picirq, int irq, int irt)
|
|
{
|
|
struct nlm_pic_irq *pic_data;
|
|
int xirq;
|
|
|
|
xirq = nlm_irq_to_xirq(node, irq);
|
|
pic_data = kzalloc(sizeof(*pic_data), GFP_KERNEL);
|
|
BUG_ON(pic_data == NULL);
|
|
pic_data->irt = irt;
|
|
pic_data->picirq = picirq;
|
|
pic_data->node = nlm_get_node(node);
|
|
irq_set_chip_and_handler(xirq, &xlp_pic, handle_level_irq);
|
|
irq_set_handler_data(xirq, pic_data);
|
|
}
|
|
|
|
void nlm_set_pic_extra_ack(int node, int irq, void (*xack)(struct irq_data *))
|
|
{
|
|
struct nlm_pic_irq *pic_data;
|
|
int xirq;
|
|
|
|
xirq = nlm_irq_to_xirq(node, irq);
|
|
pic_data = irq_get_handler_data(xirq);
|
|
pic_data->extra_ack = xack;
|
|
}
|
|
|
|
static void nlm_init_node_irqs(int node)
|
|
{
|
|
int i, irt;
|
|
uint64_t irqmask;
|
|
struct nlm_soc_info *nodep;
|
|
|
|
pr_info("Init IRQ for node %d\n", node);
|
|
nodep = nlm_get_node(node);
|
|
irqmask = PERCPU_IRQ_MASK;
|
|
for (i = PIC_IRT_FIRST_IRQ; i <= PIC_IRT_LAST_IRQ; i++) {
|
|
irt = nlm_irq_to_irt(i);
|
|
if (irt == -1)
|
|
continue;
|
|
nlm_setup_pic_irq(node, i, i, irt);
|
|
/* set interrupts to first cpu in node */
|
|
nlm_pic_init_irt(nodep->picbase, irt, i,
|
|
node * NLM_CPUS_PER_NODE);
|
|
irqmask |= (1ull << i);
|
|
}
|
|
nodep->irqmask = irqmask;
|
|
}
|
|
|
|
void __init arch_init_irq(void)
|
|
{
|
|
/* Initialize the irq descriptors */
|
|
nlm_init_percpu_irqs();
|
|
nlm_init_node_irqs(0);
|
|
write_c0_eimr(nlm_current_node()->irqmask);
|
|
#if defined(CONFIG_CPU_XLR)
|
|
nlm_setup_fmn_irq();
|
|
#endif
|
|
}
|
|
|
|
void nlm_smp_irq_init(int hwcpuid)
|
|
{
|
|
int node, cpu;
|
|
|
|
node = hwcpuid / NLM_CPUS_PER_NODE;
|
|
cpu = hwcpuid % NLM_CPUS_PER_NODE;
|
|
|
|
if (cpu == 0 && node != 0)
|
|
nlm_init_node_irqs(node);
|
|
write_c0_eimr(nlm_current_node()->irqmask);
|
|
}
|
|
|
|
asmlinkage void plat_irq_dispatch(void)
|
|
{
|
|
uint64_t eirr;
|
|
int i, node;
|
|
|
|
node = nlm_nodeid();
|
|
eirr = read_c0_eirr() & read_c0_eimr();
|
|
|
|
i = __ilog2_u64(eirr);
|
|
if (i == -1)
|
|
return;
|
|
|
|
/* per-CPU IRQs don't need translation */
|
|
if (eirr & PERCPU_IRQ_MASK) {
|
|
do_IRQ(i);
|
|
return;
|
|
}
|
|
|
|
/* top level irq handling */
|
|
do_IRQ(nlm_irq_to_xirq(node, i));
|
|
}
|