mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-27 20:25:09 +07:00
b866cc2199
Change the doorbell callers to know about their msgsnd addressing, rather than have them set a per-cpu target data tag at boot that gets sent to the cause_ipi functions. The data is only used for doorbell IPI functions, no other IPI types, so it makes sense to keep that detail local to doorbell. Have the platform code understand doorbell IPIs, rather than the interrupt controller code understand them. Platform code can look at capabilities it has available and decide which to use. Signed-off-by: Nicholas Piggin <npiggin@gmail.com> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
174 lines
4.3 KiB
C
174 lines
4.3 KiB
C
/*
|
|
* Common definitions across all variants of ICP and ICS interrupt
|
|
* controllers.
|
|
*/
|
|
|
|
#ifndef _XICS_H
|
|
#define _XICS_H
|
|
|
|
#include <linux/interrupt.h>
|
|
|
|
#define XICS_IPI 2
|
|
#define XICS_IRQ_SPURIOUS 0
|
|
|
|
/* Want a priority other than 0. Various HW issues require this. */
|
|
#define DEFAULT_PRIORITY 5
|
|
|
|
/*
|
|
* Mark IPIs as higher priority so we can take them inside interrupts
|
|
* FIXME: still true now?
|
|
*/
|
|
#define IPI_PRIORITY 4
|
|
|
|
/* The least favored priority */
|
|
#define LOWEST_PRIORITY 0xFF
|
|
|
|
/* The number of priorities defined above */
|
|
#define MAX_NUM_PRIORITIES 3
|
|
|
|
/* Native ICP */
|
|
#ifdef CONFIG_PPC_ICP_NATIVE
|
|
extern int icp_native_init(void);
|
|
extern void icp_native_flush_interrupt(void);
|
|
extern void icp_native_cause_ipi_rm(int cpu);
|
|
#else
|
|
static inline int icp_native_init(void) { return -ENODEV; }
|
|
#endif
|
|
|
|
/* PAPR ICP */
|
|
#ifdef CONFIG_PPC_ICP_HV
|
|
extern int icp_hv_init(void);
|
|
#else
|
|
static inline int icp_hv_init(void) { return -ENODEV; }
|
|
#endif
|
|
|
|
#ifdef CONFIG_PPC_POWERNV
|
|
extern int icp_opal_init(void);
|
|
extern void icp_opal_flush_interrupt(void);
|
|
#else
|
|
static inline int icp_opal_init(void) { return -ENODEV; }
|
|
#endif
|
|
|
|
/* ICP ops */
|
|
struct icp_ops {
|
|
unsigned int (*get_irq)(void);
|
|
void (*eoi)(struct irq_data *d);
|
|
void (*set_priority)(unsigned char prio);
|
|
void (*teardown_cpu)(void);
|
|
void (*flush_ipi)(void);
|
|
#ifdef CONFIG_SMP
|
|
void (*cause_ipi)(int cpu);
|
|
irq_handler_t ipi_action;
|
|
#endif
|
|
};
|
|
|
|
extern const struct icp_ops *icp_ops;
|
|
|
|
/* Native ICS */
|
|
extern int ics_native_init(void);
|
|
|
|
/* RTAS ICS */
|
|
#ifdef CONFIG_PPC_ICS_RTAS
|
|
extern int ics_rtas_init(void);
|
|
#else
|
|
static inline int ics_rtas_init(void) { return -ENODEV; }
|
|
#endif
|
|
|
|
/* HAL ICS */
|
|
#ifdef CONFIG_PPC_POWERNV
|
|
extern int ics_opal_init(void);
|
|
#else
|
|
static inline int ics_opal_init(void) { return -ENODEV; }
|
|
#endif
|
|
|
|
/* ICS instance, hooked up to chip_data of an irq */
|
|
struct ics {
|
|
struct list_head link;
|
|
int (*map)(struct ics *ics, unsigned int virq);
|
|
void (*mask_unknown)(struct ics *ics, unsigned long vec);
|
|
long (*get_server)(struct ics *ics, unsigned long vec);
|
|
int (*host_match)(struct ics *ics, struct device_node *node);
|
|
char data[];
|
|
};
|
|
|
|
/* Commons */
|
|
extern unsigned int xics_default_server;
|
|
extern unsigned int xics_default_distrib_server;
|
|
extern unsigned int xics_interrupt_server_size;
|
|
extern struct irq_domain *xics_host;
|
|
|
|
struct xics_cppr {
|
|
unsigned char stack[MAX_NUM_PRIORITIES];
|
|
int index;
|
|
};
|
|
|
|
DECLARE_PER_CPU(struct xics_cppr, xics_cppr);
|
|
|
|
static inline void xics_push_cppr(unsigned int vec)
|
|
{
|
|
struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr);
|
|
|
|
if (WARN_ON(os_cppr->index >= MAX_NUM_PRIORITIES - 1))
|
|
return;
|
|
|
|
if (vec == XICS_IPI)
|
|
os_cppr->stack[++os_cppr->index] = IPI_PRIORITY;
|
|
else
|
|
os_cppr->stack[++os_cppr->index] = DEFAULT_PRIORITY;
|
|
}
|
|
|
|
static inline unsigned char xics_pop_cppr(void)
|
|
{
|
|
struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr);
|
|
|
|
if (WARN_ON(os_cppr->index < 1))
|
|
return LOWEST_PRIORITY;
|
|
|
|
return os_cppr->stack[--os_cppr->index];
|
|
}
|
|
|
|
static inline void xics_set_base_cppr(unsigned char cppr)
|
|
{
|
|
struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr);
|
|
|
|
/* we only really want to set the priority when there's
|
|
* just one cppr value on the stack
|
|
*/
|
|
WARN_ON(os_cppr->index != 0);
|
|
|
|
os_cppr->stack[0] = cppr;
|
|
}
|
|
|
|
static inline unsigned char xics_cppr_top(void)
|
|
{
|
|
struct xics_cppr *os_cppr = this_cpu_ptr(&xics_cppr);
|
|
|
|
return os_cppr->stack[os_cppr->index];
|
|
}
|
|
|
|
DECLARE_PER_CPU_SHARED_ALIGNED(unsigned long, xics_ipi_message);
|
|
|
|
extern void xics_init(void);
|
|
extern void xics_setup_cpu(void);
|
|
extern void xics_update_irq_servers(void);
|
|
extern void xics_set_cpu_giq(unsigned int gserver, unsigned int join);
|
|
extern void xics_mask_unknown_vec(unsigned int vec);
|
|
extern irqreturn_t xics_ipi_dispatch(int cpu);
|
|
extern void xics_smp_probe(void);
|
|
extern void xics_register_ics(struct ics *ics);
|
|
extern void xics_teardown_cpu(void);
|
|
extern void xics_kexec_teardown_cpu(int secondary);
|
|
extern void xics_migrate_irqs_away(void);
|
|
extern void icp_native_eoi(struct irq_data *d);
|
|
extern int xics_set_irq_type(struct irq_data *d, unsigned int flow_type);
|
|
extern int xics_retrigger(struct irq_data *data);
|
|
#ifdef CONFIG_SMP
|
|
extern int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask,
|
|
unsigned int strict_check);
|
|
#else
|
|
#define xics_get_irq_server(virq, cpumask, strict_check) (xics_default_server)
|
|
#endif
|
|
|
|
|
|
#endif /* _XICS_H */
|