mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-25 03:30:53 +07:00
179475a3b4
* 'irq-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip: x86, sparseirq: clean up Kconfig entry x86: turn CONFIG_SPARSE_IRQ off by default sparseirq: fix numa_migrate_irq_desc dependency and comments sparseirq: add kernel-doc notation for new member in irq_desc, -v2 locking, irq: enclose irq_desc_lock_class in CONFIG_LOCKDEP sparseirq, xen: make sure irq_desc is allocated for interrupts sparseirq: fix !SMP building, #2 x86, sparseirq: move irq_desc according to smp_affinity, v7 proc: enclose desc variable of show_stat() in CONFIG_SPARSE_IRQ sparse irqs: add irqnr.h to the user headers list sparse irqs: handle !GENIRQ platforms sparseirq: fix !SMP && !PCI_MSI && !HT_IRQ build sparseirq: fix Alpha build failure sparseirq: fix typo in !CONFIG_IO_APIC case x86, MSI: pass irq_cfg and irq_desc x86: MSI start irq numbering from nr_irqs_gsi x86: use NR_IRQS_LEGACY sparse irq_desc[] array: core kernel and x86 changes genirq: record IRQ_LEVEL in irq_desc[] irq.h: remove padding from irq_desc on 64bits
159 lines
3.4 KiB
C
159 lines
3.4 KiB
C
/*
|
|
* Copyright (C) 1992, 1998 Linus Torvalds, Ingo Molnar
|
|
*
|
|
* This file contains the lowest level x86_64-specific interrupt
|
|
* entry and irq statistics code. All the remaining irq logic is
|
|
* done by the generic kernel/irq/ code and in the
|
|
* x86_64-specific irq controller code. (e.g. i8259.c and
|
|
* io_apic.c.)
|
|
*/
|
|
|
|
#include <linux/kernel_stat.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/seq_file.h>
|
|
#include <linux/module.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/ftrace.h>
|
|
#include <asm/uaccess.h>
|
|
#include <asm/io_apic.h>
|
|
#include <asm/idle.h>
|
|
#include <asm/smp.h>
|
|
|
|
/*
|
|
* Probabilistic stack overflow check:
|
|
*
|
|
* Only check the stack in process context, because everything else
|
|
* runs on the big interrupt stacks. Checking reliably is too expensive,
|
|
* so we just check from interrupts.
|
|
*/
|
|
static inline void stack_overflow_check(struct pt_regs *regs)
|
|
{
|
|
#ifdef CONFIG_DEBUG_STACKOVERFLOW
|
|
u64 curbase = (u64)task_stack_page(current);
|
|
|
|
WARN_ONCE(regs->sp >= curbase &&
|
|
regs->sp <= curbase + THREAD_SIZE &&
|
|
regs->sp < curbase + sizeof(struct thread_info) +
|
|
sizeof(struct pt_regs) + 128,
|
|
|
|
"do_IRQ: %s near stack overflow (cur:%Lx,sp:%lx)\n",
|
|
current->comm, curbase, regs->sp);
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* do_IRQ handles all normal device IRQ's (the special
|
|
* SMP cross-CPU interrupts have their own specific
|
|
* handlers).
|
|
*/
|
|
asmlinkage unsigned int __irq_entry do_IRQ(struct pt_regs *regs)
|
|
{
|
|
struct pt_regs *old_regs = set_irq_regs(regs);
|
|
struct irq_desc *desc;
|
|
|
|
/* high bit used in ret_from_ code */
|
|
unsigned vector = ~regs->orig_ax;
|
|
unsigned irq;
|
|
|
|
exit_idle();
|
|
irq_enter();
|
|
irq = __get_cpu_var(vector_irq)[vector];
|
|
|
|
stack_overflow_check(regs);
|
|
|
|
desc = irq_to_desc(irq);
|
|
if (likely(desc))
|
|
generic_handle_irq_desc(irq, desc);
|
|
else {
|
|
if (!disable_apic)
|
|
ack_APIC_irq();
|
|
|
|
if (printk_ratelimit())
|
|
printk(KERN_EMERG "%s: %d.%d No irq handler for vector\n",
|
|
__func__, smp_processor_id(), vector);
|
|
}
|
|
|
|
irq_exit();
|
|
|
|
set_irq_regs(old_regs);
|
|
return 1;
|
|
}
|
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
void fixup_irqs(cpumask_t map)
|
|
{
|
|
unsigned int irq;
|
|
static int warned;
|
|
struct irq_desc *desc;
|
|
|
|
for_each_irq_desc(irq, desc) {
|
|
cpumask_t mask;
|
|
int break_affinity = 0;
|
|
int set_affinity = 1;
|
|
|
|
if (!desc)
|
|
continue;
|
|
if (irq == 2)
|
|
continue;
|
|
|
|
/* interrupt's are disabled at this point */
|
|
spin_lock(&desc->lock);
|
|
|
|
if (!irq_has_action(irq) ||
|
|
cpus_equal(desc->affinity, map)) {
|
|
spin_unlock(&desc->lock);
|
|
continue;
|
|
}
|
|
|
|
cpus_and(mask, desc->affinity, map);
|
|
if (cpus_empty(mask)) {
|
|
break_affinity = 1;
|
|
mask = map;
|
|
}
|
|
|
|
if (desc->chip->mask)
|
|
desc->chip->mask(irq);
|
|
|
|
if (desc->chip->set_affinity)
|
|
desc->chip->set_affinity(irq, mask);
|
|
else if (!(warned++))
|
|
set_affinity = 0;
|
|
|
|
if (desc->chip->unmask)
|
|
desc->chip->unmask(irq);
|
|
|
|
spin_unlock(&desc->lock);
|
|
|
|
if (break_affinity && set_affinity)
|
|
printk("Broke affinity for irq %i\n", irq);
|
|
else if (!set_affinity)
|
|
printk("Cannot set affinity for irq %i\n", irq);
|
|
}
|
|
|
|
/* That doesn't seem sufficient. Give it 1ms. */
|
|
local_irq_enable();
|
|
mdelay(1);
|
|
local_irq_disable();
|
|
}
|
|
#endif
|
|
|
|
extern void call_softirq(void);
|
|
|
|
asmlinkage void do_softirq(void)
|
|
{
|
|
__u32 pending;
|
|
unsigned long flags;
|
|
|
|
if (in_interrupt())
|
|
return;
|
|
|
|
local_irq_save(flags);
|
|
pending = local_softirq_pending();
|
|
/* Switch to interrupt stack */
|
|
if (pending) {
|
|
call_softirq();
|
|
WARN_ON_ONCE(softirq_count());
|
|
}
|
|
local_irq_restore(flags);
|
|
}
|