mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-06 16:06:39 +07:00
b86432b42e
I got an oops when booting 2.6.19-rc5-mm1 on my ia64 machine. Below is the log. Oops 11012296146944 [1] Modules linked in: binfmt_misc dm_mirror dm_multipath dm_mod thermal processor f an container button sg eepro100 e100 mii Pid: 0, CPU 0, comm: swapper psr : 0000121008022038 ifs : 800000000000040b ip : [<a0000001000e1411>] Not tainted ip is at __do_IRQ+0x371/0x3e0 unat: 0000000000000000 pfs : 000000000000040b rsc : 0000000000000003 rnat: 656960155aa56aa5 bsps: a00000010058b890 pr : 656960155aa55a65 ldrs: 0000000000000000 ccv : 0000000000000000 fpsr: 0009804c0270033f csd : 0000000000000000 ssd : 0000000000000000 b0 : a0000001000e1390 b6 : a0000001005beac0 b7 : e00000007f01aa00 f6 : 000000000000000000000 f7 : 0ffe69090000000000000 f8 : 1000a9090000000000000 f9 : 0ffff8000000000000000 f10 : 1000a908ffffff6f70000 f11 : 1003e0000000000000909 r1 : a000000100fbbff0 r2 : 0000000000010002 r3 : 0000000000010001 r8 : fffffffffffbffff r9 : a000000100bd8060 r10 : a000000100dd83b8 r11 : fffffffffffeffff r12 : a000000100bcbbb0 r13 : a000000100bc4000 r14 : 0000000000010000 r15 : 0000000000010000 r16 : a000000100c01aa8 r17 : a000000100d2c350 r18 : 0000000000000000 r19 : a000000100d2c300 r20 : a000000100c01a88 r21 : 0000000080010100 r22 : a000000100c01ac0 r23 : a0000001000108e0 r24 : e000000477980004 r25 : 0000000000000000 r26 : 0000000000000000 r27 : e00000000913400c r28 : e0000004799ee51c r29 : e0000004778b87f0 r30 : a000000100d2c300 r31 : a00000010005c7e0 Call Trace: [<a000000100014600>] show_stack+0x40/0xa0 sp=a000000100bcb760 bsp=a000000100bc4f40 [<a000000100014f00>] show_regs+0x840/0x880 sp=a000000100bcb930 bsp=a000000100bc4ee8 [<a000000100037fb0>] die+0x250/0x320 sp=a000000100bcb930 bsp=a000000100bc4ea0 [<a00000010005e5f0>] ia64_do_page_fault+0x8d0/0xa20 sp=a000000100bcb950 bsp=a000000100bc4e50 [<a00000010000caa0>] ia64_leave_kernel+0x0/0x290 sp=a000000100bcb9e0 bsp=a000000100bc4e50 [<a0000001000e1410>] __do_IRQ+0x370/0x3e0 sp=a000000100bcbbb0 bsp=a000000100bc4df0 [<a000000100011f50>] ia64_handle_irq+0x170/0x220 sp=a000000100bcbbb0 bsp=a000000100bc4dc0 [<a00000010000caa0>] ia64_leave_kernel+0x0/0x290 sp=a000000100bcbbb0 bsp=a000000100bc4dc0 [<a000000100012390>] ia64_pal_call_static+0x90/0xc0 sp=a000000100bcbd80 bsp=a000000100bc4d78 [<a000000100015630>] default_idle+0x90/0x160 sp=a000000100bcbd80 bsp=a000000100bc4d58 [<a000000100014290>] cpu_idle+0x1f0/0x440 sp=a000000100bcbe20 bsp=a000000100bc4d18 [<a000000100009980>] rest_init+0xc0/0xe0 sp=a000000100bcbe20 bsp=a000000100bc4d00 [<a0000001009f8ea0>] start_kernel+0x6a0/0x6c0 sp=a000000100bcbe20 bsp=a000000100bc4ca0 [<a0000001000089f0>] __end_ivt_text+0x6d0/0x6f0 sp=a000000100bcbe30 bsp=a000000100bc4c00 <0>Kernel panic - not syncing: Aiee, killing interrupt handler! The root cause is that some irq_chip variables, especially ia64_msi_chip, initiate their memeber end to point to NULL. __do_IRQ doesn't check if irq_chip->end is null and just calls it after processing the interrupt. As irq_chip->end is called at many places, so I fix it by reinitiating irq_chip->end to dummy_irq_chip.end, e.g., a noop function. Signed-off-by: Zhang Yanmin <yanmin.zhang@intel.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Ingo Molnar <mingo@elte.hu> Cc: "Luck, Tony" <tony.luck@intel.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
572 lines
14 KiB
C
572 lines
14 KiB
C
/*
|
|
* linux/kernel/irq/chip.c
|
|
*
|
|
* Copyright (C) 1992, 1998-2006 Linus Torvalds, Ingo Molnar
|
|
* Copyright (C) 2005-2006, Thomas Gleixner, Russell King
|
|
*
|
|
* This file contains the core interrupt handling code, for irq-chip
|
|
* based architectures.
|
|
*
|
|
* Detailed information is available in Documentation/DocBook/genericirq
|
|
*/
|
|
|
|
#include <linux/irq.h>
|
|
#include <linux/module.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/kernel_stat.h>
|
|
|
|
#include "internals.h"
|
|
|
|
/**
|
|
* dynamic_irq_init - initialize a dynamically allocated irq
|
|
* @irq: irq number to initialize
|
|
*/
|
|
void dynamic_irq_init(unsigned int irq)
|
|
{
|
|
struct irq_desc *desc;
|
|
unsigned long flags;
|
|
|
|
if (irq >= NR_IRQS) {
|
|
printk(KERN_ERR "Trying to initialize invalid IRQ%d\n", irq);
|
|
WARN_ON(1);
|
|
return;
|
|
}
|
|
|
|
/* Ensure we don't have left over values from a previous use of this irq */
|
|
desc = irq_desc + irq;
|
|
spin_lock_irqsave(&desc->lock, flags);
|
|
desc->status = IRQ_DISABLED;
|
|
desc->chip = &no_irq_chip;
|
|
desc->handle_irq = handle_bad_irq;
|
|
desc->depth = 1;
|
|
desc->handler_data = NULL;
|
|
desc->chip_data = NULL;
|
|
desc->action = NULL;
|
|
desc->irq_count = 0;
|
|
desc->irqs_unhandled = 0;
|
|
#ifdef CONFIG_SMP
|
|
desc->affinity = CPU_MASK_ALL;
|
|
#endif
|
|
spin_unlock_irqrestore(&desc->lock, flags);
|
|
}
|
|
|
|
/**
|
|
* dynamic_irq_cleanup - cleanup a dynamically allocated irq
|
|
* @irq: irq number to initialize
|
|
*/
|
|
void dynamic_irq_cleanup(unsigned int irq)
|
|
{
|
|
struct irq_desc *desc;
|
|
unsigned long flags;
|
|
|
|
if (irq >= NR_IRQS) {
|
|
printk(KERN_ERR "Trying to cleanup invalid IRQ%d\n", irq);
|
|
WARN_ON(1);
|
|
return;
|
|
}
|
|
|
|
desc = irq_desc + irq;
|
|
spin_lock_irqsave(&desc->lock, flags);
|
|
if (desc->action) {
|
|
spin_unlock_irqrestore(&desc->lock, flags);
|
|
printk(KERN_ERR "Destroying IRQ%d without calling free_irq\n",
|
|
irq);
|
|
WARN_ON(1);
|
|
return;
|
|
}
|
|
desc->handle_irq = handle_bad_irq;
|
|
desc->chip = &no_irq_chip;
|
|
spin_unlock_irqrestore(&desc->lock, flags);
|
|
}
|
|
|
|
|
|
/**
|
|
* set_irq_chip - set the irq chip for an irq
|
|
* @irq: irq number
|
|
* @chip: pointer to irq chip description structure
|
|
*/
|
|
int set_irq_chip(unsigned int irq, struct irq_chip *chip)
|
|
{
|
|
struct irq_desc *desc;
|
|
unsigned long flags;
|
|
|
|
if (irq >= NR_IRQS) {
|
|
printk(KERN_ERR "Trying to install chip for IRQ%d\n", irq);
|
|
WARN_ON(1);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!chip)
|
|
chip = &no_irq_chip;
|
|
|
|
desc = irq_desc + irq;
|
|
spin_lock_irqsave(&desc->lock, flags);
|
|
irq_chip_set_defaults(chip);
|
|
desc->chip = chip;
|
|
spin_unlock_irqrestore(&desc->lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(set_irq_chip);
|
|
|
|
/**
|
|
* set_irq_type - set the irq type for an irq
|
|
* @irq: irq number
|
|
* @type: interrupt type - see include/linux/interrupt.h
|
|
*/
|
|
int set_irq_type(unsigned int irq, unsigned int type)
|
|
{
|
|
struct irq_desc *desc;
|
|
unsigned long flags;
|
|
int ret = -ENXIO;
|
|
|
|
if (irq >= NR_IRQS) {
|
|
printk(KERN_ERR "Trying to set irq type for IRQ%d\n", irq);
|
|
return -ENODEV;
|
|
}
|
|
|
|
desc = irq_desc + irq;
|
|
if (desc->chip->set_type) {
|
|
spin_lock_irqsave(&desc->lock, flags);
|
|
ret = desc->chip->set_type(irq, type);
|
|
spin_unlock_irqrestore(&desc->lock, flags);
|
|
}
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(set_irq_type);
|
|
|
|
/**
|
|
* set_irq_data - set irq type data for an irq
|
|
* @irq: Interrupt number
|
|
* @data: Pointer to interrupt specific data
|
|
*
|
|
* Set the hardware irq controller data for an irq
|
|
*/
|
|
int set_irq_data(unsigned int irq, void *data)
|
|
{
|
|
struct irq_desc *desc;
|
|
unsigned long flags;
|
|
|
|
if (irq >= NR_IRQS) {
|
|
printk(KERN_ERR
|
|
"Trying to install controller data for IRQ%d\n", irq);
|
|
return -EINVAL;
|
|
}
|
|
|
|
desc = irq_desc + irq;
|
|
spin_lock_irqsave(&desc->lock, flags);
|
|
desc->handler_data = data;
|
|
spin_unlock_irqrestore(&desc->lock, flags);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(set_irq_data);
|
|
|
|
/**
|
|
* set_irq_chip_data - set irq chip data for an irq
|
|
* @irq: Interrupt number
|
|
* @data: Pointer to chip specific data
|
|
*
|
|
* Set the hardware irq chip data for an irq
|
|
*/
|
|
int set_irq_chip_data(unsigned int irq, void *data)
|
|
{
|
|
struct irq_desc *desc = irq_desc + irq;
|
|
unsigned long flags;
|
|
|
|
if (irq >= NR_IRQS || !desc->chip) {
|
|
printk(KERN_ERR "BUG: bad set_irq_chip_data(IRQ#%d)\n", irq);
|
|
return -EINVAL;
|
|
}
|
|
|
|
spin_lock_irqsave(&desc->lock, flags);
|
|
desc->chip_data = data;
|
|
spin_unlock_irqrestore(&desc->lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(set_irq_chip_data);
|
|
|
|
/*
|
|
* default enable function
|
|
*/
|
|
static void default_enable(unsigned int irq)
|
|
{
|
|
struct irq_desc *desc = irq_desc + irq;
|
|
|
|
desc->chip->unmask(irq);
|
|
desc->status &= ~IRQ_MASKED;
|
|
}
|
|
|
|
/*
|
|
* default disable function
|
|
*/
|
|
static void default_disable(unsigned int irq)
|
|
{
|
|
struct irq_desc *desc = irq_desc + irq;
|
|
|
|
if (!(desc->status & IRQ_DELAYED_DISABLE))
|
|
desc->chip->mask(irq);
|
|
}
|
|
|
|
/*
|
|
* default startup function
|
|
*/
|
|
static unsigned int default_startup(unsigned int irq)
|
|
{
|
|
irq_desc[irq].chip->enable(irq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Fixup enable/disable function pointers
|
|
*/
|
|
void irq_chip_set_defaults(struct irq_chip *chip)
|
|
{
|
|
if (!chip->enable)
|
|
chip->enable = default_enable;
|
|
if (!chip->disable)
|
|
chip->disable = default_disable;
|
|
if (!chip->startup)
|
|
chip->startup = default_startup;
|
|
if (!chip->shutdown)
|
|
chip->shutdown = chip->disable;
|
|
if (!chip->name)
|
|
chip->name = chip->typename;
|
|
if (!chip->end)
|
|
chip->end = dummy_irq_chip.end;
|
|
}
|
|
|
|
static inline void mask_ack_irq(struct irq_desc *desc, int irq)
|
|
{
|
|
if (desc->chip->mask_ack)
|
|
desc->chip->mask_ack(irq);
|
|
else {
|
|
desc->chip->mask(irq);
|
|
desc->chip->ack(irq);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* handle_simple_irq - Simple and software-decoded IRQs.
|
|
* @irq: the interrupt number
|
|
* @desc: the interrupt description structure for this irq
|
|
*
|
|
* Simple interrupts are either sent from a demultiplexing interrupt
|
|
* handler or come from hardware, where no interrupt hardware control
|
|
* is necessary.
|
|
*
|
|
* Note: The caller is expected to handle the ack, clear, mask and
|
|
* unmask issues if necessary.
|
|
*/
|
|
void fastcall
|
|
handle_simple_irq(unsigned int irq, struct irq_desc *desc)
|
|
{
|
|
struct irqaction *action;
|
|
irqreturn_t action_ret;
|
|
const unsigned int cpu = smp_processor_id();
|
|
|
|
spin_lock(&desc->lock);
|
|
|
|
if (unlikely(desc->status & IRQ_INPROGRESS))
|
|
goto out_unlock;
|
|
desc->status &= ~(IRQ_REPLAY | IRQ_WAITING);
|
|
kstat_cpu(cpu).irqs[irq]++;
|
|
|
|
action = desc->action;
|
|
if (unlikely(!action || (desc->status & IRQ_DISABLED)))
|
|
goto out_unlock;
|
|
|
|
desc->status |= IRQ_INPROGRESS;
|
|
spin_unlock(&desc->lock);
|
|
|
|
action_ret = handle_IRQ_event(irq, action);
|
|
if (!noirqdebug)
|
|
note_interrupt(irq, desc, action_ret);
|
|
|
|
spin_lock(&desc->lock);
|
|
desc->status &= ~IRQ_INPROGRESS;
|
|
out_unlock:
|
|
spin_unlock(&desc->lock);
|
|
}
|
|
|
|
/**
|
|
* handle_level_irq - Level type irq handler
|
|
* @irq: the interrupt number
|
|
* @desc: the interrupt description structure for this irq
|
|
*
|
|
* Level type interrupts are active as long as the hardware line has
|
|
* the active level. This may require to mask the interrupt and unmask
|
|
* it after the associated handler has acknowledged the device, so the
|
|
* interrupt line is back to inactive.
|
|
*/
|
|
void fastcall
|
|
handle_level_irq(unsigned int irq, struct irq_desc *desc)
|
|
{
|
|
unsigned int cpu = smp_processor_id();
|
|
struct irqaction *action;
|
|
irqreturn_t action_ret;
|
|
|
|
spin_lock(&desc->lock);
|
|
mask_ack_irq(desc, irq);
|
|
|
|
if (unlikely(desc->status & IRQ_INPROGRESS))
|
|
goto out_unlock;
|
|
desc->status &= ~(IRQ_REPLAY | IRQ_WAITING);
|
|
kstat_cpu(cpu).irqs[irq]++;
|
|
|
|
/*
|
|
* If its disabled or no action available
|
|
* keep it masked and get out of here
|
|
*/
|
|
action = desc->action;
|
|
if (unlikely(!action || (desc->status & IRQ_DISABLED))) {
|
|
desc->status |= IRQ_PENDING;
|
|
goto out_unlock;
|
|
}
|
|
|
|
desc->status |= IRQ_INPROGRESS;
|
|
desc->status &= ~IRQ_PENDING;
|
|
spin_unlock(&desc->lock);
|
|
|
|
action_ret = handle_IRQ_event(irq, action);
|
|
if (!noirqdebug)
|
|
note_interrupt(irq, desc, action_ret);
|
|
|
|
spin_lock(&desc->lock);
|
|
desc->status &= ~IRQ_INPROGRESS;
|
|
if (!(desc->status & IRQ_DISABLED) && desc->chip->unmask)
|
|
desc->chip->unmask(irq);
|
|
out_unlock:
|
|
spin_unlock(&desc->lock);
|
|
}
|
|
|
|
/**
|
|
* handle_fasteoi_irq - irq handler for transparent controllers
|
|
* @irq: the interrupt number
|
|
* @desc: the interrupt description structure for this irq
|
|
*
|
|
* Only a single callback will be issued to the chip: an ->eoi()
|
|
* call when the interrupt has been serviced. This enables support
|
|
* for modern forms of interrupt handlers, which handle the flow
|
|
* details in hardware, transparently.
|
|
*/
|
|
void fastcall
|
|
handle_fasteoi_irq(unsigned int irq, struct irq_desc *desc)
|
|
{
|
|
unsigned int cpu = smp_processor_id();
|
|
struct irqaction *action;
|
|
irqreturn_t action_ret;
|
|
|
|
spin_lock(&desc->lock);
|
|
|
|
if (unlikely(desc->status & IRQ_INPROGRESS))
|
|
goto out;
|
|
|
|
desc->status &= ~(IRQ_REPLAY | IRQ_WAITING);
|
|
kstat_cpu(cpu).irqs[irq]++;
|
|
|
|
/*
|
|
* If its disabled or no action available
|
|
* keep it masked and get out of here
|
|
*/
|
|
action = desc->action;
|
|
if (unlikely(!action || (desc->status & IRQ_DISABLED))) {
|
|
desc->status |= IRQ_PENDING;
|
|
goto out;
|
|
}
|
|
|
|
desc->status |= IRQ_INPROGRESS;
|
|
desc->status &= ~IRQ_PENDING;
|
|
spin_unlock(&desc->lock);
|
|
|
|
action_ret = handle_IRQ_event(irq, action);
|
|
if (!noirqdebug)
|
|
note_interrupt(irq, desc, action_ret);
|
|
|
|
spin_lock(&desc->lock);
|
|
desc->status &= ~IRQ_INPROGRESS;
|
|
out:
|
|
desc->chip->eoi(irq);
|
|
|
|
spin_unlock(&desc->lock);
|
|
}
|
|
|
|
/**
|
|
* handle_edge_irq - edge type IRQ handler
|
|
* @irq: the interrupt number
|
|
* @desc: the interrupt description structure for this irq
|
|
*
|
|
* Interrupt occures on the falling and/or rising edge of a hardware
|
|
* signal. The occurence is latched into the irq controller hardware
|
|
* and must be acked in order to be reenabled. After the ack another
|
|
* interrupt can happen on the same source even before the first one
|
|
* is handled by the assosiacted event handler. If this happens it
|
|
* might be necessary to disable (mask) the interrupt depending on the
|
|
* controller hardware. This requires to reenable the interrupt inside
|
|
* of the loop which handles the interrupts which have arrived while
|
|
* the handler was running. If all pending interrupts are handled, the
|
|
* loop is left.
|
|
*/
|
|
void fastcall
|
|
handle_edge_irq(unsigned int irq, struct irq_desc *desc)
|
|
{
|
|
const unsigned int cpu = smp_processor_id();
|
|
|
|
spin_lock(&desc->lock);
|
|
|
|
desc->status &= ~(IRQ_REPLAY | IRQ_WAITING);
|
|
|
|
/*
|
|
* If we're currently running this IRQ, or its disabled,
|
|
* we shouldn't process the IRQ. Mark it pending, handle
|
|
* the necessary masking and go out
|
|
*/
|
|
if (unlikely((desc->status & (IRQ_INPROGRESS | IRQ_DISABLED)) ||
|
|
!desc->action)) {
|
|
desc->status |= (IRQ_PENDING | IRQ_MASKED);
|
|
mask_ack_irq(desc, irq);
|
|
goto out_unlock;
|
|
}
|
|
|
|
kstat_cpu(cpu).irqs[irq]++;
|
|
|
|
/* Start handling the irq */
|
|
desc->chip->ack(irq);
|
|
|
|
/* Mark the IRQ currently in progress.*/
|
|
desc->status |= IRQ_INPROGRESS;
|
|
|
|
do {
|
|
struct irqaction *action = desc->action;
|
|
irqreturn_t action_ret;
|
|
|
|
if (unlikely(!action)) {
|
|
desc->chip->mask(irq);
|
|
goto out_unlock;
|
|
}
|
|
|
|
/*
|
|
* When another irq arrived while we were handling
|
|
* one, we could have masked the irq.
|
|
* Renable it, if it was not disabled in meantime.
|
|
*/
|
|
if (unlikely((desc->status &
|
|
(IRQ_PENDING | IRQ_MASKED | IRQ_DISABLED)) ==
|
|
(IRQ_PENDING | IRQ_MASKED))) {
|
|
desc->chip->unmask(irq);
|
|
desc->status &= ~IRQ_MASKED;
|
|
}
|
|
|
|
desc->status &= ~IRQ_PENDING;
|
|
spin_unlock(&desc->lock);
|
|
action_ret = handle_IRQ_event(irq, action);
|
|
if (!noirqdebug)
|
|
note_interrupt(irq, desc, action_ret);
|
|
spin_lock(&desc->lock);
|
|
|
|
} while ((desc->status & (IRQ_PENDING | IRQ_DISABLED)) == IRQ_PENDING);
|
|
|
|
desc->status &= ~IRQ_INPROGRESS;
|
|
out_unlock:
|
|
spin_unlock(&desc->lock);
|
|
}
|
|
|
|
#ifdef CONFIG_SMP
|
|
/**
|
|
* handle_percpu_IRQ - Per CPU local irq handler
|
|
* @irq: the interrupt number
|
|
* @desc: the interrupt description structure for this irq
|
|
*
|
|
* Per CPU interrupts on SMP machines without locking requirements
|
|
*/
|
|
void fastcall
|
|
handle_percpu_irq(unsigned int irq, struct irq_desc *desc)
|
|
{
|
|
irqreturn_t action_ret;
|
|
|
|
kstat_this_cpu.irqs[irq]++;
|
|
|
|
if (desc->chip->ack)
|
|
desc->chip->ack(irq);
|
|
|
|
action_ret = handle_IRQ_event(irq, desc->action);
|
|
if (!noirqdebug)
|
|
note_interrupt(irq, desc, action_ret);
|
|
|
|
if (desc->chip->eoi)
|
|
desc->chip->eoi(irq);
|
|
}
|
|
|
|
#endif /* CONFIG_SMP */
|
|
|
|
void
|
|
__set_irq_handler(unsigned int irq, irq_flow_handler_t handle, int is_chained,
|
|
const char *name)
|
|
{
|
|
struct irq_desc *desc;
|
|
unsigned long flags;
|
|
|
|
if (irq >= NR_IRQS) {
|
|
printk(KERN_ERR
|
|
"Trying to install type control for IRQ%d\n", irq);
|
|
return;
|
|
}
|
|
|
|
desc = irq_desc + irq;
|
|
|
|
if (!handle)
|
|
handle = handle_bad_irq;
|
|
|
|
if (desc->chip == &no_irq_chip) {
|
|
printk(KERN_WARNING "Trying to install %sinterrupt handler "
|
|
"for IRQ%d\n", is_chained ? "chained " : " ", irq);
|
|
/*
|
|
* Some ARM implementations install a handler for really dumb
|
|
* interrupt hardware without setting an irq_chip. This worked
|
|
* with the ARM no_irq_chip but the check in setup_irq would
|
|
* prevent us to setup the interrupt at all. Switch it to
|
|
* dummy_irq_chip for easy transition.
|
|
*/
|
|
desc->chip = &dummy_irq_chip;
|
|
}
|
|
|
|
spin_lock_irqsave(&desc->lock, flags);
|
|
|
|
/* Uninstall? */
|
|
if (handle == handle_bad_irq) {
|
|
if (desc->chip != &no_irq_chip) {
|
|
desc->chip->mask(irq);
|
|
desc->chip->ack(irq);
|
|
}
|
|
desc->status |= IRQ_DISABLED;
|
|
desc->depth = 1;
|
|
}
|
|
desc->handle_irq = handle;
|
|
desc->name = name;
|
|
|
|
if (handle != handle_bad_irq && is_chained) {
|
|
desc->status &= ~IRQ_DISABLED;
|
|
desc->status |= IRQ_NOREQUEST | IRQ_NOPROBE;
|
|
desc->depth = 0;
|
|
desc->chip->unmask(irq);
|
|
}
|
|
spin_unlock_irqrestore(&desc->lock, flags);
|
|
}
|
|
|
|
void
|
|
set_irq_chip_and_handler(unsigned int irq, struct irq_chip *chip,
|
|
irq_flow_handler_t handle)
|
|
{
|
|
set_irq_chip(irq, chip);
|
|
__set_irq_handler(irq, handle, 0, NULL);
|
|
}
|
|
|
|
void
|
|
set_irq_chip_and_handler_name(unsigned int irq, struct irq_chip *chip,
|
|
irq_flow_handler_t handle, const char *name)
|
|
{
|
|
set_irq_chip(irq, chip);
|
|
__set_irq_handler(irq, handle, 0, name);
|
|
}
|