2010-09-22 22:09:43 +07:00
|
|
|
/*
|
|
|
|
* Copyright (C) 1992, 1998-2006 Linus Torvalds, Ingo Molnar
|
|
|
|
* Copyright (C) 2005-2006, Thomas Gleixner, Russell King
|
|
|
|
*
|
|
|
|
* This file contains the interrupt descriptor management code
|
|
|
|
*
|
|
|
|
* Detailed information is available in Documentation/DocBook/genericirq
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
#include <linux/irq.h>
|
|
|
|
#include <linux/slab.h>
|
2011-09-20 07:33:19 +07:00
|
|
|
#include <linux/export.h>
|
2010-09-22 22:09:43 +07:00
|
|
|
#include <linux/interrupt.h>
|
|
|
|
#include <linux/kernel_stat.h>
|
|
|
|
#include <linux/radix-tree.h>
|
2010-09-27 22:48:26 +07:00
|
|
|
#include <linux/bitmap.h>
|
2010-09-22 22:09:43 +07:00
|
|
|
|
|
|
|
#include "internals.h"
|
|
|
|
|
|
|
|
/*
|
|
|
|
* lockdep: we want to handle all irq_desc locks as a single lock-class:
|
|
|
|
*/
|
2010-09-29 22:18:47 +07:00
|
|
|
static struct lock_class_key irq_desc_lock_class;
|
2010-09-22 22:09:43 +07:00
|
|
|
|
2011-05-18 17:53:03 +07:00
|
|
|
#if defined(CONFIG_SMP)
|
2010-09-22 22:09:43 +07:00
|
|
|
static void __init init_irq_default_affinity(void)
|
|
|
|
{
|
|
|
|
alloc_cpumask_var(&irq_default_affinity, GFP_NOWAIT);
|
|
|
|
cpumask_setall(irq_default_affinity);
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
static void __init init_irq_default_affinity(void)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-09-27 22:48:26 +07:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
static int alloc_masks(struct irq_desc *desc, gfp_t gfp, int node)
|
|
|
|
{
|
|
|
|
if (!zalloc_cpumask_var_node(&desc->irq_data.affinity, gfp, node))
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
#ifdef CONFIG_GENERIC_PENDING_IRQ
|
|
|
|
if (!zalloc_cpumask_var_node(&desc->pending_mask, gfp, node)) {
|
|
|
|
free_cpumask_var(desc->irq_data.affinity);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void desc_smp_init(struct irq_desc *desc, int node)
|
|
|
|
{
|
2010-09-28 01:02:56 +07:00
|
|
|
desc->irq_data.node = node;
|
2010-09-27 22:48:26 +07:00
|
|
|
cpumask_copy(desc->irq_data.affinity, irq_default_affinity);
|
2010-09-29 23:46:55 +07:00
|
|
|
#ifdef CONFIG_GENERIC_PENDING_IRQ
|
|
|
|
cpumask_clear(desc->pending_mask);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int desc_node(struct irq_desc *desc)
|
|
|
|
{
|
|
|
|
return desc->irq_data.node;
|
2010-09-27 22:48:26 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
static inline int
|
|
|
|
alloc_masks(struct irq_desc *desc, gfp_t gfp, int node) { return 0; }
|
|
|
|
static inline void desc_smp_init(struct irq_desc *desc, int node) { }
|
2010-09-29 23:46:55 +07:00
|
|
|
static inline int desc_node(struct irq_desc *desc) { return 0; }
|
2010-09-27 22:48:26 +07:00
|
|
|
#endif
|
|
|
|
|
2011-07-11 17:17:31 +07:00
|
|
|
static void desc_set_defaults(unsigned int irq, struct irq_desc *desc, int node,
|
|
|
|
struct module *owner)
|
2010-09-27 22:48:26 +07:00
|
|
|
{
|
2011-01-14 06:45:38 +07:00
|
|
|
int cpu;
|
|
|
|
|
2010-09-27 22:48:26 +07:00
|
|
|
desc->irq_data.irq = irq;
|
|
|
|
desc->irq_data.chip = &no_irq_chip;
|
|
|
|
desc->irq_data.chip_data = NULL;
|
|
|
|
desc->irq_data.handler_data = NULL;
|
|
|
|
desc->irq_data.msi_desc = NULL;
|
2011-02-09 20:54:49 +07:00
|
|
|
irq_settings_clr_and_set(desc, ~0, _IRQ_DEFAULT_INIT_FLAGS);
|
2011-03-27 16:02:49 +07:00
|
|
|
irqd_set(&desc->irq_data, IRQD_IRQ_DISABLED);
|
2010-09-27 22:48:26 +07:00
|
|
|
desc->handle_irq = handle_bad_irq;
|
|
|
|
desc->depth = 1;
|
2010-09-29 23:46:55 +07:00
|
|
|
desc->irq_count = 0;
|
|
|
|
desc->irqs_unhandled = 0;
|
2010-09-27 22:48:26 +07:00
|
|
|
desc->name = NULL;
|
2011-07-11 17:17:31 +07:00
|
|
|
desc->owner = owner;
|
2011-01-14 06:45:38 +07:00
|
|
|
for_each_possible_cpu(cpu)
|
|
|
|
*per_cpu_ptr(desc->kstat_irqs, cpu) = 0;
|
2010-09-27 22:48:26 +07:00
|
|
|
desc_smp_init(desc, node);
|
|
|
|
}
|
|
|
|
|
2010-09-22 22:09:43 +07:00
|
|
|
int nr_irqs = NR_IRQS;
|
|
|
|
EXPORT_SYMBOL_GPL(nr_irqs);
|
|
|
|
|
2010-10-08 17:47:53 +07:00
|
|
|
static DEFINE_MUTEX(sparse_irq_lock);
|
2011-02-17 23:45:15 +07:00
|
|
|
static DECLARE_BITMAP(allocated_irqs, IRQ_BITMAP_BITS);
|
2010-09-27 22:48:26 +07:00
|
|
|
|
2010-09-22 22:09:43 +07:00
|
|
|
#ifdef CONFIG_SPARSE_IRQ
|
|
|
|
|
2010-10-05 20:14:35 +07:00
|
|
|
static RADIX_TREE(irq_desc_tree, GFP_KERNEL);
|
2010-09-22 22:09:43 +07:00
|
|
|
|
2010-09-27 22:48:26 +07:00
|
|
|
static void irq_insert_desc(unsigned int irq, struct irq_desc *desc)
|
2010-09-22 22:09:43 +07:00
|
|
|
{
|
|
|
|
radix_tree_insert(&irq_desc_tree, irq, desc);
|
|
|
|
}
|
|
|
|
|
|
|
|
struct irq_desc *irq_to_desc(unsigned int irq)
|
|
|
|
{
|
|
|
|
return radix_tree_lookup(&irq_desc_tree, irq);
|
|
|
|
}
|
2012-05-13 17:13:15 +07:00
|
|
|
EXPORT_SYMBOL(irq_to_desc);
|
2010-09-22 22:09:43 +07:00
|
|
|
|
2010-09-27 22:48:26 +07:00
|
|
|
static void delete_irq_desc(unsigned int irq)
|
|
|
|
{
|
|
|
|
radix_tree_delete(&irq_desc_tree, irq);
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
static void free_masks(struct irq_desc *desc)
|
|
|
|
{
|
|
|
|
#ifdef CONFIG_GENERIC_PENDING_IRQ
|
|
|
|
free_cpumask_var(desc->pending_mask);
|
|
|
|
#endif
|
2010-10-13 02:58:27 +07:00
|
|
|
free_cpumask_var(desc->irq_data.affinity);
|
2010-09-27 22:48:26 +07:00
|
|
|
}
|
|
|
|
#else
|
|
|
|
static inline void free_masks(struct irq_desc *desc) { }
|
|
|
|
#endif
|
|
|
|
|
2011-07-11 17:17:31 +07:00
|
|
|
static struct irq_desc *alloc_desc(int irq, int node, struct module *owner)
|
2010-09-27 22:48:26 +07:00
|
|
|
{
|
|
|
|
struct irq_desc *desc;
|
2010-10-05 20:14:35 +07:00
|
|
|
gfp_t gfp = GFP_KERNEL;
|
2010-09-27 22:48:26 +07:00
|
|
|
|
|
|
|
desc = kzalloc_node(sizeof(*desc), gfp, node);
|
|
|
|
if (!desc)
|
|
|
|
return NULL;
|
|
|
|
/* allocate based on nr_cpu_ids */
|
2011-01-14 06:45:38 +07:00
|
|
|
desc->kstat_irqs = alloc_percpu(unsigned int);
|
2010-09-27 22:48:26 +07:00
|
|
|
if (!desc->kstat_irqs)
|
|
|
|
goto err_desc;
|
|
|
|
|
|
|
|
if (alloc_masks(desc, gfp, node))
|
|
|
|
goto err_kstat;
|
|
|
|
|
|
|
|
raw_spin_lock_init(&desc->lock);
|
|
|
|
lockdep_set_class(&desc->lock, &irq_desc_lock_class);
|
|
|
|
|
2011-07-11 17:17:31 +07:00
|
|
|
desc_set_defaults(irq, desc, node, owner);
|
2010-09-27 22:48:26 +07:00
|
|
|
|
|
|
|
return desc;
|
|
|
|
|
|
|
|
err_kstat:
|
2011-01-14 06:45:38 +07:00
|
|
|
free_percpu(desc->kstat_irqs);
|
2010-09-27 22:48:26 +07:00
|
|
|
err_desc:
|
|
|
|
kfree(desc);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void free_desc(unsigned int irq)
|
|
|
|
{
|
|
|
|
struct irq_desc *desc = irq_to_desc(irq);
|
|
|
|
|
2010-09-30 07:46:07 +07:00
|
|
|
unregister_irq_proc(irq, desc);
|
|
|
|
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_lock(&sparse_irq_lock);
|
2010-09-27 22:48:26 +07:00
|
|
|
delete_irq_desc(irq);
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_unlock(&sparse_irq_lock);
|
2010-09-27 22:48:26 +07:00
|
|
|
|
|
|
|
free_masks(desc);
|
2011-01-14 06:45:38 +07:00
|
|
|
free_percpu(desc->kstat_irqs);
|
2010-09-27 22:48:26 +07:00
|
|
|
kfree(desc);
|
|
|
|
}
|
|
|
|
|
2011-07-11 17:17:31 +07:00
|
|
|
static int alloc_descs(unsigned int start, unsigned int cnt, int node,
|
|
|
|
struct module *owner)
|
2010-09-27 22:48:26 +07:00
|
|
|
{
|
|
|
|
struct irq_desc *desc;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < cnt; i++) {
|
2011-07-11 17:17:31 +07:00
|
|
|
desc = alloc_desc(start + i, node, owner);
|
2010-09-27 22:48:26 +07:00
|
|
|
if (!desc)
|
|
|
|
goto err;
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_lock(&sparse_irq_lock);
|
2010-09-27 22:48:26 +07:00
|
|
|
irq_insert_desc(start + i, desc);
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_unlock(&sparse_irq_lock);
|
2010-09-27 22:48:26 +07:00
|
|
|
}
|
|
|
|
return start;
|
|
|
|
|
|
|
|
err:
|
|
|
|
for (i--; i >= 0; i--)
|
|
|
|
free_desc(start + i);
|
|
|
|
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_lock(&sparse_irq_lock);
|
2010-09-27 22:48:26 +07:00
|
|
|
bitmap_clear(allocated_irqs, start, cnt);
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_unlock(&sparse_irq_lock);
|
2010-09-27 22:48:26 +07:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2011-02-20 02:07:37 +07:00
|
|
|
static int irq_expand_nr_irqs(unsigned int nr)
|
2011-02-16 23:12:57 +07:00
|
|
|
{
|
2011-02-20 02:07:37 +07:00
|
|
|
if (nr > IRQ_BITMAP_BITS)
|
2011-02-16 23:12:57 +07:00
|
|
|
return -ENOMEM;
|
2011-02-20 02:07:37 +07:00
|
|
|
nr_irqs = nr;
|
2011-02-16 23:12:57 +07:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2010-09-22 22:09:43 +07:00
|
|
|
int __init early_irq_init(void)
|
|
|
|
{
|
2010-09-28 01:55:03 +07:00
|
|
|
int i, initcnt, node = first_online_node;
|
2010-09-22 22:09:43 +07:00
|
|
|
struct irq_desc *desc;
|
|
|
|
|
|
|
|
init_irq_default_affinity();
|
|
|
|
|
2010-09-28 01:55:03 +07:00
|
|
|
/* Let arch update nr_irqs and return the nr of preallocated irqs */
|
|
|
|
initcnt = arch_probe_nr_irqs();
|
|
|
|
printk(KERN_INFO "NR_IRQS:%d nr_irqs:%d %d\n", NR_IRQS, nr_irqs, initcnt);
|
2010-09-22 22:09:43 +07:00
|
|
|
|
2011-02-17 23:45:15 +07:00
|
|
|
if (WARN_ON(nr_irqs > IRQ_BITMAP_BITS))
|
|
|
|
nr_irqs = IRQ_BITMAP_BITS;
|
|
|
|
|
|
|
|
if (WARN_ON(initcnt > IRQ_BITMAP_BITS))
|
|
|
|
initcnt = IRQ_BITMAP_BITS;
|
|
|
|
|
|
|
|
if (initcnt > nr_irqs)
|
|
|
|
nr_irqs = initcnt;
|
|
|
|
|
2010-09-28 01:55:03 +07:00
|
|
|
for (i = 0; i < initcnt; i++) {
|
2011-07-11 17:17:31 +07:00
|
|
|
desc = alloc_desc(i, node, NULL);
|
2010-09-28 01:02:56 +07:00
|
|
|
set_bit(i, allocated_irqs);
|
|
|
|
irq_insert_desc(i, desc);
|
2010-09-22 22:09:43 +07:00
|
|
|
}
|
|
|
|
return arch_early_irq_init();
|
|
|
|
}
|
|
|
|
|
|
|
|
#else /* !CONFIG_SPARSE_IRQ */
|
|
|
|
|
|
|
|
struct irq_desc irq_desc[NR_IRQS] __cacheline_aligned_in_smp = {
|
|
|
|
[0 ... NR_IRQS-1] = {
|
|
|
|
.handle_irq = handle_bad_irq,
|
|
|
|
.depth = 1,
|
|
|
|
.lock = __RAW_SPIN_LOCK_UNLOCKED(irq_desc->lock),
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
int __init early_irq_init(void)
|
|
|
|
{
|
2010-09-28 01:02:56 +07:00
|
|
|
int count, i, node = first_online_node;
|
2010-09-22 22:09:43 +07:00
|
|
|
struct irq_desc *desc;
|
|
|
|
|
|
|
|
init_irq_default_affinity();
|
|
|
|
|
|
|
|
printk(KERN_INFO "NR_IRQS:%d\n", NR_IRQS);
|
|
|
|
|
|
|
|
desc = irq_desc;
|
|
|
|
count = ARRAY_SIZE(irq_desc);
|
|
|
|
|
|
|
|
for (i = 0; i < count; i++) {
|
2011-01-14 06:45:38 +07:00
|
|
|
desc[i].kstat_irqs = alloc_percpu(unsigned int);
|
2011-05-31 23:14:39 +07:00
|
|
|
alloc_masks(&desc[i], GFP_KERNEL, node);
|
|
|
|
raw_spin_lock_init(&desc[i].lock);
|
2010-09-22 20:58:45 +07:00
|
|
|
lockdep_set_class(&desc[i].lock, &irq_desc_lock_class);
|
2011-07-11 17:17:31 +07:00
|
|
|
desc_set_defaults(i, &desc[i], node, NULL);
|
2010-09-22 22:09:43 +07:00
|
|
|
}
|
|
|
|
return arch_early_irq_init();
|
|
|
|
}
|
|
|
|
|
|
|
|
struct irq_desc *irq_to_desc(unsigned int irq)
|
|
|
|
{
|
|
|
|
return (irq < NR_IRQS) ? irq_desc + irq : NULL;
|
|
|
|
}
|
2014-02-11 01:39:53 +07:00
|
|
|
EXPORT_SYMBOL(irq_to_desc);
|
2010-09-22 22:09:43 +07:00
|
|
|
|
2010-09-27 22:48:26 +07:00
|
|
|
static void free_desc(unsigned int irq)
|
|
|
|
{
|
2014-05-07 22:44:23 +07:00
|
|
|
struct irq_desc *desc = irq_to_desc(irq);
|
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
raw_spin_lock_irqsave(&desc->lock, flags);
|
|
|
|
desc_set_defaults(irq, desc, desc_node(desc), NULL);
|
|
|
|
raw_spin_unlock_irqrestore(&desc->lock, flags);
|
2010-09-27 22:48:26 +07:00
|
|
|
}
|
|
|
|
|
2011-07-11 17:17:31 +07:00
|
|
|
static inline int alloc_descs(unsigned int start, unsigned int cnt, int node,
|
|
|
|
struct module *owner)
|
2010-09-27 22:48:26 +07:00
|
|
|
{
|
2011-07-11 17:17:31 +07:00
|
|
|
u32 i;
|
|
|
|
|
|
|
|
for (i = 0; i < cnt; i++) {
|
|
|
|
struct irq_desc *desc = irq_to_desc(start + i);
|
|
|
|
|
|
|
|
desc->owner = owner;
|
|
|
|
}
|
2010-09-27 22:48:26 +07:00
|
|
|
return start;
|
|
|
|
}
|
2011-02-16 23:12:57 +07:00
|
|
|
|
2011-02-20 02:07:37 +07:00
|
|
|
static int irq_expand_nr_irqs(unsigned int nr)
|
2011-02-16 23:12:57 +07:00
|
|
|
{
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2014-05-07 22:44:21 +07:00
|
|
|
void irq_mark_irq(unsigned int irq)
|
|
|
|
{
|
|
|
|
mutex_lock(&sparse_irq_lock);
|
|
|
|
bitmap_set(allocated_irqs, irq, 1);
|
|
|
|
mutex_unlock(&sparse_irq_lock);
|
|
|
|
}
|
|
|
|
|
2014-05-07 22:44:22 +07:00
|
|
|
#ifdef CONFIG_GENERIC_IRQ_LEGACY
|
|
|
|
void irq_init_desc(unsigned int irq)
|
|
|
|
{
|
2014-05-07 22:44:23 +07:00
|
|
|
free_desc(irq);
|
2014-05-07 22:44:22 +07:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2010-09-22 22:09:43 +07:00
|
|
|
#endif /* !CONFIG_SPARSE_IRQ */
|
|
|
|
|
2011-05-18 17:48:00 +07:00
|
|
|
/**
|
|
|
|
* generic_handle_irq - Invoke the handler for a particular irq
|
|
|
|
* @irq: The irq number to handle
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
int generic_handle_irq(unsigned int irq)
|
|
|
|
{
|
|
|
|
struct irq_desc *desc = irq_to_desc(irq);
|
|
|
|
|
|
|
|
if (!desc)
|
|
|
|
return -EINVAL;
|
|
|
|
generic_handle_irq_desc(irq, desc);
|
|
|
|
return 0;
|
|
|
|
}
|
2011-05-18 16:39:04 +07:00
|
|
|
EXPORT_SYMBOL_GPL(generic_handle_irq);
|
2011-05-18 17:48:00 +07:00
|
|
|
|
2010-09-27 22:48:26 +07:00
|
|
|
/* Dynamic interrupt handling */
|
|
|
|
|
|
|
|
/**
|
|
|
|
* irq_free_descs - free irq descriptors
|
|
|
|
* @from: Start of descriptor range
|
|
|
|
* @cnt: Number of consecutive irqs to free
|
|
|
|
*/
|
|
|
|
void irq_free_descs(unsigned int from, unsigned int cnt)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
if (from >= nr_irqs || (from + cnt) > nr_irqs)
|
|
|
|
return;
|
|
|
|
|
|
|
|
for (i = 0; i < cnt; i++)
|
|
|
|
free_desc(from + i);
|
|
|
|
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_lock(&sparse_irq_lock);
|
2010-09-27 22:48:26 +07:00
|
|
|
bitmap_clear(allocated_irqs, from, cnt);
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_unlock(&sparse_irq_lock);
|
2010-09-27 22:48:26 +07:00
|
|
|
}
|
2011-05-18 16:39:04 +07:00
|
|
|
EXPORT_SYMBOL_GPL(irq_free_descs);
|
2010-09-27 22:48:26 +07:00
|
|
|
|
|
|
|
/**
|
|
|
|
* irq_alloc_descs - allocate and initialize a range of irq descriptors
|
|
|
|
* @irq: Allocate for specific irq number if irq >= 0
|
|
|
|
* @from: Start the search from this irq number
|
|
|
|
* @cnt: Number of consecutive irqs to allocate.
|
|
|
|
* @node: Preferred node on which the irq descriptor should be allocated
|
2011-08-19 02:19:27 +07:00
|
|
|
* @owner: Owning module (can be NULL)
|
2010-09-27 22:48:26 +07:00
|
|
|
*
|
|
|
|
* Returns the first irq number or error code
|
|
|
|
*/
|
|
|
|
int __ref
|
2011-07-11 17:17:31 +07:00
|
|
|
__irq_alloc_descs(int irq, unsigned int from, unsigned int cnt, int node,
|
|
|
|
struct module *owner)
|
2010-09-27 22:48:26 +07:00
|
|
|
{
|
|
|
|
int start, ret;
|
|
|
|
|
|
|
|
if (!cnt)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2011-06-03 00:55:13 +07:00
|
|
|
if (irq >= 0) {
|
|
|
|
if (from > irq)
|
|
|
|
return -EINVAL;
|
|
|
|
from = irq;
|
2014-04-24 14:50:53 +07:00
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* For interrupts which are freely allocated the
|
|
|
|
* architecture can force a lower bound to the @from
|
|
|
|
* argument. x86 uses this to exclude the GSI space.
|
|
|
|
*/
|
|
|
|
from = arch_dynirq_lower_bound(from);
|
2011-06-03 00:55:13 +07:00
|
|
|
}
|
|
|
|
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_lock(&sparse_irq_lock);
|
2010-09-27 22:48:26 +07:00
|
|
|
|
2011-02-20 02:07:37 +07:00
|
|
|
start = bitmap_find_next_zero_area(allocated_irqs, IRQ_BITMAP_BITS,
|
|
|
|
from, cnt, 0);
|
2010-09-27 22:48:26 +07:00
|
|
|
ret = -EEXIST;
|
|
|
|
if (irq >=0 && start != irq)
|
|
|
|
goto err;
|
|
|
|
|
2011-02-20 02:07:37 +07:00
|
|
|
if (start + cnt > nr_irqs) {
|
|
|
|
ret = irq_expand_nr_irqs(start + cnt);
|
2011-02-16 23:12:57 +07:00
|
|
|
if (ret)
|
|
|
|
goto err;
|
|
|
|
}
|
2010-09-27 22:48:26 +07:00
|
|
|
|
|
|
|
bitmap_set(allocated_irqs, start, cnt);
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_unlock(&sparse_irq_lock);
|
2011-07-11 17:17:31 +07:00
|
|
|
return alloc_descs(start, cnt, node, owner);
|
2010-09-27 22:48:26 +07:00
|
|
|
|
|
|
|
err:
|
2010-10-08 17:47:53 +07:00
|
|
|
mutex_unlock(&sparse_irq_lock);
|
2010-09-27 22:48:26 +07:00
|
|
|
return ret;
|
|
|
|
}
|
2011-07-11 17:17:31 +07:00
|
|
|
EXPORT_SYMBOL_GPL(__irq_alloc_descs);
|
2010-09-27 22:48:26 +07:00
|
|
|
|
2014-05-07 22:44:05 +07:00
|
|
|
#ifdef CONFIG_GENERIC_IRQ_LEGACY_ALLOC_HWIRQ
|
|
|
|
/**
|
|
|
|
* irq_alloc_hwirqs - Allocate an irq descriptor and initialize the hardware
|
|
|
|
* @cnt: number of interrupts to allocate
|
|
|
|
* @node: node on which to allocate
|
|
|
|
*
|
|
|
|
* Returns an interrupt number > 0 or 0, if the allocation fails.
|
|
|
|
*/
|
|
|
|
unsigned int irq_alloc_hwirqs(int cnt, int node)
|
|
|
|
{
|
|
|
|
int i, irq = __irq_alloc_descs(-1, 0, cnt, node, NULL);
|
|
|
|
|
|
|
|
if (irq < 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
for (i = irq; cnt > 0; i++, cnt--) {
|
|
|
|
if (arch_setup_hwirq(i, node))
|
|
|
|
goto err;
|
|
|
|
irq_clear_status_flags(i, _IRQ_NOREQUEST);
|
|
|
|
}
|
|
|
|
return irq;
|
|
|
|
|
|
|
|
err:
|
|
|
|
for (i--; i >= irq; i--) {
|
|
|
|
irq_set_status_flags(i, _IRQ_NOREQUEST | _IRQ_NOPROBE);
|
|
|
|
arch_teardown_hwirq(i);
|
|
|
|
}
|
|
|
|
irq_free_descs(irq, cnt);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(irq_alloc_hwirqs);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* irq_free_hwirqs - Free irq descriptor and cleanup the hardware
|
|
|
|
* @from: Free from irq number
|
|
|
|
* @cnt: number of interrupts to free
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
void irq_free_hwirqs(unsigned int from, int cnt)
|
|
|
|
{
|
2014-07-01 05:24:44 +07:00
|
|
|
int i, j;
|
2014-05-07 22:44:05 +07:00
|
|
|
|
2014-07-01 05:24:44 +07:00
|
|
|
for (i = from, j = cnt; j > 0; i++, j--) {
|
2014-05-07 22:44:05 +07:00
|
|
|
irq_set_status_flags(i, _IRQ_NOREQUEST | _IRQ_NOPROBE);
|
|
|
|
arch_teardown_hwirq(i);
|
|
|
|
}
|
|
|
|
irq_free_descs(from, cnt);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(irq_free_hwirqs);
|
|
|
|
#endif
|
|
|
|
|
2010-09-30 15:45:07 +07:00
|
|
|
/**
|
|
|
|
* irq_get_next_irq - get next allocated irq number
|
|
|
|
* @offset: where to start the search
|
|
|
|
*
|
|
|
|
* Returns next irq number after offset or nr_irqs if none is found.
|
|
|
|
*/
|
|
|
|
unsigned int irq_get_next_irq(unsigned int offset)
|
|
|
|
{
|
|
|
|
return find_next_bit(allocated_irqs, nr_irqs, offset);
|
|
|
|
}
|
|
|
|
|
2011-02-12 18:16:16 +07:00
|
|
|
struct irq_desc *
|
genirq: Add support for per-cpu dev_id interrupts
The ARM GIC interrupt controller offers per CPU interrupts (PPIs),
which are usually used to connect local timers to each core. Each CPU
has its own private interface to the GIC, and only sees the PPIs that
are directly connect to it.
While these timers are separate devices and have a separate interrupt
line to a core, they all use the same IRQ number.
For these devices, request_irq() is not the right API as it assumes
that an IRQ number is visible by a number of CPUs (through the
affinity setting), but makes it very awkward to express that an IRQ
number can be handled by all CPUs, and yet be a different interrupt
line on each CPU, requiring a different dev_id cookie to be passed
back to the handler.
The *_percpu_irq() functions is designed to overcome these
limitations, by providing a per-cpu dev_id vector:
int request_percpu_irq(unsigned int irq, irq_handler_t handler,
const char *devname, void __percpu *percpu_dev_id);
void free_percpu_irq(unsigned int, void __percpu *);
int setup_percpu_irq(unsigned int irq, struct irqaction *new);
void remove_percpu_irq(unsigned int irq, struct irqaction *act);
void enable_percpu_irq(unsigned int irq);
void disable_percpu_irq(unsigned int irq);
The API has a number of limitations:
- no interrupt sharing
- no threading
- common handler across all the CPUs
Once the interrupt is requested using setup_percpu_irq() or
request_percpu_irq(), it must be enabled by each core that wishes its
local interrupt to be delivered.
Based on an initial patch by Thomas Gleixner.
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
Cc: linux-arm-kernel@lists.infradead.org
Link: http://lkml.kernel.org/r/1316793788-14500-2-git-send-email-marc.zyngier@arm.com
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
2011-09-23 23:03:06 +07:00
|
|
|
__irq_get_desc_lock(unsigned int irq, unsigned long *flags, bool bus,
|
|
|
|
unsigned int check)
|
2011-02-12 18:16:16 +07:00
|
|
|
{
|
|
|
|
struct irq_desc *desc = irq_to_desc(irq);
|
|
|
|
|
|
|
|
if (desc) {
|
genirq: Add support for per-cpu dev_id interrupts
The ARM GIC interrupt controller offers per CPU interrupts (PPIs),
which are usually used to connect local timers to each core. Each CPU
has its own private interface to the GIC, and only sees the PPIs that
are directly connect to it.
While these timers are separate devices and have a separate interrupt
line to a core, they all use the same IRQ number.
For these devices, request_irq() is not the right API as it assumes
that an IRQ number is visible by a number of CPUs (through the
affinity setting), but makes it very awkward to express that an IRQ
number can be handled by all CPUs, and yet be a different interrupt
line on each CPU, requiring a different dev_id cookie to be passed
back to the handler.
The *_percpu_irq() functions is designed to overcome these
limitations, by providing a per-cpu dev_id vector:
int request_percpu_irq(unsigned int irq, irq_handler_t handler,
const char *devname, void __percpu *percpu_dev_id);
void free_percpu_irq(unsigned int, void __percpu *);
int setup_percpu_irq(unsigned int irq, struct irqaction *new);
void remove_percpu_irq(unsigned int irq, struct irqaction *act);
void enable_percpu_irq(unsigned int irq);
void disable_percpu_irq(unsigned int irq);
The API has a number of limitations:
- no interrupt sharing
- no threading
- common handler across all the CPUs
Once the interrupt is requested using setup_percpu_irq() or
request_percpu_irq(), it must be enabled by each core that wishes its
local interrupt to be delivered.
Based on an initial patch by Thomas Gleixner.
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
Cc: linux-arm-kernel@lists.infradead.org
Link: http://lkml.kernel.org/r/1316793788-14500-2-git-send-email-marc.zyngier@arm.com
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
2011-09-23 23:03:06 +07:00
|
|
|
if (check & _IRQ_DESC_CHECK) {
|
|
|
|
if ((check & _IRQ_DESC_PERCPU) &&
|
|
|
|
!irq_settings_is_per_cpu_devid(desc))
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
if (!(check & _IRQ_DESC_PERCPU) &&
|
|
|
|
irq_settings_is_per_cpu_devid(desc))
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2011-02-12 18:16:16 +07:00
|
|
|
if (bus)
|
|
|
|
chip_bus_lock(desc);
|
|
|
|
raw_spin_lock_irqsave(&desc->lock, *flags);
|
|
|
|
}
|
|
|
|
return desc;
|
|
|
|
}
|
|
|
|
|
|
|
|
void __irq_put_desc_unlock(struct irq_desc *desc, unsigned long flags, bool bus)
|
|
|
|
{
|
|
|
|
raw_spin_unlock_irqrestore(&desc->lock, flags);
|
|
|
|
if (bus)
|
|
|
|
chip_bus_sync_unlock(desc);
|
|
|
|
}
|
|
|
|
|
genirq: Add support for per-cpu dev_id interrupts
The ARM GIC interrupt controller offers per CPU interrupts (PPIs),
which are usually used to connect local timers to each core. Each CPU
has its own private interface to the GIC, and only sees the PPIs that
are directly connect to it.
While these timers are separate devices and have a separate interrupt
line to a core, they all use the same IRQ number.
For these devices, request_irq() is not the right API as it assumes
that an IRQ number is visible by a number of CPUs (through the
affinity setting), but makes it very awkward to express that an IRQ
number can be handled by all CPUs, and yet be a different interrupt
line on each CPU, requiring a different dev_id cookie to be passed
back to the handler.
The *_percpu_irq() functions is designed to overcome these
limitations, by providing a per-cpu dev_id vector:
int request_percpu_irq(unsigned int irq, irq_handler_t handler,
const char *devname, void __percpu *percpu_dev_id);
void free_percpu_irq(unsigned int, void __percpu *);
int setup_percpu_irq(unsigned int irq, struct irqaction *new);
void remove_percpu_irq(unsigned int irq, struct irqaction *act);
void enable_percpu_irq(unsigned int irq);
void disable_percpu_irq(unsigned int irq);
The API has a number of limitations:
- no interrupt sharing
- no threading
- common handler across all the CPUs
Once the interrupt is requested using setup_percpu_irq() or
request_percpu_irq(), it must be enabled by each core that wishes its
local interrupt to be delivered.
Based on an initial patch by Thomas Gleixner.
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
Cc: linux-arm-kernel@lists.infradead.org
Link: http://lkml.kernel.org/r/1316793788-14500-2-git-send-email-marc.zyngier@arm.com
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
2011-09-23 23:03:06 +07:00
|
|
|
int irq_set_percpu_devid(unsigned int irq)
|
|
|
|
{
|
|
|
|
struct irq_desc *desc = irq_to_desc(irq);
|
|
|
|
|
|
|
|
if (!desc)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (desc->percpu_enabled)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
desc->percpu_enabled = kzalloc(sizeof(*desc->percpu_enabled), GFP_KERNEL);
|
|
|
|
|
|
|
|
if (!desc->percpu_enabled)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
irq_set_percpu_devid_flags(irq);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2014-02-24 04:40:14 +07:00
|
|
|
void kstat_incr_irq_this_cpu(unsigned int irq)
|
|
|
|
{
|
|
|
|
kstat_incr_irqs_this_cpu(irq, irq_to_desc(irq));
|
|
|
|
}
|
|
|
|
|
2010-09-22 22:09:43 +07:00
|
|
|
unsigned int kstat_irqs_cpu(unsigned int irq, int cpu)
|
|
|
|
{
|
|
|
|
struct irq_desc *desc = irq_to_desc(irq);
|
2011-01-14 06:45:38 +07:00
|
|
|
|
|
|
|
return desc && desc->kstat_irqs ?
|
|
|
|
*per_cpu_ptr(desc->kstat_irqs, cpu) : 0;
|
2010-09-22 22:09:43 +07:00
|
|
|
}
|
2010-10-28 05:34:15 +07:00
|
|
|
|
|
|
|
unsigned int kstat_irqs(unsigned int irq)
|
|
|
|
{
|
|
|
|
struct irq_desc *desc = irq_to_desc(irq);
|
|
|
|
int cpu;
|
|
|
|
int sum = 0;
|
|
|
|
|
2011-01-14 06:45:38 +07:00
|
|
|
if (!desc || !desc->kstat_irqs)
|
2010-10-28 05:34:15 +07:00
|
|
|
return 0;
|
|
|
|
for_each_possible_cpu(cpu)
|
2011-01-14 06:45:38 +07:00
|
|
|
sum += *per_cpu_ptr(desc->kstat_irqs, cpu);
|
2010-10-28 05:34:15 +07:00
|
|
|
return sum;
|
|
|
|
}
|