mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-19 21:07:56 +07:00
d7f3d47818
Currently x86 support identity mapping between GSI(IOAPIC pin) and IRQ number, so continous IRQs at low end are statically allocated to IOAPICs at boot time. This design causes trouble to support IOAPIC hotplug. This patch implements basic mechanism to dynamically allocate IRQ on demand for IOAPIC pins by using irqdomain framework. It first adds several fields into struct ioapic to support irqdomain. Then it implements an algorithm to dynamically allocate IRQ number for IOAPIC pins on demand. Currently it supports three types of irqdomain: 1) LEGACY: used to support IOAPIC hosting legacy IRQs and building identity mapping for legacy IRQs. A speical case, we dynamically allocate IRQ number for IOAPIC pin which has GSI number below nr_legacy_irqs() but isn't legacy IRQ. This is for backward compatibility and avoid regression. 2) STRICT: build identity mapping between GSI and IRQ nubmer. 3) DYNAMIC: dynamically allocate IRQ number for IOAPIC pin on demand. Legacy(ISA) IRQs is not managed by irqdomain because there may be multiple pins sharing the same IRQ number and current irqdomain only supports 1:1 mapping between pins and IRQ. Signed-off-by: Jiang Liu <jiang.liu@linux.intel.com> Cc: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com> Cc: Tony Luck <tony.luck@intel.com> Cc: Joerg Roedel <joro@8bytes.org> Cc: Paul Gortmaker <paul.gortmaker@windriver.com> Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org> Cc: Grant Likely <grant.likely@linaro.org> Cc: Rafael J. Wysocki <rjw@rjwysocki.net> Cc: Bjorn Helgaas <bhelgaas@google.com> Cc: Randy Dunlap <rdunlap@infradead.org> Cc: Yinghai Lu <yinghai@kernel.org> Cc: Len Brown <len.brown@intel.com> Cc: Pavel Machek <pavel@ucw.cz> Link: http://lkml.kernel.org/r/1402302011-23642-24-git-send-email-jiang.liu@linux.intel.com Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
1630 lines
38 KiB
C
1630 lines
38 KiB
C
/*
|
|
* boot.c - Architecture-Specific Low-Level ACPI Boot Support
|
|
*
|
|
* Copyright (C) 2001, 2002 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com>
|
|
* Copyright (C) 2001 Jun Nakajima <jun.nakajima@intel.com>
|
|
*
|
|
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*
|
|
* ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
*/
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/acpi.h>
|
|
#include <linux/acpi_pmtmr.h>
|
|
#include <linux/efi.h>
|
|
#include <linux/cpumask.h>
|
|
#include <linux/module.h>
|
|
#include <linux/dmi.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/bootmem.h>
|
|
#include <linux/ioport.h>
|
|
#include <linux/pci.h>
|
|
|
|
#include <asm/pci_x86.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/io_apic.h>
|
|
#include <asm/apic.h>
|
|
#include <asm/io.h>
|
|
#include <asm/mpspec.h>
|
|
#include <asm/smp.h>
|
|
#include <asm/i8259.h>
|
|
|
|
#include "sleep.h" /* To include x86_acpi_suspend_lowlevel */
|
|
static int __initdata acpi_force = 0;
|
|
int acpi_disabled;
|
|
EXPORT_SYMBOL(acpi_disabled);
|
|
|
|
#ifdef CONFIG_X86_64
|
|
# include <asm/proto.h>
|
|
#endif /* X86 */
|
|
|
|
#define PREFIX "ACPI: "
|
|
|
|
int acpi_noirq; /* skip ACPI IRQ initialization */
|
|
int acpi_pci_disabled; /* skip ACPI PCI scan and IRQ initialization */
|
|
EXPORT_SYMBOL(acpi_pci_disabled);
|
|
|
|
int acpi_lapic;
|
|
int acpi_ioapic;
|
|
int acpi_strict;
|
|
int acpi_disable_cmcff;
|
|
|
|
u8 acpi_sci_flags __initdata;
|
|
int acpi_sci_override_gsi __initdata;
|
|
int acpi_skip_timer_override __initdata;
|
|
int acpi_use_timer_override __initdata;
|
|
int acpi_fix_pin2_polarity __initdata;
|
|
|
|
#ifdef CONFIG_X86_LOCAL_APIC
|
|
static u64 acpi_lapic_addr __initdata = APIC_DEFAULT_PHYS_BASE;
|
|
#endif
|
|
|
|
#ifndef __HAVE_ARCH_CMPXCHG
|
|
#warning ACPI uses CMPXCHG, i486 and later hardware
|
|
#endif
|
|
|
|
/* --------------------------------------------------------------------------
|
|
Boot-time Configuration
|
|
-------------------------------------------------------------------------- */
|
|
|
|
/*
|
|
* The default interrupt routing model is PIC (8259). This gets
|
|
* overridden if IOAPICs are enumerated (below).
|
|
*/
|
|
enum acpi_irq_model_id acpi_irq_model = ACPI_IRQ_MODEL_PIC;
|
|
|
|
|
|
/*
|
|
* ISA irqs by default are the first 16 gsis but can be
|
|
* any gsi as specified by an interrupt source override.
|
|
*/
|
|
static u32 isa_irq_to_gsi[NR_IRQS_LEGACY] __read_mostly = {
|
|
0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15
|
|
};
|
|
|
|
#define ACPI_INVALID_GSI INT_MIN
|
|
|
|
static int map_gsi_to_irq(unsigned int gsi, unsigned int flags)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < nr_legacy_irqs(); i++)
|
|
if (isa_irq_to_gsi[i] == gsi)
|
|
return i;
|
|
|
|
return mp_map_gsi_to_irq(gsi, flags);
|
|
}
|
|
|
|
/*
|
|
* This is just a simple wrapper around early_ioremap(),
|
|
* with sanity checks for phys == 0 and size == 0.
|
|
*/
|
|
char *__init __acpi_map_table(unsigned long phys, unsigned long size)
|
|
{
|
|
|
|
if (!phys || !size)
|
|
return NULL;
|
|
|
|
return early_ioremap(phys, size);
|
|
}
|
|
|
|
void __init __acpi_unmap_table(char *map, unsigned long size)
|
|
{
|
|
if (!map || !size)
|
|
return;
|
|
|
|
early_iounmap(map, size);
|
|
}
|
|
|
|
#ifdef CONFIG_X86_LOCAL_APIC
|
|
static int __init acpi_parse_madt(struct acpi_table_header *table)
|
|
{
|
|
struct acpi_table_madt *madt = NULL;
|
|
|
|
if (!cpu_has_apic)
|
|
return -EINVAL;
|
|
|
|
madt = (struct acpi_table_madt *)table;
|
|
if (!madt) {
|
|
printk(KERN_WARNING PREFIX "Unable to map MADT\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
if (madt->address) {
|
|
acpi_lapic_addr = (u64) madt->address;
|
|
|
|
printk(KERN_DEBUG PREFIX "Local APIC address 0x%08x\n",
|
|
madt->address);
|
|
}
|
|
|
|
default_acpi_madt_oem_check(madt->header.oem_id,
|
|
madt->header.oem_table_id);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* acpi_register_lapic - register a local apic and generates a logic cpu number
|
|
* @id: local apic id to register
|
|
* @enabled: this cpu is enabled or not
|
|
*
|
|
* Returns the logic cpu number which maps to the local apic
|
|
*/
|
|
static int acpi_register_lapic(int id, u8 enabled)
|
|
{
|
|
unsigned int ver = 0;
|
|
|
|
if (id >= MAX_LOCAL_APIC) {
|
|
printk(KERN_INFO PREFIX "skipped apicid that is too big\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!enabled) {
|
|
++disabled_cpus;
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (boot_cpu_physical_apicid != -1U)
|
|
ver = apic_version[boot_cpu_physical_apicid];
|
|
|
|
return generic_processor_info(id, ver);
|
|
}
|
|
|
|
static int __init
|
|
acpi_parse_x2apic(struct acpi_subtable_header *header, const unsigned long end)
|
|
{
|
|
struct acpi_madt_local_x2apic *processor = NULL;
|
|
int apic_id;
|
|
u8 enabled;
|
|
|
|
processor = (struct acpi_madt_local_x2apic *)header;
|
|
|
|
if (BAD_MADT_ENTRY(processor, end))
|
|
return -EINVAL;
|
|
|
|
acpi_table_print_madt_entry(header);
|
|
|
|
apic_id = processor->local_apic_id;
|
|
enabled = processor->lapic_flags & ACPI_MADT_ENABLED;
|
|
#ifdef CONFIG_X86_X2APIC
|
|
/*
|
|
* We need to register disabled CPU as well to permit
|
|
* counting disabled CPUs. This allows us to size
|
|
* cpus_possible_map more accurately, to permit
|
|
* to not preallocating memory for all NR_CPUS
|
|
* when we use CPU hotplug.
|
|
*/
|
|
if (!apic->apic_id_valid(apic_id) && enabled)
|
|
printk(KERN_WARNING PREFIX "x2apic entry ignored\n");
|
|
else
|
|
acpi_register_lapic(apic_id, enabled);
|
|
#else
|
|
printk(KERN_WARNING PREFIX "x2apic entry ignored\n");
|
|
#endif
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init
|
|
acpi_parse_lapic(struct acpi_subtable_header * header, const unsigned long end)
|
|
{
|
|
struct acpi_madt_local_apic *processor = NULL;
|
|
|
|
processor = (struct acpi_madt_local_apic *)header;
|
|
|
|
if (BAD_MADT_ENTRY(processor, end))
|
|
return -EINVAL;
|
|
|
|
acpi_table_print_madt_entry(header);
|
|
|
|
/*
|
|
* We need to register disabled CPU as well to permit
|
|
* counting disabled CPUs. This allows us to size
|
|
* cpus_possible_map more accurately, to permit
|
|
* to not preallocating memory for all NR_CPUS
|
|
* when we use CPU hotplug.
|
|
*/
|
|
acpi_register_lapic(processor->id, /* APIC ID */
|
|
processor->lapic_flags & ACPI_MADT_ENABLED);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init
|
|
acpi_parse_sapic(struct acpi_subtable_header *header, const unsigned long end)
|
|
{
|
|
struct acpi_madt_local_sapic *processor = NULL;
|
|
|
|
processor = (struct acpi_madt_local_sapic *)header;
|
|
|
|
if (BAD_MADT_ENTRY(processor, end))
|
|
return -EINVAL;
|
|
|
|
acpi_table_print_madt_entry(header);
|
|
|
|
acpi_register_lapic((processor->id << 8) | processor->eid,/* APIC ID */
|
|
processor->lapic_flags & ACPI_MADT_ENABLED);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init
|
|
acpi_parse_lapic_addr_ovr(struct acpi_subtable_header * header,
|
|
const unsigned long end)
|
|
{
|
|
struct acpi_madt_local_apic_override *lapic_addr_ovr = NULL;
|
|
|
|
lapic_addr_ovr = (struct acpi_madt_local_apic_override *)header;
|
|
|
|
if (BAD_MADT_ENTRY(lapic_addr_ovr, end))
|
|
return -EINVAL;
|
|
|
|
acpi_lapic_addr = lapic_addr_ovr->address;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init
|
|
acpi_parse_x2apic_nmi(struct acpi_subtable_header *header,
|
|
const unsigned long end)
|
|
{
|
|
struct acpi_madt_local_x2apic_nmi *x2apic_nmi = NULL;
|
|
|
|
x2apic_nmi = (struct acpi_madt_local_x2apic_nmi *)header;
|
|
|
|
if (BAD_MADT_ENTRY(x2apic_nmi, end))
|
|
return -EINVAL;
|
|
|
|
acpi_table_print_madt_entry(header);
|
|
|
|
if (x2apic_nmi->lint != 1)
|
|
printk(KERN_WARNING PREFIX "NMI not connected to LINT 1!\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init
|
|
acpi_parse_lapic_nmi(struct acpi_subtable_header * header, const unsigned long end)
|
|
{
|
|
struct acpi_madt_local_apic_nmi *lapic_nmi = NULL;
|
|
|
|
lapic_nmi = (struct acpi_madt_local_apic_nmi *)header;
|
|
|
|
if (BAD_MADT_ENTRY(lapic_nmi, end))
|
|
return -EINVAL;
|
|
|
|
acpi_table_print_madt_entry(header);
|
|
|
|
if (lapic_nmi->lint != 1)
|
|
printk(KERN_WARNING PREFIX "NMI not connected to LINT 1!\n");
|
|
|
|
return 0;
|
|
}
|
|
|
|
#endif /*CONFIG_X86_LOCAL_APIC */
|
|
|
|
#ifdef CONFIG_X86_IO_APIC
|
|
#define MP_ISA_BUS 0
|
|
|
|
static void __init mp_override_legacy_irq(u8 bus_irq, u8 polarity, u8 trigger,
|
|
u32 gsi)
|
|
{
|
|
int ioapic;
|
|
int pin;
|
|
struct mpc_intsrc mp_irq;
|
|
|
|
/*
|
|
* Convert 'gsi' to 'ioapic.pin'.
|
|
*/
|
|
ioapic = mp_find_ioapic(gsi);
|
|
if (ioapic < 0)
|
|
return;
|
|
pin = mp_find_ioapic_pin(ioapic, gsi);
|
|
|
|
/*
|
|
* TBD: This check is for faulty timer entries, where the override
|
|
* erroneously sets the trigger to level, resulting in a HUGE
|
|
* increase of timer interrupts!
|
|
*/
|
|
if ((bus_irq == 0) && (trigger == 3))
|
|
trigger = 1;
|
|
|
|
mp_irq.type = MP_INTSRC;
|
|
mp_irq.irqtype = mp_INT;
|
|
mp_irq.irqflag = (trigger << 2) | polarity;
|
|
mp_irq.srcbus = MP_ISA_BUS;
|
|
mp_irq.srcbusirq = bus_irq; /* IRQ */
|
|
mp_irq.dstapic = mpc_ioapic_id(ioapic); /* APIC ID */
|
|
mp_irq.dstirq = pin; /* INTIN# */
|
|
|
|
mp_save_irq(&mp_irq);
|
|
|
|
/*
|
|
* Reset default identity mapping if gsi is also an legacy IRQ,
|
|
* otherwise there will be more than one entry with the same GSI
|
|
* and acpi_isa_irq_to_gsi() may give wrong result.
|
|
*/
|
|
if (gsi < nr_legacy_irqs() && isa_irq_to_gsi[gsi] == gsi)
|
|
isa_irq_to_gsi[gsi] = ACPI_INVALID_GSI;
|
|
isa_irq_to_gsi[bus_irq] = gsi;
|
|
}
|
|
|
|
static int mp_config_acpi_gsi(struct device *dev, u32 gsi, int trigger,
|
|
int polarity)
|
|
{
|
|
#ifdef CONFIG_X86_MPPARSE
|
|
struct mpc_intsrc mp_irq;
|
|
struct pci_dev *pdev;
|
|
unsigned char number;
|
|
unsigned int devfn;
|
|
int ioapic;
|
|
u8 pin;
|
|
|
|
if (!acpi_ioapic)
|
|
return 0;
|
|
if (!dev || !dev_is_pci(dev))
|
|
return 0;
|
|
|
|
pdev = to_pci_dev(dev);
|
|
number = pdev->bus->number;
|
|
devfn = pdev->devfn;
|
|
pin = pdev->pin;
|
|
/* print the entry should happen on mptable identically */
|
|
mp_irq.type = MP_INTSRC;
|
|
mp_irq.irqtype = mp_INT;
|
|
mp_irq.irqflag = (trigger == ACPI_EDGE_SENSITIVE ? 4 : 0x0c) |
|
|
(polarity == ACPI_ACTIVE_HIGH ? 1 : 3);
|
|
mp_irq.srcbus = number;
|
|
mp_irq.srcbusirq = (((devfn >> 3) & 0x1f) << 2) | ((pin - 1) & 3);
|
|
ioapic = mp_find_ioapic(gsi);
|
|
mp_irq.dstapic = mpc_ioapic_id(ioapic);
|
|
mp_irq.dstirq = mp_find_ioapic_pin(ioapic, gsi);
|
|
|
|
mp_save_irq(&mp_irq);
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
static int mp_register_gsi(struct device *dev, u32 gsi, int trigger,
|
|
int polarity)
|
|
{
|
|
int irq;
|
|
int ioapic;
|
|
int ioapic_pin;
|
|
struct io_apic_irq_attr irq_attr;
|
|
int ret;
|
|
|
|
if (acpi_irq_model != ACPI_IRQ_MODEL_IOAPIC)
|
|
return gsi;
|
|
|
|
/* Don't set up the ACPI SCI because it's already set up */
|
|
if (acpi_gbl_FADT.sci_interrupt == gsi)
|
|
return gsi;
|
|
|
|
irq = map_gsi_to_irq(gsi, IOAPIC_MAP_ALLOC);
|
|
if (irq < 0)
|
|
return irq;
|
|
|
|
ioapic = mp_find_ioapic(gsi);
|
|
if (ioapic < 0) {
|
|
printk(KERN_WARNING "No IOAPIC for GSI %u\n", gsi);
|
|
return gsi;
|
|
}
|
|
|
|
ioapic_pin = mp_find_ioapic_pin(ioapic, gsi);
|
|
|
|
if (ioapic_pin > MP_MAX_IOAPIC_PIN) {
|
|
printk(KERN_ERR "Invalid reference to IOAPIC pin "
|
|
"%d-%d\n", mpc_ioapic_id(ioapic),
|
|
ioapic_pin);
|
|
return gsi;
|
|
}
|
|
|
|
if (enable_update_mptable)
|
|
mp_config_acpi_gsi(dev, gsi, trigger, polarity);
|
|
|
|
set_io_apic_irq_attr(&irq_attr, ioapic, ioapic_pin,
|
|
trigger == ACPI_EDGE_SENSITIVE ? 0 : 1,
|
|
polarity == ACPI_ACTIVE_HIGH ? 0 : 1);
|
|
ret = io_apic_set_pci_routing(dev, irq, &irq_attr);
|
|
if (ret < 0)
|
|
irq = -1;
|
|
|
|
return irq;
|
|
}
|
|
|
|
static int __init
|
|
acpi_parse_ioapic(struct acpi_subtable_header * header, const unsigned long end)
|
|
{
|
|
struct acpi_madt_io_apic *ioapic = NULL;
|
|
|
|
ioapic = (struct acpi_madt_io_apic *)header;
|
|
|
|
if (BAD_MADT_ENTRY(ioapic, end))
|
|
return -EINVAL;
|
|
|
|
acpi_table_print_madt_entry(header);
|
|
|
|
mp_register_ioapic(ioapic->id,
|
|
ioapic->address, ioapic->global_irq_base);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Parse Interrupt Source Override for the ACPI SCI
|
|
*/
|
|
static void __init acpi_sci_ioapic_setup(u8 bus_irq, u16 polarity, u16 trigger, u32 gsi)
|
|
{
|
|
if (trigger == 0) /* compatible SCI trigger is level */
|
|
trigger = 3;
|
|
|
|
if (polarity == 0) /* compatible SCI polarity is low */
|
|
polarity = 3;
|
|
|
|
/* Command-line over-ride via acpi_sci= */
|
|
if (acpi_sci_flags & ACPI_MADT_TRIGGER_MASK)
|
|
trigger = (acpi_sci_flags & ACPI_MADT_TRIGGER_MASK) >> 2;
|
|
|
|
if (acpi_sci_flags & ACPI_MADT_POLARITY_MASK)
|
|
polarity = acpi_sci_flags & ACPI_MADT_POLARITY_MASK;
|
|
|
|
mp_override_legacy_irq(bus_irq, polarity, trigger, gsi);
|
|
|
|
/*
|
|
* stash over-ride to indicate we've been here
|
|
* and for later update of acpi_gbl_FADT
|
|
*/
|
|
acpi_sci_override_gsi = gsi;
|
|
return;
|
|
}
|
|
|
|
static int __init
|
|
acpi_parse_int_src_ovr(struct acpi_subtable_header * header,
|
|
const unsigned long end)
|
|
{
|
|
struct acpi_madt_interrupt_override *intsrc = NULL;
|
|
|
|
intsrc = (struct acpi_madt_interrupt_override *)header;
|
|
|
|
if (BAD_MADT_ENTRY(intsrc, end))
|
|
return -EINVAL;
|
|
|
|
acpi_table_print_madt_entry(header);
|
|
|
|
if (intsrc->source_irq == acpi_gbl_FADT.sci_interrupt) {
|
|
acpi_sci_ioapic_setup(intsrc->source_irq,
|
|
intsrc->inti_flags & ACPI_MADT_POLARITY_MASK,
|
|
(intsrc->inti_flags & ACPI_MADT_TRIGGER_MASK) >> 2,
|
|
intsrc->global_irq);
|
|
return 0;
|
|
}
|
|
|
|
if (intsrc->source_irq == 0) {
|
|
if (acpi_skip_timer_override) {
|
|
printk(PREFIX "BIOS IRQ0 override ignored.\n");
|
|
return 0;
|
|
}
|
|
|
|
if ((intsrc->global_irq == 2) && acpi_fix_pin2_polarity
|
|
&& (intsrc->inti_flags & ACPI_MADT_POLARITY_MASK)) {
|
|
intsrc->inti_flags &= ~ACPI_MADT_POLARITY_MASK;
|
|
printk(PREFIX "BIOS IRQ0 pin2 override: forcing polarity to high active.\n");
|
|
}
|
|
}
|
|
|
|
mp_override_legacy_irq(intsrc->source_irq,
|
|
intsrc->inti_flags & ACPI_MADT_POLARITY_MASK,
|
|
(intsrc->inti_flags & ACPI_MADT_TRIGGER_MASK) >> 2,
|
|
intsrc->global_irq);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init
|
|
acpi_parse_nmi_src(struct acpi_subtable_header * header, const unsigned long end)
|
|
{
|
|
struct acpi_madt_nmi_source *nmi_src = NULL;
|
|
|
|
nmi_src = (struct acpi_madt_nmi_source *)header;
|
|
|
|
if (BAD_MADT_ENTRY(nmi_src, end))
|
|
return -EINVAL;
|
|
|
|
acpi_table_print_madt_entry(header);
|
|
|
|
/* TBD: Support nimsrc entries? */
|
|
|
|
return 0;
|
|
}
|
|
|
|
#endif /* CONFIG_X86_IO_APIC */
|
|
|
|
/*
|
|
* acpi_pic_sci_set_trigger()
|
|
*
|
|
* use ELCR to set PIC-mode trigger type for SCI
|
|
*
|
|
* If a PIC-mode SCI is not recognized or gives spurious IRQ7's
|
|
* it may require Edge Trigger -- use "acpi_sci=edge"
|
|
*
|
|
* Port 0x4d0-4d1 are ECLR1 and ECLR2, the Edge/Level Control Registers
|
|
* for the 8259 PIC. bit[n] = 1 means irq[n] is Level, otherwise Edge.
|
|
* ECLR1 is IRQs 0-7 (IRQ 0, 1, 2 must be 0)
|
|
* ECLR2 is IRQs 8-15 (IRQ 8, 13 must be 0)
|
|
*/
|
|
|
|
void __init acpi_pic_sci_set_trigger(unsigned int irq, u16 trigger)
|
|
{
|
|
unsigned int mask = 1 << irq;
|
|
unsigned int old, new;
|
|
|
|
/* Real old ELCR mask */
|
|
old = inb(0x4d0) | (inb(0x4d1) << 8);
|
|
|
|
/*
|
|
* If we use ACPI to set PCI IRQs, then we should clear ELCR
|
|
* since we will set it correctly as we enable the PCI irq
|
|
* routing.
|
|
*/
|
|
new = acpi_noirq ? old : 0;
|
|
|
|
/*
|
|
* Update SCI information in the ELCR, it isn't in the PCI
|
|
* routing tables..
|
|
*/
|
|
switch (trigger) {
|
|
case 1: /* Edge - clear */
|
|
new &= ~mask;
|
|
break;
|
|
case 3: /* Level - set */
|
|
new |= mask;
|
|
break;
|
|
}
|
|
|
|
if (old == new)
|
|
return;
|
|
|
|
printk(PREFIX "setting ELCR to %04x (from %04x)\n", new, old);
|
|
outb(new, 0x4d0);
|
|
outb(new >> 8, 0x4d1);
|
|
}
|
|
|
|
int acpi_gsi_to_irq(u32 gsi, unsigned int *irqp)
|
|
{
|
|
int irq = map_gsi_to_irq(gsi, IOAPIC_MAP_ALLOC | IOAPIC_MAP_CHECK);
|
|
|
|
if (irq >= 0) {
|
|
#ifdef CONFIG_X86_IO_APIC
|
|
if (acpi_irq_model == ACPI_IRQ_MODEL_IOAPIC)
|
|
setup_IO_APIC_irq_extra(gsi);
|
|
#endif
|
|
*irqp = irq;
|
|
return 0;
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
EXPORT_SYMBOL_GPL(acpi_gsi_to_irq);
|
|
|
|
int acpi_isa_irq_to_gsi(unsigned isa_irq, u32 *gsi)
|
|
{
|
|
if (isa_irq < nr_legacy_irqs() &&
|
|
isa_irq_to_gsi[isa_irq] != ACPI_INVALID_GSI) {
|
|
*gsi = isa_irq_to_gsi[isa_irq];
|
|
return 0;
|
|
}
|
|
|
|
return -1;
|
|
}
|
|
|
|
static int acpi_register_gsi_pic(struct device *dev, u32 gsi,
|
|
int trigger, int polarity)
|
|
{
|
|
#ifdef CONFIG_PCI
|
|
/*
|
|
* Make sure all (legacy) PCI IRQs are set as level-triggered.
|
|
*/
|
|
if (trigger == ACPI_LEVEL_SENSITIVE)
|
|
eisa_set_level_irq(gsi);
|
|
#endif
|
|
|
|
return gsi;
|
|
}
|
|
|
|
static int acpi_register_gsi_ioapic(struct device *dev, u32 gsi,
|
|
int trigger, int polarity)
|
|
{
|
|
int irq = gsi;
|
|
|
|
#ifdef CONFIG_X86_IO_APIC
|
|
irq = mp_register_gsi(dev, gsi, trigger, polarity);
|
|
#endif
|
|
|
|
return irq;
|
|
}
|
|
|
|
int (*__acpi_register_gsi)(struct device *dev, u32 gsi,
|
|
int trigger, int polarity) = acpi_register_gsi_pic;
|
|
|
|
#ifdef CONFIG_ACPI_SLEEP
|
|
int (*acpi_suspend_lowlevel)(void) = x86_acpi_suspend_lowlevel;
|
|
#else
|
|
int (*acpi_suspend_lowlevel)(void);
|
|
#endif
|
|
|
|
/*
|
|
* success: return IRQ number (>=0)
|
|
* failure: return < 0
|
|
*/
|
|
int acpi_register_gsi(struct device *dev, u32 gsi, int trigger, int polarity)
|
|
{
|
|
return __acpi_register_gsi(dev, gsi, trigger, polarity);
|
|
}
|
|
EXPORT_SYMBOL_GPL(acpi_register_gsi);
|
|
|
|
void acpi_unregister_gsi(u32 gsi)
|
|
{
|
|
}
|
|
EXPORT_SYMBOL_GPL(acpi_unregister_gsi);
|
|
|
|
static void __init acpi_set_irq_model_ioapic(void)
|
|
{
|
|
acpi_irq_model = ACPI_IRQ_MODEL_IOAPIC;
|
|
__acpi_register_gsi = acpi_register_gsi_ioapic;
|
|
acpi_ioapic = 1;
|
|
}
|
|
|
|
/*
|
|
* ACPI based hotplug support for CPU
|
|
*/
|
|
#ifdef CONFIG_ACPI_HOTPLUG_CPU
|
|
#include <acpi/processor.h>
|
|
|
|
static void acpi_map_cpu2node(acpi_handle handle, int cpu, int physid)
|
|
{
|
|
#ifdef CONFIG_ACPI_NUMA
|
|
int nid;
|
|
|
|
nid = acpi_get_node(handle);
|
|
if (nid != -1) {
|
|
set_apicid_to_node(physid, nid);
|
|
numa_set_node(cpu, nid);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
static int _acpi_map_lsapic(acpi_handle handle, int physid, int *pcpu)
|
|
{
|
|
int cpu;
|
|
|
|
cpu = acpi_register_lapic(physid, ACPI_MADT_ENABLED);
|
|
if (cpu < 0) {
|
|
pr_info(PREFIX "Unable to map lapic to logical cpu number\n");
|
|
return cpu;
|
|
}
|
|
|
|
acpi_processor_set_pdc(handle);
|
|
acpi_map_cpu2node(handle, cpu, physid);
|
|
|
|
*pcpu = cpu;
|
|
return 0;
|
|
}
|
|
|
|
/* wrapper to silence section mismatch warning */
|
|
int __ref acpi_map_lsapic(acpi_handle handle, int physid, int *pcpu)
|
|
{
|
|
return _acpi_map_lsapic(handle, physid, pcpu);
|
|
}
|
|
EXPORT_SYMBOL(acpi_map_lsapic);
|
|
|
|
int acpi_unmap_lsapic(int cpu)
|
|
{
|
|
#ifdef CONFIG_ACPI_NUMA
|
|
set_apicid_to_node(per_cpu(x86_cpu_to_apicid, cpu), NUMA_NO_NODE);
|
|
#endif
|
|
|
|
per_cpu(x86_cpu_to_apicid, cpu) = -1;
|
|
set_cpu_present(cpu, false);
|
|
num_processors--;
|
|
|
|
return (0);
|
|
}
|
|
|
|
EXPORT_SYMBOL(acpi_unmap_lsapic);
|
|
#endif /* CONFIG_ACPI_HOTPLUG_CPU */
|
|
|
|
int acpi_register_ioapic(acpi_handle handle, u64 phys_addr, u32 gsi_base)
|
|
{
|
|
/* TBD */
|
|
return -EINVAL;
|
|
}
|
|
|
|
EXPORT_SYMBOL(acpi_register_ioapic);
|
|
|
|
int acpi_unregister_ioapic(acpi_handle handle, u32 gsi_base)
|
|
{
|
|
/* TBD */
|
|
return -EINVAL;
|
|
}
|
|
|
|
EXPORT_SYMBOL(acpi_unregister_ioapic);
|
|
|
|
static int __init acpi_parse_sbf(struct acpi_table_header *table)
|
|
{
|
|
struct acpi_table_boot *sb;
|
|
|
|
sb = (struct acpi_table_boot *)table;
|
|
if (!sb) {
|
|
printk(KERN_WARNING PREFIX "Unable to map SBF\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
sbf_port = sb->cmos_index; /* Save CMOS port */
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_HPET_TIMER
|
|
#include <asm/hpet.h>
|
|
|
|
static struct resource *hpet_res __initdata;
|
|
|
|
static int __init acpi_parse_hpet(struct acpi_table_header *table)
|
|
{
|
|
struct acpi_table_hpet *hpet_tbl;
|
|
|
|
hpet_tbl = (struct acpi_table_hpet *)table;
|
|
if (!hpet_tbl) {
|
|
printk(KERN_WARNING PREFIX "Unable to map HPET\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
if (hpet_tbl->address.space_id != ACPI_SPACE_MEM) {
|
|
printk(KERN_WARNING PREFIX "HPET timers must be located in "
|
|
"memory.\n");
|
|
return -1;
|
|
}
|
|
|
|
hpet_address = hpet_tbl->address.address;
|
|
hpet_blockid = hpet_tbl->sequence;
|
|
|
|
/*
|
|
* Some broken BIOSes advertise HPET at 0x0. We really do not
|
|
* want to allocate a resource there.
|
|
*/
|
|
if (!hpet_address) {
|
|
printk(KERN_WARNING PREFIX
|
|
"HPET id: %#x base: %#lx is invalid\n",
|
|
hpet_tbl->id, hpet_address);
|
|
return 0;
|
|
}
|
|
#ifdef CONFIG_X86_64
|
|
/*
|
|
* Some even more broken BIOSes advertise HPET at
|
|
* 0xfed0000000000000 instead of 0xfed00000. Fix it up and add
|
|
* some noise:
|
|
*/
|
|
if (hpet_address == 0xfed0000000000000UL) {
|
|
if (!hpet_force_user) {
|
|
printk(KERN_WARNING PREFIX "HPET id: %#x "
|
|
"base: 0xfed0000000000000 is bogus\n "
|
|
"try hpet=force on the kernel command line to "
|
|
"fix it up to 0xfed00000.\n", hpet_tbl->id);
|
|
hpet_address = 0;
|
|
return 0;
|
|
}
|
|
printk(KERN_WARNING PREFIX
|
|
"HPET id: %#x base: 0xfed0000000000000 fixed up "
|
|
"to 0xfed00000.\n", hpet_tbl->id);
|
|
hpet_address >>= 32;
|
|
}
|
|
#endif
|
|
printk(KERN_INFO PREFIX "HPET id: %#x base: %#lx\n",
|
|
hpet_tbl->id, hpet_address);
|
|
|
|
/*
|
|
* Allocate and initialize the HPET firmware resource for adding into
|
|
* the resource tree during the lateinit timeframe.
|
|
*/
|
|
#define HPET_RESOURCE_NAME_SIZE 9
|
|
hpet_res = alloc_bootmem(sizeof(*hpet_res) + HPET_RESOURCE_NAME_SIZE);
|
|
|
|
hpet_res->name = (void *)&hpet_res[1];
|
|
hpet_res->flags = IORESOURCE_MEM;
|
|
snprintf((char *)hpet_res->name, HPET_RESOURCE_NAME_SIZE, "HPET %u",
|
|
hpet_tbl->sequence);
|
|
|
|
hpet_res->start = hpet_address;
|
|
hpet_res->end = hpet_address + (1 * 1024) - 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* hpet_insert_resource inserts the HPET resources used into the resource
|
|
* tree.
|
|
*/
|
|
static __init int hpet_insert_resource(void)
|
|
{
|
|
if (!hpet_res)
|
|
return 1;
|
|
|
|
return insert_resource(&iomem_resource, hpet_res);
|
|
}
|
|
|
|
late_initcall(hpet_insert_resource);
|
|
|
|
#else
|
|
#define acpi_parse_hpet NULL
|
|
#endif
|
|
|
|
static int __init acpi_parse_fadt(struct acpi_table_header *table)
|
|
{
|
|
|
|
#ifdef CONFIG_X86_PM_TIMER
|
|
/* detect the location of the ACPI PM Timer */
|
|
if (acpi_gbl_FADT.header.revision >= FADT2_REVISION_ID) {
|
|
/* FADT rev. 2 */
|
|
if (acpi_gbl_FADT.xpm_timer_block.space_id !=
|
|
ACPI_ADR_SPACE_SYSTEM_IO)
|
|
return 0;
|
|
|
|
pmtmr_ioport = acpi_gbl_FADT.xpm_timer_block.address;
|
|
/*
|
|
* "X" fields are optional extensions to the original V1.0
|
|
* fields, so we must selectively expand V1.0 fields if the
|
|
* corresponding X field is zero.
|
|
*/
|
|
if (!pmtmr_ioport)
|
|
pmtmr_ioport = acpi_gbl_FADT.pm_timer_block;
|
|
} else {
|
|
/* FADT rev. 1 */
|
|
pmtmr_ioport = acpi_gbl_FADT.pm_timer_block;
|
|
}
|
|
if (pmtmr_ioport)
|
|
printk(KERN_INFO PREFIX "PM-Timer IO Port: %#x\n",
|
|
pmtmr_ioport);
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_X86_LOCAL_APIC
|
|
/*
|
|
* Parse LAPIC entries in MADT
|
|
* returns 0 on success, < 0 on error
|
|
*/
|
|
|
|
static int __init early_acpi_parse_madt_lapic_addr_ovr(void)
|
|
{
|
|
int count;
|
|
|
|
if (!cpu_has_apic)
|
|
return -ENODEV;
|
|
|
|
/*
|
|
* Note that the LAPIC address is obtained from the MADT (32-bit value)
|
|
* and (optionally) overriden by a LAPIC_ADDR_OVR entry (64-bit value).
|
|
*/
|
|
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_LOCAL_APIC_OVERRIDE,
|
|
acpi_parse_lapic_addr_ovr, 0);
|
|
if (count < 0) {
|
|
printk(KERN_ERR PREFIX
|
|
"Error parsing LAPIC address override entry\n");
|
|
return count;
|
|
}
|
|
|
|
register_lapic_address(acpi_lapic_addr);
|
|
|
|
return count;
|
|
}
|
|
|
|
static int __init acpi_parse_madt_lapic_entries(void)
|
|
{
|
|
int count;
|
|
int x2count = 0;
|
|
|
|
if (!cpu_has_apic)
|
|
return -ENODEV;
|
|
|
|
/*
|
|
* Note that the LAPIC address is obtained from the MADT (32-bit value)
|
|
* and (optionally) overriden by a LAPIC_ADDR_OVR entry (64-bit value).
|
|
*/
|
|
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_LOCAL_APIC_OVERRIDE,
|
|
acpi_parse_lapic_addr_ovr, 0);
|
|
if (count < 0) {
|
|
printk(KERN_ERR PREFIX
|
|
"Error parsing LAPIC address override entry\n");
|
|
return count;
|
|
}
|
|
|
|
register_lapic_address(acpi_lapic_addr);
|
|
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_LOCAL_SAPIC,
|
|
acpi_parse_sapic, MAX_LOCAL_APIC);
|
|
|
|
if (!count) {
|
|
x2count = acpi_table_parse_madt(ACPI_MADT_TYPE_LOCAL_X2APIC,
|
|
acpi_parse_x2apic, MAX_LOCAL_APIC);
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_LOCAL_APIC,
|
|
acpi_parse_lapic, MAX_LOCAL_APIC);
|
|
}
|
|
if (!count && !x2count) {
|
|
printk(KERN_ERR PREFIX "No LAPIC entries present\n");
|
|
/* TBD: Cleanup to allow fallback to MPS */
|
|
return -ENODEV;
|
|
} else if (count < 0 || x2count < 0) {
|
|
printk(KERN_ERR PREFIX "Error parsing LAPIC entry\n");
|
|
/* TBD: Cleanup to allow fallback to MPS */
|
|
return count;
|
|
}
|
|
|
|
x2count = acpi_table_parse_madt(ACPI_MADT_TYPE_LOCAL_X2APIC_NMI,
|
|
acpi_parse_x2apic_nmi, 0);
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_LOCAL_APIC_NMI,
|
|
acpi_parse_lapic_nmi, 0);
|
|
if (count < 0 || x2count < 0) {
|
|
printk(KERN_ERR PREFIX "Error parsing LAPIC NMI entry\n");
|
|
/* TBD: Cleanup to allow fallback to MPS */
|
|
return count;
|
|
}
|
|
return 0;
|
|
}
|
|
#endif /* CONFIG_X86_LOCAL_APIC */
|
|
|
|
#ifdef CONFIG_X86_IO_APIC
|
|
static void __init mp_config_acpi_legacy_irqs(void)
|
|
{
|
|
int i;
|
|
struct mpc_intsrc mp_irq;
|
|
|
|
#ifdef CONFIG_EISA
|
|
/*
|
|
* Fabricate the legacy ISA bus (bus #31).
|
|
*/
|
|
mp_bus_id_to_type[MP_ISA_BUS] = MP_BUS_ISA;
|
|
#endif
|
|
set_bit(MP_ISA_BUS, mp_bus_not_pci);
|
|
pr_debug("Bus #%d is ISA\n", MP_ISA_BUS);
|
|
|
|
/*
|
|
* Use the default configuration for the IRQs 0-15. Unless
|
|
* overridden by (MADT) interrupt source override entries.
|
|
*/
|
|
for (i = 0; i < nr_legacy_irqs(); i++) {
|
|
int ioapic, pin;
|
|
unsigned int dstapic;
|
|
int idx;
|
|
u32 gsi;
|
|
|
|
/* Locate the gsi that irq i maps to. */
|
|
if (acpi_isa_irq_to_gsi(i, &gsi))
|
|
continue;
|
|
|
|
/*
|
|
* Locate the IOAPIC that manages the ISA IRQ.
|
|
*/
|
|
ioapic = mp_find_ioapic(gsi);
|
|
if (ioapic < 0)
|
|
continue;
|
|
pin = mp_find_ioapic_pin(ioapic, gsi);
|
|
dstapic = mpc_ioapic_id(ioapic);
|
|
|
|
for (idx = 0; idx < mp_irq_entries; idx++) {
|
|
struct mpc_intsrc *irq = mp_irqs + idx;
|
|
|
|
/* Do we already have a mapping for this ISA IRQ? */
|
|
if (irq->srcbus == MP_ISA_BUS && irq->srcbusirq == i)
|
|
break;
|
|
|
|
/* Do we already have a mapping for this IOAPIC pin */
|
|
if (irq->dstapic == dstapic && irq->dstirq == pin)
|
|
break;
|
|
}
|
|
|
|
if (idx != mp_irq_entries) {
|
|
printk(KERN_DEBUG "ACPI: IRQ%d used by override.\n", i);
|
|
continue; /* IRQ already used */
|
|
}
|
|
|
|
mp_irq.type = MP_INTSRC;
|
|
mp_irq.irqflag = 0; /* Conforming */
|
|
mp_irq.srcbus = MP_ISA_BUS;
|
|
mp_irq.dstapic = dstapic;
|
|
mp_irq.irqtype = mp_INT;
|
|
mp_irq.srcbusirq = i; /* Identity mapped */
|
|
mp_irq.dstirq = pin;
|
|
|
|
mp_save_irq(&mp_irq);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Parse IOAPIC related entries in MADT
|
|
* returns 0 on success, < 0 on error
|
|
*/
|
|
static int __init acpi_parse_madt_ioapic_entries(void)
|
|
{
|
|
int count;
|
|
|
|
/*
|
|
* ACPI interpreter is required to complete interrupt setup,
|
|
* so if it is off, don't enumerate the io-apics with ACPI.
|
|
* If MPS is present, it will handle them,
|
|
* otherwise the system will stay in PIC mode
|
|
*/
|
|
if (acpi_disabled || acpi_noirq)
|
|
return -ENODEV;
|
|
|
|
if (!cpu_has_apic)
|
|
return -ENODEV;
|
|
|
|
/*
|
|
* if "noapic" boot option, don't look for IO-APICs
|
|
*/
|
|
if (skip_ioapic_setup) {
|
|
printk(KERN_INFO PREFIX "Skipping IOAPIC probe "
|
|
"due to 'noapic' option.\n");
|
|
return -ENODEV;
|
|
}
|
|
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_IO_APIC, acpi_parse_ioapic,
|
|
MAX_IO_APICS);
|
|
if (!count) {
|
|
printk(KERN_ERR PREFIX "No IOAPIC entries present\n");
|
|
return -ENODEV;
|
|
} else if (count < 0) {
|
|
printk(KERN_ERR PREFIX "Error parsing IOAPIC entry\n");
|
|
return count;
|
|
}
|
|
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_INTERRUPT_OVERRIDE,
|
|
acpi_parse_int_src_ovr, nr_irqs);
|
|
if (count < 0) {
|
|
printk(KERN_ERR PREFIX
|
|
"Error parsing interrupt source overrides entry\n");
|
|
/* TBD: Cleanup to allow fallback to MPS */
|
|
return count;
|
|
}
|
|
|
|
/*
|
|
* If BIOS did not supply an INT_SRC_OVR for the SCI
|
|
* pretend we got one so we can set the SCI flags.
|
|
*/
|
|
if (!acpi_sci_override_gsi)
|
|
acpi_sci_ioapic_setup(acpi_gbl_FADT.sci_interrupt, 0, 0,
|
|
acpi_gbl_FADT.sci_interrupt);
|
|
|
|
/* Fill in identity legacy mappings where no override */
|
|
mp_config_acpi_legacy_irqs();
|
|
|
|
count = acpi_table_parse_madt(ACPI_MADT_TYPE_NMI_SOURCE,
|
|
acpi_parse_nmi_src, nr_irqs);
|
|
if (count < 0) {
|
|
printk(KERN_ERR PREFIX "Error parsing NMI SRC entry\n");
|
|
/* TBD: Cleanup to allow fallback to MPS */
|
|
return count;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
#else
|
|
static inline int acpi_parse_madt_ioapic_entries(void)
|
|
{
|
|
return -1;
|
|
}
|
|
#endif /* !CONFIG_X86_IO_APIC */
|
|
|
|
static void __init early_acpi_process_madt(void)
|
|
{
|
|
#ifdef CONFIG_X86_LOCAL_APIC
|
|
int error;
|
|
|
|
if (!acpi_table_parse(ACPI_SIG_MADT, acpi_parse_madt)) {
|
|
|
|
/*
|
|
* Parse MADT LAPIC entries
|
|
*/
|
|
error = early_acpi_parse_madt_lapic_addr_ovr();
|
|
if (!error) {
|
|
acpi_lapic = 1;
|
|
smp_found_config = 1;
|
|
}
|
|
if (error == -EINVAL) {
|
|
/*
|
|
* Dell Precision Workstation 410, 610 come here.
|
|
*/
|
|
printk(KERN_ERR PREFIX
|
|
"Invalid BIOS MADT, disabling ACPI\n");
|
|
disable_acpi();
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
|
|
static void __init acpi_process_madt(void)
|
|
{
|
|
#ifdef CONFIG_X86_LOCAL_APIC
|
|
int error;
|
|
|
|
if (!acpi_table_parse(ACPI_SIG_MADT, acpi_parse_madt)) {
|
|
|
|
/*
|
|
* Parse MADT LAPIC entries
|
|
*/
|
|
error = acpi_parse_madt_lapic_entries();
|
|
if (!error) {
|
|
acpi_lapic = 1;
|
|
|
|
/*
|
|
* Parse MADT IO-APIC entries
|
|
*/
|
|
error = acpi_parse_madt_ioapic_entries();
|
|
if (!error) {
|
|
acpi_set_irq_model_ioapic();
|
|
|
|
smp_found_config = 1;
|
|
}
|
|
}
|
|
if (error == -EINVAL) {
|
|
/*
|
|
* Dell Precision Workstation 410, 610 come here.
|
|
*/
|
|
printk(KERN_ERR PREFIX
|
|
"Invalid BIOS MADT, disabling ACPI\n");
|
|
disable_acpi();
|
|
}
|
|
} else {
|
|
/*
|
|
* ACPI found no MADT, and so ACPI wants UP PIC mode.
|
|
* In the event an MPS table was found, forget it.
|
|
* Boot with "acpi=off" to use MPS on such a system.
|
|
*/
|
|
if (smp_found_config) {
|
|
printk(KERN_WARNING PREFIX
|
|
"No APIC-table, disabling MPS\n");
|
|
smp_found_config = 0;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* ACPI supports both logical (e.g. Hyper-Threading) and physical
|
|
* processors, where MPS only supports physical.
|
|
*/
|
|
if (acpi_lapic && acpi_ioapic)
|
|
printk(KERN_INFO "Using ACPI (MADT) for SMP configuration "
|
|
"information\n");
|
|
else if (acpi_lapic)
|
|
printk(KERN_INFO "Using ACPI for processor (LAPIC) "
|
|
"configuration information\n");
|
|
#endif
|
|
return;
|
|
}
|
|
|
|
static int __init disable_acpi_irq(const struct dmi_system_id *d)
|
|
{
|
|
if (!acpi_force) {
|
|
printk(KERN_NOTICE "%s detected: force use of acpi=noirq\n",
|
|
d->ident);
|
|
acpi_noirq_set();
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int __init disable_acpi_pci(const struct dmi_system_id *d)
|
|
{
|
|
if (!acpi_force) {
|
|
printk(KERN_NOTICE "%s detected: force use of pci=noacpi\n",
|
|
d->ident);
|
|
acpi_disable_pci();
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int __init dmi_disable_acpi(const struct dmi_system_id *d)
|
|
{
|
|
if (!acpi_force) {
|
|
printk(KERN_NOTICE "%s detected: acpi off\n", d->ident);
|
|
disable_acpi();
|
|
} else {
|
|
printk(KERN_NOTICE
|
|
"Warning: DMI blacklist says broken, but acpi forced\n");
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Force ignoring BIOS IRQ0 override
|
|
*/
|
|
static int __init dmi_ignore_irq0_timer_override(const struct dmi_system_id *d)
|
|
{
|
|
if (!acpi_skip_timer_override) {
|
|
pr_notice("%s detected: Ignoring BIOS IRQ0 override\n",
|
|
d->ident);
|
|
acpi_skip_timer_override = 1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* If your system is blacklisted here, but you find that acpi=force
|
|
* works for you, please contact linux-acpi@vger.kernel.org
|
|
*/
|
|
static struct dmi_system_id __initdata acpi_dmi_table[] = {
|
|
/*
|
|
* Boxes that need ACPI disabled
|
|
*/
|
|
{
|
|
.callback = dmi_disable_acpi,
|
|
.ident = "IBM Thinkpad",
|
|
.matches = {
|
|
DMI_MATCH(DMI_BOARD_VENDOR, "IBM"),
|
|
DMI_MATCH(DMI_BOARD_NAME, "2629H1G"),
|
|
},
|
|
},
|
|
|
|
/*
|
|
* Boxes that need ACPI PCI IRQ routing disabled
|
|
*/
|
|
{
|
|
.callback = disable_acpi_irq,
|
|
.ident = "ASUS A7V",
|
|
.matches = {
|
|
DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTeK Computer INC"),
|
|
DMI_MATCH(DMI_BOARD_NAME, "<A7V>"),
|
|
/* newer BIOS, Revision 1011, does work */
|
|
DMI_MATCH(DMI_BIOS_VERSION,
|
|
"ASUS A7V ACPI BIOS Revision 1007"),
|
|
},
|
|
},
|
|
{
|
|
/*
|
|
* Latest BIOS for IBM 600E (1.16) has bad pcinum
|
|
* for LPC bridge, which is needed for the PCI
|
|
* interrupt links to work. DSDT fix is in bug 5966.
|
|
* 2645, 2646 model numbers are shared with 600/600E/600X
|
|
*/
|
|
.callback = disable_acpi_irq,
|
|
.ident = "IBM Thinkpad 600 Series 2645",
|
|
.matches = {
|
|
DMI_MATCH(DMI_BOARD_VENDOR, "IBM"),
|
|
DMI_MATCH(DMI_BOARD_NAME, "2645"),
|
|
},
|
|
},
|
|
{
|
|
.callback = disable_acpi_irq,
|
|
.ident = "IBM Thinkpad 600 Series 2646",
|
|
.matches = {
|
|
DMI_MATCH(DMI_BOARD_VENDOR, "IBM"),
|
|
DMI_MATCH(DMI_BOARD_NAME, "2646"),
|
|
},
|
|
},
|
|
/*
|
|
* Boxes that need ACPI PCI IRQ routing and PCI scan disabled
|
|
*/
|
|
{ /* _BBN 0 bug */
|
|
.callback = disable_acpi_pci,
|
|
.ident = "ASUS PR-DLS",
|
|
.matches = {
|
|
DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTeK Computer INC."),
|
|
DMI_MATCH(DMI_BOARD_NAME, "PR-DLS"),
|
|
DMI_MATCH(DMI_BIOS_VERSION,
|
|
"ASUS PR-DLS ACPI BIOS Revision 1010"),
|
|
DMI_MATCH(DMI_BIOS_DATE, "03/21/2003")
|
|
},
|
|
},
|
|
{
|
|
.callback = disable_acpi_pci,
|
|
.ident = "Acer TravelMate 36x Laptop",
|
|
.matches = {
|
|
DMI_MATCH(DMI_SYS_VENDOR, "Acer"),
|
|
DMI_MATCH(DMI_PRODUCT_NAME, "TravelMate 360"),
|
|
},
|
|
},
|
|
{}
|
|
};
|
|
|
|
/* second table for DMI checks that should run after early-quirks */
|
|
static struct dmi_system_id __initdata acpi_dmi_table_late[] = {
|
|
/*
|
|
* HP laptops which use a DSDT reporting as HP/SB400/10000,
|
|
* which includes some code which overrides all temperature
|
|
* trip points to 16C if the INTIN2 input of the I/O APIC
|
|
* is enabled. This input is incorrectly designated the
|
|
* ISA IRQ 0 via an interrupt source override even though
|
|
* it is wired to the output of the master 8259A and INTIN0
|
|
* is not connected at all. Force ignoring BIOS IRQ0
|
|
* override in that cases.
|
|
*/
|
|
{
|
|
.callback = dmi_ignore_irq0_timer_override,
|
|
.ident = "HP nx6115 laptop",
|
|
.matches = {
|
|
DMI_MATCH(DMI_SYS_VENDOR, "Hewlett-Packard"),
|
|
DMI_MATCH(DMI_PRODUCT_NAME, "HP Compaq nx6115"),
|
|
},
|
|
},
|
|
{
|
|
.callback = dmi_ignore_irq0_timer_override,
|
|
.ident = "HP NX6125 laptop",
|
|
.matches = {
|
|
DMI_MATCH(DMI_SYS_VENDOR, "Hewlett-Packard"),
|
|
DMI_MATCH(DMI_PRODUCT_NAME, "HP Compaq nx6125"),
|
|
},
|
|
},
|
|
{
|
|
.callback = dmi_ignore_irq0_timer_override,
|
|
.ident = "HP NX6325 laptop",
|
|
.matches = {
|
|
DMI_MATCH(DMI_SYS_VENDOR, "Hewlett-Packard"),
|
|
DMI_MATCH(DMI_PRODUCT_NAME, "HP Compaq nx6325"),
|
|
},
|
|
},
|
|
{
|
|
.callback = dmi_ignore_irq0_timer_override,
|
|
.ident = "HP 6715b laptop",
|
|
.matches = {
|
|
DMI_MATCH(DMI_SYS_VENDOR, "Hewlett-Packard"),
|
|
DMI_MATCH(DMI_PRODUCT_NAME, "HP Compaq 6715b"),
|
|
},
|
|
},
|
|
{
|
|
.callback = dmi_ignore_irq0_timer_override,
|
|
.ident = "FUJITSU SIEMENS",
|
|
.matches = {
|
|
DMI_MATCH(DMI_SYS_VENDOR, "FUJITSU SIEMENS"),
|
|
DMI_MATCH(DMI_PRODUCT_NAME, "AMILO PRO V2030"),
|
|
},
|
|
},
|
|
{}
|
|
};
|
|
|
|
/*
|
|
* acpi_boot_table_init() and acpi_boot_init()
|
|
* called from setup_arch(), always.
|
|
* 1. checksums all tables
|
|
* 2. enumerates lapics
|
|
* 3. enumerates io-apics
|
|
*
|
|
* acpi_table_init() is separate to allow reading SRAT without
|
|
* other side effects.
|
|
*
|
|
* side effects of acpi_boot_init:
|
|
* acpi_lapic = 1 if LAPIC found
|
|
* acpi_ioapic = 1 if IOAPIC found
|
|
* if (acpi_lapic && acpi_ioapic) smp_found_config = 1;
|
|
* if acpi_blacklisted() acpi_disabled = 1;
|
|
* acpi_irq_model=...
|
|
* ...
|
|
*/
|
|
|
|
void __init acpi_boot_table_init(void)
|
|
{
|
|
dmi_check_system(acpi_dmi_table);
|
|
|
|
/*
|
|
* If acpi_disabled, bail out
|
|
*/
|
|
if (acpi_disabled)
|
|
return;
|
|
|
|
/*
|
|
* Initialize the ACPI boot-time table parser.
|
|
*/
|
|
if (acpi_table_init()) {
|
|
disable_acpi();
|
|
return;
|
|
}
|
|
|
|
acpi_table_parse(ACPI_SIG_BOOT, acpi_parse_sbf);
|
|
|
|
/*
|
|
* blacklist may disable ACPI entirely
|
|
*/
|
|
if (acpi_blacklisted()) {
|
|
if (acpi_force) {
|
|
printk(KERN_WARNING PREFIX "acpi=force override\n");
|
|
} else {
|
|
printk(KERN_WARNING PREFIX "Disabling ACPI support\n");
|
|
disable_acpi();
|
|
return;
|
|
}
|
|
}
|
|
}
|
|
|
|
int __init early_acpi_boot_init(void)
|
|
{
|
|
/*
|
|
* If acpi_disabled, bail out
|
|
*/
|
|
if (acpi_disabled)
|
|
return 1;
|
|
|
|
/*
|
|
* Process the Multiple APIC Description Table (MADT), if present
|
|
*/
|
|
early_acpi_process_madt();
|
|
|
|
return 0;
|
|
}
|
|
|
|
int __init acpi_boot_init(void)
|
|
{
|
|
/* those are executed after early-quirks are executed */
|
|
dmi_check_system(acpi_dmi_table_late);
|
|
|
|
/*
|
|
* If acpi_disabled, bail out
|
|
*/
|
|
if (acpi_disabled)
|
|
return 1;
|
|
|
|
acpi_table_parse(ACPI_SIG_BOOT, acpi_parse_sbf);
|
|
|
|
/*
|
|
* set sci_int and PM timer address
|
|
*/
|
|
acpi_table_parse(ACPI_SIG_FADT, acpi_parse_fadt);
|
|
|
|
/*
|
|
* Process the Multiple APIC Description Table (MADT), if present
|
|
*/
|
|
acpi_process_madt();
|
|
|
|
acpi_table_parse(ACPI_SIG_HPET, acpi_parse_hpet);
|
|
|
|
if (!acpi_noirq)
|
|
x86_init.pci.init = pci_acpi_init;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int __init parse_acpi(char *arg)
|
|
{
|
|
if (!arg)
|
|
return -EINVAL;
|
|
|
|
/* "acpi=off" disables both ACPI table parsing and interpreter */
|
|
if (strcmp(arg, "off") == 0) {
|
|
disable_acpi();
|
|
}
|
|
/* acpi=force to over-ride black-list */
|
|
else if (strcmp(arg, "force") == 0) {
|
|
acpi_force = 1;
|
|
acpi_disabled = 0;
|
|
}
|
|
/* acpi=strict disables out-of-spec workarounds */
|
|
else if (strcmp(arg, "strict") == 0) {
|
|
acpi_strict = 1;
|
|
}
|
|
/* acpi=rsdt use RSDT instead of XSDT */
|
|
else if (strcmp(arg, "rsdt") == 0) {
|
|
acpi_gbl_do_not_use_xsdt = TRUE;
|
|
}
|
|
/* "acpi=noirq" disables ACPI interrupt routing */
|
|
else if (strcmp(arg, "noirq") == 0) {
|
|
acpi_noirq_set();
|
|
}
|
|
/* "acpi=copy_dsdt" copys DSDT */
|
|
else if (strcmp(arg, "copy_dsdt") == 0) {
|
|
acpi_gbl_copy_dsdt_locally = 1;
|
|
}
|
|
/* "acpi=nocmcff" disables FF mode for corrected errors */
|
|
else if (strcmp(arg, "nocmcff") == 0) {
|
|
acpi_disable_cmcff = 1;
|
|
} else {
|
|
/* Core will printk when we return error. */
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
early_param("acpi", parse_acpi);
|
|
|
|
/* FIXME: Using pci= for an ACPI parameter is a travesty. */
|
|
static int __init parse_pci(char *arg)
|
|
{
|
|
if (arg && strcmp(arg, "noacpi") == 0)
|
|
acpi_disable_pci();
|
|
return 0;
|
|
}
|
|
early_param("pci", parse_pci);
|
|
|
|
int __init acpi_mps_check(void)
|
|
{
|
|
#if defined(CONFIG_X86_LOCAL_APIC) && !defined(CONFIG_X86_MPPARSE)
|
|
/* mptable code is not built-in*/
|
|
if (acpi_disabled || acpi_noirq) {
|
|
printk(KERN_WARNING "MPS support code is not built-in.\n"
|
|
"Using acpi=off or acpi=noirq or pci=noacpi "
|
|
"may have problem\n");
|
|
return 1;
|
|
}
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_X86_IO_APIC
|
|
static int __init parse_acpi_skip_timer_override(char *arg)
|
|
{
|
|
acpi_skip_timer_override = 1;
|
|
return 0;
|
|
}
|
|
early_param("acpi_skip_timer_override", parse_acpi_skip_timer_override);
|
|
|
|
static int __init parse_acpi_use_timer_override(char *arg)
|
|
{
|
|
acpi_use_timer_override = 1;
|
|
return 0;
|
|
}
|
|
early_param("acpi_use_timer_override", parse_acpi_use_timer_override);
|
|
#endif /* CONFIG_X86_IO_APIC */
|
|
|
|
static int __init setup_acpi_sci(char *s)
|
|
{
|
|
if (!s)
|
|
return -EINVAL;
|
|
if (!strcmp(s, "edge"))
|
|
acpi_sci_flags = ACPI_MADT_TRIGGER_EDGE |
|
|
(acpi_sci_flags & ~ACPI_MADT_TRIGGER_MASK);
|
|
else if (!strcmp(s, "level"))
|
|
acpi_sci_flags = ACPI_MADT_TRIGGER_LEVEL |
|
|
(acpi_sci_flags & ~ACPI_MADT_TRIGGER_MASK);
|
|
else if (!strcmp(s, "high"))
|
|
acpi_sci_flags = ACPI_MADT_POLARITY_ACTIVE_HIGH |
|
|
(acpi_sci_flags & ~ACPI_MADT_POLARITY_MASK);
|
|
else if (!strcmp(s, "low"))
|
|
acpi_sci_flags = ACPI_MADT_POLARITY_ACTIVE_LOW |
|
|
(acpi_sci_flags & ~ACPI_MADT_POLARITY_MASK);
|
|
else
|
|
return -EINVAL;
|
|
return 0;
|
|
}
|
|
early_param("acpi_sci", setup_acpi_sci);
|
|
|
|
int __acpi_acquire_global_lock(unsigned int *lock)
|
|
{
|
|
unsigned int old, new, val;
|
|
do {
|
|
old = *lock;
|
|
new = (((old & ~0x3) + 2) + ((old >> 1) & 0x1));
|
|
val = cmpxchg(lock, old, new);
|
|
} while (unlikely (val != old));
|
|
return (new < 3) ? -1 : 0;
|
|
}
|
|
|
|
int __acpi_release_global_lock(unsigned int *lock)
|
|
{
|
|
unsigned int old, new, val;
|
|
do {
|
|
old = *lock;
|
|
new = old & ~0x3;
|
|
val = cmpxchg(lock, old, new);
|
|
} while (unlikely (val != old));
|
|
return old & 0x1;
|
|
}
|
|
|
|
void __init arch_reserve_mem_area(acpi_physical_address addr, size_t size)
|
|
{
|
|
e820_add_region(addr, size, E820_ACPI);
|
|
update_e820();
|
|
}
|