mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
e24442733e
The "powernv" platform of the powerpc architecture needs to assign PCI resources using a specific algorithm to fit some HW constraints of the IBM "IODA" architecture (related to the ability to create error handling domains that encompass specific segments of MMIO space). For doing so, it wants to call pci_setup_bridge() from architecture specific resource management in order to configure bridges after all resources have been assigned. So make it non-static. Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org> Signed-off-by: Jesse Barnes <jbarnes@virtuousgeek.org>
1270 lines
34 KiB
C
1270 lines
34 KiB
C
/*
|
|
* drivers/pci/setup-bus.c
|
|
*
|
|
* Extruded from code written by
|
|
* Dave Rusling (david.rusling@reo.mts.dec.com)
|
|
* David Mosberger (davidm@cs.arizona.edu)
|
|
* David Miller (davem@redhat.com)
|
|
*
|
|
* Support routines for initializing a PCI subsystem.
|
|
*/
|
|
|
|
/*
|
|
* Nov 2000, Ivan Kokshaysky <ink@jurassic.park.msu.ru>
|
|
* PCI-PCI bridges cleanup, sorted resource allocation.
|
|
* Feb 2002, Ivan Kokshaysky <ink@jurassic.park.msu.ru>
|
|
* Converted to allocation in 3 passes, which gives
|
|
* tighter packing. Prefetchable range support.
|
|
*/
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/ioport.h>
|
|
#include <linux/cache.h>
|
|
#include <linux/slab.h>
|
|
#include "pci.h"
|
|
|
|
struct resource_list_x {
|
|
struct resource_list_x *next;
|
|
struct resource *res;
|
|
struct pci_dev *dev;
|
|
resource_size_t start;
|
|
resource_size_t end;
|
|
resource_size_t add_size;
|
|
resource_size_t min_align;
|
|
unsigned long flags;
|
|
};
|
|
|
|
#define free_list(type, head) do { \
|
|
struct type *list, *tmp; \
|
|
for (list = (head)->next; list;) { \
|
|
tmp = list; \
|
|
list = list->next; \
|
|
kfree(tmp); \
|
|
} \
|
|
(head)->next = NULL; \
|
|
} while (0)
|
|
|
|
int pci_realloc_enable = 0;
|
|
#define pci_realloc_enabled() pci_realloc_enable
|
|
void pci_realloc(void)
|
|
{
|
|
pci_realloc_enable = 1;
|
|
}
|
|
|
|
/**
|
|
* add_to_list() - add a new resource tracker to the list
|
|
* @head: Head of the list
|
|
* @dev: device corresponding to which the resource
|
|
* belongs
|
|
* @res: The resource to be tracked
|
|
* @add_size: additional size to be optionally added
|
|
* to the resource
|
|
*/
|
|
static void add_to_list(struct resource_list_x *head,
|
|
struct pci_dev *dev, struct resource *res,
|
|
resource_size_t add_size, resource_size_t min_align)
|
|
{
|
|
struct resource_list_x *list = head;
|
|
struct resource_list_x *ln = list->next;
|
|
struct resource_list_x *tmp;
|
|
|
|
tmp = kmalloc(sizeof(*tmp), GFP_KERNEL);
|
|
if (!tmp) {
|
|
pr_warning("add_to_list: kmalloc() failed!\n");
|
|
return;
|
|
}
|
|
|
|
tmp->next = ln;
|
|
tmp->res = res;
|
|
tmp->dev = dev;
|
|
tmp->start = res->start;
|
|
tmp->end = res->end;
|
|
tmp->flags = res->flags;
|
|
tmp->add_size = add_size;
|
|
tmp->min_align = min_align;
|
|
list->next = tmp;
|
|
}
|
|
|
|
static void add_to_failed_list(struct resource_list_x *head,
|
|
struct pci_dev *dev, struct resource *res)
|
|
{
|
|
add_to_list(head, dev, res,
|
|
0 /* dont care */,
|
|
0 /* dont care */);
|
|
}
|
|
|
|
static void __dev_sort_resources(struct pci_dev *dev,
|
|
struct resource_list *head)
|
|
{
|
|
u16 class = dev->class >> 8;
|
|
|
|
/* Don't touch classless devices or host bridges or ioapics. */
|
|
if (class == PCI_CLASS_NOT_DEFINED || class == PCI_CLASS_BRIDGE_HOST)
|
|
return;
|
|
|
|
/* Don't touch ioapic devices already enabled by firmware */
|
|
if (class == PCI_CLASS_SYSTEM_PIC) {
|
|
u16 command;
|
|
pci_read_config_word(dev, PCI_COMMAND, &command);
|
|
if (command & (PCI_COMMAND_IO | PCI_COMMAND_MEMORY))
|
|
return;
|
|
}
|
|
|
|
pdev_sort_resources(dev, head);
|
|
}
|
|
|
|
static inline void reset_resource(struct resource *res)
|
|
{
|
|
res->start = 0;
|
|
res->end = 0;
|
|
res->flags = 0;
|
|
}
|
|
|
|
/**
|
|
* reassign_resources_sorted() - satisfy any additional resource requests
|
|
*
|
|
* @realloc_head : head of the list tracking requests requiring additional
|
|
* resources
|
|
* @head : head of the list tracking requests with allocated
|
|
* resources
|
|
*
|
|
* Walk through each element of the realloc_head and try to procure
|
|
* additional resources for the element, provided the element
|
|
* is in the head list.
|
|
*/
|
|
static void reassign_resources_sorted(struct resource_list_x *realloc_head,
|
|
struct resource_list *head)
|
|
{
|
|
struct resource *res;
|
|
struct resource_list_x *list, *tmp, *prev;
|
|
struct resource_list *hlist;
|
|
resource_size_t add_size;
|
|
int idx;
|
|
|
|
prev = realloc_head;
|
|
for (list = realloc_head->next; list;) {
|
|
res = list->res;
|
|
/* skip resource that has been reset */
|
|
if (!res->flags)
|
|
goto out;
|
|
|
|
/* skip this resource if not found in head list */
|
|
for (hlist = head->next; hlist && hlist->res != res;
|
|
hlist = hlist->next);
|
|
if (!hlist) { /* just skip */
|
|
prev = list;
|
|
list = list->next;
|
|
continue;
|
|
}
|
|
|
|
idx = res - &list->dev->resource[0];
|
|
add_size=list->add_size;
|
|
if (!resource_size(res)) {
|
|
res->start = list->start;
|
|
res->end = res->start + add_size - 1;
|
|
if(pci_assign_resource(list->dev, idx))
|
|
reset_resource(res);
|
|
} else {
|
|
resource_size_t align = list->min_align;
|
|
res->flags |= list->flags & (IORESOURCE_STARTALIGN|IORESOURCE_SIZEALIGN);
|
|
if (pci_reassign_resource(list->dev, idx, add_size, align))
|
|
dev_printk(KERN_DEBUG, &list->dev->dev, "failed to add optional resources res=%pR\n",
|
|
res);
|
|
}
|
|
out:
|
|
tmp = list;
|
|
prev->next = list = list->next;
|
|
kfree(tmp);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* assign_requested_resources_sorted() - satisfy resource requests
|
|
*
|
|
* @head : head of the list tracking requests for resources
|
|
* @failed_list : head of the list tracking requests that could
|
|
* not be allocated
|
|
*
|
|
* Satisfy resource requests of each element in the list. Add
|
|
* requests that could not satisfied to the failed_list.
|
|
*/
|
|
static void assign_requested_resources_sorted(struct resource_list *head,
|
|
struct resource_list_x *fail_head)
|
|
{
|
|
struct resource *res;
|
|
struct resource_list *list;
|
|
int idx;
|
|
|
|
for (list = head->next; list; list = list->next) {
|
|
res = list->res;
|
|
idx = res - &list->dev->resource[0];
|
|
if (resource_size(res) && pci_assign_resource(list->dev, idx)) {
|
|
if (fail_head && !pci_is_root_bus(list->dev->bus)) {
|
|
/*
|
|
* if the failed res is for ROM BAR, and it will
|
|
* be enabled later, don't add it to the list
|
|
*/
|
|
if (!((idx == PCI_ROM_RESOURCE) &&
|
|
(!(res->flags & IORESOURCE_ROM_ENABLE))))
|
|
add_to_failed_list(fail_head, list->dev, res);
|
|
}
|
|
reset_resource(res);
|
|
}
|
|
}
|
|
}
|
|
|
|
static void __assign_resources_sorted(struct resource_list *head,
|
|
struct resource_list_x *realloc_head,
|
|
struct resource_list_x *fail_head)
|
|
{
|
|
/* Satisfy the must-have resource requests */
|
|
assign_requested_resources_sorted(head, fail_head);
|
|
|
|
/* Try to satisfy any additional optional resource
|
|
requests */
|
|
if (realloc_head)
|
|
reassign_resources_sorted(realloc_head, head);
|
|
free_list(resource_list, head);
|
|
}
|
|
|
|
static void pdev_assign_resources_sorted(struct pci_dev *dev,
|
|
struct resource_list_x *fail_head)
|
|
{
|
|
struct resource_list head;
|
|
|
|
head.next = NULL;
|
|
__dev_sort_resources(dev, &head);
|
|
__assign_resources_sorted(&head, NULL, fail_head);
|
|
|
|
}
|
|
|
|
static void pbus_assign_resources_sorted(const struct pci_bus *bus,
|
|
struct resource_list_x *realloc_head,
|
|
struct resource_list_x *fail_head)
|
|
{
|
|
struct pci_dev *dev;
|
|
struct resource_list head;
|
|
|
|
head.next = NULL;
|
|
list_for_each_entry(dev, &bus->devices, bus_list)
|
|
__dev_sort_resources(dev, &head);
|
|
|
|
__assign_resources_sorted(&head, realloc_head, fail_head);
|
|
}
|
|
|
|
void pci_setup_cardbus(struct pci_bus *bus)
|
|
{
|
|
struct pci_dev *bridge = bus->self;
|
|
struct resource *res;
|
|
struct pci_bus_region region;
|
|
|
|
dev_info(&bridge->dev, "CardBus bridge to [bus %02x-%02x]\n",
|
|
bus->secondary, bus->subordinate);
|
|
|
|
res = bus->resource[0];
|
|
pcibios_resource_to_bus(bridge, ®ion, res);
|
|
if (res->flags & IORESOURCE_IO) {
|
|
/*
|
|
* The IO resource is allocated a range twice as large as it
|
|
* would normally need. This allows us to set both IO regs.
|
|
*/
|
|
dev_info(&bridge->dev, " bridge window %pR\n", res);
|
|
pci_write_config_dword(bridge, PCI_CB_IO_BASE_0,
|
|
region.start);
|
|
pci_write_config_dword(bridge, PCI_CB_IO_LIMIT_0,
|
|
region.end);
|
|
}
|
|
|
|
res = bus->resource[1];
|
|
pcibios_resource_to_bus(bridge, ®ion, res);
|
|
if (res->flags & IORESOURCE_IO) {
|
|
dev_info(&bridge->dev, " bridge window %pR\n", res);
|
|
pci_write_config_dword(bridge, PCI_CB_IO_BASE_1,
|
|
region.start);
|
|
pci_write_config_dword(bridge, PCI_CB_IO_LIMIT_1,
|
|
region.end);
|
|
}
|
|
|
|
res = bus->resource[2];
|
|
pcibios_resource_to_bus(bridge, ®ion, res);
|
|
if (res->flags & IORESOURCE_MEM) {
|
|
dev_info(&bridge->dev, " bridge window %pR\n", res);
|
|
pci_write_config_dword(bridge, PCI_CB_MEMORY_BASE_0,
|
|
region.start);
|
|
pci_write_config_dword(bridge, PCI_CB_MEMORY_LIMIT_0,
|
|
region.end);
|
|
}
|
|
|
|
res = bus->resource[3];
|
|
pcibios_resource_to_bus(bridge, ®ion, res);
|
|
if (res->flags & IORESOURCE_MEM) {
|
|
dev_info(&bridge->dev, " bridge window %pR\n", res);
|
|
pci_write_config_dword(bridge, PCI_CB_MEMORY_BASE_1,
|
|
region.start);
|
|
pci_write_config_dword(bridge, PCI_CB_MEMORY_LIMIT_1,
|
|
region.end);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(pci_setup_cardbus);
|
|
|
|
/* Initialize bridges with base/limit values we have collected.
|
|
PCI-to-PCI Bridge Architecture Specification rev. 1.1 (1998)
|
|
requires that if there is no I/O ports or memory behind the
|
|
bridge, corresponding range must be turned off by writing base
|
|
value greater than limit to the bridge's base/limit registers.
|
|
|
|
Note: care must be taken when updating I/O base/limit registers
|
|
of bridges which support 32-bit I/O. This update requires two
|
|
config space writes, so it's quite possible that an I/O window of
|
|
the bridge will have some undesirable address (e.g. 0) after the
|
|
first write. Ditto 64-bit prefetchable MMIO. */
|
|
static void pci_setup_bridge_io(struct pci_bus *bus)
|
|
{
|
|
struct pci_dev *bridge = bus->self;
|
|
struct resource *res;
|
|
struct pci_bus_region region;
|
|
u32 l, io_upper16;
|
|
|
|
/* Set up the top and bottom of the PCI I/O segment for this bus. */
|
|
res = bus->resource[0];
|
|
pcibios_resource_to_bus(bridge, ®ion, res);
|
|
if (res->flags & IORESOURCE_IO) {
|
|
pci_read_config_dword(bridge, PCI_IO_BASE, &l);
|
|
l &= 0xffff0000;
|
|
l |= (region.start >> 8) & 0x00f0;
|
|
l |= region.end & 0xf000;
|
|
/* Set up upper 16 bits of I/O base/limit. */
|
|
io_upper16 = (region.end & 0xffff0000) | (region.start >> 16);
|
|
dev_info(&bridge->dev, " bridge window %pR\n", res);
|
|
} else {
|
|
/* Clear upper 16 bits of I/O base/limit. */
|
|
io_upper16 = 0;
|
|
l = 0x00f0;
|
|
}
|
|
/* Temporarily disable the I/O range before updating PCI_IO_BASE. */
|
|
pci_write_config_dword(bridge, PCI_IO_BASE_UPPER16, 0x0000ffff);
|
|
/* Update lower 16 bits of I/O base/limit. */
|
|
pci_write_config_dword(bridge, PCI_IO_BASE, l);
|
|
/* Update upper 16 bits of I/O base/limit. */
|
|
pci_write_config_dword(bridge, PCI_IO_BASE_UPPER16, io_upper16);
|
|
}
|
|
|
|
static void pci_setup_bridge_mmio(struct pci_bus *bus)
|
|
{
|
|
struct pci_dev *bridge = bus->self;
|
|
struct resource *res;
|
|
struct pci_bus_region region;
|
|
u32 l;
|
|
|
|
/* Set up the top and bottom of the PCI Memory segment for this bus. */
|
|
res = bus->resource[1];
|
|
pcibios_resource_to_bus(bridge, ®ion, res);
|
|
if (res->flags & IORESOURCE_MEM) {
|
|
l = (region.start >> 16) & 0xfff0;
|
|
l |= region.end & 0xfff00000;
|
|
dev_info(&bridge->dev, " bridge window %pR\n", res);
|
|
} else {
|
|
l = 0x0000fff0;
|
|
}
|
|
pci_write_config_dword(bridge, PCI_MEMORY_BASE, l);
|
|
}
|
|
|
|
static void pci_setup_bridge_mmio_pref(struct pci_bus *bus)
|
|
{
|
|
struct pci_dev *bridge = bus->self;
|
|
struct resource *res;
|
|
struct pci_bus_region region;
|
|
u32 l, bu, lu;
|
|
|
|
/* Clear out the upper 32 bits of PREF limit.
|
|
If PCI_PREF_BASE_UPPER32 was non-zero, this temporarily
|
|
disables PREF range, which is ok. */
|
|
pci_write_config_dword(bridge, PCI_PREF_LIMIT_UPPER32, 0);
|
|
|
|
/* Set up PREF base/limit. */
|
|
bu = lu = 0;
|
|
res = bus->resource[2];
|
|
pcibios_resource_to_bus(bridge, ®ion, res);
|
|
if (res->flags & IORESOURCE_PREFETCH) {
|
|
l = (region.start >> 16) & 0xfff0;
|
|
l |= region.end & 0xfff00000;
|
|
if (res->flags & IORESOURCE_MEM_64) {
|
|
bu = upper_32_bits(region.start);
|
|
lu = upper_32_bits(region.end);
|
|
}
|
|
dev_info(&bridge->dev, " bridge window %pR\n", res);
|
|
} else {
|
|
l = 0x0000fff0;
|
|
}
|
|
pci_write_config_dword(bridge, PCI_PREF_MEMORY_BASE, l);
|
|
|
|
/* Set the upper 32 bits of PREF base & limit. */
|
|
pci_write_config_dword(bridge, PCI_PREF_BASE_UPPER32, bu);
|
|
pci_write_config_dword(bridge, PCI_PREF_LIMIT_UPPER32, lu);
|
|
}
|
|
|
|
static void __pci_setup_bridge(struct pci_bus *bus, unsigned long type)
|
|
{
|
|
struct pci_dev *bridge = bus->self;
|
|
|
|
dev_info(&bridge->dev, "PCI bridge to [bus %02x-%02x]\n",
|
|
bus->secondary, bus->subordinate);
|
|
|
|
if (type & IORESOURCE_IO)
|
|
pci_setup_bridge_io(bus);
|
|
|
|
if (type & IORESOURCE_MEM)
|
|
pci_setup_bridge_mmio(bus);
|
|
|
|
if (type & IORESOURCE_PREFETCH)
|
|
pci_setup_bridge_mmio_pref(bus);
|
|
|
|
pci_write_config_word(bridge, PCI_BRIDGE_CONTROL, bus->bridge_ctl);
|
|
}
|
|
|
|
void pci_setup_bridge(struct pci_bus *bus)
|
|
{
|
|
unsigned long type = IORESOURCE_IO | IORESOURCE_MEM |
|
|
IORESOURCE_PREFETCH;
|
|
|
|
__pci_setup_bridge(bus, type);
|
|
}
|
|
|
|
/* Check whether the bridge supports optional I/O and
|
|
prefetchable memory ranges. If not, the respective
|
|
base/limit registers must be read-only and read as 0. */
|
|
static void pci_bridge_check_ranges(struct pci_bus *bus)
|
|
{
|
|
u16 io;
|
|
u32 pmem;
|
|
struct pci_dev *bridge = bus->self;
|
|
struct resource *b_res;
|
|
|
|
b_res = &bridge->resource[PCI_BRIDGE_RESOURCES];
|
|
b_res[1].flags |= IORESOURCE_MEM;
|
|
|
|
pci_read_config_word(bridge, PCI_IO_BASE, &io);
|
|
if (!io) {
|
|
pci_write_config_word(bridge, PCI_IO_BASE, 0xf0f0);
|
|
pci_read_config_word(bridge, PCI_IO_BASE, &io);
|
|
pci_write_config_word(bridge, PCI_IO_BASE, 0x0);
|
|
}
|
|
if (io)
|
|
b_res[0].flags |= IORESOURCE_IO;
|
|
/* DECchip 21050 pass 2 errata: the bridge may miss an address
|
|
disconnect boundary by one PCI data phase.
|
|
Workaround: do not use prefetching on this device. */
|
|
if (bridge->vendor == PCI_VENDOR_ID_DEC && bridge->device == 0x0001)
|
|
return;
|
|
pci_read_config_dword(bridge, PCI_PREF_MEMORY_BASE, &pmem);
|
|
if (!pmem) {
|
|
pci_write_config_dword(bridge, PCI_PREF_MEMORY_BASE,
|
|
0xfff0fff0);
|
|
pci_read_config_dword(bridge, PCI_PREF_MEMORY_BASE, &pmem);
|
|
pci_write_config_dword(bridge, PCI_PREF_MEMORY_BASE, 0x0);
|
|
}
|
|
if (pmem) {
|
|
b_res[2].flags |= IORESOURCE_MEM | IORESOURCE_PREFETCH;
|
|
if ((pmem & PCI_PREF_RANGE_TYPE_MASK) ==
|
|
PCI_PREF_RANGE_TYPE_64) {
|
|
b_res[2].flags |= IORESOURCE_MEM_64;
|
|
b_res[2].flags |= PCI_PREF_RANGE_TYPE_64;
|
|
}
|
|
}
|
|
|
|
/* double check if bridge does support 64 bit pref */
|
|
if (b_res[2].flags & IORESOURCE_MEM_64) {
|
|
u32 mem_base_hi, tmp;
|
|
pci_read_config_dword(bridge, PCI_PREF_BASE_UPPER32,
|
|
&mem_base_hi);
|
|
pci_write_config_dword(bridge, PCI_PREF_BASE_UPPER32,
|
|
0xffffffff);
|
|
pci_read_config_dword(bridge, PCI_PREF_BASE_UPPER32, &tmp);
|
|
if (!tmp)
|
|
b_res[2].flags &= ~IORESOURCE_MEM_64;
|
|
pci_write_config_dword(bridge, PCI_PREF_BASE_UPPER32,
|
|
mem_base_hi);
|
|
}
|
|
}
|
|
|
|
/* Helper function for sizing routines: find first available
|
|
bus resource of a given type. Note: we intentionally skip
|
|
the bus resources which have already been assigned (that is,
|
|
have non-NULL parent resource). */
|
|
static struct resource *find_free_bus_resource(struct pci_bus *bus, unsigned long type)
|
|
{
|
|
int i;
|
|
struct resource *r;
|
|
unsigned long type_mask = IORESOURCE_IO | IORESOURCE_MEM |
|
|
IORESOURCE_PREFETCH;
|
|
|
|
pci_bus_for_each_resource(bus, r, i) {
|
|
if (r == &ioport_resource || r == &iomem_resource)
|
|
continue;
|
|
if (r && (r->flags & type_mask) == type && !r->parent)
|
|
return r;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static resource_size_t calculate_iosize(resource_size_t size,
|
|
resource_size_t min_size,
|
|
resource_size_t size1,
|
|
resource_size_t old_size,
|
|
resource_size_t align)
|
|
{
|
|
if (size < min_size)
|
|
size = min_size;
|
|
if (old_size == 1 )
|
|
old_size = 0;
|
|
/* To be fixed in 2.5: we should have sort of HAVE_ISA
|
|
flag in the struct pci_bus. */
|
|
#if defined(CONFIG_ISA) || defined(CONFIG_EISA)
|
|
size = (size & 0xff) + ((size & ~0xffUL) << 2);
|
|
#endif
|
|
size = ALIGN(size + size1, align);
|
|
if (size < old_size)
|
|
size = old_size;
|
|
return size;
|
|
}
|
|
|
|
static resource_size_t calculate_memsize(resource_size_t size,
|
|
resource_size_t min_size,
|
|
resource_size_t size1,
|
|
resource_size_t old_size,
|
|
resource_size_t align)
|
|
{
|
|
if (size < min_size)
|
|
size = min_size;
|
|
if (old_size == 1 )
|
|
old_size = 0;
|
|
if (size < old_size)
|
|
size = old_size;
|
|
size = ALIGN(size + size1, align);
|
|
return size;
|
|
}
|
|
|
|
static resource_size_t get_res_add_size(struct resource_list_x *realloc_head,
|
|
struct resource *res)
|
|
{
|
|
struct resource_list_x *list;
|
|
|
|
/* check if it is in realloc_head list */
|
|
for (list = realloc_head->next; list && list->res != res;
|
|
list = list->next);
|
|
if (list)
|
|
return list->add_size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* pbus_size_io() - size the io window of a given bus
|
|
*
|
|
* @bus : the bus
|
|
* @min_size : the minimum io window that must to be allocated
|
|
* @add_size : additional optional io window
|
|
* @realloc_head : track the additional io window on this list
|
|
*
|
|
* Sizing the IO windows of the PCI-PCI bridge is trivial,
|
|
* since these windows have 4K granularity and the IO ranges
|
|
* of non-bridge PCI devices are limited to 256 bytes.
|
|
* We must be careful with the ISA aliasing though.
|
|
*/
|
|
static void pbus_size_io(struct pci_bus *bus, resource_size_t min_size,
|
|
resource_size_t add_size, struct resource_list_x *realloc_head)
|
|
{
|
|
struct pci_dev *dev;
|
|
struct resource *b_res = find_free_bus_resource(bus, IORESOURCE_IO);
|
|
unsigned long size = 0, size0 = 0, size1 = 0;
|
|
resource_size_t children_add_size = 0;
|
|
|
|
if (!b_res)
|
|
return;
|
|
|
|
list_for_each_entry(dev, &bus->devices, bus_list) {
|
|
int i;
|
|
|
|
for (i = 0; i < PCI_NUM_RESOURCES; i++) {
|
|
struct resource *r = &dev->resource[i];
|
|
unsigned long r_size;
|
|
|
|
if (r->parent || !(r->flags & IORESOURCE_IO))
|
|
continue;
|
|
r_size = resource_size(r);
|
|
|
|
if (r_size < 0x400)
|
|
/* Might be re-aligned for ISA */
|
|
size += r_size;
|
|
else
|
|
size1 += r_size;
|
|
|
|
if (realloc_head)
|
|
children_add_size += get_res_add_size(realloc_head, r);
|
|
}
|
|
}
|
|
size0 = calculate_iosize(size, min_size, size1,
|
|
resource_size(b_res), 4096);
|
|
if (children_add_size > add_size)
|
|
add_size = children_add_size;
|
|
size1 = (!realloc_head || (realloc_head && !add_size)) ? size0 :
|
|
calculate_iosize(size, min_size+add_size, size1,
|
|
resource_size(b_res), 4096);
|
|
if (!size0 && !size1) {
|
|
if (b_res->start || b_res->end)
|
|
dev_info(&bus->self->dev, "disabling bridge window "
|
|
"%pR to [bus %02x-%02x] (unused)\n", b_res,
|
|
bus->secondary, bus->subordinate);
|
|
b_res->flags = 0;
|
|
return;
|
|
}
|
|
/* Alignment of the IO window is always 4K */
|
|
b_res->start = 4096;
|
|
b_res->end = b_res->start + size0 - 1;
|
|
b_res->flags |= IORESOURCE_STARTALIGN;
|
|
if (size1 > size0 && realloc_head)
|
|
add_to_list(realloc_head, bus->self, b_res, size1-size0, 4096);
|
|
}
|
|
|
|
/**
|
|
* pbus_size_mem() - size the memory window of a given bus
|
|
*
|
|
* @bus : the bus
|
|
* @min_size : the minimum memory window that must to be allocated
|
|
* @add_size : additional optional memory window
|
|
* @realloc_head : track the additional memory window on this list
|
|
*
|
|
* Calculate the size of the bus and minimal alignment which
|
|
* guarantees that all child resources fit in this size.
|
|
*/
|
|
static int pbus_size_mem(struct pci_bus *bus, unsigned long mask,
|
|
unsigned long type, resource_size_t min_size,
|
|
resource_size_t add_size,
|
|
struct resource_list_x *realloc_head)
|
|
{
|
|
struct pci_dev *dev;
|
|
resource_size_t min_align, align, size, size0, size1;
|
|
resource_size_t aligns[12]; /* Alignments from 1Mb to 2Gb */
|
|
int order, max_order;
|
|
struct resource *b_res = find_free_bus_resource(bus, type);
|
|
unsigned int mem64_mask = 0;
|
|
resource_size_t children_add_size = 0;
|
|
|
|
if (!b_res)
|
|
return 0;
|
|
|
|
memset(aligns, 0, sizeof(aligns));
|
|
max_order = 0;
|
|
size = 0;
|
|
|
|
mem64_mask = b_res->flags & IORESOURCE_MEM_64;
|
|
b_res->flags &= ~IORESOURCE_MEM_64;
|
|
|
|
list_for_each_entry(dev, &bus->devices, bus_list) {
|
|
int i;
|
|
|
|
for (i = 0; i < PCI_NUM_RESOURCES; i++) {
|
|
struct resource *r = &dev->resource[i];
|
|
resource_size_t r_size;
|
|
|
|
if (r->parent || (r->flags & mask) != type)
|
|
continue;
|
|
r_size = resource_size(r);
|
|
#ifdef CONFIG_PCI_IOV
|
|
/* put SRIOV requested res to the optional list */
|
|
if (realloc_head && i >= PCI_IOV_RESOURCES &&
|
|
i <= PCI_IOV_RESOURCE_END) {
|
|
r->end = r->start - 1;
|
|
add_to_list(realloc_head, dev, r, r_size, 0/* dont' care */);
|
|
children_add_size += r_size;
|
|
continue;
|
|
}
|
|
#endif
|
|
/* For bridges size != alignment */
|
|
align = pci_resource_alignment(dev, r);
|
|
order = __ffs(align) - 20;
|
|
if (order > 11) {
|
|
dev_warn(&dev->dev, "disabling BAR %d: %pR "
|
|
"(bad alignment %#llx)\n", i, r,
|
|
(unsigned long long) align);
|
|
r->flags = 0;
|
|
continue;
|
|
}
|
|
size += r_size;
|
|
if (order < 0)
|
|
order = 0;
|
|
/* Exclude ranges with size > align from
|
|
calculation of the alignment. */
|
|
if (r_size == align)
|
|
aligns[order] += align;
|
|
if (order > max_order)
|
|
max_order = order;
|
|
mem64_mask &= r->flags & IORESOURCE_MEM_64;
|
|
|
|
if (realloc_head)
|
|
children_add_size += get_res_add_size(realloc_head, r);
|
|
}
|
|
}
|
|
align = 0;
|
|
min_align = 0;
|
|
for (order = 0; order <= max_order; order++) {
|
|
resource_size_t align1 = 1;
|
|
|
|
align1 <<= (order + 20);
|
|
|
|
if (!align)
|
|
min_align = align1;
|
|
else if (ALIGN(align + min_align, min_align) < align1)
|
|
min_align = align1 >> 1;
|
|
align += aligns[order];
|
|
}
|
|
size0 = calculate_memsize(size, min_size, 0, resource_size(b_res), min_align);
|
|
if (children_add_size > add_size)
|
|
add_size = children_add_size;
|
|
size1 = (!realloc_head || (realloc_head && !add_size)) ? size0 :
|
|
calculate_memsize(size, min_size+add_size, 0,
|
|
resource_size(b_res), min_align);
|
|
if (!size0 && !size1) {
|
|
if (b_res->start || b_res->end)
|
|
dev_info(&bus->self->dev, "disabling bridge window "
|
|
"%pR to [bus %02x-%02x] (unused)\n", b_res,
|
|
bus->secondary, bus->subordinate);
|
|
b_res->flags = 0;
|
|
return 1;
|
|
}
|
|
b_res->start = min_align;
|
|
b_res->end = size0 + min_align - 1;
|
|
b_res->flags |= IORESOURCE_STARTALIGN | mem64_mask;
|
|
if (size1 > size0 && realloc_head)
|
|
add_to_list(realloc_head, bus->self, b_res, size1-size0, min_align);
|
|
return 1;
|
|
}
|
|
|
|
unsigned long pci_cardbus_resource_alignment(struct resource *res)
|
|
{
|
|
if (res->flags & IORESOURCE_IO)
|
|
return pci_cardbus_io_size;
|
|
if (res->flags & IORESOURCE_MEM)
|
|
return pci_cardbus_mem_size;
|
|
return 0;
|
|
}
|
|
|
|
static void pci_bus_size_cardbus(struct pci_bus *bus,
|
|
struct resource_list_x *realloc_head)
|
|
{
|
|
struct pci_dev *bridge = bus->self;
|
|
struct resource *b_res = &bridge->resource[PCI_BRIDGE_RESOURCES];
|
|
u16 ctrl;
|
|
|
|
/*
|
|
* Reserve some resources for CardBus. We reserve
|
|
* a fixed amount of bus space for CardBus bridges.
|
|
*/
|
|
b_res[0].start = 0;
|
|
b_res[0].flags |= IORESOURCE_IO | IORESOURCE_SIZEALIGN;
|
|
if (realloc_head)
|
|
add_to_list(realloc_head, bridge, b_res, pci_cardbus_io_size, 0 /* dont care */);
|
|
|
|
b_res[1].start = 0;
|
|
b_res[1].flags |= IORESOURCE_IO | IORESOURCE_SIZEALIGN;
|
|
if (realloc_head)
|
|
add_to_list(realloc_head, bridge, b_res+1, pci_cardbus_io_size, 0 /* dont care */);
|
|
|
|
/*
|
|
* Check whether prefetchable memory is supported
|
|
* by this bridge.
|
|
*/
|
|
pci_read_config_word(bridge, PCI_CB_BRIDGE_CONTROL, &ctrl);
|
|
if (!(ctrl & PCI_CB_BRIDGE_CTL_PREFETCH_MEM0)) {
|
|
ctrl |= PCI_CB_BRIDGE_CTL_PREFETCH_MEM0;
|
|
pci_write_config_word(bridge, PCI_CB_BRIDGE_CONTROL, ctrl);
|
|
pci_read_config_word(bridge, PCI_CB_BRIDGE_CONTROL, &ctrl);
|
|
}
|
|
|
|
/*
|
|
* If we have prefetchable memory support, allocate
|
|
* two regions. Otherwise, allocate one region of
|
|
* twice the size.
|
|
*/
|
|
if (ctrl & PCI_CB_BRIDGE_CTL_PREFETCH_MEM0) {
|
|
b_res[2].start = 0;
|
|
b_res[2].flags |= IORESOURCE_MEM | IORESOURCE_PREFETCH | IORESOURCE_SIZEALIGN;
|
|
if (realloc_head)
|
|
add_to_list(realloc_head, bridge, b_res+2, pci_cardbus_mem_size, 0 /* dont care */);
|
|
|
|
b_res[3].start = 0;
|
|
b_res[3].flags |= IORESOURCE_MEM | IORESOURCE_SIZEALIGN;
|
|
if (realloc_head)
|
|
add_to_list(realloc_head, bridge, b_res+3, pci_cardbus_mem_size, 0 /* dont care */);
|
|
} else {
|
|
b_res[3].start = 0;
|
|
b_res[3].flags |= IORESOURCE_MEM | IORESOURCE_SIZEALIGN;
|
|
if (realloc_head)
|
|
add_to_list(realloc_head, bridge, b_res+3, pci_cardbus_mem_size * 2, 0 /* dont care */);
|
|
}
|
|
|
|
/* set the size of the resource to zero, so that the resource does not
|
|
* get assigned during required-resource allocation cycle but gets assigned
|
|
* during the optional-resource allocation cycle.
|
|
*/
|
|
b_res[0].start = b_res[1].start = b_res[2].start = b_res[3].start = 1;
|
|
b_res[0].end = b_res[1].end = b_res[2].end = b_res[3].end = 0;
|
|
}
|
|
|
|
void __ref __pci_bus_size_bridges(struct pci_bus *bus,
|
|
struct resource_list_x *realloc_head)
|
|
{
|
|
struct pci_dev *dev;
|
|
unsigned long mask, prefmask;
|
|
resource_size_t additional_mem_size = 0, additional_io_size = 0;
|
|
|
|
list_for_each_entry(dev, &bus->devices, bus_list) {
|
|
struct pci_bus *b = dev->subordinate;
|
|
if (!b)
|
|
continue;
|
|
|
|
switch (dev->class >> 8) {
|
|
case PCI_CLASS_BRIDGE_CARDBUS:
|
|
pci_bus_size_cardbus(b, realloc_head);
|
|
break;
|
|
|
|
case PCI_CLASS_BRIDGE_PCI:
|
|
default:
|
|
__pci_bus_size_bridges(b, realloc_head);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* The root bus? */
|
|
if (!bus->self)
|
|
return;
|
|
|
|
switch (bus->self->class >> 8) {
|
|
case PCI_CLASS_BRIDGE_CARDBUS:
|
|
/* don't size cardbuses yet. */
|
|
break;
|
|
|
|
case PCI_CLASS_BRIDGE_PCI:
|
|
pci_bridge_check_ranges(bus);
|
|
if (bus->self->is_hotplug_bridge) {
|
|
additional_io_size = pci_hotplug_io_size;
|
|
additional_mem_size = pci_hotplug_mem_size;
|
|
}
|
|
/*
|
|
* Follow thru
|
|
*/
|
|
default:
|
|
pbus_size_io(bus, 0, additional_io_size, realloc_head);
|
|
/* If the bridge supports prefetchable range, size it
|
|
separately. If it doesn't, or its prefetchable window
|
|
has already been allocated by arch code, try
|
|
non-prefetchable range for both types of PCI memory
|
|
resources. */
|
|
mask = IORESOURCE_MEM;
|
|
prefmask = IORESOURCE_MEM | IORESOURCE_PREFETCH;
|
|
if (pbus_size_mem(bus, prefmask, prefmask, 0, additional_mem_size, realloc_head))
|
|
mask = prefmask; /* Success, size non-prefetch only. */
|
|
else
|
|
additional_mem_size += additional_mem_size;
|
|
pbus_size_mem(bus, mask, IORESOURCE_MEM, 0, additional_mem_size, realloc_head);
|
|
break;
|
|
}
|
|
}
|
|
|
|
void __ref pci_bus_size_bridges(struct pci_bus *bus)
|
|
{
|
|
__pci_bus_size_bridges(bus, NULL);
|
|
}
|
|
EXPORT_SYMBOL(pci_bus_size_bridges);
|
|
|
|
static void __ref __pci_bus_assign_resources(const struct pci_bus *bus,
|
|
struct resource_list_x *realloc_head,
|
|
struct resource_list_x *fail_head)
|
|
{
|
|
struct pci_bus *b;
|
|
struct pci_dev *dev;
|
|
|
|
pbus_assign_resources_sorted(bus, realloc_head, fail_head);
|
|
|
|
list_for_each_entry(dev, &bus->devices, bus_list) {
|
|
b = dev->subordinate;
|
|
if (!b)
|
|
continue;
|
|
|
|
__pci_bus_assign_resources(b, realloc_head, fail_head);
|
|
|
|
switch (dev->class >> 8) {
|
|
case PCI_CLASS_BRIDGE_PCI:
|
|
if (!pci_is_enabled(dev))
|
|
pci_setup_bridge(b);
|
|
break;
|
|
|
|
case PCI_CLASS_BRIDGE_CARDBUS:
|
|
pci_setup_cardbus(b);
|
|
break;
|
|
|
|
default:
|
|
dev_info(&dev->dev, "not setting up bridge for bus "
|
|
"%04x:%02x\n", pci_domain_nr(b), b->number);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
void __ref pci_bus_assign_resources(const struct pci_bus *bus)
|
|
{
|
|
__pci_bus_assign_resources(bus, NULL, NULL);
|
|
}
|
|
EXPORT_SYMBOL(pci_bus_assign_resources);
|
|
|
|
static void __ref __pci_bridge_assign_resources(const struct pci_dev *bridge,
|
|
struct resource_list_x *fail_head)
|
|
{
|
|
struct pci_bus *b;
|
|
|
|
pdev_assign_resources_sorted((struct pci_dev *)bridge, fail_head);
|
|
|
|
b = bridge->subordinate;
|
|
if (!b)
|
|
return;
|
|
|
|
__pci_bus_assign_resources(b, NULL, fail_head);
|
|
|
|
switch (bridge->class >> 8) {
|
|
case PCI_CLASS_BRIDGE_PCI:
|
|
pci_setup_bridge(b);
|
|
break;
|
|
|
|
case PCI_CLASS_BRIDGE_CARDBUS:
|
|
pci_setup_cardbus(b);
|
|
break;
|
|
|
|
default:
|
|
dev_info(&bridge->dev, "not setting up bridge for bus "
|
|
"%04x:%02x\n", pci_domain_nr(b), b->number);
|
|
break;
|
|
}
|
|
}
|
|
static void pci_bridge_release_resources(struct pci_bus *bus,
|
|
unsigned long type)
|
|
{
|
|
int idx;
|
|
bool changed = false;
|
|
struct pci_dev *dev;
|
|
struct resource *r;
|
|
unsigned long type_mask = IORESOURCE_IO | IORESOURCE_MEM |
|
|
IORESOURCE_PREFETCH;
|
|
|
|
dev = bus->self;
|
|
for (idx = PCI_BRIDGE_RESOURCES; idx <= PCI_BRIDGE_RESOURCE_END;
|
|
idx++) {
|
|
r = &dev->resource[idx];
|
|
if ((r->flags & type_mask) != type)
|
|
continue;
|
|
if (!r->parent)
|
|
continue;
|
|
/*
|
|
* if there are children under that, we should release them
|
|
* all
|
|
*/
|
|
release_child_resources(r);
|
|
if (!release_resource(r)) {
|
|
dev_printk(KERN_DEBUG, &dev->dev,
|
|
"resource %d %pR released\n", idx, r);
|
|
/* keep the old size */
|
|
r->end = resource_size(r) - 1;
|
|
r->start = 0;
|
|
r->flags = 0;
|
|
changed = true;
|
|
}
|
|
}
|
|
|
|
if (changed) {
|
|
/* avoiding touch the one without PREF */
|
|
if (type & IORESOURCE_PREFETCH)
|
|
type = IORESOURCE_PREFETCH;
|
|
__pci_setup_bridge(bus, type);
|
|
}
|
|
}
|
|
|
|
enum release_type {
|
|
leaf_only,
|
|
whole_subtree,
|
|
};
|
|
/*
|
|
* try to release pci bridge resources that is from leaf bridge,
|
|
* so we can allocate big new one later
|
|
*/
|
|
static void __ref pci_bus_release_bridge_resources(struct pci_bus *bus,
|
|
unsigned long type,
|
|
enum release_type rel_type)
|
|
{
|
|
struct pci_dev *dev;
|
|
bool is_leaf_bridge = true;
|
|
|
|
list_for_each_entry(dev, &bus->devices, bus_list) {
|
|
struct pci_bus *b = dev->subordinate;
|
|
if (!b)
|
|
continue;
|
|
|
|
is_leaf_bridge = false;
|
|
|
|
if ((dev->class >> 8) != PCI_CLASS_BRIDGE_PCI)
|
|
continue;
|
|
|
|
if (rel_type == whole_subtree)
|
|
pci_bus_release_bridge_resources(b, type,
|
|
whole_subtree);
|
|
}
|
|
|
|
if (pci_is_root_bus(bus))
|
|
return;
|
|
|
|
if ((bus->self->class >> 8) != PCI_CLASS_BRIDGE_PCI)
|
|
return;
|
|
|
|
if ((rel_type == whole_subtree) || is_leaf_bridge)
|
|
pci_bridge_release_resources(bus, type);
|
|
}
|
|
|
|
static void pci_bus_dump_res(struct pci_bus *bus)
|
|
{
|
|
struct resource *res;
|
|
int i;
|
|
|
|
pci_bus_for_each_resource(bus, res, i) {
|
|
if (!res || !res->end || !res->flags)
|
|
continue;
|
|
|
|
dev_printk(KERN_DEBUG, &bus->dev, "resource %d %pR\n", i, res);
|
|
}
|
|
}
|
|
|
|
static void pci_bus_dump_resources(struct pci_bus *bus)
|
|
{
|
|
struct pci_bus *b;
|
|
struct pci_dev *dev;
|
|
|
|
|
|
pci_bus_dump_res(bus);
|
|
|
|
list_for_each_entry(dev, &bus->devices, bus_list) {
|
|
b = dev->subordinate;
|
|
if (!b)
|
|
continue;
|
|
|
|
pci_bus_dump_resources(b);
|
|
}
|
|
}
|
|
|
|
static int __init pci_bus_get_depth(struct pci_bus *bus)
|
|
{
|
|
int depth = 0;
|
|
struct pci_dev *dev;
|
|
|
|
list_for_each_entry(dev, &bus->devices, bus_list) {
|
|
int ret;
|
|
struct pci_bus *b = dev->subordinate;
|
|
if (!b)
|
|
continue;
|
|
|
|
ret = pci_bus_get_depth(b);
|
|
if (ret + 1 > depth)
|
|
depth = ret + 1;
|
|
}
|
|
|
|
return depth;
|
|
}
|
|
static int __init pci_get_max_depth(void)
|
|
{
|
|
int depth = 0;
|
|
struct pci_bus *bus;
|
|
|
|
list_for_each_entry(bus, &pci_root_buses, node) {
|
|
int ret;
|
|
|
|
ret = pci_bus_get_depth(bus);
|
|
if (ret > depth)
|
|
depth = ret;
|
|
}
|
|
|
|
return depth;
|
|
}
|
|
|
|
|
|
/*
|
|
* first try will not touch pci bridge res
|
|
* second and later try will clear small leaf bridge res
|
|
* will stop till to the max deepth if can not find good one
|
|
*/
|
|
void __init
|
|
pci_assign_unassigned_resources(void)
|
|
{
|
|
struct pci_bus *bus;
|
|
struct resource_list_x realloc_list; /* list of resources that
|
|
want additional resources */
|
|
int tried_times = 0;
|
|
enum release_type rel_type = leaf_only;
|
|
struct resource_list_x head, *list;
|
|
unsigned long type_mask = IORESOURCE_IO | IORESOURCE_MEM |
|
|
IORESOURCE_PREFETCH;
|
|
unsigned long failed_type;
|
|
int max_depth = pci_get_max_depth();
|
|
int pci_try_num;
|
|
|
|
|
|
head.next = NULL;
|
|
realloc_list.next = NULL;
|
|
|
|
pci_try_num = max_depth + 1;
|
|
printk(KERN_DEBUG "PCI: max bus depth: %d pci_try_num: %d\n",
|
|
max_depth, pci_try_num);
|
|
|
|
again:
|
|
/* Depth first, calculate sizes and alignments of all
|
|
subordinate buses. */
|
|
list_for_each_entry(bus, &pci_root_buses, node)
|
|
__pci_bus_size_bridges(bus, &realloc_list);
|
|
|
|
/* Depth last, allocate resources and update the hardware. */
|
|
list_for_each_entry(bus, &pci_root_buses, node)
|
|
__pci_bus_assign_resources(bus, &realloc_list, &head);
|
|
BUG_ON(realloc_list.next);
|
|
tried_times++;
|
|
|
|
/* any device complain? */
|
|
if (!head.next)
|
|
goto enable_and_dump;
|
|
|
|
/* don't realloc if asked to do so */
|
|
if (!pci_realloc_enabled()) {
|
|
free_list(resource_list_x, &head);
|
|
goto enable_and_dump;
|
|
}
|
|
|
|
failed_type = 0;
|
|
for (list = head.next; list;) {
|
|
failed_type |= list->flags;
|
|
list = list->next;
|
|
}
|
|
/*
|
|
* io port are tight, don't try extra
|
|
* or if reach the limit, don't want to try more
|
|
*/
|
|
failed_type &= type_mask;
|
|
if ((failed_type == IORESOURCE_IO) || (tried_times >= pci_try_num)) {
|
|
free_list(resource_list_x, &head);
|
|
goto enable_and_dump;
|
|
}
|
|
|
|
printk(KERN_DEBUG "PCI: No. %d try to assign unassigned res\n",
|
|
tried_times + 1);
|
|
|
|
/* third times and later will not check if it is leaf */
|
|
if ((tried_times + 1) > 2)
|
|
rel_type = whole_subtree;
|
|
|
|
/*
|
|
* Try to release leaf bridge's resources that doesn't fit resource of
|
|
* child device under that bridge
|
|
*/
|
|
for (list = head.next; list;) {
|
|
bus = list->dev->bus;
|
|
pci_bus_release_bridge_resources(bus, list->flags & type_mask,
|
|
rel_type);
|
|
list = list->next;
|
|
}
|
|
/* restore size and flags */
|
|
for (list = head.next; list;) {
|
|
struct resource *res = list->res;
|
|
|
|
res->start = list->start;
|
|
res->end = list->end;
|
|
res->flags = list->flags;
|
|
if (list->dev->subordinate)
|
|
res->flags = 0;
|
|
|
|
list = list->next;
|
|
}
|
|
free_list(resource_list_x, &head);
|
|
|
|
goto again;
|
|
|
|
enable_and_dump:
|
|
/* Depth last, update the hardware. */
|
|
list_for_each_entry(bus, &pci_root_buses, node)
|
|
pci_enable_bridges(bus);
|
|
|
|
/* dump the resource on buses */
|
|
list_for_each_entry(bus, &pci_root_buses, node)
|
|
pci_bus_dump_resources(bus);
|
|
}
|
|
|
|
void pci_assign_unassigned_bridge_resources(struct pci_dev *bridge)
|
|
{
|
|
struct pci_bus *parent = bridge->subordinate;
|
|
int tried_times = 0;
|
|
struct resource_list_x head, *list;
|
|
int retval;
|
|
unsigned long type_mask = IORESOURCE_IO | IORESOURCE_MEM |
|
|
IORESOURCE_PREFETCH;
|
|
|
|
head.next = NULL;
|
|
|
|
again:
|
|
pci_bus_size_bridges(parent);
|
|
__pci_bridge_assign_resources(bridge, &head);
|
|
|
|
tried_times++;
|
|
|
|
if (!head.next)
|
|
goto enable_all;
|
|
|
|
if (tried_times >= 2) {
|
|
/* still fail, don't need to try more */
|
|
free_list(resource_list_x, &head);
|
|
goto enable_all;
|
|
}
|
|
|
|
printk(KERN_DEBUG "PCI: No. %d try to assign unassigned res\n",
|
|
tried_times + 1);
|
|
|
|
/*
|
|
* Try to release leaf bridge's resources that doesn't fit resource of
|
|
* child device under that bridge
|
|
*/
|
|
for (list = head.next; list;) {
|
|
struct pci_bus *bus = list->dev->bus;
|
|
unsigned long flags = list->flags;
|
|
|
|
pci_bus_release_bridge_resources(bus, flags & type_mask,
|
|
whole_subtree);
|
|
list = list->next;
|
|
}
|
|
/* restore size and flags */
|
|
for (list = head.next; list;) {
|
|
struct resource *res = list->res;
|
|
|
|
res->start = list->start;
|
|
res->end = list->end;
|
|
res->flags = list->flags;
|
|
if (list->dev->subordinate)
|
|
res->flags = 0;
|
|
|
|
list = list->next;
|
|
}
|
|
free_list(resource_list_x, &head);
|
|
|
|
goto again;
|
|
|
|
enable_all:
|
|
retval = pci_reenable_device(bridge);
|
|
pci_set_master(bridge);
|
|
pci_enable_bridges(parent);
|
|
}
|
|
EXPORT_SYMBOL_GPL(pci_assign_unassigned_bridge_resources);
|