mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-05 07:56:48 +07:00
085ae41f66
There were three changes necessary in order to allow sparc64 to use setup-res.c: 1) Sparc64 roots the PCI I/O and MEM address space using parent resources contained in the PCI controller structure. I'm actually surprised no other platforms do this, especially ones like Alpha and PPC{,64}. These resources get linked into the iomem/ioport tree when PCI controllers are probed. So the hierarchy looks like this: iomem --| PCI controller 1 MEM space --| device 1 device 2 etc. PCI controller 2 MEM space --| ... ioport --| PCI controller 1 IO space --| ... PCI controller 2 IO space --| ... You get the idea. The drivers/pci/setup-res.c code allocates using plain iomem_space and ioport_space as the root, so that wouldn't work with the above setup. So I added a pcibios_select_root() that is used to handle this. It uses the PCI controller struct's io_space and mem_space on sparc64, and io{port,mem}_resource on every other platform to keep current behavior. 2) quirk_io_region() is buggy. It takes in raw BUS view addresses and tries to use them as a PCI resource. pci_claim_resource() expects the resource to be fully formed when it gets called. The sparc64 implementation would do the translation but that's absolutely wrong, because if the same resource gets released then re-claimed we'll adjust things twice. So I fixed up quirk_io_region() to do the proper pcibios_bus_to_resource() conversion before passing it on to pci_claim_resource(). 3) I was mistakedly __init'ing the function methods the PCI controller drivers provide on sparc64 to implement some parts of these routines. This was, of course, easy to fix. So we end up with the following, and that nasty SPARC64 makefile ifdef in drivers/pci/Makefile is finally zapped. Signed-off-by: David S. Miller <davem@davemloft.net> Signed-off-by: Greg Kroah-Hartman <gregkh@suse.de>
175 lines
5.2 KiB
C
175 lines
5.2 KiB
C
#ifndef _ASM_IA64_PCI_H
|
|
#define _ASM_IA64_PCI_H
|
|
|
|
#include <linux/mm.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/string.h>
|
|
#include <linux/types.h>
|
|
|
|
#include <asm/io.h>
|
|
#include <asm/scatterlist.h>
|
|
|
|
/*
|
|
* Can be used to override the logic in pci_scan_bus for skipping already-configured bus
|
|
* numbers - to be used for buggy BIOSes or architectures with incomplete PCI setup by the
|
|
* loader.
|
|
*/
|
|
#define pcibios_assign_all_busses() 0
|
|
#define pcibios_scan_all_fns(a, b) 0
|
|
|
|
#define PCIBIOS_MIN_IO 0x1000
|
|
#define PCIBIOS_MIN_MEM 0x10000000
|
|
|
|
void pcibios_config_init(void);
|
|
|
|
struct pci_dev;
|
|
|
|
/*
|
|
* PCI_DMA_BUS_IS_PHYS should be set to 1 if there is _necessarily_ a direct correspondence
|
|
* between device bus addresses and CPU physical addresses. Platforms with a hardware I/O
|
|
* MMU _must_ turn this off to suppress the bounce buffer handling code in the block and
|
|
* network device layers. Platforms with separate bus address spaces _must_ turn this off
|
|
* and provide a device DMA mapping implementation that takes care of the necessary
|
|
* address translation.
|
|
*
|
|
* For now, the ia64 platforms which may have separate/multiple bus address spaces all
|
|
* have I/O MMUs which support the merging of physically discontiguous buffers, so we can
|
|
* use that as the sole factor to determine the setting of PCI_DMA_BUS_IS_PHYS.
|
|
*/
|
|
extern unsigned long ia64_max_iommu_merge_mask;
|
|
#define PCI_DMA_BUS_IS_PHYS (ia64_max_iommu_merge_mask == ~0UL)
|
|
|
|
static inline void
|
|
pcibios_set_master (struct pci_dev *dev)
|
|
{
|
|
/* No special bus mastering setup handling */
|
|
}
|
|
|
|
static inline void
|
|
pcibios_penalize_isa_irq (int irq, int active)
|
|
{
|
|
/* We don't do dynamic PCI IRQ allocation */
|
|
}
|
|
|
|
#define HAVE_ARCH_PCI_MWI 1
|
|
extern int pcibios_prep_mwi (struct pci_dev *);
|
|
|
|
#include <asm-generic/pci-dma-compat.h>
|
|
|
|
/* pci_unmap_{single,page} is not a nop, thus... */
|
|
#define DECLARE_PCI_UNMAP_ADDR(ADDR_NAME) \
|
|
dma_addr_t ADDR_NAME;
|
|
#define DECLARE_PCI_UNMAP_LEN(LEN_NAME) \
|
|
__u32 LEN_NAME;
|
|
#define pci_unmap_addr(PTR, ADDR_NAME) \
|
|
((PTR)->ADDR_NAME)
|
|
#define pci_unmap_addr_set(PTR, ADDR_NAME, VAL) \
|
|
(((PTR)->ADDR_NAME) = (VAL))
|
|
#define pci_unmap_len(PTR, LEN_NAME) \
|
|
((PTR)->LEN_NAME)
|
|
#define pci_unmap_len_set(PTR, LEN_NAME, VAL) \
|
|
(((PTR)->LEN_NAME) = (VAL))
|
|
|
|
/* The ia64 platform always supports 64-bit addressing. */
|
|
#define pci_dac_dma_supported(pci_dev, mask) (1)
|
|
#define pci_dac_page_to_dma(dev,pg,off,dir) ((dma_addr_t) page_to_bus(pg) + (off))
|
|
#define pci_dac_dma_to_page(dev,dma_addr) (virt_to_page(bus_to_virt(dma_addr)))
|
|
#define pci_dac_dma_to_offset(dev,dma_addr) offset_in_page(dma_addr)
|
|
#define pci_dac_dma_sync_single_for_cpu(dev,dma_addr,len,dir) do { } while (0)
|
|
#define pci_dac_dma_sync_single_for_device(dev,dma_addr,len,dir) do { mb(); } while (0)
|
|
|
|
#define sg_dma_len(sg) ((sg)->dma_length)
|
|
#define sg_dma_address(sg) ((sg)->dma_address)
|
|
|
|
#ifdef CONFIG_PCI
|
|
static inline void pci_dma_burst_advice(struct pci_dev *pdev,
|
|
enum pci_dma_burst_strategy *strat,
|
|
unsigned long *strategy_parameter)
|
|
{
|
|
unsigned long cacheline_size;
|
|
u8 byte;
|
|
|
|
pci_read_config_byte(pdev, PCI_CACHE_LINE_SIZE, &byte);
|
|
if (byte == 0)
|
|
cacheline_size = 1024;
|
|
else
|
|
cacheline_size = (int) byte * 4;
|
|
|
|
*strat = PCI_DMA_BURST_MULTIPLE;
|
|
*strategy_parameter = cacheline_size;
|
|
}
|
|
#endif
|
|
|
|
#define HAVE_PCI_MMAP
|
|
extern int pci_mmap_page_range (struct pci_dev *dev, struct vm_area_struct *vma,
|
|
enum pci_mmap_state mmap_state, int write_combine);
|
|
#define HAVE_PCI_LEGACY
|
|
extern int pci_mmap_legacy_page_range(struct pci_bus *bus,
|
|
struct vm_area_struct *vma);
|
|
extern ssize_t pci_read_legacy_io(struct kobject *kobj, char *buf, loff_t off,
|
|
size_t count);
|
|
extern ssize_t pci_write_legacy_io(struct kobject *kobj, char *buf, loff_t off,
|
|
size_t count);
|
|
extern int pci_mmap_legacy_mem(struct kobject *kobj,
|
|
struct bin_attribute *attr,
|
|
struct vm_area_struct *vma);
|
|
|
|
#define pci_get_legacy_mem platform_pci_get_legacy_mem
|
|
#define pci_legacy_read platform_pci_legacy_read
|
|
#define pci_legacy_write platform_pci_legacy_write
|
|
|
|
struct pci_window {
|
|
struct resource resource;
|
|
u64 offset;
|
|
};
|
|
|
|
struct pci_controller {
|
|
void *acpi_handle;
|
|
void *iommu;
|
|
int segment;
|
|
int node; /* nearest node with memory or -1 for global allocation */
|
|
|
|
unsigned int windows;
|
|
struct pci_window *window;
|
|
|
|
void *platform_data;
|
|
};
|
|
|
|
#define PCI_CONTROLLER(busdev) ((struct pci_controller *) busdev->sysdata)
|
|
#define pci_domain_nr(busdev) (PCI_CONTROLLER(busdev)->segment)
|
|
|
|
extern struct pci_ops pci_root_ops;
|
|
|
|
static inline int pci_proc_domain(struct pci_bus *bus)
|
|
{
|
|
return (pci_domain_nr(bus) != 0);
|
|
}
|
|
|
|
static inline void pcibios_add_platform_entries(struct pci_dev *dev)
|
|
{
|
|
}
|
|
|
|
extern void pcibios_resource_to_bus(struct pci_dev *dev,
|
|
struct pci_bus_region *region, struct resource *res);
|
|
|
|
extern void pcibios_bus_to_resource(struct pci_dev *dev,
|
|
struct resource *res, struct pci_bus_region *region);
|
|
|
|
static inline struct resource *
|
|
pcibios_select_root(struct pci_dev *pdev, struct resource *res)
|
|
{
|
|
struct resource *root = NULL;
|
|
|
|
if (res->flags & IORESOURCE_IO)
|
|
root = &ioport_resource;
|
|
if (res->flags & IORESOURCE_MEM)
|
|
root = &iomem_resource;
|
|
|
|
return root;
|
|
}
|
|
|
|
#define pcibios_scan_all_fns(a, b) 0
|
|
|
|
#endif /* _ASM_IA64_PCI_H */
|