mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-25 15:30:54 +07:00
ace2dc7d12
This plugs in the alignment and emulation fault reporting for perf sw events. Signed-off-by: Paul Mundt <lethal@linux-sh.org>
298 lines
6.4 KiB
C
298 lines
6.4 KiB
C
/*
|
|
* Trapped io support
|
|
*
|
|
* Copyright (C) 2008 Magnus Damm
|
|
*
|
|
* Intercept io operations by trapping.
|
|
*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*/
|
|
#include <linux/kernel.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/bitops.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <asm/system.h>
|
|
#include <asm/mmu_context.h>
|
|
#include <asm/uaccess.h>
|
|
#include <asm/io.h>
|
|
#include <asm/io_trapped.h>
|
|
|
|
#define TRAPPED_PAGES_MAX 16
|
|
|
|
#ifdef CONFIG_HAS_IOPORT
|
|
LIST_HEAD(trapped_io);
|
|
EXPORT_SYMBOL_GPL(trapped_io);
|
|
#endif
|
|
#ifdef CONFIG_HAS_IOMEM
|
|
LIST_HEAD(trapped_mem);
|
|
EXPORT_SYMBOL_GPL(trapped_mem);
|
|
#endif
|
|
static DEFINE_SPINLOCK(trapped_lock);
|
|
|
|
static int trapped_io_disable __read_mostly;
|
|
|
|
static int __init trapped_io_setup(char *__unused)
|
|
{
|
|
trapped_io_disable = 1;
|
|
return 1;
|
|
}
|
|
__setup("noiotrap", trapped_io_setup);
|
|
|
|
int register_trapped_io(struct trapped_io *tiop)
|
|
{
|
|
struct resource *res;
|
|
unsigned long len = 0, flags = 0;
|
|
struct page *pages[TRAPPED_PAGES_MAX];
|
|
int k, n;
|
|
|
|
if (unlikely(trapped_io_disable))
|
|
return 0;
|
|
|
|
/* structure must be page aligned */
|
|
if ((unsigned long)tiop & (PAGE_SIZE - 1))
|
|
goto bad;
|
|
|
|
for (k = 0; k < tiop->num_resources; k++) {
|
|
res = tiop->resource + k;
|
|
len += roundup((res->end - res->start) + 1, PAGE_SIZE);
|
|
flags |= res->flags;
|
|
}
|
|
|
|
/* support IORESOURCE_IO _or_ MEM, not both */
|
|
if (hweight_long(flags) != 1)
|
|
goto bad;
|
|
|
|
n = len >> PAGE_SHIFT;
|
|
|
|
if (n >= TRAPPED_PAGES_MAX)
|
|
goto bad;
|
|
|
|
for (k = 0; k < n; k++)
|
|
pages[k] = virt_to_page(tiop);
|
|
|
|
tiop->virt_base = vmap(pages, n, VM_MAP, PAGE_NONE);
|
|
if (!tiop->virt_base)
|
|
goto bad;
|
|
|
|
len = 0;
|
|
for (k = 0; k < tiop->num_resources; k++) {
|
|
res = tiop->resource + k;
|
|
pr_info("trapped io 0x%08lx overrides %s 0x%08lx\n",
|
|
(unsigned long)(tiop->virt_base + len),
|
|
res->flags & IORESOURCE_IO ? "io" : "mmio",
|
|
(unsigned long)res->start);
|
|
len += roundup((res->end - res->start) + 1, PAGE_SIZE);
|
|
}
|
|
|
|
tiop->magic = IO_TRAPPED_MAGIC;
|
|
INIT_LIST_HEAD(&tiop->list);
|
|
spin_lock_irq(&trapped_lock);
|
|
#ifdef CONFIG_HAS_IOPORT
|
|
if (flags & IORESOURCE_IO)
|
|
list_add(&tiop->list, &trapped_io);
|
|
#endif
|
|
#ifdef CONFIG_HAS_IOMEM
|
|
if (flags & IORESOURCE_MEM)
|
|
list_add(&tiop->list, &trapped_mem);
|
|
#endif
|
|
spin_unlock_irq(&trapped_lock);
|
|
|
|
return 0;
|
|
bad:
|
|
pr_warning("unable to install trapped io filter\n");
|
|
return -1;
|
|
}
|
|
EXPORT_SYMBOL_GPL(register_trapped_io);
|
|
|
|
void __iomem *match_trapped_io_handler(struct list_head *list,
|
|
unsigned long offset,
|
|
unsigned long size)
|
|
{
|
|
unsigned long voffs;
|
|
struct trapped_io *tiop;
|
|
struct resource *res;
|
|
int k, len;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&trapped_lock, flags);
|
|
list_for_each_entry(tiop, list, list) {
|
|
voffs = 0;
|
|
for (k = 0; k < tiop->num_resources; k++) {
|
|
res = tiop->resource + k;
|
|
if (res->start == offset) {
|
|
spin_unlock_irqrestore(&trapped_lock, flags);
|
|
return tiop->virt_base + voffs;
|
|
}
|
|
|
|
len = (res->end - res->start) + 1;
|
|
voffs += roundup(len, PAGE_SIZE);
|
|
}
|
|
}
|
|
spin_unlock_irqrestore(&trapped_lock, flags);
|
|
return NULL;
|
|
}
|
|
EXPORT_SYMBOL_GPL(match_trapped_io_handler);
|
|
|
|
static struct trapped_io *lookup_tiop(unsigned long address)
|
|
{
|
|
pgd_t *pgd_k;
|
|
pud_t *pud_k;
|
|
pmd_t *pmd_k;
|
|
pte_t *pte_k;
|
|
pte_t entry;
|
|
|
|
pgd_k = swapper_pg_dir + pgd_index(address);
|
|
if (!pgd_present(*pgd_k))
|
|
return NULL;
|
|
|
|
pud_k = pud_offset(pgd_k, address);
|
|
if (!pud_present(*pud_k))
|
|
return NULL;
|
|
|
|
pmd_k = pmd_offset(pud_k, address);
|
|
if (!pmd_present(*pmd_k))
|
|
return NULL;
|
|
|
|
pte_k = pte_offset_kernel(pmd_k, address);
|
|
entry = *pte_k;
|
|
|
|
return pfn_to_kaddr(pte_pfn(entry));
|
|
}
|
|
|
|
static unsigned long lookup_address(struct trapped_io *tiop,
|
|
unsigned long address)
|
|
{
|
|
struct resource *res;
|
|
unsigned long vaddr = (unsigned long)tiop->virt_base;
|
|
unsigned long len;
|
|
int k;
|
|
|
|
for (k = 0; k < tiop->num_resources; k++) {
|
|
res = tiop->resource + k;
|
|
len = roundup((res->end - res->start) + 1, PAGE_SIZE);
|
|
if (address < (vaddr + len))
|
|
return res->start + (address - vaddr);
|
|
vaddr += len;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static unsigned long long copy_word(unsigned long src_addr, int src_len,
|
|
unsigned long dst_addr, int dst_len)
|
|
{
|
|
unsigned long long tmp = 0;
|
|
|
|
switch (src_len) {
|
|
case 1:
|
|
tmp = __raw_readb(src_addr);
|
|
break;
|
|
case 2:
|
|
tmp = __raw_readw(src_addr);
|
|
break;
|
|
case 4:
|
|
tmp = __raw_readl(src_addr);
|
|
break;
|
|
case 8:
|
|
tmp = __raw_readq(src_addr);
|
|
break;
|
|
}
|
|
|
|
switch (dst_len) {
|
|
case 1:
|
|
__raw_writeb(tmp, dst_addr);
|
|
break;
|
|
case 2:
|
|
__raw_writew(tmp, dst_addr);
|
|
break;
|
|
case 4:
|
|
__raw_writel(tmp, dst_addr);
|
|
break;
|
|
case 8:
|
|
__raw_writeq(tmp, dst_addr);
|
|
break;
|
|
}
|
|
|
|
return tmp;
|
|
}
|
|
|
|
static unsigned long from_device(void *dst, const void *src, unsigned long cnt)
|
|
{
|
|
struct trapped_io *tiop;
|
|
unsigned long src_addr = (unsigned long)src;
|
|
unsigned long long tmp;
|
|
|
|
pr_debug("trapped io read 0x%08lx (%ld)\n", src_addr, cnt);
|
|
tiop = lookup_tiop(src_addr);
|
|
WARN_ON(!tiop || (tiop->magic != IO_TRAPPED_MAGIC));
|
|
|
|
src_addr = lookup_address(tiop, src_addr);
|
|
if (!src_addr)
|
|
return cnt;
|
|
|
|
tmp = copy_word(src_addr,
|
|
max_t(unsigned long, cnt,
|
|
(tiop->minimum_bus_width / 8)),
|
|
(unsigned long)dst, cnt);
|
|
|
|
pr_debug("trapped io read 0x%08lx -> 0x%08llx\n", src_addr, tmp);
|
|
return 0;
|
|
}
|
|
|
|
static unsigned long to_device(void *dst, const void *src, unsigned long cnt)
|
|
{
|
|
struct trapped_io *tiop;
|
|
unsigned long dst_addr = (unsigned long)dst;
|
|
unsigned long long tmp;
|
|
|
|
pr_debug("trapped io write 0x%08lx (%ld)\n", dst_addr, cnt);
|
|
tiop = lookup_tiop(dst_addr);
|
|
WARN_ON(!tiop || (tiop->magic != IO_TRAPPED_MAGIC));
|
|
|
|
dst_addr = lookup_address(tiop, dst_addr);
|
|
if (!dst_addr)
|
|
return cnt;
|
|
|
|
tmp = copy_word((unsigned long)src, cnt,
|
|
dst_addr, max_t(unsigned long, cnt,
|
|
(tiop->minimum_bus_width / 8)));
|
|
|
|
pr_debug("trapped io write 0x%08lx -> 0x%08llx\n", dst_addr, tmp);
|
|
return 0;
|
|
}
|
|
|
|
static struct mem_access trapped_io_access = {
|
|
from_device,
|
|
to_device,
|
|
};
|
|
|
|
int handle_trapped_io(struct pt_regs *regs, unsigned long address)
|
|
{
|
|
mm_segment_t oldfs;
|
|
insn_size_t instruction;
|
|
int tmp;
|
|
|
|
if (trapped_io_disable)
|
|
return 0;
|
|
if (!lookup_tiop(address))
|
|
return 0;
|
|
|
|
WARN_ON(user_mode(regs));
|
|
|
|
oldfs = get_fs();
|
|
set_fs(KERNEL_DS);
|
|
if (copy_from_user(&instruction, (void *)(regs->pc),
|
|
sizeof(instruction))) {
|
|
set_fs(oldfs);
|
|
return 0;
|
|
}
|
|
|
|
tmp = handle_unaligned_access(instruction, regs,
|
|
&trapped_io_access, 1, address);
|
|
set_fs(oldfs);
|
|
return tmp == 0;
|
|
}
|