2019-05-27 13:55:05 +07:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-or-later
|
2005-04-17 05:20:36 +07:00
|
|
|
/*
|
|
|
|
* Copyright (C) 2001 Mike Corrigan & Dave Engebretsen, IBM Corporation
|
|
|
|
*
|
2006-04-29 10:51:59 +07:00
|
|
|
* Rewrite, cleanup:
|
2005-04-17 05:20:36 +07:00
|
|
|
*
|
2005-11-21 15:12:32 +07:00
|
|
|
* Copyright (C) 2004 Olof Johansson <olof@lixom.net>, IBM Corporation
|
2006-04-29 10:51:59 +07:00
|
|
|
* Copyright (C) 2006 Olof Johansson <olof@lixom.net>
|
2005-04-17 05:20:36 +07:00
|
|
|
*
|
|
|
|
* Dynamic DMA mapping support, pSeries-specific parts, both SMP and LPAR.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/types.h>
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/mm.h>
|
2012-07-26 04:20:03 +07:00
|
|
|
#include <linux/memblock.h>
|
2005-04-17 05:20:36 +07:00
|
|
|
#include <linux/spinlock.h>
|
|
|
|
#include <linux/string.h>
|
|
|
|
#include <linux/pci.h>
|
|
|
|
#include <linux/dma-mapping.h>
|
2008-10-23 03:39:04 +07:00
|
|
|
#include <linux/crash_dump.h>
|
2011-02-10 16:10:47 +07:00
|
|
|
#include <linux/memory.h>
|
2012-10-02 23:57:57 +07:00
|
|
|
#include <linux/of.h>
|
2015-06-05 13:34:56 +07:00
|
|
|
#include <linux/iommu.h>
|
2015-06-05 13:35:09 +07:00
|
|
|
#include <linux/rculist.h>
|
2005-04-17 05:20:36 +07:00
|
|
|
#include <asm/io.h>
|
|
|
|
#include <asm/prom.h>
|
|
|
|
#include <asm/rtas.h>
|
|
|
|
#include <asm/iommu.h>
|
|
|
|
#include <asm/pci-bridge.h>
|
|
|
|
#include <asm/machdep.h>
|
2005-08-03 11:35:25 +07:00
|
|
|
#include <asm/firmware.h>
|
2005-09-20 10:45:41 +07:00
|
|
|
#include <asm/tce.h>
|
2005-09-27 23:50:25 +07:00
|
|
|
#include <asm/ppc-pci.h>
|
2005-11-07 09:18:13 +07:00
|
|
|
#include <asm/udbg.h>
|
2011-02-10 16:10:47 +07:00
|
|
|
#include <asm/mmzone.h>
|
2013-08-22 16:53:52 +07:00
|
|
|
#include <asm/plpar_wrappers.h>
|
2005-11-03 11:33:31 +07:00
|
|
|
|
2015-03-31 12:00:50 +07:00
|
|
|
#include "pseries.h"
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2015-06-05 13:35:08 +07:00
|
|
|
static struct iommu_table_group *iommu_pseries_alloc_group(int node)
|
|
|
|
{
|
2017-10-19 01:48:52 +07:00
|
|
|
struct iommu_table_group *table_group;
|
|
|
|
struct iommu_table *tbl;
|
2015-06-05 13:35:08 +07:00
|
|
|
|
|
|
|
table_group = kzalloc_node(sizeof(struct iommu_table_group), GFP_KERNEL,
|
|
|
|
node);
|
|
|
|
if (!table_group)
|
2017-10-19 01:48:52 +07:00
|
|
|
return NULL;
|
2015-06-05 13:35:08 +07:00
|
|
|
|
|
|
|
tbl = kzalloc_node(sizeof(struct iommu_table), GFP_KERNEL, node);
|
|
|
|
if (!tbl)
|
2017-10-19 01:48:52 +07:00
|
|
|
goto free_group;
|
2015-06-05 13:35:08 +07:00
|
|
|
|
2015-06-05 13:35:09 +07:00
|
|
|
INIT_LIST_HEAD_RCU(&tbl->it_group_list);
|
2017-03-22 11:21:50 +07:00
|
|
|
kref_init(&tbl->it_kref);
|
2015-06-05 13:35:09 +07:00
|
|
|
|
2015-06-05 13:35:08 +07:00
|
|
|
table_group->tables[0] = tbl;
|
|
|
|
|
|
|
|
return table_group;
|
|
|
|
|
2017-10-19 01:48:52 +07:00
|
|
|
free_group:
|
|
|
|
kfree(table_group);
|
2015-06-05 13:35:08 +07:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void iommu_pseries_free_group(struct iommu_table_group *table_group,
|
2015-06-05 13:34:56 +07:00
|
|
|
const char *node_name)
|
|
|
|
{
|
2015-06-05 13:35:08 +07:00
|
|
|
struct iommu_table *tbl;
|
|
|
|
|
|
|
|
if (!table_group)
|
|
|
|
return;
|
|
|
|
|
2015-06-05 13:35:09 +07:00
|
|
|
tbl = table_group->tables[0];
|
2015-06-05 13:34:56 +07:00
|
|
|
#ifdef CONFIG_IOMMU_API
|
2015-06-05 13:35:08 +07:00
|
|
|
if (table_group->group) {
|
|
|
|
iommu_group_put(table_group->group);
|
|
|
|
BUG_ON(table_group->group);
|
2015-06-05 13:34:56 +07:00
|
|
|
}
|
|
|
|
#endif
|
2017-03-22 11:21:50 +07:00
|
|
|
iommu_tce_table_put(tbl);
|
2015-06-05 13:35:08 +07:00
|
|
|
|
|
|
|
kfree(table_group);
|
2015-06-05 13:34:56 +07:00
|
|
|
}
|
|
|
|
|
2008-07-24 01:31:16 +07:00
|
|
|
static int tce_build_pSeries(struct iommu_table *tbl, long index,
|
2006-04-29 10:51:59 +07:00
|
|
|
long npages, unsigned long uaddr,
|
2008-07-16 02:51:47 +07:00
|
|
|
enum dma_data_direction direction,
|
2016-08-04 03:46:00 +07:00
|
|
|
unsigned long attrs)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2006-04-29 10:51:59 +07:00
|
|
|
u64 proto_tce;
|
2019-04-07 09:48:08 +07:00
|
|
|
__be64 *tcep;
|
2006-04-29 10:51:59 +07:00
|
|
|
u64 rpn;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2006-04-29 10:51:59 +07:00
|
|
|
proto_tce = TCE_PCI_READ; // Read allowed
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
if (direction != DMA_TO_DEVICE)
|
2006-04-29 10:51:59 +07:00
|
|
|
proto_tce |= TCE_PCI_WRITE;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2019-04-07 09:48:08 +07:00
|
|
|
tcep = ((__be64 *)tbl->it_base) + index;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
while (npages--) {
|
2010-07-12 11:36:09 +07:00
|
|
|
/* can't move this out since we might cross MEMBLOCK boundary */
|
2012-07-26 04:19:57 +07:00
|
|
|
rpn = __pa(uaddr) >> TCE_SHIFT;
|
2013-10-17 19:21:15 +07:00
|
|
|
*tcep = cpu_to_be64(proto_tce | (rpn & TCE_RPN_MASK) << TCE_RPN_SHIFT);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2005-09-20 10:46:44 +07:00
|
|
|
uaddr += TCE_PAGE_SIZE;
|
2006-04-29 10:51:59 +07:00
|
|
|
tcep++;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
2008-07-24 01:31:16 +07:00
|
|
|
return 0;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static void tce_free_pSeries(struct iommu_table *tbl, long index, long npages)
|
|
|
|
{
|
2019-04-07 09:48:08 +07:00
|
|
|
__be64 *tcep;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2019-04-07 09:48:08 +07:00
|
|
|
tcep = ((__be64 *)tbl->it_base) + index;
|
2006-04-29 10:51:59 +07:00
|
|
|
|
|
|
|
while (npages--)
|
|
|
|
*(tcep++) = 0;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2006-06-23 13:35:10 +07:00
|
|
|
static unsigned long tce_get_pseries(struct iommu_table *tbl, long index)
|
|
|
|
{
|
2013-10-17 19:21:15 +07:00
|
|
|
__be64 *tcep;
|
2006-06-23 13:35:10 +07:00
|
|
|
|
2013-10-17 19:21:15 +07:00
|
|
|
tcep = ((__be64 *)tbl->it_base) + index;
|
2006-06-23 13:35:10 +07:00
|
|
|
|
2013-10-17 19:21:15 +07:00
|
|
|
return be64_to_cpu(*tcep);
|
2006-06-23 13:35:10 +07:00
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2019-12-16 11:19:22 +07:00
|
|
|
static void tce_free_pSeriesLP(unsigned long liobn, long, long);
|
2008-07-24 01:31:16 +07:00
|
|
|
static void tce_freemulti_pSeriesLP(struct iommu_table*, long, long);
|
|
|
|
|
2019-12-16 11:19:22 +07:00
|
|
|
static int tce_build_pSeriesLP(unsigned long liobn, long tcenum, long tceshift,
|
2005-04-17 05:20:36 +07:00
|
|
|
long npages, unsigned long uaddr,
|
2008-07-16 02:51:47 +07:00
|
|
|
enum dma_data_direction direction,
|
2016-08-04 03:46:00 +07:00
|
|
|
unsigned long attrs)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2008-07-24 01:31:16 +07:00
|
|
|
u64 rc = 0;
|
2006-04-29 10:51:59 +07:00
|
|
|
u64 proto_tce, tce;
|
|
|
|
u64 rpn;
|
2008-07-24 01:31:16 +07:00
|
|
|
int ret = 0;
|
|
|
|
long tcenum_start = tcenum, npages_start = npages;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2019-12-16 11:19:22 +07:00
|
|
|
rpn = __pa(uaddr) >> tceshift;
|
2006-04-29 10:51:59 +07:00
|
|
|
proto_tce = TCE_PCI_READ;
|
2005-04-17 05:20:36 +07:00
|
|
|
if (direction != DMA_TO_DEVICE)
|
2006-04-29 10:51:59 +07:00
|
|
|
proto_tce |= TCE_PCI_WRITE;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
while (npages--) {
|
2019-12-16 11:19:22 +07:00
|
|
|
tce = proto_tce | (rpn & TCE_RPN_MASK) << tceshift;
|
|
|
|
rc = plpar_tce_put((u64)liobn, (u64)tcenum << tceshift, tce);
|
2006-04-29 10:51:59 +07:00
|
|
|
|
2008-07-24 01:31:16 +07:00
|
|
|
if (unlikely(rc == H_NOT_ENOUGH_RESOURCES)) {
|
|
|
|
ret = (int)rc;
|
2019-12-16 11:19:22 +07:00
|
|
|
tce_free_pSeriesLP(liobn, tcenum_start,
|
2008-07-24 01:31:16 +07:00
|
|
|
(npages_start - (npages + 1)));
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
if (rc && printk_ratelimit()) {
|
2009-01-06 21:26:03 +07:00
|
|
|
printk("tce_build_pSeriesLP: plpar_tce_put failed. rc=%lld\n", rc);
|
2019-12-16 11:19:22 +07:00
|
|
|
printk("\tindex = 0x%llx\n", (u64)liobn);
|
2009-01-06 21:26:03 +07:00
|
|
|
printk("\ttcenum = 0x%llx\n", (u64)tcenum);
|
|
|
|
printk("\ttce val = 0x%llx\n", tce );
|
2014-10-13 15:41:40 +07:00
|
|
|
dump_stack();
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
2006-04-29 10:51:59 +07:00
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
tcenum++;
|
2006-04-29 10:51:59 +07:00
|
|
|
rpn++;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
2008-07-24 01:31:16 +07:00
|
|
|
return ret;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2013-10-17 19:21:15 +07:00
|
|
|
static DEFINE_PER_CPU(__be64 *, tce_page);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2008-07-24 01:31:16 +07:00
|
|
|
static int tce_buildmulti_pSeriesLP(struct iommu_table *tbl, long tcenum,
|
2005-04-17 05:20:36 +07:00
|
|
|
long npages, unsigned long uaddr,
|
2008-07-16 02:51:47 +07:00
|
|
|
enum dma_data_direction direction,
|
2016-08-04 03:46:00 +07:00
|
|
|
unsigned long attrs)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2008-07-24 01:31:16 +07:00
|
|
|
u64 rc = 0;
|
2006-04-29 10:51:59 +07:00
|
|
|
u64 proto_tce;
|
2013-10-17 19:21:15 +07:00
|
|
|
__be64 *tcep;
|
2006-04-29 10:51:59 +07:00
|
|
|
u64 rpn;
|
2005-04-17 05:20:36 +07:00
|
|
|
long l, limit;
|
2008-07-24 01:31:16 +07:00
|
|
|
long tcenum_start = tcenum, npages_start = npages;
|
|
|
|
int ret = 0;
|
2012-06-04 02:42:13 +07:00
|
|
|
unsigned long flags;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2019-12-16 11:19:23 +07:00
|
|
|
if ((npages == 1) || !firmware_has_feature(FW_FEATURE_PUT_TCE_IND)) {
|
2019-12-16 11:19:22 +07:00
|
|
|
return tce_build_pSeriesLP(tbl->it_index, tcenum,
|
|
|
|
tbl->it_page_shift, npages, uaddr,
|
2008-07-24 01:31:16 +07:00
|
|
|
direction, attrs);
|
2008-05-08 11:27:23 +07:00
|
|
|
}
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2012-06-04 02:42:13 +07:00
|
|
|
local_irq_save(flags); /* to protect tcep and the page behind it */
|
|
|
|
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 03:23:25 +07:00
|
|
|
tcep = __this_cpu_read(tce_page);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
/* This is safe to do since interrupts are off when we're called
|
|
|
|
* from iommu_alloc{,_sg}()
|
|
|
|
*/
|
|
|
|
if (!tcep) {
|
2013-10-17 19:21:15 +07:00
|
|
|
tcep = (__be64 *)__get_free_page(GFP_ATOMIC);
|
2005-04-17 05:20:36 +07:00
|
|
|
/* If allocation fails, fall back to the loop implementation */
|
2008-05-08 11:27:23 +07:00
|
|
|
if (!tcep) {
|
2012-06-04 02:42:13 +07:00
|
|
|
local_irq_restore(flags);
|
2019-12-16 11:19:22 +07:00
|
|
|
return tce_build_pSeriesLP(tbl->it_index, tcenum,
|
|
|
|
tbl->it_page_shift,
|
|
|
|
npages, uaddr, direction, attrs);
|
2008-05-08 11:27:23 +07:00
|
|
|
}
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 03:23:25 +07:00
|
|
|
__this_cpu_write(tce_page, tcep);
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2012-07-26 04:19:57 +07:00
|
|
|
rpn = __pa(uaddr) >> TCE_SHIFT;
|
2006-04-29 10:51:59 +07:00
|
|
|
proto_tce = TCE_PCI_READ;
|
2005-04-17 05:20:36 +07:00
|
|
|
if (direction != DMA_TO_DEVICE)
|
2006-04-29 10:51:59 +07:00
|
|
|
proto_tce |= TCE_PCI_WRITE;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
/* We can map max one pageful of TCEs at a time */
|
|
|
|
do {
|
|
|
|
/*
|
|
|
|
* Set up the page with TCE data, looping through and setting
|
|
|
|
* the values.
|
|
|
|
*/
|
2006-04-29 10:51:59 +07:00
|
|
|
limit = min_t(long, npages, 4096/TCE_ENTRY_SIZE);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
for (l = 0; l < limit; l++) {
|
2013-10-17 19:21:15 +07:00
|
|
|
tcep[l] = cpu_to_be64(proto_tce | (rpn & TCE_RPN_MASK) << TCE_RPN_SHIFT);
|
2006-04-29 10:51:59 +07:00
|
|
|
rpn++;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
rc = plpar_tce_put_indirect((u64)tbl->it_index,
|
|
|
|
(u64)tcenum << 12,
|
2012-07-26 04:19:57 +07:00
|
|
|
(u64)__pa(tcep),
|
2005-04-17 05:20:36 +07:00
|
|
|
limit);
|
|
|
|
|
|
|
|
npages -= limit;
|
|
|
|
tcenum += limit;
|
|
|
|
} while (npages > 0 && !rc);
|
|
|
|
|
2012-06-04 02:42:13 +07:00
|
|
|
local_irq_restore(flags);
|
|
|
|
|
2008-07-24 01:31:16 +07:00
|
|
|
if (unlikely(rc == H_NOT_ENOUGH_RESOURCES)) {
|
|
|
|
ret = (int)rc;
|
|
|
|
tce_freemulti_pSeriesLP(tbl, tcenum_start,
|
|
|
|
(npages_start - (npages + limit)));
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
if (rc && printk_ratelimit()) {
|
2009-01-06 21:26:03 +07:00
|
|
|
printk("tce_buildmulti_pSeriesLP: plpar_tce_put failed. rc=%lld\n", rc);
|
|
|
|
printk("\tindex = 0x%llx\n", (u64)tbl->it_index);
|
|
|
|
printk("\tnpages = 0x%llx\n", (u64)npages);
|
|
|
|
printk("\ttce[0] val = 0x%llx\n", tcep[0]);
|
2014-10-13 15:41:40 +07:00
|
|
|
dump_stack();
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
2008-07-24 01:31:16 +07:00
|
|
|
return ret;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2019-12-16 11:19:22 +07:00
|
|
|
static void tce_free_pSeriesLP(unsigned long liobn, long tcenum, long npages)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
|
|
|
u64 rc;
|
|
|
|
|
|
|
|
while (npages--) {
|
2019-12-16 11:19:22 +07:00
|
|
|
rc = plpar_tce_put((u64)liobn, (u64)tcenum << 12, 0);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
if (rc && printk_ratelimit()) {
|
2009-01-06 21:26:03 +07:00
|
|
|
printk("tce_free_pSeriesLP: plpar_tce_put failed. rc=%lld\n", rc);
|
2019-12-16 11:19:22 +07:00
|
|
|
printk("\tindex = 0x%llx\n", (u64)liobn);
|
2009-01-06 21:26:03 +07:00
|
|
|
printk("\ttcenum = 0x%llx\n", (u64)tcenum);
|
2014-10-13 15:41:40 +07:00
|
|
|
dump_stack();
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
tcenum++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static void tce_freemulti_pSeriesLP(struct iommu_table *tbl, long tcenum, long npages)
|
|
|
|
{
|
|
|
|
u64 rc;
|
|
|
|
|
2019-12-16 11:19:23 +07:00
|
|
|
if (!firmware_has_feature(FW_FEATURE_STUFF_TCE))
|
2019-12-16 11:19:22 +07:00
|
|
|
return tce_free_pSeriesLP(tbl->it_index, tcenum, npages);
|
2015-06-05 13:35:06 +07:00
|
|
|
|
2006-04-29 10:51:59 +07:00
|
|
|
rc = plpar_tce_stuff((u64)tbl->it_index, (u64)tcenum << 12, 0, npages);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
if (rc && printk_ratelimit()) {
|
|
|
|
printk("tce_freemulti_pSeriesLP: plpar_tce_stuff failed\n");
|
2009-01-06 21:26:03 +07:00
|
|
|
printk("\trc = %lld\n", rc);
|
|
|
|
printk("\tindex = 0x%llx\n", (u64)tbl->it_index);
|
|
|
|
printk("\tnpages = 0x%llx\n", (u64)npages);
|
2014-10-13 15:41:40 +07:00
|
|
|
dump_stack();
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2006-06-23 13:35:10 +07:00
|
|
|
static unsigned long tce_get_pSeriesLP(struct iommu_table *tbl, long tcenum)
|
|
|
|
{
|
|
|
|
u64 rc;
|
|
|
|
unsigned long tce_ret;
|
|
|
|
|
|
|
|
rc = plpar_tce_get((u64)tbl->it_index, (u64)tcenum << 12, &tce_ret);
|
|
|
|
|
|
|
|
if (rc && printk_ratelimit()) {
|
2009-01-06 21:26:03 +07:00
|
|
|
printk("tce_get_pSeriesLP: plpar_tce_get failed. rc=%lld\n", rc);
|
|
|
|
printk("\tindex = 0x%llx\n", (u64)tbl->it_index);
|
|
|
|
printk("\ttcenum = 0x%llx\n", (u64)tcenum);
|
2014-10-13 15:41:40 +07:00
|
|
|
dump_stack();
|
2006-06-23 13:35:10 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
return tce_ret;
|
|
|
|
}
|
|
|
|
|
2011-03-31 08:57:33 +07:00
|
|
|
/* this is compatible with cells for the device tree property */
|
2011-02-10 16:10:47 +07:00
|
|
|
struct dynamic_dma_window_prop {
|
|
|
|
__be32 liobn; /* tce table number */
|
|
|
|
__be64 dma_base; /* address hi,lo */
|
|
|
|
__be32 tce_shift; /* ilog2(tce_page_size) */
|
|
|
|
__be32 window_shift; /* ilog2(tce_window_size) */
|
|
|
|
};
|
|
|
|
|
|
|
|
struct direct_window {
|
|
|
|
struct device_node *device;
|
|
|
|
const struct dynamic_dma_window_prop *prop;
|
|
|
|
struct list_head list;
|
|
|
|
};
|
|
|
|
|
|
|
|
/* Dynamic DMA Window support */
|
|
|
|
struct ddw_query_response {
|
2014-09-25 13:39:18 +07:00
|
|
|
u32 windows_available;
|
|
|
|
u32 largest_available_block;
|
|
|
|
u32 page_size;
|
|
|
|
u32 migration_capable;
|
2011-02-10 16:10:47 +07:00
|
|
|
};
|
|
|
|
|
|
|
|
struct ddw_create_response {
|
2014-09-25 13:39:18 +07:00
|
|
|
u32 liobn;
|
|
|
|
u32 addr_hi;
|
|
|
|
u32 addr_lo;
|
2011-02-10 16:10:47 +07:00
|
|
|
};
|
|
|
|
|
|
|
|
static LIST_HEAD(direct_window_list);
|
|
|
|
/* prevents races between memory on/offline and window creation */
|
|
|
|
static DEFINE_SPINLOCK(direct_window_list_lock);
|
|
|
|
/* protects initializing window twice for same device */
|
|
|
|
static DEFINE_MUTEX(direct_window_init_mutex);
|
|
|
|
#define DIRECT64_PROPNAME "linux,direct64-ddr-window-info"
|
|
|
|
|
|
|
|
static int tce_clearrange_multi_pSeriesLP(unsigned long start_pfn,
|
|
|
|
unsigned long num_pfn, const void *arg)
|
|
|
|
{
|
|
|
|
const struct dynamic_dma_window_prop *maprange = arg;
|
|
|
|
int rc;
|
|
|
|
u64 tce_size, num_tce, dma_offset, next;
|
|
|
|
u32 tce_shift;
|
|
|
|
long limit;
|
|
|
|
|
|
|
|
tce_shift = be32_to_cpu(maprange->tce_shift);
|
|
|
|
tce_size = 1ULL << tce_shift;
|
|
|
|
next = start_pfn << PAGE_SHIFT;
|
|
|
|
num_tce = num_pfn << PAGE_SHIFT;
|
|
|
|
|
|
|
|
/* round back to the beginning of the tce page size */
|
|
|
|
num_tce += next & (tce_size - 1);
|
|
|
|
next &= ~(tce_size - 1);
|
|
|
|
|
|
|
|
/* covert to number of tces */
|
|
|
|
num_tce |= tce_size - 1;
|
|
|
|
num_tce >>= tce_shift;
|
|
|
|
|
|
|
|
do {
|
|
|
|
/*
|
|
|
|
* Set up the page with TCE data, looping through and setting
|
|
|
|
* the values.
|
|
|
|
*/
|
|
|
|
limit = min_t(long, num_tce, 512);
|
|
|
|
dma_offset = next + be64_to_cpu(maprange->dma_base);
|
|
|
|
|
|
|
|
rc = plpar_tce_stuff((u64)be32_to_cpu(maprange->liobn),
|
|
|
|
dma_offset,
|
|
|
|
0, limit);
|
2013-01-18 16:16:24 +07:00
|
|
|
next += limit * tce_size;
|
2011-02-10 16:10:47 +07:00
|
|
|
num_tce -= limit;
|
|
|
|
} while (num_tce > 0 && !rc);
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int tce_setrange_multi_pSeriesLP(unsigned long start_pfn,
|
|
|
|
unsigned long num_pfn, const void *arg)
|
|
|
|
{
|
|
|
|
const struct dynamic_dma_window_prop *maprange = arg;
|
2013-10-17 19:21:15 +07:00
|
|
|
u64 tce_size, num_tce, dma_offset, next, proto_tce, liobn;
|
|
|
|
__be64 *tcep;
|
2011-02-10 16:10:47 +07:00
|
|
|
u32 tce_shift;
|
|
|
|
u64 rc = 0;
|
|
|
|
long l, limit;
|
|
|
|
|
2019-12-16 11:19:23 +07:00
|
|
|
if (!firmware_has_feature(FW_FEATURE_PUT_TCE_IND)) {
|
2019-12-16 11:19:22 +07:00
|
|
|
unsigned long tceshift = be32_to_cpu(maprange->tce_shift);
|
|
|
|
unsigned long dmastart = (start_pfn << PAGE_SHIFT) +
|
|
|
|
be64_to_cpu(maprange->dma_base);
|
|
|
|
unsigned long tcenum = dmastart >> tceshift;
|
|
|
|
unsigned long npages = num_pfn << PAGE_SHIFT >> tceshift;
|
|
|
|
void *uaddr = __va(start_pfn << PAGE_SHIFT);
|
|
|
|
|
|
|
|
return tce_build_pSeriesLP(be32_to_cpu(maprange->liobn),
|
|
|
|
tcenum, tceshift, npages, (unsigned long) uaddr,
|
|
|
|
DMA_BIDIRECTIONAL, 0);
|
|
|
|
}
|
|
|
|
|
2011-02-10 16:10:47 +07:00
|
|
|
local_irq_disable(); /* to protect tcep and the page behind it */
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 03:23:25 +07:00
|
|
|
tcep = __this_cpu_read(tce_page);
|
2011-02-10 16:10:47 +07:00
|
|
|
|
|
|
|
if (!tcep) {
|
2013-10-17 19:21:15 +07:00
|
|
|
tcep = (__be64 *)__get_free_page(GFP_ATOMIC);
|
2011-02-10 16:10:47 +07:00
|
|
|
if (!tcep) {
|
|
|
|
local_irq_enable();
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
powerpc: Replace __get_cpu_var uses
This still has not been merged and now powerpc is the only arch that does
not have this change. Sorry about missing linuxppc-dev before.
V2->V2
- Fix up to work against 3.18-rc1
__get_cpu_var() is used for multiple purposes in the kernel source. One of
them is address calculation via the form &__get_cpu_var(x). This calculates
the address for the instance of the percpu variable of the current processor
based on an offset.
Other use cases are for storing and retrieving data from the current
processors percpu area. __get_cpu_var() can be used as an lvalue when
writing data or on the right side of an assignment.
__get_cpu_var() is defined as :
__get_cpu_var() always only does an address determination. However, store
and retrieve operations could use a segment prefix (or global register on
other platforms) to avoid the address calculation.
this_cpu_write() and this_cpu_read() can directly take an offset into a
percpu area and use optimized assembly code to read and write per cpu
variables.
This patch converts __get_cpu_var into either an explicit address
calculation using this_cpu_ptr() or into a use of this_cpu operations that
use the offset. Thereby address calculations are avoided and less registers
are used when code is generated.
At the end of the patch set all uses of __get_cpu_var have been removed so
the macro is removed too.
The patch set includes passes over all arches as well. Once these operations
are used throughout then specialized macros can be defined in non -x86
arches as well in order to optimize per cpu access by f.e. using a global
register that may be set to the per cpu base.
Transformations done to __get_cpu_var()
1. Determine the address of the percpu instance of the current processor.
DEFINE_PER_CPU(int, y);
int *x = &__get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(&y);
2. Same as #1 but this time an array structure is involved.
DEFINE_PER_CPU(int, y[20]);
int *x = __get_cpu_var(y);
Converts to
int *x = this_cpu_ptr(y);
3. Retrieve the content of the current processors instance of a per cpu
variable.
DEFINE_PER_CPU(int, y);
int x = __get_cpu_var(y)
Converts to
int x = __this_cpu_read(y);
4. Retrieve the content of a percpu struct
DEFINE_PER_CPU(struct mystruct, y);
struct mystruct x = __get_cpu_var(y);
Converts to
memcpy(&x, this_cpu_ptr(&y), sizeof(x));
5. Assignment to a per cpu variable
DEFINE_PER_CPU(int, y)
__get_cpu_var(y) = x;
Converts to
__this_cpu_write(y, x);
6. Increment/Decrement etc of a per cpu variable
DEFINE_PER_CPU(int, y);
__get_cpu_var(y)++
Converts to
__this_cpu_inc(y)
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
CC: Paul Mackerras <paulus@samba.org>
Signed-off-by: Christoph Lameter <cl@linux.com>
[mpe: Fix build errors caused by set/or_softirq_pending(), and rework
assignment in __set_breakpoint() to use memcpy().]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2014-10-22 03:23:25 +07:00
|
|
|
__this_cpu_write(tce_page, tcep);
|
2011-02-10 16:10:47 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
proto_tce = TCE_PCI_READ | TCE_PCI_WRITE;
|
|
|
|
|
|
|
|
liobn = (u64)be32_to_cpu(maprange->liobn);
|
|
|
|
tce_shift = be32_to_cpu(maprange->tce_shift);
|
|
|
|
tce_size = 1ULL << tce_shift;
|
|
|
|
next = start_pfn << PAGE_SHIFT;
|
|
|
|
num_tce = num_pfn << PAGE_SHIFT;
|
|
|
|
|
|
|
|
/* round back to the beginning of the tce page size */
|
|
|
|
num_tce += next & (tce_size - 1);
|
|
|
|
next &= ~(tce_size - 1);
|
|
|
|
|
|
|
|
/* covert to number of tces */
|
|
|
|
num_tce |= tce_size - 1;
|
|
|
|
num_tce >>= tce_shift;
|
|
|
|
|
|
|
|
/* We can map max one pageful of TCEs at a time */
|
|
|
|
do {
|
|
|
|
/*
|
|
|
|
* Set up the page with TCE data, looping through and setting
|
|
|
|
* the values.
|
|
|
|
*/
|
|
|
|
limit = min_t(long, num_tce, 4096/TCE_ENTRY_SIZE);
|
|
|
|
dma_offset = next + be64_to_cpu(maprange->dma_base);
|
|
|
|
|
|
|
|
for (l = 0; l < limit; l++) {
|
2013-10-17 19:21:15 +07:00
|
|
|
tcep[l] = cpu_to_be64(proto_tce | next);
|
2011-02-10 16:10:47 +07:00
|
|
|
next += tce_size;
|
|
|
|
}
|
|
|
|
|
|
|
|
rc = plpar_tce_put_indirect(liobn,
|
|
|
|
dma_offset,
|
2012-07-26 04:19:57 +07:00
|
|
|
(u64)__pa(tcep),
|
2011-02-10 16:10:47 +07:00
|
|
|
limit);
|
|
|
|
|
|
|
|
num_tce -= limit;
|
|
|
|
} while (num_tce > 0 && !rc);
|
|
|
|
|
|
|
|
/* error cleanup: caller will clear whole range */
|
|
|
|
|
|
|
|
local_irq_enable();
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int tce_setrange_multi_pSeriesLP_walk(unsigned long start_pfn,
|
|
|
|
unsigned long num_pfn, void *arg)
|
|
|
|
{
|
|
|
|
return tce_setrange_multi_pSeriesLP(start_pfn, num_pfn, arg);
|
|
|
|
}
|
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
static void iommu_table_setparms(struct pci_controller *phb,
|
|
|
|
struct device_node *dn,
|
2006-04-29 10:51:59 +07:00
|
|
|
struct iommu_table *tbl)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
|
|
|
struct device_node *node;
|
2016-07-08 13:37:10 +07:00
|
|
|
const unsigned long *basep;
|
2006-10-05 10:28:00 +07:00
|
|
|
const u32 *sizep;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2007-12-10 10:33:21 +07:00
|
|
|
node = phb->dn;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2007-04-03 19:26:41 +07:00
|
|
|
basep = of_get_property(node, "linux,tce-base", NULL);
|
|
|
|
sizep = of_get_property(node, "linux,tce-size", NULL);
|
2005-04-17 05:20:36 +07:00
|
|
|
if (basep == NULL || sizep == NULL) {
|
2017-08-21 22:16:47 +07:00
|
|
|
printk(KERN_ERR "PCI_DMA: iommu_table_setparms: %pOF has "
|
|
|
|
"missing tce entries !\n", dn);
|
2005-04-17 05:20:36 +07:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
tbl->it_base = (unsigned long)__va(*basep);
|
2006-06-23 13:35:10 +07:00
|
|
|
|
2008-10-23 03:39:04 +07:00
|
|
|
if (!is_kdump_kernel())
|
2008-10-22 00:38:10 +07:00
|
|
|
memset((void *)tbl->it_base, 0, *sizep);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
tbl->it_busno = phb->bus->number;
|
2013-12-09 14:17:02 +07:00
|
|
|
tbl->it_page_shift = IOMMU_PAGE_SHIFT_4K;
|
2006-04-29 10:51:59 +07:00
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
/* Units of tce entries */
|
2013-12-09 14:17:02 +07:00
|
|
|
tbl->it_offset = phb->dma_window_base_cur >> tbl->it_page_shift;
|
2006-04-29 10:51:59 +07:00
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
/* Test if we are going over 2GB of DMA space */
|
2005-09-21 23:55:31 +07:00
|
|
|
if (phb->dma_window_base_cur + phb->dma_window_size > 0x80000000ul) {
|
|
|
|
udbg_printf("PCI_DMA: Unexpected number of IOAs under this PHB.\n");
|
2006-04-29 10:51:59 +07:00
|
|
|
panic("PCI_DMA: Unexpected number of IOAs under this PHB.\n");
|
2005-09-21 23:55:31 +07:00
|
|
|
}
|
2006-04-29 10:51:59 +07:00
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
phb->dma_window_base_cur += phb->dma_window_size;
|
|
|
|
|
|
|
|
/* Set the tce table size - measured in entries */
|
2013-12-09 14:17:02 +07:00
|
|
|
tbl->it_size = phb->dma_window_size >> tbl->it_page_shift;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
tbl->it_index = 0;
|
|
|
|
tbl->it_blocksize = 16;
|
|
|
|
tbl->it_type = TCE_PCI;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* iommu_table_setparms_lpar
|
|
|
|
*
|
|
|
|
* Function: On pSeries LPAR systems, return TCE table info, given a pci bus.
|
|
|
|
*/
|
|
|
|
static void iommu_table_setparms_lpar(struct pci_controller *phb,
|
|
|
|
struct device_node *dn,
|
|
|
|
struct iommu_table *tbl,
|
2017-03-24 13:37:21 +07:00
|
|
|
struct iommu_table_group *table_group,
|
2013-08-06 23:01:36 +07:00
|
|
|
const __be32 *dma_window)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2006-05-18 15:06:37 +07:00
|
|
|
unsigned long offset, size;
|
|
|
|
|
|
|
|
of_parse_dma_window(dn, dma_window, &tbl->it_index, &offset, &size);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2010-12-09 11:24:01 +07:00
|
|
|
tbl->it_busno = phb->bus->number;
|
2013-12-09 14:17:02 +07:00
|
|
|
tbl->it_page_shift = IOMMU_PAGE_SHIFT_4K;
|
2005-04-17 05:20:36 +07:00
|
|
|
tbl->it_base = 0;
|
|
|
|
tbl->it_blocksize = 16;
|
|
|
|
tbl->it_type = TCE_PCI;
|
2013-12-09 14:17:02 +07:00
|
|
|
tbl->it_offset = offset >> tbl->it_page_shift;
|
|
|
|
tbl->it_size = size >> tbl->it_page_shift;
|
2017-03-24 13:37:21 +07:00
|
|
|
|
|
|
|
table_group->tce32_start = offset;
|
|
|
|
table_group->tce32_size = size;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2015-06-05 13:35:06 +07:00
|
|
|
struct iommu_table_ops iommu_table_pseries_ops = {
|
|
|
|
.set = tce_build_pSeries,
|
|
|
|
.clear = tce_free_pSeries,
|
|
|
|
.get = tce_get_pseries
|
|
|
|
};
|
|
|
|
|
2006-11-11 13:25:02 +07:00
|
|
|
static void pci_dma_bus_setup_pSeries(struct pci_bus *bus)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2005-09-21 23:55:31 +07:00
|
|
|
struct device_node *dn;
|
2005-04-17 05:20:36 +07:00
|
|
|
struct iommu_table *tbl;
|
2005-09-21 23:55:31 +07:00
|
|
|
struct device_node *isa_dn, *isa_dn_orig;
|
|
|
|
struct device_node *tmp;
|
|
|
|
struct pci_dn *pci;
|
|
|
|
int children;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2005-09-21 23:55:31 +07:00
|
|
|
dn = pci_bus_to_OF_node(bus);
|
2006-11-11 13:25:02 +07:00
|
|
|
|
2017-08-21 22:16:47 +07:00
|
|
|
pr_debug("pci_dma_bus_setup_pSeries: setting up bus %pOF\n", dn);
|
2005-09-21 23:55:31 +07:00
|
|
|
|
|
|
|
if (bus->self) {
|
|
|
|
/* This is not a root bus, any setup will be done for the
|
|
|
|
* device-side of the bridge in iommu_dev_setup_pSeries().
|
|
|
|
*/
|
|
|
|
return;
|
|
|
|
}
|
2006-11-11 13:25:02 +07:00
|
|
|
pci = PCI_DN(dn);
|
2005-09-21 23:55:31 +07:00
|
|
|
|
|
|
|
/* Check if the ISA bus on the system is under
|
|
|
|
* this PHB.
|
2005-04-17 05:20:36 +07:00
|
|
|
*/
|
2005-09-21 23:55:31 +07:00
|
|
|
isa_dn = isa_dn_orig = of_find_node_by_type(NULL, "isa");
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2005-09-21 23:55:31 +07:00
|
|
|
while (isa_dn && isa_dn != dn)
|
|
|
|
isa_dn = isa_dn->parent;
|
|
|
|
|
2014-08-08 17:07:47 +07:00
|
|
|
of_node_put(isa_dn_orig);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2006-06-20 15:00:30 +07:00
|
|
|
/* Count number of direct PCI children of the PHB. */
|
2005-09-21 23:55:31 +07:00
|
|
|
for (children = 0, tmp = dn->child; tmp; tmp = tmp->sibling)
|
2006-06-20 15:00:30 +07:00
|
|
|
children++;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2008-04-24 12:13:19 +07:00
|
|
|
pr_debug("Children: %d\n", children);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2005-09-21 23:55:31 +07:00
|
|
|
/* Calculate amount of DMA window per slot. Each window must be
|
|
|
|
* a power of two (due to pci_alloc_consistent requirements).
|
|
|
|
*
|
|
|
|
* Keep 256MB aside for PHBs with ISA.
|
|
|
|
*/
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2005-09-21 23:55:31 +07:00
|
|
|
if (!isa_dn) {
|
|
|
|
/* No ISA/IDE - just set window size and return */
|
|
|
|
pci->phb->dma_window_size = 0x80000000ul; /* To be divided */
|
|
|
|
|
|
|
|
while (pci->phb->dma_window_size * children > 0x80000000ul)
|
|
|
|
pci->phb->dma_window_size >>= 1;
|
2009-06-03 01:21:30 +07:00
|
|
|
pr_debug("No ISA/IDE, window size is 0x%llx\n",
|
2008-04-24 12:13:19 +07:00
|
|
|
pci->phb->dma_window_size);
|
2005-09-21 23:55:31 +07:00
|
|
|
pci->phb->dma_window_base_cur = 0;
|
|
|
|
|
|
|
|
return;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
2005-09-21 23:55:31 +07:00
|
|
|
|
|
|
|
/* If we have ISA, then we probably have an IDE
|
|
|
|
* controller too. Allocate a 128MB table but
|
|
|
|
* skip the first 128MB to avoid stepping on ISA
|
|
|
|
* space.
|
|
|
|
*/
|
|
|
|
pci->phb->dma_window_size = 0x8000000ul;
|
|
|
|
pci->phb->dma_window_base_cur = 0x8000000ul;
|
|
|
|
|
2015-06-05 13:35:08 +07:00
|
|
|
pci->table_group = iommu_pseries_alloc_group(pci->phb->node);
|
|
|
|
tbl = pci->table_group->tables[0];
|
2005-09-21 23:55:31 +07:00
|
|
|
|
|
|
|
iommu_table_setparms(pci->phb, dn, tbl);
|
2015-06-05 13:35:06 +07:00
|
|
|
tbl->it_ops = &iommu_table_pseries_ops;
|
2019-07-18 12:11:39 +07:00
|
|
|
iommu_init_table(tbl, pci->phb->node, 0, 0);
|
2005-09-21 23:55:31 +07:00
|
|
|
|
|
|
|
/* Divide the rest (1.75GB) among the children */
|
|
|
|
pci->phb->dma_window_size = 0x80000000ul;
|
|
|
|
while (pci->phb->dma_window_size * children > 0x70000000ul)
|
|
|
|
pci->phb->dma_window_size >>= 1;
|
|
|
|
|
2009-06-03 01:21:30 +07:00
|
|
|
pr_debug("ISA/IDE, window size is 0x%llx\n", pci->phb->dma_window_size);
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2017-03-24 13:37:21 +07:00
|
|
|
#ifdef CONFIG_IOMMU_API
|
|
|
|
static int tce_exchange_pseries(struct iommu_table *tbl, long index, unsigned
|
2019-08-29 15:52:51 +07:00
|
|
|
long *tce, enum dma_data_direction *direction,
|
|
|
|
bool realmode)
|
2017-03-24 13:37:21 +07:00
|
|
|
{
|
|
|
|
long rc;
|
|
|
|
unsigned long ioba = (unsigned long) index << tbl->it_page_shift;
|
|
|
|
unsigned long flags, oldtce = 0;
|
|
|
|
u64 proto_tce = iommu_direction_to_tce_perm(*direction);
|
|
|
|
unsigned long newtce = *tce | proto_tce;
|
|
|
|
|
|
|
|
spin_lock_irqsave(&tbl->large_pool.lock, flags);
|
|
|
|
|
|
|
|
rc = plpar_tce_get((u64)tbl->it_index, ioba, &oldtce);
|
|
|
|
if (!rc)
|
|
|
|
rc = plpar_tce_put((u64)tbl->it_index, ioba, newtce);
|
|
|
|
|
|
|
|
if (!rc) {
|
|
|
|
*direction = iommu_tce_direction(oldtce);
|
|
|
|
*tce = oldtce & ~(TCE_PCI_READ | TCE_PCI_WRITE);
|
|
|
|
}
|
|
|
|
|
|
|
|
spin_unlock_irqrestore(&tbl->large_pool.lock, flags);
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2015-06-05 13:35:06 +07:00
|
|
|
struct iommu_table_ops iommu_table_lpar_multi_ops = {
|
|
|
|
.set = tce_buildmulti_pSeriesLP,
|
2017-03-24 13:37:21 +07:00
|
|
|
#ifdef CONFIG_IOMMU_API
|
2019-08-29 15:52:51 +07:00
|
|
|
.xchg_no_kill = tce_exchange_pseries,
|
2017-03-24 13:37:21 +07:00
|
|
|
#endif
|
2015-06-05 13:35:06 +07:00
|
|
|
.clear = tce_freemulti_pSeriesLP,
|
|
|
|
.get = tce_get_pSeriesLP
|
|
|
|
};
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2006-11-11 13:25:02 +07:00
|
|
|
static void pci_dma_bus_setup_pSeriesLP(struct pci_bus *bus)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
|
|
|
struct iommu_table *tbl;
|
|
|
|
struct device_node *dn, *pdn;
|
2005-09-06 10:17:54 +07:00
|
|
|
struct pci_dn *ppci;
|
2013-08-06 23:01:36 +07:00
|
|
|
const __be32 *dma_window = NULL;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
dn = pci_bus_to_OF_node(bus);
|
|
|
|
|
2017-08-21 22:16:47 +07:00
|
|
|
pr_debug("pci_dma_bus_setup_pSeriesLP: setting up bus %pOF\n",
|
|
|
|
dn);
|
2006-11-11 13:25:02 +07:00
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
/* Find nearest ibm,dma-window, walking up the device tree */
|
|
|
|
for (pdn = dn; pdn != NULL; pdn = pdn->parent) {
|
2007-04-03 19:26:41 +07:00
|
|
|
dma_window = of_get_property(pdn, "ibm,dma-window", NULL);
|
2005-04-17 05:20:36 +07:00
|
|
|
if (dma_window != NULL)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (dma_window == NULL) {
|
2008-04-24 12:13:19 +07:00
|
|
|
pr_debug(" no ibm,dma-window property !\n");
|
2005-04-17 05:20:36 +07:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2005-12-06 08:37:35 +07:00
|
|
|
ppci = PCI_DN(pdn);
|
2006-11-11 13:25:02 +07:00
|
|
|
|
2017-08-21 22:16:47 +07:00
|
|
|
pr_debug(" parent is %pOF, iommu_table: 0x%p\n",
|
|
|
|
pdn, ppci->table_group);
|
2006-11-11 13:25:02 +07:00
|
|
|
|
2015-06-05 13:35:08 +07:00
|
|
|
if (!ppci->table_group) {
|
|
|
|
ppci->table_group = iommu_pseries_alloc_group(ppci->phb->node);
|
|
|
|
tbl = ppci->table_group->tables[0];
|
2017-03-24 13:37:21 +07:00
|
|
|
iommu_table_setparms_lpar(ppci->phb, pdn, tbl,
|
|
|
|
ppci->table_group, dma_window);
|
2015-06-05 13:35:06 +07:00
|
|
|
tbl->it_ops = &iommu_table_lpar_multi_ops;
|
2019-07-18 12:11:39 +07:00
|
|
|
iommu_init_table(tbl, ppci->phb->node, 0, 0);
|
2015-06-05 13:35:08 +07:00
|
|
|
iommu_register_group(ppci->table_group,
|
|
|
|
pci_domain_nr(bus), 0);
|
|
|
|
pr_debug(" created table: %p\n", ppci->table_group);
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2006-11-11 13:25:02 +07:00
|
|
|
static void pci_dma_dev_setup_pSeries(struct pci_dev *dev)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
2006-11-11 13:25:02 +07:00
|
|
|
struct device_node *dn;
|
2005-09-21 23:55:31 +07:00
|
|
|
struct iommu_table *tbl;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2008-04-24 12:13:19 +07:00
|
|
|
pr_debug("pci_dma_dev_setup_pSeries: %s\n", pci_name(dev));
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2010-04-14 06:12:56 +07:00
|
|
|
dn = dev->dev.of_node;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2005-09-21 23:55:31 +07:00
|
|
|
/* If we're the direct child of a root bus, then we need to allocate
|
|
|
|
* an iommu table ourselves. The bus setup code should have setup
|
|
|
|
* the window sizes already.
|
|
|
|
*/
|
|
|
|
if (!dev->bus->self) {
|
2006-11-11 13:25:02 +07:00
|
|
|
struct pci_controller *phb = PCI_DN(dn)->phb;
|
|
|
|
|
2008-04-24 12:13:19 +07:00
|
|
|
pr_debug(" --> first child, no bridge. Allocating iommu table.\n");
|
2015-06-05 13:35:08 +07:00
|
|
|
PCI_DN(dn)->table_group = iommu_pseries_alloc_group(phb->node);
|
|
|
|
tbl = PCI_DN(dn)->table_group->tables[0];
|
2006-11-11 13:25:02 +07:00
|
|
|
iommu_table_setparms(phb, dn, tbl);
|
2015-06-05 13:35:06 +07:00
|
|
|
tbl->it_ops = &iommu_table_pseries_ops;
|
2019-07-18 12:11:39 +07:00
|
|
|
iommu_init_table(tbl, phb->node, 0, 0);
|
2015-06-05 13:34:54 +07:00
|
|
|
set_iommu_table_base(&dev->dev, tbl);
|
2005-09-21 23:55:31 +07:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* If this device is further down the bus tree, search upwards until
|
|
|
|
* an already allocated iommu table is found and use that.
|
|
|
|
*/
|
|
|
|
|
2015-06-05 13:35:08 +07:00
|
|
|
while (dn && PCI_DN(dn) && PCI_DN(dn)->table_group == NULL)
|
2005-04-17 05:20:36 +07:00
|
|
|
dn = dn->parent;
|
|
|
|
|
2018-12-19 15:52:20 +07:00
|
|
|
if (dn && PCI_DN(dn))
|
2015-06-05 13:35:08 +07:00
|
|
|
set_iommu_table_base(&dev->dev,
|
|
|
|
PCI_DN(dn)->table_group->tables[0]);
|
2018-12-19 15:52:20 +07:00
|
|
|
else
|
2006-11-11 13:25:02 +07:00
|
|
|
printk(KERN_WARNING "iommu: Device %s has no iommu table\n",
|
|
|
|
pci_name(dev));
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2011-02-10 16:10:47 +07:00
|
|
|
static int __read_mostly disable_ddw;
|
|
|
|
|
|
|
|
static int __init disable_ddw_setup(char *str)
|
|
|
|
{
|
|
|
|
disable_ddw = 1;
|
|
|
|
printk(KERN_INFO "ppc iommu: disabling ddw.\n");
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
early_param("disable_ddw", disable_ddw_setup);
|
|
|
|
|
2014-08-11 16:16:20 +07:00
|
|
|
static void remove_ddw(struct device_node *np, bool remove_prop)
|
2011-02-10 16:10:47 +07:00
|
|
|
{
|
|
|
|
struct dynamic_dma_window_prop *dwp;
|
|
|
|
struct property *win64;
|
2014-09-25 13:39:18 +07:00
|
|
|
u32 ddw_avail[3];
|
2011-02-10 16:10:47 +07:00
|
|
|
u64 liobn;
|
2014-09-25 13:39:18 +07:00
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
ret = of_property_read_u32_array(np, "ibm,ddw-applicable",
|
|
|
|
&ddw_avail[0], 3);
|
2011-02-10 16:10:47 +07:00
|
|
|
|
|
|
|
win64 = of_find_property(np, DIRECT64_PROPNAME, NULL);
|
2011-05-11 19:24:58 +07:00
|
|
|
if (!win64)
|
2011-02-10 16:10:47 +07:00
|
|
|
return;
|
|
|
|
|
2014-09-25 13:39:18 +07:00
|
|
|
if (ret || win64->length < sizeof(*dwp))
|
2011-05-11 19:24:58 +07:00
|
|
|
goto delprop;
|
|
|
|
|
2011-02-10 16:10:47 +07:00
|
|
|
dwp = win64->value;
|
|
|
|
liobn = (u64)be32_to_cpu(dwp->liobn);
|
|
|
|
|
|
|
|
/* clear the whole window, note the arg is in kernel pages */
|
|
|
|
ret = tce_clearrange_multi_pSeriesLP(0,
|
|
|
|
1ULL << (be32_to_cpu(dwp->window_shift) - PAGE_SHIFT), dwp);
|
|
|
|
if (ret)
|
2016-10-25 11:00:08 +07:00
|
|
|
pr_warn("%pOF failed to clear tces in window.\n",
|
|
|
|
np);
|
2011-02-10 16:10:47 +07:00
|
|
|
else
|
2017-08-21 22:16:47 +07:00
|
|
|
pr_debug("%pOF successfully cleared tces in window.\n",
|
|
|
|
np);
|
2011-02-10 16:10:47 +07:00
|
|
|
|
2014-01-11 06:09:38 +07:00
|
|
|
ret = rtas_call(ddw_avail[2], 1, 1, NULL, liobn);
|
|
|
|
if (ret)
|
2016-10-25 11:00:08 +07:00
|
|
|
pr_warn("%pOF: failed to remove direct window: rtas returned "
|
2014-01-11 06:09:38 +07:00
|
|
|
"%d to ibm,remove-pe-dma-window(%x) %llx\n",
|
2017-08-21 22:16:47 +07:00
|
|
|
np, ret, ddw_avail[2], liobn);
|
2014-01-11 06:09:38 +07:00
|
|
|
else
|
2017-08-21 22:16:47 +07:00
|
|
|
pr_debug("%pOF: successfully removed direct window: rtas returned "
|
2014-01-11 06:09:38 +07:00
|
|
|
"%d to ibm,remove-pe-dma-window(%x) %llx\n",
|
2017-08-21 22:16:47 +07:00
|
|
|
np, ret, ddw_avail[2], liobn);
|
2011-02-10 16:10:47 +07:00
|
|
|
|
2011-05-11 19:24:58 +07:00
|
|
|
delprop:
|
2014-08-11 16:16:20 +07:00
|
|
|
if (remove_prop)
|
|
|
|
ret = of_remove_property(np, win64);
|
2011-05-11 19:24:58 +07:00
|
|
|
if (ret)
|
2016-10-25 11:00:08 +07:00
|
|
|
pr_warn("%pOF: failed to remove direct window property: %d\n",
|
2017-08-21 22:16:47 +07:00
|
|
|
np, ret);
|
2011-05-11 19:24:58 +07:00
|
|
|
}
|
2011-02-10 16:10:47 +07:00
|
|
|
|
2011-05-11 19:25:00 +07:00
|
|
|
static u64 find_existing_ddw(struct device_node *pdn)
|
2011-02-10 16:10:47 +07:00
|
|
|
{
|
|
|
|
struct direct_window *window;
|
|
|
|
const struct dynamic_dma_window_prop *direct64;
|
|
|
|
u64 dma_addr = 0;
|
|
|
|
|
|
|
|
spin_lock(&direct_window_list_lock);
|
|
|
|
/* check if we already created a window and dupe that config if so */
|
|
|
|
list_for_each_entry(window, &direct_window_list, list) {
|
|
|
|
if (window->device == pdn) {
|
|
|
|
direct64 = window->prop;
|
2013-10-17 19:21:15 +07:00
|
|
|
dma_addr = be64_to_cpu(direct64->dma_base);
|
2011-02-10 16:10:47 +07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
spin_unlock(&direct_window_list_lock);
|
|
|
|
|
|
|
|
return dma_addr;
|
|
|
|
}
|
|
|
|
|
2011-05-11 19:24:59 +07:00
|
|
|
static int find_existing_ddw_windows(void)
|
2011-02-10 16:10:47 +07:00
|
|
|
{
|
2014-01-11 06:10:41 +07:00
|
|
|
int len;
|
2011-05-11 19:24:59 +07:00
|
|
|
struct device_node *pdn;
|
2014-01-11 06:10:41 +07:00
|
|
|
struct direct_window *window;
|
2011-02-10 16:10:47 +07:00
|
|
|
const struct dynamic_dma_window_prop *direct64;
|
|
|
|
|
2011-05-11 19:24:59 +07:00
|
|
|
if (!firmware_has_feature(FW_FEATURE_LPAR))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
for_each_node_with_property(pdn, DIRECT64_PROPNAME) {
|
2014-01-11 06:10:41 +07:00
|
|
|
direct64 = of_get_property(pdn, DIRECT64_PROPNAME, &len);
|
2011-05-11 19:24:59 +07:00
|
|
|
if (!direct64)
|
|
|
|
continue;
|
|
|
|
|
2014-01-11 06:10:41 +07:00
|
|
|
window = kzalloc(sizeof(*window), GFP_KERNEL);
|
|
|
|
if (!window || len < sizeof(struct dynamic_dma_window_prop)) {
|
|
|
|
kfree(window);
|
2014-08-11 16:16:20 +07:00
|
|
|
remove_ddw(pdn, true);
|
2014-01-11 06:10:41 +07:00
|
|
|
continue;
|
|
|
|
}
|
2011-05-11 19:24:59 +07:00
|
|
|
|
2014-01-11 06:10:41 +07:00
|
|
|
window->device = pdn;
|
|
|
|
window->prop = direct64;
|
|
|
|
spin_lock(&direct_window_list_lock);
|
|
|
|
list_add(&window->list, &direct_window_list);
|
|
|
|
spin_unlock(&direct_window_list_lock);
|
2011-02-10 16:10:47 +07:00
|
|
|
}
|
|
|
|
|
2011-05-11 19:24:59 +07:00
|
|
|
return 0;
|
2011-02-10 16:10:47 +07:00
|
|
|
}
|
2011-05-11 19:24:59 +07:00
|
|
|
machine_arch_initcall(pseries, find_existing_ddw_windows);
|
2011-02-10 16:10:47 +07:00
|
|
|
|
2011-05-11 19:25:00 +07:00
|
|
|
static int query_ddw(struct pci_dev *dev, const u32 *ddw_avail,
|
2011-02-10 16:10:47 +07:00
|
|
|
struct ddw_query_response *query)
|
|
|
|
{
|
2016-04-12 02:17:23 +07:00
|
|
|
struct device_node *dn;
|
|
|
|
struct pci_dn *pdn;
|
2011-02-10 16:10:47 +07:00
|
|
|
u32 cfg_addr;
|
|
|
|
u64 buid;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Get the config address and phb buid of the PE window.
|
|
|
|
* Rely on eeh to retrieve this for us.
|
|
|
|
* Retrieve them from the pci device, not the node with the
|
|
|
|
* dma-window property
|
|
|
|
*/
|
2016-04-12 02:17:23 +07:00
|
|
|
dn = pci_device_to_OF_node(dev);
|
|
|
|
pdn = PCI_DN(dn);
|
|
|
|
buid = pdn->phb->buid;
|
2016-05-26 06:56:07 +07:00
|
|
|
cfg_addr = ((pdn->busno << 16) | (pdn->devfn << 8));
|
2012-03-21 04:30:28 +07:00
|
|
|
|
2011-05-11 19:25:00 +07:00
|
|
|
ret = rtas_call(ddw_avail[0], 3, 5, (u32 *)query,
|
2011-02-10 16:10:47 +07:00
|
|
|
cfg_addr, BUID_HI(buid), BUID_LO(buid));
|
|
|
|
dev_info(&dev->dev, "ibm,query-pe-dma-windows(%x) %x %x %x"
|
2011-05-11 19:25:00 +07:00
|
|
|
" returned %d\n", ddw_avail[0], cfg_addr, BUID_HI(buid),
|
2011-02-10 16:10:47 +07:00
|
|
|
BUID_LO(buid), ret);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2011-05-11 19:25:00 +07:00
|
|
|
static int create_ddw(struct pci_dev *dev, const u32 *ddw_avail,
|
2011-02-10 16:10:47 +07:00
|
|
|
struct ddw_create_response *create, int page_shift,
|
|
|
|
int window_shift)
|
|
|
|
{
|
2016-04-12 02:17:23 +07:00
|
|
|
struct device_node *dn;
|
|
|
|
struct pci_dn *pdn;
|
2011-02-10 16:10:47 +07:00
|
|
|
u32 cfg_addr;
|
|
|
|
u64 buid;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Get the config address and phb buid of the PE window.
|
|
|
|
* Rely on eeh to retrieve this for us.
|
|
|
|
* Retrieve them from the pci device, not the node with the
|
|
|
|
* dma-window property
|
|
|
|
*/
|
2016-04-12 02:17:23 +07:00
|
|
|
dn = pci_device_to_OF_node(dev);
|
|
|
|
pdn = PCI_DN(dn);
|
|
|
|
buid = pdn->phb->buid;
|
2016-05-26 06:56:07 +07:00
|
|
|
cfg_addr = ((pdn->busno << 16) | (pdn->devfn << 8));
|
2011-02-10 16:10:47 +07:00
|
|
|
|
|
|
|
do {
|
|
|
|
/* extra outputs are LIOBN and dma-addr (hi, lo) */
|
2014-09-25 13:39:18 +07:00
|
|
|
ret = rtas_call(ddw_avail[1], 5, 4, (u32 *)create,
|
|
|
|
cfg_addr, BUID_HI(buid), BUID_LO(buid),
|
|
|
|
page_shift, window_shift);
|
2011-02-10 16:10:47 +07:00
|
|
|
} while (rtas_busy_delay(ret));
|
|
|
|
dev_info(&dev->dev,
|
|
|
|
"ibm,create-pe-dma-window(%x) %x %x %x %x %x returned %d "
|
2011-05-11 19:25:00 +07:00
|
|
|
"(liobn = 0x%x starting addr = %x %x)\n", ddw_avail[1],
|
2011-02-10 16:10:47 +07:00
|
|
|
cfg_addr, BUID_HI(buid), BUID_LO(buid), page_shift,
|
|
|
|
window_shift, ret, create->liobn, create->addr_hi, create->addr_lo);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
powerpc/pseries: close DDW race between functions of adapter
Given a PCI device with multiple functions in a DDW capable slot, the
following situation can be encountered: When the first function sets a
64-bit DMA mask, enable_ddw() will be called and we can fail to properly
configure DDW (the most common reason being the new DMA window's size is
not large enough to map all of an LPAR's memory). With the recent
changes to DDW, we remove the base window in order to determine if the
new window is of sufficient size to cover an LPAR's memory. We correctly
replace the base window if we find that not to be the case. However,
once we go through and re-configured 32-bit DMA via the IOMMU, the next
function of the adapter will go through the same process. And since DDW
is a characteristic of the slot itself, we are most likely going to fail
again. But to determine we are going to fail the second slot, we again
remove the base window -- but that is now in-use by the first
function/driver, which might be issuing I/O already.
To close this window, keep a list of all the failed struct device_nodes
that have failed to configure DDW. If the current device_node is in that
list, just fail out immediately and fall back to 32-bit DMA without
doing any DDW manipulation.
Signed-off-by: Nishanth Aravamudan <nacc@linux.vnet.ibm.com>
Signed-off-by: Michael Ellerman <michael@ellerman.id.au>
2013-03-07 19:33:03 +07:00
|
|
|
struct failed_ddw_pdn {
|
|
|
|
struct device_node *pdn;
|
|
|
|
struct list_head list;
|
|
|
|
};
|
|
|
|
|
|
|
|
static LIST_HEAD(failed_ddw_pdn_list);
|
|
|
|
|
2018-12-19 15:52:18 +07:00
|
|
|
static phys_addr_t ddw_memory_hotplug_max(void)
|
|
|
|
{
|
|
|
|
phys_addr_t max_addr = memory_hotplug_max();
|
|
|
|
struct device_node *memory;
|
|
|
|
|
|
|
|
for_each_node_by_type(memory, "memory") {
|
|
|
|
unsigned long start, size;
|
2019-04-07 09:48:08 +07:00
|
|
|
int n_mem_addr_cells, n_mem_size_cells, len;
|
2018-12-19 15:52:18 +07:00
|
|
|
const __be32 *memcell_buf;
|
|
|
|
|
|
|
|
memcell_buf = of_get_property(memory, "reg", &len);
|
|
|
|
if (!memcell_buf || len <= 0)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
n_mem_addr_cells = of_n_addr_cells(memory);
|
|
|
|
n_mem_size_cells = of_n_size_cells(memory);
|
|
|
|
|
|
|
|
start = of_read_number(memcell_buf, n_mem_addr_cells);
|
|
|
|
memcell_buf += n_mem_addr_cells;
|
|
|
|
size = of_read_number(memcell_buf, n_mem_size_cells);
|
|
|
|
memcell_buf += n_mem_size_cells;
|
|
|
|
|
|
|
|
max_addr = max_t(phys_addr_t, max_addr, start + size);
|
|
|
|
}
|
|
|
|
|
|
|
|
return max_addr;
|
|
|
|
}
|
|
|
|
|
2011-02-10 16:10:47 +07:00
|
|
|
/*
|
|
|
|
* If the PE supports dynamic dma windows, and there is space for a table
|
|
|
|
* that can map all pages in a linear offset, then setup such a table,
|
|
|
|
* and record the dma-offset in the struct device.
|
|
|
|
*
|
|
|
|
* dev: the pci device we are checking
|
|
|
|
* pdn: the parent pe node with the ibm,dma_window property
|
|
|
|
* Future: also check if we can remap the base window for our base page size
|
|
|
|
*
|
2019-02-13 14:01:07 +07:00
|
|
|
* returns the dma offset for use by the direct mapped DMA code.
|
2011-02-10 16:10:47 +07:00
|
|
|
*/
|
|
|
|
static u64 enable_ddw(struct pci_dev *dev, struct device_node *pdn)
|
|
|
|
{
|
|
|
|
int len, ret;
|
|
|
|
struct ddw_query_response query;
|
|
|
|
struct ddw_create_response create;
|
|
|
|
int page_shift;
|
|
|
|
u64 dma_addr, max_addr;
|
|
|
|
struct device_node *dn;
|
2014-09-25 13:39:18 +07:00
|
|
|
u32 ddw_avail[3];
|
2011-02-10 16:10:47 +07:00
|
|
|
struct direct_window *window;
|
2011-05-06 20:27:30 +07:00
|
|
|
struct property *win64;
|
2011-02-10 16:10:47 +07:00
|
|
|
struct dynamic_dma_window_prop *ddwprop;
|
powerpc/pseries: close DDW race between functions of adapter
Given a PCI device with multiple functions in a DDW capable slot, the
following situation can be encountered: When the first function sets a
64-bit DMA mask, enable_ddw() will be called and we can fail to properly
configure DDW (the most common reason being the new DMA window's size is
not large enough to map all of an LPAR's memory). With the recent
changes to DDW, we remove the base window in order to determine if the
new window is of sufficient size to cover an LPAR's memory. We correctly
replace the base window if we find that not to be the case. However,
once we go through and re-configured 32-bit DMA via the IOMMU, the next
function of the adapter will go through the same process. And since DDW
is a characteristic of the slot itself, we are most likely going to fail
again. But to determine we are going to fail the second slot, we again
remove the base window -- but that is now in-use by the first
function/driver, which might be issuing I/O already.
To close this window, keep a list of all the failed struct device_nodes
that have failed to configure DDW. If the current device_node is in that
list, just fail out immediately and fall back to 32-bit DMA without
doing any DDW manipulation.
Signed-off-by: Nishanth Aravamudan <nacc@linux.vnet.ibm.com>
Signed-off-by: Michael Ellerman <michael@ellerman.id.au>
2013-03-07 19:33:03 +07:00
|
|
|
struct failed_ddw_pdn *fpdn;
|
2011-02-10 16:10:47 +07:00
|
|
|
|
|
|
|
mutex_lock(&direct_window_init_mutex);
|
|
|
|
|
2011-05-11 19:25:00 +07:00
|
|
|
dma_addr = find_existing_ddw(pdn);
|
2011-02-10 16:10:47 +07:00
|
|
|
if (dma_addr != 0)
|
|
|
|
goto out_unlock;
|
|
|
|
|
powerpc/pseries: close DDW race between functions of adapter
Given a PCI device with multiple functions in a DDW capable slot, the
following situation can be encountered: When the first function sets a
64-bit DMA mask, enable_ddw() will be called and we can fail to properly
configure DDW (the most common reason being the new DMA window's size is
not large enough to map all of an LPAR's memory). With the recent
changes to DDW, we remove the base window in order to determine if the
new window is of sufficient size to cover an LPAR's memory. We correctly
replace the base window if we find that not to be the case. However,
once we go through and re-configured 32-bit DMA via the IOMMU, the next
function of the adapter will go through the same process. And since DDW
is a characteristic of the slot itself, we are most likely going to fail
again. But to determine we are going to fail the second slot, we again
remove the base window -- but that is now in-use by the first
function/driver, which might be issuing I/O already.
To close this window, keep a list of all the failed struct device_nodes
that have failed to configure DDW. If the current device_node is in that
list, just fail out immediately and fall back to 32-bit DMA without
doing any DDW manipulation.
Signed-off-by: Nishanth Aravamudan <nacc@linux.vnet.ibm.com>
Signed-off-by: Michael Ellerman <michael@ellerman.id.au>
2013-03-07 19:33:03 +07:00
|
|
|
/*
|
|
|
|
* If we already went through this for a previous function of
|
|
|
|
* the same device and failed, we don't want to muck with the
|
|
|
|
* DMA window again, as it will race with in-flight operations
|
|
|
|
* and can lead to EEHs. The above mutex protects access to the
|
|
|
|
* list.
|
|
|
|
*/
|
|
|
|
list_for_each_entry(fpdn, &failed_ddw_pdn_list, list) {
|
2017-08-21 22:16:47 +07:00
|
|
|
if (fpdn->pdn == pdn)
|
powerpc/pseries: close DDW race between functions of adapter
Given a PCI device with multiple functions in a DDW capable slot, the
following situation can be encountered: When the first function sets a
64-bit DMA mask, enable_ddw() will be called and we can fail to properly
configure DDW (the most common reason being the new DMA window's size is
not large enough to map all of an LPAR's memory). With the recent
changes to DDW, we remove the base window in order to determine if the
new window is of sufficient size to cover an LPAR's memory. We correctly
replace the base window if we find that not to be the case. However,
once we go through and re-configured 32-bit DMA via the IOMMU, the next
function of the adapter will go through the same process. And since DDW
is a characteristic of the slot itself, we are most likely going to fail
again. But to determine we are going to fail the second slot, we again
remove the base window -- but that is now in-use by the first
function/driver, which might be issuing I/O already.
To close this window, keep a list of all the failed struct device_nodes
that have failed to configure DDW. If the current device_node is in that
list, just fail out immediately and fall back to 32-bit DMA without
doing any DDW manipulation.
Signed-off-by: Nishanth Aravamudan <nacc@linux.vnet.ibm.com>
Signed-off-by: Michael Ellerman <michael@ellerman.id.au>
2013-03-07 19:33:03 +07:00
|
|
|
goto out_unlock;
|
|
|
|
}
|
|
|
|
|
2011-02-10 16:10:47 +07:00
|
|
|
/*
|
|
|
|
* the ibm,ddw-applicable property holds the tokens for:
|
|
|
|
* ibm,query-pe-dma-window
|
|
|
|
* ibm,create-pe-dma-window
|
|
|
|
* ibm,remove-pe-dma-window
|
|
|
|
* for the given node in that order.
|
|
|
|
* the property is actually in the parent, not the PE
|
|
|
|
*/
|
2014-09-25 13:39:18 +07:00
|
|
|
ret = of_property_read_u32_array(pdn, "ibm,ddw-applicable",
|
|
|
|
&ddw_avail[0], 3);
|
|
|
|
if (ret)
|
2014-01-11 06:09:38 +07:00
|
|
|
goto out_failed;
|
2012-05-15 14:04:32 +07:00
|
|
|
|
2014-01-11 06:09:38 +07:00
|
|
|
/*
|
2011-02-10 16:10:47 +07:00
|
|
|
* Query if there is a second window of size to map the
|
|
|
|
* whole partition. Query returns number of windows, largest
|
|
|
|
* block assigned to PE (partition endpoint), and two bitmasks
|
|
|
|
* of page sizes: supported and supported for migrate-dma.
|
|
|
|
*/
|
|
|
|
dn = pci_device_to_OF_node(dev);
|
2011-05-11 19:25:00 +07:00
|
|
|
ret = query_ddw(dev, ddw_avail, &query);
|
2011-02-10 16:10:47 +07:00
|
|
|
if (ret != 0)
|
2014-01-11 06:09:38 +07:00
|
|
|
goto out_failed;
|
2011-02-10 16:10:47 +07:00
|
|
|
|
|
|
|
if (query.windows_available == 0) {
|
|
|
|
/*
|
|
|
|
* no additional windows are available for this device.
|
|
|
|
* We might be able to reallocate the existing window,
|
|
|
|
* trading in for a larger page size.
|
|
|
|
*/
|
|
|
|
dev_dbg(&dev->dev, "no free dynamic windows");
|
2014-01-11 06:09:38 +07:00
|
|
|
goto out_failed;
|
2011-02-10 16:10:47 +07:00
|
|
|
}
|
2014-09-25 13:39:18 +07:00
|
|
|
if (query.page_size & 4) {
|
2011-02-10 16:10:47 +07:00
|
|
|
page_shift = 24; /* 16MB */
|
2014-09-25 13:39:18 +07:00
|
|
|
} else if (query.page_size & 2) {
|
2011-02-10 16:10:47 +07:00
|
|
|
page_shift = 16; /* 64kB */
|
2014-09-25 13:39:18 +07:00
|
|
|
} else if (query.page_size & 1) {
|
2011-02-10 16:10:47 +07:00
|
|
|
page_shift = 12; /* 4kB */
|
|
|
|
} else {
|
|
|
|
dev_dbg(&dev->dev, "no supported direct page size in mask %x",
|
|
|
|
query.page_size);
|
2014-01-11 06:09:38 +07:00
|
|
|
goto out_failed;
|
2011-02-10 16:10:47 +07:00
|
|
|
}
|
|
|
|
/* verify the window * number of ptes will map the partition */
|
|
|
|
/* check largest block * page size > max memory hotplug addr */
|
2018-12-19 15:52:18 +07:00
|
|
|
max_addr = ddw_memory_hotplug_max();
|
2014-09-25 13:39:18 +07:00
|
|
|
if (query.largest_available_block < (max_addr >> page_shift)) {
|
2017-02-28 05:28:55 +07:00
|
|
|
dev_dbg(&dev->dev, "can't map partition max 0x%llx with %u "
|
2011-02-10 16:10:47 +07:00
|
|
|
"%llu-sized pages\n", max_addr, query.largest_available_block,
|
|
|
|
1ULL << page_shift);
|
2014-01-11 06:09:38 +07:00
|
|
|
goto out_failed;
|
2011-02-10 16:10:47 +07:00
|
|
|
}
|
|
|
|
len = order_base_2(max_addr);
|
|
|
|
win64 = kzalloc(sizeof(struct property), GFP_KERNEL);
|
|
|
|
if (!win64) {
|
|
|
|
dev_info(&dev->dev,
|
|
|
|
"couldn't allocate property for 64bit dma window\n");
|
2014-01-11 06:09:38 +07:00
|
|
|
goto out_failed;
|
2011-02-10 16:10:47 +07:00
|
|
|
}
|
|
|
|
win64->name = kstrdup(DIRECT64_PROPNAME, GFP_KERNEL);
|
|
|
|
win64->value = ddwprop = kmalloc(sizeof(*ddwprop), GFP_KERNEL);
|
2011-05-06 20:27:30 +07:00
|
|
|
win64->length = sizeof(*ddwprop);
|
2011-02-10 16:10:47 +07:00
|
|
|
if (!win64->name || !win64->value) {
|
|
|
|
dev_info(&dev->dev,
|
|
|
|
"couldn't allocate property name and value\n");
|
|
|
|
goto out_free_prop;
|
|
|
|
}
|
|
|
|
|
2011-05-11 19:25:00 +07:00
|
|
|
ret = create_ddw(dev, ddw_avail, &create, page_shift, len);
|
2011-02-10 16:10:47 +07:00
|
|
|
if (ret != 0)
|
|
|
|
goto out_free_prop;
|
|
|
|
|
2014-09-25 13:39:18 +07:00
|
|
|
ddwprop->liobn = cpu_to_be32(create.liobn);
|
|
|
|
ddwprop->dma_base = cpu_to_be64(((u64)create.addr_hi << 32) |
|
|
|
|
create.addr_lo);
|
2011-02-10 16:10:47 +07:00
|
|
|
ddwprop->tce_shift = cpu_to_be32(page_shift);
|
|
|
|
ddwprop->window_shift = cpu_to_be32(len);
|
|
|
|
|
2017-08-21 22:16:47 +07:00
|
|
|
dev_dbg(&dev->dev, "created tce table LIOBN 0x%x for %pOF\n",
|
|
|
|
create.liobn, dn);
|
2011-02-10 16:10:47 +07:00
|
|
|
|
|
|
|
window = kzalloc(sizeof(*window), GFP_KERNEL);
|
|
|
|
if (!window)
|
|
|
|
goto out_clear_window;
|
|
|
|
|
|
|
|
ret = walk_system_ram_range(0, memblock_end_of_DRAM() >> PAGE_SHIFT,
|
|
|
|
win64->value, tce_setrange_multi_pSeriesLP_walk);
|
|
|
|
if (ret) {
|
2017-08-21 22:16:47 +07:00
|
|
|
dev_info(&dev->dev, "failed to map direct window for %pOF: %d\n",
|
|
|
|
dn, ret);
|
2011-08-08 08:18:00 +07:00
|
|
|
goto out_free_window;
|
2011-02-10 16:10:47 +07:00
|
|
|
}
|
|
|
|
|
2012-10-02 23:58:46 +07:00
|
|
|
ret = of_add_property(pdn, win64);
|
2011-02-10 16:10:47 +07:00
|
|
|
if (ret) {
|
2017-08-21 22:16:47 +07:00
|
|
|
dev_err(&dev->dev, "unable to add dma window property for %pOF: %d",
|
|
|
|
pdn, ret);
|
2011-08-08 08:18:00 +07:00
|
|
|
goto out_free_window;
|
2011-02-10 16:10:47 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
window->device = pdn;
|
|
|
|
window->prop = ddwprop;
|
|
|
|
spin_lock(&direct_window_list_lock);
|
|
|
|
list_add(&window->list, &direct_window_list);
|
|
|
|
spin_unlock(&direct_window_list_lock);
|
|
|
|
|
2014-09-25 13:39:18 +07:00
|
|
|
dma_addr = be64_to_cpu(ddwprop->dma_base);
|
2011-02-10 16:10:47 +07:00
|
|
|
goto out_unlock;
|
|
|
|
|
2011-08-08 08:18:00 +07:00
|
|
|
out_free_window:
|
|
|
|
kfree(window);
|
|
|
|
|
2011-02-10 16:10:47 +07:00
|
|
|
out_clear_window:
|
2014-08-11 16:16:20 +07:00
|
|
|
remove_ddw(pdn, true);
|
2011-02-10 16:10:47 +07:00
|
|
|
|
|
|
|
out_free_prop:
|
|
|
|
kfree(win64->name);
|
|
|
|
kfree(win64->value);
|
|
|
|
kfree(win64);
|
|
|
|
|
2014-01-11 06:09:38 +07:00
|
|
|
out_failed:
|
2012-05-15 14:04:32 +07:00
|
|
|
|
powerpc/pseries: close DDW race between functions of adapter
Given a PCI device with multiple functions in a DDW capable slot, the
following situation can be encountered: When the first function sets a
64-bit DMA mask, enable_ddw() will be called and we can fail to properly
configure DDW (the most common reason being the new DMA window's size is
not large enough to map all of an LPAR's memory). With the recent
changes to DDW, we remove the base window in order to determine if the
new window is of sufficient size to cover an LPAR's memory. We correctly
replace the base window if we find that not to be the case. However,
once we go through and re-configured 32-bit DMA via the IOMMU, the next
function of the adapter will go through the same process. And since DDW
is a characteristic of the slot itself, we are most likely going to fail
again. But to determine we are going to fail the second slot, we again
remove the base window -- but that is now in-use by the first
function/driver, which might be issuing I/O already.
To close this window, keep a list of all the failed struct device_nodes
that have failed to configure DDW. If the current device_node is in that
list, just fail out immediately and fall back to 32-bit DMA without
doing any DDW manipulation.
Signed-off-by: Nishanth Aravamudan <nacc@linux.vnet.ibm.com>
Signed-off-by: Michael Ellerman <michael@ellerman.id.au>
2013-03-07 19:33:03 +07:00
|
|
|
fpdn = kzalloc(sizeof(*fpdn), GFP_KERNEL);
|
|
|
|
if (!fpdn)
|
|
|
|
goto out_unlock;
|
|
|
|
fpdn->pdn = pdn;
|
|
|
|
list_add(&fpdn->list, &failed_ddw_pdn_list);
|
|
|
|
|
2011-02-10 16:10:47 +07:00
|
|
|
out_unlock:
|
|
|
|
mutex_unlock(&direct_window_init_mutex);
|
|
|
|
return dma_addr;
|
|
|
|
}
|
|
|
|
|
2006-11-11 13:25:02 +07:00
|
|
|
static void pci_dma_dev_setup_pSeriesLP(struct pci_dev *dev)
|
2005-04-17 05:20:36 +07:00
|
|
|
{
|
|
|
|
struct device_node *pdn, *dn;
|
|
|
|
struct iommu_table *tbl;
|
2013-08-06 23:01:36 +07:00
|
|
|
const __be32 *dma_window = NULL;
|
2005-09-06 10:17:54 +07:00
|
|
|
struct pci_dn *pci;
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2008-04-24 12:13:19 +07:00
|
|
|
pr_debug("pci_dma_dev_setup_pSeriesLP: %s\n", pci_name(dev));
|
2006-11-11 13:25:02 +07:00
|
|
|
|
2005-04-17 05:20:36 +07:00
|
|
|
/* dev setup for LPAR is a little tricky, since the device tree might
|
2011-03-31 08:57:33 +07:00
|
|
|
* contain the dma-window properties per-device and not necessarily
|
2005-04-17 05:20:36 +07:00
|
|
|
* for the bus. So we need to search upwards in the tree until we
|
|
|
|
* either hit a dma-window property, OR find a parent with a table
|
|
|
|
* already allocated.
|
|
|
|
*/
|
|
|
|
dn = pci_device_to_OF_node(dev);
|
2017-08-21 22:16:47 +07:00
|
|
|
pr_debug(" node is %pOF\n", dn);
|
2006-10-30 12:15:59 +07:00
|
|
|
|
2015-06-05 13:35:08 +07:00
|
|
|
for (pdn = dn; pdn && PCI_DN(pdn) && !PCI_DN(pdn)->table_group;
|
2005-09-06 10:17:54 +07:00
|
|
|
pdn = pdn->parent) {
|
2007-04-03 19:26:41 +07:00
|
|
|
dma_window = of_get_property(pdn, "ibm,dma-window", NULL);
|
2005-04-17 05:20:36 +07:00
|
|
|
if (dma_window)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2007-04-11 03:11:23 +07:00
|
|
|
if (!pdn || !PCI_DN(pdn)) {
|
|
|
|
printk(KERN_WARNING "pci_dma_dev_setup_pSeriesLP: "
|
2017-08-21 22:16:47 +07:00
|
|
|
"no DMA window found for pci dev=%s dn=%pOF\n",
|
|
|
|
pci_name(dev), dn);
|
2007-04-11 03:11:23 +07:00
|
|
|
return;
|
|
|
|
}
|
2017-08-21 22:16:47 +07:00
|
|
|
pr_debug(" parent is %pOF\n", pdn);
|
2006-11-11 13:25:02 +07:00
|
|
|
|
2005-12-06 08:37:35 +07:00
|
|
|
pci = PCI_DN(pdn);
|
2015-06-05 13:35:08 +07:00
|
|
|
if (!pci->table_group) {
|
|
|
|
pci->table_group = iommu_pseries_alloc_group(pci->phb->node);
|
|
|
|
tbl = pci->table_group->tables[0];
|
2017-03-24 13:37:21 +07:00
|
|
|
iommu_table_setparms_lpar(pci->phb, pdn, tbl,
|
|
|
|
pci->table_group, dma_window);
|
2015-06-05 13:35:06 +07:00
|
|
|
tbl->it_ops = &iommu_table_lpar_multi_ops;
|
2019-07-18 12:11:39 +07:00
|
|
|
iommu_init_table(tbl, pci->phb->node, 0, 0);
|
2015-06-05 13:35:08 +07:00
|
|
|
iommu_register_group(pci->table_group,
|
|
|
|
pci_domain_nr(pci->phb->bus), 0);
|
|
|
|
pr_debug(" created table: %p\n", pci->table_group);
|
2007-05-10 12:16:27 +07:00
|
|
|
} else {
|
2015-06-05 13:35:08 +07:00
|
|
|
pr_debug(" found DMA window, table: %p\n", pci->table_group);
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2015-06-05 13:35:08 +07:00
|
|
|
set_iommu_table_base(&dev->dev, pci->table_group->tables[0]);
|
2018-12-19 15:52:21 +07:00
|
|
|
iommu_add_device(pci->table_group, &dev->dev);
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
2011-02-10 16:10:47 +07:00
|
|
|
|
2019-02-13 14:01:07 +07:00
|
|
|
static bool iommu_bypass_supported_pSeriesLP(struct pci_dev *pdev, u64 dma_mask)
|
2011-02-10 16:10:47 +07:00
|
|
|
{
|
2019-02-13 14:01:07 +07:00
|
|
|
struct device_node *dn = pci_device_to_OF_node(pdev), *pdn;
|
2013-08-06 23:01:36 +07:00
|
|
|
const __be32 *dma_window = NULL;
|
2011-02-10 16:10:47 +07:00
|
|
|
|
|
|
|
/* only attempt to use a new window if 64-bit DMA is requested */
|
2019-02-13 14:01:07 +07:00
|
|
|
if (dma_mask < DMA_BIT_MASK(64))
|
|
|
|
return false;
|
2011-02-10 16:10:47 +07:00
|
|
|
|
2019-02-13 14:01:07 +07:00
|
|
|
dev_dbg(&pdev->dev, "node is %pOF\n", dn);
|
2011-02-10 16:10:47 +07:00
|
|
|
|
2019-02-13 14:01:07 +07:00
|
|
|
/*
|
|
|
|
* the device tree might contain the dma-window properties
|
|
|
|
* per-device and not necessarily for the bus. So we need to
|
|
|
|
* search upwards in the tree until we either hit a dma-window
|
|
|
|
* property, OR find a parent with a table already allocated.
|
|
|
|
*/
|
|
|
|
for (pdn = dn; pdn && PCI_DN(pdn) && !PCI_DN(pdn)->table_group;
|
|
|
|
pdn = pdn->parent) {
|
|
|
|
dma_window = of_get_property(pdn, "ibm,dma-window", NULL);
|
|
|
|
if (dma_window)
|
|
|
|
break;
|
2011-02-10 16:10:47 +07:00
|
|
|
}
|
|
|
|
|
2019-02-13 14:01:07 +07:00
|
|
|
if (pdn && PCI_DN(pdn)) {
|
2019-02-13 14:01:32 +07:00
|
|
|
pdev->dev.archdata.dma_offset = enable_ddw(pdev, pdn);
|
|
|
|
if (pdev->dev.archdata.dma_offset)
|
2019-02-13 14:01:07 +07:00
|
|
|
return true;
|
2011-06-24 16:05:22 +07:00
|
|
|
}
|
|
|
|
|
2019-02-13 14:01:07 +07:00
|
|
|
return false;
|
2011-06-24 16:05:22 +07:00
|
|
|
}
|
|
|
|
|
2011-02-10 16:10:47 +07:00
|
|
|
static int iommu_mem_notifier(struct notifier_block *nb, unsigned long action,
|
|
|
|
void *data)
|
|
|
|
{
|
|
|
|
struct direct_window *window;
|
|
|
|
struct memory_notify *arg = data;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
switch (action) {
|
|
|
|
case MEM_GOING_ONLINE:
|
|
|
|
spin_lock(&direct_window_list_lock);
|
|
|
|
list_for_each_entry(window, &direct_window_list, list) {
|
|
|
|
ret |= tce_setrange_multi_pSeriesLP(arg->start_pfn,
|
|
|
|
arg->nr_pages, window->prop);
|
|
|
|
/* XXX log error */
|
|
|
|
}
|
|
|
|
spin_unlock(&direct_window_list_lock);
|
|
|
|
break;
|
|
|
|
case MEM_CANCEL_ONLINE:
|
|
|
|
case MEM_OFFLINE:
|
|
|
|
spin_lock(&direct_window_list_lock);
|
|
|
|
list_for_each_entry(window, &direct_window_list, list) {
|
|
|
|
ret |= tce_clearrange_multi_pSeriesLP(arg->start_pfn,
|
|
|
|
arg->nr_pages, window->prop);
|
|
|
|
/* XXX log error */
|
|
|
|
}
|
|
|
|
spin_unlock(&direct_window_list_lock);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (ret && action != MEM_CANCEL_ONLINE)
|
|
|
|
return NOTIFY_BAD;
|
|
|
|
|
|
|
|
return NOTIFY_OK;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct notifier_block iommu_mem_nb = {
|
|
|
|
.notifier_call = iommu_mem_notifier,
|
|
|
|
};
|
|
|
|
|
2014-11-25 00:58:01 +07:00
|
|
|
static int iommu_reconfig_notifier(struct notifier_block *nb, unsigned long action, void *data)
|
2007-03-04 13:04:44 +07:00
|
|
|
{
|
|
|
|
int err = NOTIFY_OK;
|
2014-11-25 00:58:01 +07:00
|
|
|
struct of_reconfig_data *rd = data;
|
|
|
|
struct device_node *np = rd->dn;
|
2007-03-04 13:04:44 +07:00
|
|
|
struct pci_dn *pci = PCI_DN(np);
|
2011-02-10 16:10:47 +07:00
|
|
|
struct direct_window *window;
|
2007-03-04 13:04:44 +07:00
|
|
|
|
|
|
|
switch (action) {
|
2012-10-02 23:57:57 +07:00
|
|
|
case OF_RECONFIG_DETACH_NODE:
|
2014-08-11 16:16:20 +07:00
|
|
|
/*
|
|
|
|
* Removing the property will invoke the reconfig
|
|
|
|
* notifier again, which causes dead-lock on the
|
|
|
|
* read-write semaphore of the notifier chain. So
|
|
|
|
* we have to remove the property when releasing
|
|
|
|
* the device node.
|
|
|
|
*/
|
|
|
|
remove_ddw(np, false);
|
2015-06-05 13:35:08 +07:00
|
|
|
if (pci && pci->table_group)
|
|
|
|
iommu_pseries_free_group(pci->table_group,
|
2015-06-05 13:34:56 +07:00
|
|
|
np->full_name);
|
2011-02-10 16:10:47 +07:00
|
|
|
|
|
|
|
spin_lock(&direct_window_list_lock);
|
|
|
|
list_for_each_entry(window, &direct_window_list, list) {
|
|
|
|
if (window->device == np) {
|
|
|
|
list_del(&window->list);
|
|
|
|
kfree(window);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
spin_unlock(&direct_window_list_lock);
|
2007-03-04 13:04:44 +07:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
err = NOTIFY_DONE;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct notifier_block iommu_reconfig_nb = {
|
|
|
|
.notifier_call = iommu_reconfig_notifier,
|
|
|
|
};
|
2005-04-17 05:20:36 +07:00
|
|
|
|
|
|
|
/* These are called very early. */
|
|
|
|
void iommu_init_early_pSeries(void)
|
|
|
|
{
|
2010-10-18 14:27:03 +07:00
|
|
|
if (of_chosen && of_get_property(of_chosen, "linux,iommu-off", NULL))
|
2005-04-17 05:20:36 +07:00
|
|
|
return;
|
|
|
|
|
2006-03-21 16:45:59 +07:00
|
|
|
if (firmware_has_feature(FW_FEATURE_LPAR)) {
|
2015-03-31 12:00:50 +07:00
|
|
|
pseries_pci_controller_ops.dma_bus_setup = pci_dma_bus_setup_pSeriesLP;
|
|
|
|
pseries_pci_controller_ops.dma_dev_setup = pci_dma_dev_setup_pSeriesLP;
|
2019-02-13 14:01:07 +07:00
|
|
|
if (!disable_ddw)
|
|
|
|
pseries_pci_controller_ops.iommu_bypass_supported =
|
|
|
|
iommu_bypass_supported_pSeriesLP;
|
2005-04-17 05:20:36 +07:00
|
|
|
} else {
|
2015-03-31 12:00:50 +07:00
|
|
|
pseries_pci_controller_ops.dma_bus_setup = pci_dma_bus_setup_pSeries;
|
|
|
|
pseries_pci_controller_ops.dma_dev_setup = pci_dma_dev_setup_pSeries;
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
|
2012-10-02 23:57:57 +07:00
|
|
|
of_reconfig_notifier_register(&iommu_reconfig_nb);
|
2011-02-10 16:10:47 +07:00
|
|
|
register_memory_notifier(&iommu_mem_nb);
|
2005-04-17 05:20:36 +07:00
|
|
|
|
2019-12-16 11:19:21 +07:00
|
|
|
set_pci_dma_ops(&dma_iommu_ops);
|
2005-04-17 05:20:36 +07:00
|
|
|
}
|
|
|
|
|
2010-09-28 22:33:12 +07:00
|
|
|
static int __init disable_multitce(char *str)
|
|
|
|
{
|
|
|
|
if (strcmp(str, "off") == 0 &&
|
|
|
|
firmware_has_feature(FW_FEATURE_LPAR) &&
|
2019-12-16 11:19:23 +07:00
|
|
|
(firmware_has_feature(FW_FEATURE_PUT_TCE_IND) ||
|
|
|
|
firmware_has_feature(FW_FEATURE_STUFF_TCE))) {
|
2010-09-28 22:33:12 +07:00
|
|
|
printk(KERN_INFO "Disabling MULTITCE firmware feature\n");
|
2019-12-16 11:19:23 +07:00
|
|
|
powerpc_firmware_features &=
|
|
|
|
~(FW_FEATURE_PUT_TCE_IND | FW_FEATURE_STUFF_TCE);
|
2010-09-28 22:33:12 +07:00
|
|
|
}
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
__setup("multitce=", disable_multitce);
|
2015-02-22 02:00:50 +07:00
|
|
|
|
2018-12-19 15:52:21 +07:00
|
|
|
static int tce_iommu_bus_notifier(struct notifier_block *nb,
|
|
|
|
unsigned long action, void *data)
|
|
|
|
{
|
|
|
|
struct device *dev = data;
|
|
|
|
|
|
|
|
switch (action) {
|
|
|
|
case BUS_NOTIFY_DEL_DEVICE:
|
|
|
|
iommu_del_device(dev);
|
|
|
|
return 0;
|
|
|
|
default:
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct notifier_block tce_iommu_bus_nb = {
|
|
|
|
.notifier_call = tce_iommu_bus_notifier,
|
|
|
|
};
|
|
|
|
|
|
|
|
static int __init tce_iommu_bus_notifier_init(void)
|
|
|
|
{
|
|
|
|
bus_register_notifier(&pci_bus_type, &tce_iommu_bus_nb);
|
|
|
|
return 0;
|
|
|
|
}
|
2015-02-22 02:00:50 +07:00
|
|
|
machine_subsys_initcall_sync(pseries, tce_iommu_bus_notifier_init);
|