mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
a4574f63ed
The 'struct resource' in 'struct dev_pagemap' is only used for holding resource span information. The other fields, 'name', 'flags', 'desc', 'parent', 'sibling', and 'child' are all unused wasted space. This is in preparation for introducing a multi-range extension of devm_memremap_pages(). The bulk of this change is unwinding all the places internal to libnvdimm that used 'struct resource' unnecessarily, and replacing instances of 'struct dev_pagemap'.res with 'struct dev_pagemap'.range. P2PDMA had a minor usage of the resource flags field, but only to report failures with "%pR". That is replaced with an open coded print of the range. [dan.carpenter@oracle.com: mm/hmm/test: use after free in dmirror_allocate_chunk()] Link: https://lkml.kernel.org/r/20200926121402.GA7467@kadam Signed-off-by: Dan Williams <dan.j.williams@intel.com> Signed-off-by: Dan Carpenter <dan.carpenter@oracle.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Reviewed-by: Boris Ostrovsky <boris.ostrovsky@oracle.com> [xen] Cc: Paul Mackerras <paulus@ozlabs.org> Cc: Michael Ellerman <mpe@ellerman.id.au> Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org> Cc: Vishal Verma <vishal.l.verma@intel.com> Cc: Vivek Goyal <vgoyal@redhat.com> Cc: Dave Jiang <dave.jiang@intel.com> Cc: Ben Skeggs <bskeggs@redhat.com> Cc: David Airlie <airlied@linux.ie> Cc: Daniel Vetter <daniel@ffwll.ch> Cc: Ira Weiny <ira.weiny@intel.com> Cc: Bjorn Helgaas <bhelgaas@google.com> Cc: Juergen Gross <jgross@suse.com> Cc: Stefano Stabellini <sstabellini@kernel.org> Cc: "Jérôme Glisse" <jglisse@redhat.com> Cc: Andy Lutomirski <luto@kernel.org> Cc: Ard Biesheuvel <ard.biesheuvel@linaro.org> Cc: Ard Biesheuvel <ardb@kernel.org> Cc: Borislav Petkov <bp@alien8.de> Cc: Brice Goglin <Brice.Goglin@inria.fr> Cc: Catalin Marinas <catalin.marinas@arm.com> Cc: Dave Hansen <dave.hansen@linux.intel.com> Cc: David Hildenbrand <david@redhat.com> Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Cc: "H. Peter Anvin" <hpa@zytor.com> Cc: Hulk Robot <hulkci@huawei.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: Jason Gunthorpe <jgg@mellanox.com> Cc: Jason Yan <yanaijie@huawei.com> Cc: Jeff Moyer <jmoyer@redhat.com> Cc: Jia He <justin.he@arm.com> Cc: Joao Martins <joao.m.martins@oracle.com> Cc: Jonathan Cameron <Jonathan.Cameron@huawei.com> Cc: kernel test robot <lkp@intel.com> Cc: Mike Rapoport <rppt@linux.ibm.com> Cc: Pavel Tatashin <pasha.tatashin@soleen.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: "Rafael J. Wysocki" <rafael.j.wysocki@intel.com> Cc: Randy Dunlap <rdunlap@infradead.org> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Tom Lendacky <thomas.lendacky@amd.com> Cc: Wei Yang <richard.weiyang@linux.alibaba.com> Cc: Will Deacon <will@kernel.org> Link: https://lkml.kernel.org/r/159643103173.4062302.768998885691711532.stgit@dwillia2-desk3.amr.corp.intel.com Link: https://lkml.kernel.org/r/160106115761.30709.13539840236873663620.stgit@dwillia2-desk3.amr.corp.intel.com Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
432 lines
11 KiB
C
432 lines
11 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright(c) 2013-2015 Intel Corporation. All rights reserved.
|
|
*/
|
|
#include <linux/memremap.h>
|
|
#include <linux/rculist.h>
|
|
#include <linux/export.h>
|
|
#include <linux/ioport.h>
|
|
#include <linux/module.h>
|
|
#include <linux/types.h>
|
|
#include <linux/pfn_t.h>
|
|
#include <linux/acpi.h>
|
|
#include <linux/io.h>
|
|
#include <linux/mm.h>
|
|
#include "nfit_test.h"
|
|
|
|
static LIST_HEAD(iomap_head);
|
|
|
|
static struct iomap_ops {
|
|
nfit_test_lookup_fn nfit_test_lookup;
|
|
nfit_test_evaluate_dsm_fn evaluate_dsm;
|
|
struct list_head list;
|
|
} iomap_ops = {
|
|
.list = LIST_HEAD_INIT(iomap_ops.list),
|
|
};
|
|
|
|
void nfit_test_setup(nfit_test_lookup_fn lookup,
|
|
nfit_test_evaluate_dsm_fn evaluate)
|
|
{
|
|
iomap_ops.nfit_test_lookup = lookup;
|
|
iomap_ops.evaluate_dsm = evaluate;
|
|
list_add_rcu(&iomap_ops.list, &iomap_head);
|
|
}
|
|
EXPORT_SYMBOL(nfit_test_setup);
|
|
|
|
void nfit_test_teardown(void)
|
|
{
|
|
list_del_rcu(&iomap_ops.list);
|
|
synchronize_rcu();
|
|
}
|
|
EXPORT_SYMBOL(nfit_test_teardown);
|
|
|
|
static struct nfit_test_resource *__get_nfit_res(resource_size_t resource)
|
|
{
|
|
struct iomap_ops *ops;
|
|
|
|
ops = list_first_or_null_rcu(&iomap_head, typeof(*ops), list);
|
|
if (ops)
|
|
return ops->nfit_test_lookup(resource);
|
|
return NULL;
|
|
}
|
|
|
|
struct nfit_test_resource *get_nfit_res(resource_size_t resource)
|
|
{
|
|
struct nfit_test_resource *res;
|
|
|
|
rcu_read_lock();
|
|
res = __get_nfit_res(resource);
|
|
rcu_read_unlock();
|
|
|
|
return res;
|
|
}
|
|
EXPORT_SYMBOL(get_nfit_res);
|
|
|
|
void __iomem *__nfit_test_ioremap(resource_size_t offset, unsigned long size,
|
|
void __iomem *(*fallback_fn)(resource_size_t, unsigned long))
|
|
{
|
|
struct nfit_test_resource *nfit_res = get_nfit_res(offset);
|
|
|
|
if (nfit_res)
|
|
return (void __iomem *) nfit_res->buf + offset
|
|
- nfit_res->res.start;
|
|
return fallback_fn(offset, size);
|
|
}
|
|
|
|
void __iomem *__wrap_devm_ioremap(struct device *dev,
|
|
resource_size_t offset, unsigned long size)
|
|
{
|
|
struct nfit_test_resource *nfit_res = get_nfit_res(offset);
|
|
|
|
if (nfit_res)
|
|
return (void __iomem *) nfit_res->buf + offset
|
|
- nfit_res->res.start;
|
|
return devm_ioremap(dev, offset, size);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_devm_ioremap);
|
|
|
|
void *__wrap_devm_memremap(struct device *dev, resource_size_t offset,
|
|
size_t size, unsigned long flags)
|
|
{
|
|
struct nfit_test_resource *nfit_res = get_nfit_res(offset);
|
|
|
|
if (nfit_res)
|
|
return nfit_res->buf + offset - nfit_res->res.start;
|
|
return devm_memremap(dev, offset, size, flags);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_devm_memremap);
|
|
|
|
static void nfit_test_kill(void *_pgmap)
|
|
{
|
|
struct dev_pagemap *pgmap = _pgmap;
|
|
|
|
WARN_ON(!pgmap || !pgmap->ref);
|
|
|
|
if (pgmap->ops && pgmap->ops->kill)
|
|
pgmap->ops->kill(pgmap);
|
|
else
|
|
percpu_ref_kill(pgmap->ref);
|
|
|
|
if (pgmap->ops && pgmap->ops->cleanup) {
|
|
pgmap->ops->cleanup(pgmap);
|
|
} else {
|
|
wait_for_completion(&pgmap->done);
|
|
percpu_ref_exit(pgmap->ref);
|
|
}
|
|
}
|
|
|
|
static void dev_pagemap_percpu_release(struct percpu_ref *ref)
|
|
{
|
|
struct dev_pagemap *pgmap =
|
|
container_of(ref, struct dev_pagemap, internal_ref);
|
|
|
|
complete(&pgmap->done);
|
|
}
|
|
|
|
void *__wrap_devm_memremap_pages(struct device *dev, struct dev_pagemap *pgmap)
|
|
{
|
|
int error;
|
|
resource_size_t offset = pgmap->range.start;
|
|
struct nfit_test_resource *nfit_res = get_nfit_res(offset);
|
|
|
|
if (!nfit_res)
|
|
return devm_memremap_pages(dev, pgmap);
|
|
|
|
if (!pgmap->ref) {
|
|
if (pgmap->ops && (pgmap->ops->kill || pgmap->ops->cleanup))
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
init_completion(&pgmap->done);
|
|
error = percpu_ref_init(&pgmap->internal_ref,
|
|
dev_pagemap_percpu_release, 0, GFP_KERNEL);
|
|
if (error)
|
|
return ERR_PTR(error);
|
|
pgmap->ref = &pgmap->internal_ref;
|
|
} else {
|
|
if (!pgmap->ops || !pgmap->ops->kill || !pgmap->ops->cleanup) {
|
|
WARN(1, "Missing reference count teardown definition\n");
|
|
return ERR_PTR(-EINVAL);
|
|
}
|
|
}
|
|
|
|
error = devm_add_action_or_reset(dev, nfit_test_kill, pgmap);
|
|
if (error)
|
|
return ERR_PTR(error);
|
|
return nfit_res->buf + offset - nfit_res->res.start;
|
|
}
|
|
EXPORT_SYMBOL_GPL(__wrap_devm_memremap_pages);
|
|
|
|
pfn_t __wrap_phys_to_pfn_t(phys_addr_t addr, unsigned long flags)
|
|
{
|
|
struct nfit_test_resource *nfit_res = get_nfit_res(addr);
|
|
|
|
if (nfit_res)
|
|
flags &= ~PFN_MAP;
|
|
return phys_to_pfn_t(addr, flags);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_phys_to_pfn_t);
|
|
|
|
void *__wrap_memremap(resource_size_t offset, size_t size,
|
|
unsigned long flags)
|
|
{
|
|
struct nfit_test_resource *nfit_res = get_nfit_res(offset);
|
|
|
|
if (nfit_res)
|
|
return nfit_res->buf + offset - nfit_res->res.start;
|
|
return memremap(offset, size, flags);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_memremap);
|
|
|
|
void __wrap_devm_memunmap(struct device *dev, void *addr)
|
|
{
|
|
struct nfit_test_resource *nfit_res = get_nfit_res((long) addr);
|
|
|
|
if (nfit_res)
|
|
return;
|
|
return devm_memunmap(dev, addr);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_devm_memunmap);
|
|
|
|
void __iomem *__wrap_ioremap(resource_size_t offset, unsigned long size)
|
|
{
|
|
return __nfit_test_ioremap(offset, size, ioremap);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_ioremap);
|
|
|
|
void __iomem *__wrap_ioremap_wc(resource_size_t offset, unsigned long size)
|
|
{
|
|
return __nfit_test_ioremap(offset, size, ioremap_wc);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_ioremap_wc);
|
|
|
|
void __wrap_iounmap(volatile void __iomem *addr)
|
|
{
|
|
struct nfit_test_resource *nfit_res = get_nfit_res((long) addr);
|
|
if (nfit_res)
|
|
return;
|
|
return iounmap(addr);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_iounmap);
|
|
|
|
void __wrap_memunmap(void *addr)
|
|
{
|
|
struct nfit_test_resource *nfit_res = get_nfit_res((long) addr);
|
|
|
|
if (nfit_res)
|
|
return;
|
|
return memunmap(addr);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_memunmap);
|
|
|
|
static bool nfit_test_release_region(struct device *dev,
|
|
struct resource *parent, resource_size_t start,
|
|
resource_size_t n);
|
|
|
|
static void nfit_devres_release(struct device *dev, void *data)
|
|
{
|
|
struct resource *res = *((struct resource **) data);
|
|
|
|
WARN_ON(!nfit_test_release_region(NULL, &iomem_resource, res->start,
|
|
resource_size(res)));
|
|
}
|
|
|
|
static int match(struct device *dev, void *__res, void *match_data)
|
|
{
|
|
struct resource *res = *((struct resource **) __res);
|
|
resource_size_t start = *((resource_size_t *) match_data);
|
|
|
|
return res->start == start;
|
|
}
|
|
|
|
static bool nfit_test_release_region(struct device *dev,
|
|
struct resource *parent, resource_size_t start,
|
|
resource_size_t n)
|
|
{
|
|
if (parent == &iomem_resource) {
|
|
struct nfit_test_resource *nfit_res = get_nfit_res(start);
|
|
|
|
if (nfit_res) {
|
|
struct nfit_test_request *req;
|
|
struct resource *res = NULL;
|
|
|
|
if (dev) {
|
|
devres_release(dev, nfit_devres_release, match,
|
|
&start);
|
|
return true;
|
|
}
|
|
|
|
spin_lock(&nfit_res->lock);
|
|
list_for_each_entry(req, &nfit_res->requests, list)
|
|
if (req->res.start == start) {
|
|
res = &req->res;
|
|
list_del(&req->list);
|
|
break;
|
|
}
|
|
spin_unlock(&nfit_res->lock);
|
|
|
|
WARN(!res || resource_size(res) != n,
|
|
"%s: start: %llx n: %llx mismatch: %pr\n",
|
|
__func__, start, n, res);
|
|
if (res)
|
|
kfree(req);
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
static struct resource *nfit_test_request_region(struct device *dev,
|
|
struct resource *parent, resource_size_t start,
|
|
resource_size_t n, const char *name, int flags)
|
|
{
|
|
struct nfit_test_resource *nfit_res;
|
|
|
|
if (parent == &iomem_resource) {
|
|
nfit_res = get_nfit_res(start);
|
|
if (nfit_res) {
|
|
struct nfit_test_request *req;
|
|
struct resource *res = NULL;
|
|
|
|
if (start + n > nfit_res->res.start
|
|
+ resource_size(&nfit_res->res)) {
|
|
pr_debug("%s: start: %llx n: %llx overflow: %pr\n",
|
|
__func__, start, n,
|
|
&nfit_res->res);
|
|
return NULL;
|
|
}
|
|
|
|
spin_lock(&nfit_res->lock);
|
|
list_for_each_entry(req, &nfit_res->requests, list)
|
|
if (start == req->res.start) {
|
|
res = &req->res;
|
|
break;
|
|
}
|
|
spin_unlock(&nfit_res->lock);
|
|
|
|
if (res) {
|
|
WARN(1, "%pr already busy\n", res);
|
|
return NULL;
|
|
}
|
|
|
|
req = kzalloc(sizeof(*req), GFP_KERNEL);
|
|
if (!req)
|
|
return NULL;
|
|
INIT_LIST_HEAD(&req->list);
|
|
res = &req->res;
|
|
|
|
res->start = start;
|
|
res->end = start + n - 1;
|
|
res->name = name;
|
|
res->flags = resource_type(parent);
|
|
res->flags |= IORESOURCE_BUSY | flags;
|
|
spin_lock(&nfit_res->lock);
|
|
list_add(&req->list, &nfit_res->requests);
|
|
spin_unlock(&nfit_res->lock);
|
|
|
|
if (dev) {
|
|
struct resource **d;
|
|
|
|
d = devres_alloc(nfit_devres_release,
|
|
sizeof(struct resource *),
|
|
GFP_KERNEL);
|
|
if (!d)
|
|
return NULL;
|
|
*d = res;
|
|
devres_add(dev, d);
|
|
}
|
|
|
|
pr_debug("%s: %pr\n", __func__, res);
|
|
return res;
|
|
}
|
|
}
|
|
if (dev)
|
|
return __devm_request_region(dev, parent, start, n, name);
|
|
return __request_region(parent, start, n, name, flags);
|
|
}
|
|
|
|
struct resource *__wrap___request_region(struct resource *parent,
|
|
resource_size_t start, resource_size_t n, const char *name,
|
|
int flags)
|
|
{
|
|
return nfit_test_request_region(NULL, parent, start, n, name, flags);
|
|
}
|
|
EXPORT_SYMBOL(__wrap___request_region);
|
|
|
|
int __wrap_insert_resource(struct resource *parent, struct resource *res)
|
|
{
|
|
if (get_nfit_res(res->start))
|
|
return 0;
|
|
return insert_resource(parent, res);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_insert_resource);
|
|
|
|
int __wrap_remove_resource(struct resource *res)
|
|
{
|
|
if (get_nfit_res(res->start))
|
|
return 0;
|
|
return remove_resource(res);
|
|
}
|
|
EXPORT_SYMBOL(__wrap_remove_resource);
|
|
|
|
struct resource *__wrap___devm_request_region(struct device *dev,
|
|
struct resource *parent, resource_size_t start,
|
|
resource_size_t n, const char *name)
|
|
{
|
|
if (!dev)
|
|
return NULL;
|
|
return nfit_test_request_region(dev, parent, start, n, name, 0);
|
|
}
|
|
EXPORT_SYMBOL(__wrap___devm_request_region);
|
|
|
|
void __wrap___release_region(struct resource *parent, resource_size_t start,
|
|
resource_size_t n)
|
|
{
|
|
if (!nfit_test_release_region(NULL, parent, start, n))
|
|
__release_region(parent, start, n);
|
|
}
|
|
EXPORT_SYMBOL(__wrap___release_region);
|
|
|
|
void __wrap___devm_release_region(struct device *dev, struct resource *parent,
|
|
resource_size_t start, resource_size_t n)
|
|
{
|
|
if (!nfit_test_release_region(dev, parent, start, n))
|
|
__devm_release_region(dev, parent, start, n);
|
|
}
|
|
EXPORT_SYMBOL(__wrap___devm_release_region);
|
|
|
|
acpi_status __wrap_acpi_evaluate_object(acpi_handle handle, acpi_string path,
|
|
struct acpi_object_list *p, struct acpi_buffer *buf)
|
|
{
|
|
struct nfit_test_resource *nfit_res = get_nfit_res((long) handle);
|
|
union acpi_object **obj;
|
|
|
|
if (!nfit_res || strcmp(path, "_FIT") || !buf)
|
|
return acpi_evaluate_object(handle, path, p, buf);
|
|
|
|
obj = nfit_res->buf;
|
|
buf->length = sizeof(union acpi_object);
|
|
buf->pointer = *obj;
|
|
return AE_OK;
|
|
}
|
|
EXPORT_SYMBOL(__wrap_acpi_evaluate_object);
|
|
|
|
union acpi_object * __wrap_acpi_evaluate_dsm(acpi_handle handle, const guid_t *guid,
|
|
u64 rev, u64 func, union acpi_object *argv4)
|
|
{
|
|
union acpi_object *obj = ERR_PTR(-ENXIO);
|
|
struct iomap_ops *ops;
|
|
|
|
rcu_read_lock();
|
|
ops = list_first_or_null_rcu(&iomap_head, typeof(*ops), list);
|
|
if (ops)
|
|
obj = ops->evaluate_dsm(handle, guid, rev, func, argv4);
|
|
rcu_read_unlock();
|
|
|
|
if (IS_ERR(obj))
|
|
return acpi_evaluate_dsm(handle, guid, rev, func, argv4);
|
|
return obj;
|
|
}
|
|
EXPORT_SYMBOL(__wrap_acpi_evaluate_dsm);
|
|
|
|
MODULE_LICENSE("GPL v2");
|