linux_dsm_epyc7002/drivers/pci/hotplug/pnv_php.c
Linus Torvalds 685f7e4f16 powerpc updates for 4.20
Notable changes:
 
  - A large series to rewrite our SLB miss handling, replacing a lot of fairly
    complicated asm with much fewer lines of C.
 
  - Following on from that, we now maintain a cache of SLB entries for each
    process and preload them on context switch. Leading to a 27% speedup for our
    context switch benchmark on Power9.
 
  - Improvements to our handling of SLB multi-hit errors. We now print more debug
    information when they occur, and try to continue running by flushing the SLB
    and reloading, rather than treating them as fatal.
 
  - Enable THP migration on 64-bit Book3S machines (eg. Power7/8/9).
 
  - Add support for physical memory up to 2PB in the linear mapping on 64-bit
    Book3S. We only support up to 512TB as regular system memory, otherwise the
    percpu allocator runs out of vmalloc space.
 
  - Add stack protector support for 32 and 64-bit, with a per-task canary.
 
  - Add support for PTRACE_SYSEMU and PTRACE_SYSEMU_SINGLESTEP.
 
  - Support recognising "big cores" on Power9, where two SMT4 cores are presented
    to us as a single SMT8 core.
 
  - A large series to cleanup some of our ioremap handling and PTE flags.
 
  - Add a driver for the PAPR SCM (storage class memory) interface, allowing
    guests to operate on SCM devices (acked by Dan).
 
  - Changes to our ftrace code to handle very large kernels, where we need to use
    a trampoline to get to ftrace_caller().
 
 Many other smaller enhancements and cleanups.
 
 Thanks to:
   Alan Modra, Alistair Popple, Aneesh Kumar K.V, Anton Blanchard, Aravinda
   Prasad, Bartlomiej Zolnierkiewicz, Benjamin Herrenschmidt, Breno Leitao,
   Cédric Le Goater, Christophe Leroy, Christophe Lombard, Dan Carpenter, Daniel
   Axtens, Finn Thain, Gautham R. Shenoy, Gustavo Romero, Haren Myneni, Hari
   Bathini, Jia Hongtao, Joel Stanley, John Allen, Laurent Dufour, Madhavan
   Srinivasan, Mahesh Salgaonkar, Mark Hairgrove, Masahiro Yamada, Michael
   Bringmann, Michael Neuling, Michal Suchanek, Murilo Opsfelder Araujo, Nathan
   Fontenot, Naveen N. Rao, Nicholas Piggin, Nick Desaulniers, Oliver O'Halloran,
   Paul Mackerras, Petr Vorel, Rashmica Gupta, Reza Arbab, Rob Herring, Sam
   Bobroff, Samuel Mendoza-Jonas, Scott Wood, Stan Johnson, Stephen Rothwell,
   Stewart Smith, Suraj Jitindar Singh, Tyrel Datwyler, Vaibhav Jain, Vasant
   Hegde, YueHaibing, zhong jiang,
 -----BEGIN PGP SIGNATURE-----
 
 iQIcBAABAgAGBQJb01vTAAoJEFHr6jzI4aWADsEP/jqL3+2qxs098ra80tpXCpXJ
 tgXCosEs4b35sGtyHeUWZZZfWXeisaPAIlP8zTx1n50HACZduDYRAl0Ew9XB7Xdw
 enDHRVccD21FsmHBOx/Ii1rVJlovWlj6EQCWHKeZmNjeRoFuClVZ7CYmf+mBifKR
 sw2Db2fKA/59wMTq2zIMy5pqYgqlAs4jTWS6uN5hKPoBmO/82ARnNG+qgLuloD3Z
 O8zSDM9QQ7PpuyDgTjO9SAo2YjmEfXlEG6cOCCejsU3DMctaEAK5PUZ+blsHYHBH
 BYZYKs/x4pcw0SO41GtTh0M2YqDYBVuBIpRw8lLZap97Xo9ucSkAm5WD3rGxk4CY
 YeZKEPUql6MHN3+DKl8mx2F0V+Et/tio2HNqc9KReR1tfoolZAbe+SFZHfgmc/Rq
 RD9nnG8KRd4K2K1BTqpkTmI1EtE7jPtPJPSV8gMGhgL/N5vPmH3mql/qyOtYx48E
 6/hPzWESgs16VRZ/opLh8VvjlY1HBDODQhehhhl+o23/Vb8qEgRf8Uqhq50rQW1H
 EeOqyyYQ90txSU31Sgy1kQkvOgIFAsBObWT1ZCJ3RbfGbB4/tdEAvZqTZRlXo2OY
 7P0Sqcw/9Le5eJkHIlLtBv0TF7y1OYemCbLgRQzFlcRP+UKtYyg8eFnFjqbPEEmP
 ulwhn/BfFVSgaYKQ503u
 =I0pj
 -----END PGP SIGNATURE-----

Merge tag 'powerpc-4.20-1' of git://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux

Pull powerpc updates from Michael Ellerman:
 "Notable changes:

   - A large series to rewrite our SLB miss handling, replacing a lot of
     fairly complicated asm with much fewer lines of C.

   - Following on from that, we now maintain a cache of SLB entries for
     each process and preload them on context switch. Leading to a 27%
     speedup for our context switch benchmark on Power9.

   - Improvements to our handling of SLB multi-hit errors. We now print
     more debug information when they occur, and try to continue running
     by flushing the SLB and reloading, rather than treating them as
     fatal.

   - Enable THP migration on 64-bit Book3S machines (eg. Power7/8/9).

   - Add support for physical memory up to 2PB in the linear mapping on
     64-bit Book3S. We only support up to 512TB as regular system
     memory, otherwise the percpu allocator runs out of vmalloc space.

   - Add stack protector support for 32 and 64-bit, with a per-task
     canary.

   - Add support for PTRACE_SYSEMU and PTRACE_SYSEMU_SINGLESTEP.

   - Support recognising "big cores" on Power9, where two SMT4 cores are
     presented to us as a single SMT8 core.

   - A large series to cleanup some of our ioremap handling and PTE
     flags.

   - Add a driver for the PAPR SCM (storage class memory) interface,
     allowing guests to operate on SCM devices (acked by Dan).

   - Changes to our ftrace code to handle very large kernels, where we
     need to use a trampoline to get to ftrace_caller().

  And many other smaller enhancements and cleanups.

  Thanks to: Alan Modra, Alistair Popple, Aneesh Kumar K.V, Anton
  Blanchard, Aravinda Prasad, Bartlomiej Zolnierkiewicz, Benjamin
  Herrenschmidt, Breno Leitao, Cédric Le Goater, Christophe Leroy,
  Christophe Lombard, Dan Carpenter, Daniel Axtens, Finn Thain, Gautham
  R. Shenoy, Gustavo Romero, Haren Myneni, Hari Bathini, Jia Hongtao,
  Joel Stanley, John Allen, Laurent Dufour, Madhavan Srinivasan, Mahesh
  Salgaonkar, Mark Hairgrove, Masahiro Yamada, Michael Bringmann,
  Michael Neuling, Michal Suchanek, Murilo Opsfelder Araujo, Nathan
  Fontenot, Naveen N. Rao, Nicholas Piggin, Nick Desaulniers, Oliver
  O'Halloran, Paul Mackerras, Petr Vorel, Rashmica Gupta, Reza Arbab,
  Rob Herring, Sam Bobroff, Samuel Mendoza-Jonas, Scott Wood, Stan
  Johnson, Stephen Rothwell, Stewart Smith, Suraj Jitindar Singh, Tyrel
  Datwyler, Vaibhav Jain, Vasant Hegde, YueHaibing, zhong jiang"

* tag 'powerpc-4.20-1' of git://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux: (221 commits)
  Revert "selftests/powerpc: Fix out-of-tree build errors"
  powerpc/msi: Fix compile error on mpc83xx
  powerpc: Fix stack protector crashes on CPU hotplug
  powerpc/traps: restore recoverability of machine_check interrupts
  powerpc/64/module: REL32 relocation range check
  powerpc/64s/radix: Fix radix__flush_tlb_collapsed_pmd double flushing pmd
  selftests/powerpc: Add a test of wild bctr
  powerpc/mm: Fix page table dump to work on Radix
  powerpc/mm/radix: Display if mappings are exec or not
  powerpc/mm/radix: Simplify split mapping logic
  powerpc/mm/radix: Remove the retry in the split mapping logic
  powerpc/mm/radix: Fix small page at boundary when splitting
  powerpc/mm/radix: Fix overuse of small pages in splitting logic
  powerpc/mm/radix: Fix off-by-one in split mapping logic
  powerpc/ftrace: Handle large kernel configs
  powerpc/mm: Fix WARN_ON with THP NUMA migration
  selftests/powerpc: Fix out-of-tree build errors
  powerpc/time: no steal_time when CONFIG_PPC_SPLPAR is not selected
  powerpc/time: Only set CONFIG_ARCH_HAS_SCALED_CPUTIME on PPC64
  powerpc/time: isolate scaled cputime accounting in dedicated functions.
  ...
2018-10-26 14:36:21 -07:00

976 lines
23 KiB
C

// SPDX-License-Identifier: GPL-2.0+
/*
* PCI Hotplug Driver for PowerPC PowerNV platform.
*
* Copyright Gavin Shan, IBM Corporation 2016.
*/
#include <linux/libfdt.h>
#include <linux/module.h>
#include <linux/pci.h>
#include <linux/pci_hotplug.h>
#include <asm/opal.h>
#include <asm/pnv-pci.h>
#include <asm/ppc-pci.h>
#define DRIVER_VERSION "0.1"
#define DRIVER_AUTHOR "Gavin Shan, IBM Corporation"
#define DRIVER_DESC "PowerPC PowerNV PCI Hotplug Driver"
struct pnv_php_event {
bool added;
struct pnv_php_slot *php_slot;
struct work_struct work;
};
static LIST_HEAD(pnv_php_slot_list);
static DEFINE_SPINLOCK(pnv_php_lock);
static void pnv_php_register(struct device_node *dn);
static void pnv_php_unregister_one(struct device_node *dn);
static void pnv_php_unregister(struct device_node *dn);
static void pnv_php_disable_irq(struct pnv_php_slot *php_slot,
bool disable_device)
{
struct pci_dev *pdev = php_slot->pdev;
int irq = php_slot->irq;
u16 ctrl;
if (php_slot->irq > 0) {
pcie_capability_read_word(pdev, PCI_EXP_SLTCTL, &ctrl);
ctrl &= ~(PCI_EXP_SLTCTL_HPIE |
PCI_EXP_SLTCTL_PDCE |
PCI_EXP_SLTCTL_DLLSCE);
pcie_capability_write_word(pdev, PCI_EXP_SLTCTL, ctrl);
free_irq(php_slot->irq, php_slot);
php_slot->irq = 0;
}
if (php_slot->wq) {
destroy_workqueue(php_slot->wq);
php_slot->wq = NULL;
}
if (disable_device || irq > 0) {
if (pdev->msix_enabled)
pci_disable_msix(pdev);
else if (pdev->msi_enabled)
pci_disable_msi(pdev);
pci_disable_device(pdev);
}
}
static void pnv_php_free_slot(struct kref *kref)
{
struct pnv_php_slot *php_slot = container_of(kref,
struct pnv_php_slot, kref);
WARN_ON(!list_empty(&php_slot->children));
pnv_php_disable_irq(php_slot, false);
kfree(php_slot->name);
kfree(php_slot);
}
static inline void pnv_php_put_slot(struct pnv_php_slot *php_slot)
{
if (!php_slot)
return;
kref_put(&php_slot->kref, pnv_php_free_slot);
}
static struct pnv_php_slot *pnv_php_match(struct device_node *dn,
struct pnv_php_slot *php_slot)
{
struct pnv_php_slot *target, *tmp;
if (php_slot->dn == dn) {
kref_get(&php_slot->kref);
return php_slot;
}
list_for_each_entry(tmp, &php_slot->children, link) {
target = pnv_php_match(dn, tmp);
if (target)
return target;
}
return NULL;
}
struct pnv_php_slot *pnv_php_find_slot(struct device_node *dn)
{
struct pnv_php_slot *php_slot, *tmp;
unsigned long flags;
spin_lock_irqsave(&pnv_php_lock, flags);
list_for_each_entry(tmp, &pnv_php_slot_list, link) {
php_slot = pnv_php_match(dn, tmp);
if (php_slot) {
spin_unlock_irqrestore(&pnv_php_lock, flags);
return php_slot;
}
}
spin_unlock_irqrestore(&pnv_php_lock, flags);
return NULL;
}
EXPORT_SYMBOL_GPL(pnv_php_find_slot);
/*
* Remove pdn for all children of the indicated device node.
* The function should remove pdn in a depth-first manner.
*/
static void pnv_php_rmv_pdns(struct device_node *dn)
{
struct device_node *child;
for_each_child_of_node(dn, child) {
pnv_php_rmv_pdns(child);
pci_remove_device_node_info(child);
}
}
/*
* Detach all child nodes of the indicated device nodes. The
* function should handle device nodes in depth-first manner.
*
* We should not invoke of_node_release() as the memory for
* individual device node is part of large memory block. The
* large block is allocated from memblock (system bootup) or
* kmalloc() when unflattening the device tree by OF changeset.
* We can not free the large block allocated from memblock. For
* later case, it should be released at once.
*/
static void pnv_php_detach_device_nodes(struct device_node *parent)
{
struct device_node *dn;
int refcount;
for_each_child_of_node(parent, dn) {
pnv_php_detach_device_nodes(dn);
of_node_put(dn);
refcount = kref_read(&dn->kobj.kref);
if (refcount != 1)
pr_warn("Invalid refcount %d on <%pOF>\n",
refcount, dn);
of_detach_node(dn);
}
}
static void pnv_php_rmv_devtree(struct pnv_php_slot *php_slot)
{
pnv_php_rmv_pdns(php_slot->dn);
/*
* Decrease the refcount if the device nodes were created
* through OF changeset before detaching them.
*/
if (php_slot->fdt)
of_changeset_destroy(&php_slot->ocs);
pnv_php_detach_device_nodes(php_slot->dn);
if (php_slot->fdt) {
kfree(php_slot->dt);
kfree(php_slot->fdt);
php_slot->dt = NULL;
php_slot->dn->child = NULL;
php_slot->fdt = NULL;
}
}
/*
* As the nodes in OF changeset are applied in reverse order, we
* need revert the nodes in advance so that we have correct node
* order after the changeset is applied.
*/
static void pnv_php_reverse_nodes(struct device_node *parent)
{
struct device_node *child, *next;
/* In-depth first */
for_each_child_of_node(parent, child)
pnv_php_reverse_nodes(child);
/* Reverse the nodes in the child list */
child = parent->child;
parent->child = NULL;
while (child) {
next = child->sibling;
child->sibling = parent->child;
parent->child = child;
child = next;
}
}
static int pnv_php_populate_changeset(struct of_changeset *ocs,
struct device_node *dn)
{
struct device_node *child;
int ret = 0;
for_each_child_of_node(dn, child) {
ret = of_changeset_attach_node(ocs, child);
if (ret) {
of_node_put(child);
break;
}
ret = pnv_php_populate_changeset(ocs, child);
if (ret) {
of_node_put(child);
break;
}
}
return ret;
}
static void *pnv_php_add_one_pdn(struct device_node *dn, void *data)
{
struct pci_controller *hose = (struct pci_controller *)data;
struct pci_dn *pdn;
pdn = pci_add_device_node_info(hose, dn);
if (!pdn)
return ERR_PTR(-ENOMEM);
return NULL;
}
static void pnv_php_add_pdns(struct pnv_php_slot *slot)
{
struct pci_controller *hose = pci_bus_to_host(slot->bus);
pci_traverse_device_nodes(slot->dn, pnv_php_add_one_pdn, hose);
}
static int pnv_php_add_devtree(struct pnv_php_slot *php_slot)
{
void *fdt, *fdt1, *dt;
int ret;
/* We don't know the FDT blob size. We try to get it through
* maximal memory chunk and then copy it to another chunk that
* fits the real size.
*/
fdt1 = kzalloc(0x10000, GFP_KERNEL);
if (!fdt1) {
ret = -ENOMEM;
goto out;
}
ret = pnv_pci_get_device_tree(php_slot->dn->phandle, fdt1, 0x10000);
if (ret) {
pci_warn(php_slot->pdev, "Error %d getting FDT blob\n", ret);
goto free_fdt1;
}
fdt = kmemdup(fdt1, fdt_totalsize(fdt1), GFP_KERNEL);
if (!fdt) {
ret = -ENOMEM;
goto free_fdt1;
}
/* Unflatten device tree blob */
dt = of_fdt_unflatten_tree(fdt, php_slot->dn, NULL);
if (!dt) {
ret = -EINVAL;
pci_warn(php_slot->pdev, "Cannot unflatten FDT\n");
goto free_fdt;
}
/* Initialize and apply the changeset */
of_changeset_init(&php_slot->ocs);
pnv_php_reverse_nodes(php_slot->dn);
ret = pnv_php_populate_changeset(&php_slot->ocs, php_slot->dn);
if (ret) {
pnv_php_reverse_nodes(php_slot->dn);
pci_warn(php_slot->pdev, "Error %d populating changeset\n",
ret);
goto free_dt;
}
php_slot->dn->child = NULL;
ret = of_changeset_apply(&php_slot->ocs);
if (ret) {
pci_warn(php_slot->pdev, "Error %d applying changeset\n", ret);
goto destroy_changeset;
}
/* Add device node firmware data */
pnv_php_add_pdns(php_slot);
php_slot->fdt = fdt;
php_slot->dt = dt;
kfree(fdt1);
goto out;
destroy_changeset:
of_changeset_destroy(&php_slot->ocs);
free_dt:
kfree(dt);
php_slot->dn->child = NULL;
free_fdt:
kfree(fdt);
free_fdt1:
kfree(fdt1);
out:
return ret;
}
static inline struct pnv_php_slot *to_pnv_php_slot(struct hotplug_slot *slot)
{
return container_of(slot, struct pnv_php_slot, slot);
}
int pnv_php_set_slot_power_state(struct hotplug_slot *slot,
uint8_t state)
{
struct pnv_php_slot *php_slot = to_pnv_php_slot(slot);
struct opal_msg msg;
int ret;
ret = pnv_pci_set_power_state(php_slot->id, state, &msg);
if (ret > 0) {
if (be64_to_cpu(msg.params[1]) != php_slot->dn->phandle ||
be64_to_cpu(msg.params[2]) != state ||
be64_to_cpu(msg.params[3]) != OPAL_SUCCESS) {
pci_warn(php_slot->pdev, "Wrong msg (%lld, %lld, %lld)\n",
be64_to_cpu(msg.params[1]),
be64_to_cpu(msg.params[2]),
be64_to_cpu(msg.params[3]));
return -ENOMSG;
}
} else if (ret < 0) {
pci_warn(php_slot->pdev, "Error %d powering %s\n",
ret, (state == OPAL_PCI_SLOT_POWER_ON) ? "on" : "off");
return ret;
}
if (state == OPAL_PCI_SLOT_POWER_OFF || state == OPAL_PCI_SLOT_OFFLINE)
pnv_php_rmv_devtree(php_slot);
else
ret = pnv_php_add_devtree(php_slot);
return ret;
}
EXPORT_SYMBOL_GPL(pnv_php_set_slot_power_state);
static int pnv_php_get_power_state(struct hotplug_slot *slot, u8 *state)
{
struct pnv_php_slot *php_slot = to_pnv_php_slot(slot);
uint8_t power_state = OPAL_PCI_SLOT_POWER_ON;
int ret;
/*
* Retrieve power status from firmware. If we fail
* getting that, the power status fails back to
* be on.
*/
ret = pnv_pci_get_power_state(php_slot->id, &power_state);
if (ret) {
pci_warn(php_slot->pdev, "Error %d getting power status\n",
ret);
} else {
*state = power_state;
}
return 0;
}
static int pnv_php_get_adapter_state(struct hotplug_slot *slot, u8 *state)
{
struct pnv_php_slot *php_slot = to_pnv_php_slot(slot);
uint8_t presence = OPAL_PCI_SLOT_EMPTY;
int ret;
/*
* Retrieve presence status from firmware. If we can't
* get that, it will fail back to be empty.
*/
ret = pnv_pci_get_presence_state(php_slot->id, &presence);
if (ret >= 0) {
*state = presence;
ret = 0;
} else {
pci_warn(php_slot->pdev, "Error %d getting presence\n", ret);
}
return ret;
}
static int pnv_php_get_attention_state(struct hotplug_slot *slot, u8 *state)
{
struct pnv_php_slot *php_slot = to_pnv_php_slot(slot);
*state = php_slot->attention_state;
return 0;
}
static int pnv_php_set_attention_state(struct hotplug_slot *slot, u8 state)
{
struct pnv_php_slot *php_slot = to_pnv_php_slot(slot);
/* FIXME: Make it real once firmware supports it */
php_slot->attention_state = state;
return 0;
}
static int pnv_php_enable(struct pnv_php_slot *php_slot, bool rescan)
{
struct hotplug_slot *slot = &php_slot->slot;
uint8_t presence = OPAL_PCI_SLOT_EMPTY;
uint8_t power_status = OPAL_PCI_SLOT_POWER_ON;
int ret;
/* Check if the slot has been configured */
if (php_slot->state != PNV_PHP_STATE_REGISTERED)
return 0;
/* Retrieve slot presence status */
ret = pnv_php_get_adapter_state(slot, &presence);
if (ret)
return ret;
/*
* Proceed if there have nothing behind the slot. However,
* we should leave the slot in registered state at the
* beginning. Otherwise, the PCI devices inserted afterwards
* won't be probed and populated.
*/
if (presence == OPAL_PCI_SLOT_EMPTY) {
if (!php_slot->power_state_check) {
php_slot->power_state_check = true;
return 0;
}
goto scan;
}
/*
* If the power supply to the slot is off, we can't detect
* adapter presence state. That means we have to turn the
* slot on before going to probe slot's presence state.
*
* On the first time, we don't change the power status to
* boost system boot with assumption that the firmware
* supplies consistent slot power status: empty slot always
* has its power off and non-empty slot has its power on.
*/
if (!php_slot->power_state_check) {
php_slot->power_state_check = true;
ret = pnv_php_get_power_state(slot, &power_status);
if (ret)
return ret;
if (power_status != OPAL_PCI_SLOT_POWER_ON)
return 0;
}
/* Check the power status. Scan the slot if it is already on */
ret = pnv_php_get_power_state(slot, &power_status);
if (ret)
return ret;
if (power_status == OPAL_PCI_SLOT_POWER_ON)
goto scan;
/* Power is off, turn it on and then scan the slot */
ret = pnv_php_set_slot_power_state(slot, OPAL_PCI_SLOT_POWER_ON);
if (ret)
return ret;
scan:
if (presence == OPAL_PCI_SLOT_PRESENT) {
if (rescan) {
pci_lock_rescan_remove();
pci_hp_add_devices(php_slot->bus);
pci_unlock_rescan_remove();
}
/* Rescan for child hotpluggable slots */
php_slot->state = PNV_PHP_STATE_POPULATED;
if (rescan)
pnv_php_register(php_slot->dn);
} else {
php_slot->state = PNV_PHP_STATE_POPULATED;
}
return 0;
}
static int pnv_php_enable_slot(struct hotplug_slot *slot)
{
struct pnv_php_slot *php_slot = to_pnv_php_slot(slot);
return pnv_php_enable(php_slot, true);
}
static int pnv_php_disable_slot(struct hotplug_slot *slot)
{
struct pnv_php_slot *php_slot = to_pnv_php_slot(slot);
int ret;
if (php_slot->state != PNV_PHP_STATE_POPULATED)
return 0;
/* Remove all devices behind the slot */
pci_lock_rescan_remove();
pci_hp_remove_devices(php_slot->bus);
pci_unlock_rescan_remove();
/* Detach the child hotpluggable slots */
pnv_php_unregister(php_slot->dn);
/* Notify firmware and remove device nodes */
ret = pnv_php_set_slot_power_state(slot, OPAL_PCI_SLOT_POWER_OFF);
php_slot->state = PNV_PHP_STATE_REGISTERED;
return ret;
}
static const struct hotplug_slot_ops php_slot_ops = {
.get_power_status = pnv_php_get_power_state,
.get_adapter_status = pnv_php_get_adapter_state,
.get_attention_status = pnv_php_get_attention_state,
.set_attention_status = pnv_php_set_attention_state,
.enable_slot = pnv_php_enable_slot,
.disable_slot = pnv_php_disable_slot,
};
static void pnv_php_release(struct pnv_php_slot *php_slot)
{
unsigned long flags;
/* Remove from global or child list */
spin_lock_irqsave(&pnv_php_lock, flags);
list_del(&php_slot->link);
spin_unlock_irqrestore(&pnv_php_lock, flags);
/* Detach from parent */
pnv_php_put_slot(php_slot);
pnv_php_put_slot(php_slot->parent);
}
static struct pnv_php_slot *pnv_php_alloc_slot(struct device_node *dn)
{
struct pnv_php_slot *php_slot;
struct pci_bus *bus;
const char *label;
uint64_t id;
int ret;
ret = of_property_read_string(dn, "ibm,slot-label", &label);
if (ret)
return NULL;
if (pnv_pci_get_slot_id(dn, &id))
return NULL;
bus = pci_find_bus_by_node(dn);
if (!bus)
return NULL;
php_slot = kzalloc(sizeof(*php_slot), GFP_KERNEL);
if (!php_slot)
return NULL;
php_slot->name = kstrdup(label, GFP_KERNEL);
if (!php_slot->name) {
kfree(php_slot);
return NULL;
}
if (dn->child && PCI_DN(dn->child))
php_slot->slot_no = PCI_SLOT(PCI_DN(dn->child)->devfn);
else
php_slot->slot_no = -1; /* Placeholder slot */
kref_init(&php_slot->kref);
php_slot->state = PNV_PHP_STATE_INITIALIZED;
php_slot->dn = dn;
php_slot->pdev = bus->self;
php_slot->bus = bus;
php_slot->id = id;
php_slot->power_state_check = false;
php_slot->slot.ops = &php_slot_ops;
INIT_LIST_HEAD(&php_slot->children);
INIT_LIST_HEAD(&php_slot->link);
return php_slot;
}
static int pnv_php_register_slot(struct pnv_php_slot *php_slot)
{
struct pnv_php_slot *parent;
struct device_node *dn = php_slot->dn;
unsigned long flags;
int ret;
/* Check if the slot is registered or not */
parent = pnv_php_find_slot(php_slot->dn);
if (parent) {
pnv_php_put_slot(parent);
return -EEXIST;
}
/* Register PCI slot */
ret = pci_hp_register(&php_slot->slot, php_slot->bus,
php_slot->slot_no, php_slot->name);
if (ret) {
pci_warn(php_slot->pdev, "Error %d registering slot\n", ret);
return ret;
}
/* Attach to the parent's child list or global list */
while ((dn = of_get_parent(dn))) {
if (!PCI_DN(dn)) {
of_node_put(dn);
break;
}
parent = pnv_php_find_slot(dn);
if (parent) {
of_node_put(dn);
break;
}
of_node_put(dn);
}
spin_lock_irqsave(&pnv_php_lock, flags);
php_slot->parent = parent;
if (parent)
list_add_tail(&php_slot->link, &parent->children);
else
list_add_tail(&php_slot->link, &pnv_php_slot_list);
spin_unlock_irqrestore(&pnv_php_lock, flags);
php_slot->state = PNV_PHP_STATE_REGISTERED;
return 0;
}
static int pnv_php_enable_msix(struct pnv_php_slot *php_slot)
{
struct pci_dev *pdev = php_slot->pdev;
struct msix_entry entry;
int nr_entries, ret;
u16 pcie_flag;
/* Get total number of MSIx entries */
nr_entries = pci_msix_vec_count(pdev);
if (nr_entries < 0)
return nr_entries;
/* Check hotplug MSIx entry is in range */
pcie_capability_read_word(pdev, PCI_EXP_FLAGS, &pcie_flag);
entry.entry = (pcie_flag & PCI_EXP_FLAGS_IRQ) >> 9;
if (entry.entry >= nr_entries)
return -ERANGE;
/* Enable MSIx */
ret = pci_enable_msix_exact(pdev, &entry, 1);
if (ret) {
pci_warn(pdev, "Error %d enabling MSIx\n", ret);
return ret;
}
return entry.vector;
}
static void pnv_php_event_handler(struct work_struct *work)
{
struct pnv_php_event *event =
container_of(work, struct pnv_php_event, work);
struct pnv_php_slot *php_slot = event->php_slot;
if (event->added)
pnv_php_enable_slot(&php_slot->slot);
else
pnv_php_disable_slot(&php_slot->slot);
kfree(event);
}
static irqreturn_t pnv_php_interrupt(int irq, void *data)
{
struct pnv_php_slot *php_slot = data;
struct pci_dev *pchild, *pdev = php_slot->pdev;
struct eeh_dev *edev;
struct eeh_pe *pe;
struct pnv_php_event *event;
u16 sts, lsts;
u8 presence;
bool added;
unsigned long flags;
int ret;
pcie_capability_read_word(pdev, PCI_EXP_SLTSTA, &sts);
sts &= (PCI_EXP_SLTSTA_PDC | PCI_EXP_SLTSTA_DLLSC);
pcie_capability_write_word(pdev, PCI_EXP_SLTSTA, sts);
if (sts & PCI_EXP_SLTSTA_DLLSC) {
pcie_capability_read_word(pdev, PCI_EXP_LNKSTA, &lsts);
added = !!(lsts & PCI_EXP_LNKSTA_DLLLA);
} else if (!(php_slot->flags & PNV_PHP_FLAG_BROKEN_PDC) &&
(sts & PCI_EXP_SLTSTA_PDC)) {
ret = pnv_pci_get_presence_state(php_slot->id, &presence);
if (ret) {
pci_warn(pdev, "PCI slot [%s] error %d getting presence (0x%04x), to retry the operation.\n",
php_slot->name, ret, sts);
return IRQ_HANDLED;
}
added = !!(presence == OPAL_PCI_SLOT_PRESENT);
} else {
return IRQ_NONE;
}
/* Freeze the removed PE to avoid unexpected error reporting */
if (!added) {
pchild = list_first_entry_or_null(&php_slot->bus->devices,
struct pci_dev, bus_list);
edev = pchild ? pci_dev_to_eeh_dev(pchild) : NULL;
pe = edev ? edev->pe : NULL;
if (pe) {
eeh_serialize_lock(&flags);
eeh_pe_mark_isolated(pe);
eeh_serialize_unlock(flags);
eeh_pe_set_option(pe, EEH_OPT_FREEZE_PE);
}
}
/*
* The PE is left in frozen state if the event is missed. It's
* fine as the PCI devices (PE) aren't functional any more.
*/
event = kzalloc(sizeof(*event), GFP_ATOMIC);
if (!event) {
pci_warn(pdev, "PCI slot [%s] missed hotplug event 0x%04x\n",
php_slot->name, sts);
return IRQ_HANDLED;
}
pci_info(pdev, "PCI slot [%s] %s (IRQ: %d)\n",
php_slot->name, added ? "added" : "removed", irq);
INIT_WORK(&event->work, pnv_php_event_handler);
event->added = added;
event->php_slot = php_slot;
queue_work(php_slot->wq, &event->work);
return IRQ_HANDLED;
}
static void pnv_php_init_irq(struct pnv_php_slot *php_slot, int irq)
{
struct pci_dev *pdev = php_slot->pdev;
u32 broken_pdc = 0;
u16 sts, ctrl;
int ret;
/* Allocate workqueue */
php_slot->wq = alloc_workqueue("pciehp-%s", 0, 0, php_slot->name);
if (!php_slot->wq) {
pci_warn(pdev, "Cannot alloc workqueue\n");
pnv_php_disable_irq(php_slot, true);
return;
}
/* Check PDC (Presence Detection Change) is broken or not */
ret = of_property_read_u32(php_slot->dn, "ibm,slot-broken-pdc",
&broken_pdc);
if (!ret && broken_pdc)
php_slot->flags |= PNV_PHP_FLAG_BROKEN_PDC;
/* Clear pending interrupts */
pcie_capability_read_word(pdev, PCI_EXP_SLTSTA, &sts);
if (php_slot->flags & PNV_PHP_FLAG_BROKEN_PDC)
sts |= PCI_EXP_SLTSTA_DLLSC;
else
sts |= (PCI_EXP_SLTSTA_PDC | PCI_EXP_SLTSTA_DLLSC);
pcie_capability_write_word(pdev, PCI_EXP_SLTSTA, sts);
/* Request the interrupt */
ret = request_irq(irq, pnv_php_interrupt, IRQF_SHARED,
php_slot->name, php_slot);
if (ret) {
pnv_php_disable_irq(php_slot, true);
pci_warn(pdev, "Error %d enabling IRQ %d\n", ret, irq);
return;
}
/* Enable the interrupts */
pcie_capability_read_word(pdev, PCI_EXP_SLTCTL, &ctrl);
if (php_slot->flags & PNV_PHP_FLAG_BROKEN_PDC) {
ctrl &= ~PCI_EXP_SLTCTL_PDCE;
ctrl |= (PCI_EXP_SLTCTL_HPIE |
PCI_EXP_SLTCTL_DLLSCE);
} else {
ctrl |= (PCI_EXP_SLTCTL_HPIE |
PCI_EXP_SLTCTL_PDCE |
PCI_EXP_SLTCTL_DLLSCE);
}
pcie_capability_write_word(pdev, PCI_EXP_SLTCTL, ctrl);
/* The interrupt is initialized successfully when @irq is valid */
php_slot->irq = irq;
}
static void pnv_php_enable_irq(struct pnv_php_slot *php_slot)
{
struct pci_dev *pdev = php_slot->pdev;
int irq, ret;
/*
* The MSI/MSIx interrupt might have been occupied by other
* drivers. Don't populate the surprise hotplug capability
* in that case.
*/
if (pci_dev_msi_enabled(pdev))
return;
ret = pci_enable_device(pdev);
if (ret) {
pci_warn(pdev, "Error %d enabling device\n", ret);
return;
}
pci_set_master(pdev);
/* Enable MSIx interrupt */
irq = pnv_php_enable_msix(php_slot);
if (irq > 0) {
pnv_php_init_irq(php_slot, irq);
return;
}
/*
* Use MSI if MSIx doesn't work. Fail back to legacy INTx
* if MSI doesn't work either
*/
ret = pci_enable_msi(pdev);
if (!ret || pdev->irq) {
irq = pdev->irq;
pnv_php_init_irq(php_slot, irq);
}
}
static int pnv_php_register_one(struct device_node *dn)
{
struct pnv_php_slot *php_slot;
u32 prop32;
int ret;
/* Check if it's hotpluggable slot */
ret = of_property_read_u32(dn, "ibm,slot-pluggable", &prop32);
if (ret || !prop32)
return -ENXIO;
ret = of_property_read_u32(dn, "ibm,reset-by-firmware", &prop32);
if (ret || !prop32)
return -ENXIO;
php_slot = pnv_php_alloc_slot(dn);
if (!php_slot)
return -ENODEV;
ret = pnv_php_register_slot(php_slot);
if (ret)
goto free_slot;
ret = pnv_php_enable(php_slot, false);
if (ret)
goto unregister_slot;
/* Enable interrupt if the slot supports surprise hotplug */
ret = of_property_read_u32(dn, "ibm,slot-surprise-pluggable", &prop32);
if (!ret && prop32)
pnv_php_enable_irq(php_slot);
return 0;
unregister_slot:
pnv_php_unregister_one(php_slot->dn);
free_slot:
pnv_php_put_slot(php_slot);
return ret;
}
static void pnv_php_register(struct device_node *dn)
{
struct device_node *child;
/*
* The parent slots should be registered before their
* child slots.
*/
for_each_child_of_node(dn, child) {
pnv_php_register_one(child);
pnv_php_register(child);
}
}
static void pnv_php_unregister_one(struct device_node *dn)
{
struct pnv_php_slot *php_slot;
php_slot = pnv_php_find_slot(dn);
if (!php_slot)
return;
php_slot->state = PNV_PHP_STATE_OFFLINE;
pci_hp_deregister(&php_slot->slot);
pnv_php_release(php_slot);
pnv_php_put_slot(php_slot);
}
static void pnv_php_unregister(struct device_node *dn)
{
struct device_node *child;
/* The child slots should go before their parent slots */
for_each_child_of_node(dn, child) {
pnv_php_unregister(child);
pnv_php_unregister_one(child);
}
}
static int __init pnv_php_init(void)
{
struct device_node *dn;
pr_info(DRIVER_DESC " version: " DRIVER_VERSION "\n");
for_each_compatible_node(dn, NULL, "ibm,ioda2-phb")
pnv_php_register(dn);
return 0;
}
static void __exit pnv_php_exit(void)
{
struct device_node *dn;
for_each_compatible_node(dn, NULL, "ibm,ioda2-phb")
pnv_php_unregister(dn);
}
module_init(pnv_php_init);
module_exit(pnv_php_exit);
MODULE_VERSION(DRIVER_VERSION);
MODULE_LICENSE("GPL v2");
MODULE_AUTHOR(DRIVER_AUTHOR);
MODULE_DESCRIPTION(DRIVER_DESC);