mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-27 12:25:08 +07:00
a1e537b3f0
This patch increases the timeout for PCI ELBI configuration to support low frequency Palladium images. Signed-off-by: Omer Shpigelman <oshpigelman@habana.ai> Signed-off-by: Oded Gabbay <oded.gabbay@gmail.com>
417 lines
10 KiB
C
417 lines
10 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
/*
|
|
* Copyright 2016-2019 HabanaLabs, Ltd.
|
|
* All Rights Reserved.
|
|
*/
|
|
|
|
#include "habanalabs.h"
|
|
#include "include/hw_ip/pci/pci_general.h"
|
|
|
|
#include <linux/pci.h>
|
|
|
|
#define HL_PLDM_PCI_ELBI_TIMEOUT_MSEC (HL_PCI_ELBI_TIMEOUT_MSEC * 10)
|
|
|
|
/**
|
|
* hl_pci_bars_map() - Map PCI BARs.
|
|
* @hdev: Pointer to hl_device structure.
|
|
* @bar_name: Array of BAR names.
|
|
* @is_wc: Array with flag per BAR whether a write-combined mapping is needed.
|
|
*
|
|
* Request PCI regions and map them to kernel virtual addresses.
|
|
*
|
|
* Return: 0 on success, non-zero for failure.
|
|
*/
|
|
int hl_pci_bars_map(struct hl_device *hdev, const char * const name[3],
|
|
bool is_wc[3])
|
|
{
|
|
struct pci_dev *pdev = hdev->pdev;
|
|
int rc, i, bar;
|
|
|
|
rc = pci_request_regions(pdev, HL_NAME);
|
|
if (rc) {
|
|
dev_err(hdev->dev, "Cannot obtain PCI resources\n");
|
|
return rc;
|
|
}
|
|
|
|
for (i = 0 ; i < 3 ; i++) {
|
|
bar = i * 2; /* 64-bit BARs */
|
|
hdev->pcie_bar[bar] = is_wc[i] ?
|
|
pci_ioremap_wc_bar(pdev, bar) :
|
|
pci_ioremap_bar(pdev, bar);
|
|
if (!hdev->pcie_bar[bar]) {
|
|
dev_err(hdev->dev, "pci_ioremap%s_bar failed for %s\n",
|
|
is_wc[i] ? "_wc" : "", name[i]);
|
|
rc = -ENODEV;
|
|
goto err;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
|
|
err:
|
|
for (i = 2 ; i >= 0 ; i--) {
|
|
bar = i * 2; /* 64-bit BARs */
|
|
if (hdev->pcie_bar[bar])
|
|
iounmap(hdev->pcie_bar[bar]);
|
|
}
|
|
|
|
pci_release_regions(pdev);
|
|
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* hl_pci_bars_unmap() - Unmap PCI BARS.
|
|
* @hdev: Pointer to hl_device structure.
|
|
*
|
|
* Release all PCI BARs and unmap their virtual addresses.
|
|
*/
|
|
static void hl_pci_bars_unmap(struct hl_device *hdev)
|
|
{
|
|
struct pci_dev *pdev = hdev->pdev;
|
|
int i, bar;
|
|
|
|
for (i = 2 ; i >= 0 ; i--) {
|
|
bar = i * 2; /* 64-bit BARs */
|
|
iounmap(hdev->pcie_bar[bar]);
|
|
}
|
|
|
|
pci_release_regions(pdev);
|
|
}
|
|
|
|
/*
|
|
* hl_pci_elbi_write() - Write through the ELBI interface.
|
|
* @hdev: Pointer to hl_device structure.
|
|
*
|
|
* Return: 0 on success, negative value for failure.
|
|
*/
|
|
static int hl_pci_elbi_write(struct hl_device *hdev, u64 addr, u32 data)
|
|
{
|
|
struct pci_dev *pdev = hdev->pdev;
|
|
ktime_t timeout;
|
|
u64 msec;
|
|
u32 val;
|
|
|
|
if (hdev->pldm)
|
|
msec = HL_PLDM_PCI_ELBI_TIMEOUT_MSEC;
|
|
else
|
|
msec = HL_PCI_ELBI_TIMEOUT_MSEC;
|
|
|
|
/* Clear previous status */
|
|
pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_STS, 0);
|
|
|
|
pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_ADDR, (u32) addr);
|
|
pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_DATA, data);
|
|
pci_write_config_dword(pdev, mmPCI_CONFIG_ELBI_CTRL,
|
|
PCI_CONFIG_ELBI_CTRL_WRITE);
|
|
|
|
timeout = ktime_add_ms(ktime_get(), msec);
|
|
for (;;) {
|
|
pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_STS, &val);
|
|
if (val & PCI_CONFIG_ELBI_STS_MASK)
|
|
break;
|
|
if (ktime_compare(ktime_get(), timeout) > 0) {
|
|
pci_read_config_dword(pdev, mmPCI_CONFIG_ELBI_STS,
|
|
&val);
|
|
break;
|
|
}
|
|
|
|
usleep_range(300, 500);
|
|
}
|
|
|
|
if ((val & PCI_CONFIG_ELBI_STS_MASK) == PCI_CONFIG_ELBI_STS_DONE)
|
|
return 0;
|
|
|
|
if (val & PCI_CONFIG_ELBI_STS_ERR) {
|
|
dev_err(hdev->dev, "Error writing to ELBI\n");
|
|
return -EIO;
|
|
}
|
|
|
|
if (!(val & PCI_CONFIG_ELBI_STS_MASK)) {
|
|
dev_err(hdev->dev, "ELBI write didn't finish in time\n");
|
|
return -EIO;
|
|
}
|
|
|
|
dev_err(hdev->dev, "ELBI write has undefined bits in status\n");
|
|
return -EIO;
|
|
}
|
|
|
|
/**
|
|
* hl_pci_iatu_write() - iatu write routine.
|
|
* @hdev: Pointer to hl_device structure.
|
|
*
|
|
* Return: 0 on success, negative value for failure.
|
|
*/
|
|
int hl_pci_iatu_write(struct hl_device *hdev, u32 addr, u32 data)
|
|
{
|
|
struct asic_fixed_properties *prop = &hdev->asic_prop;
|
|
u32 dbi_offset;
|
|
int rc;
|
|
|
|
dbi_offset = addr & 0xFFF;
|
|
|
|
rc = hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr, 0x00300000);
|
|
rc |= hl_pci_elbi_write(hdev, prop->pcie_dbi_base_address + dbi_offset,
|
|
data);
|
|
|
|
if (rc)
|
|
return -EIO;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* hl_pci_reset_link_through_bridge() - Reset PCI link.
|
|
* @hdev: Pointer to hl_device structure.
|
|
*/
|
|
static void hl_pci_reset_link_through_bridge(struct hl_device *hdev)
|
|
{
|
|
struct pci_dev *pdev = hdev->pdev;
|
|
struct pci_dev *parent_port;
|
|
u16 val;
|
|
|
|
parent_port = pdev->bus->self;
|
|
pci_read_config_word(parent_port, PCI_BRIDGE_CONTROL, &val);
|
|
val |= PCI_BRIDGE_CTL_BUS_RESET;
|
|
pci_write_config_word(parent_port, PCI_BRIDGE_CONTROL, val);
|
|
ssleep(1);
|
|
|
|
val &= ~(PCI_BRIDGE_CTL_BUS_RESET);
|
|
pci_write_config_word(parent_port, PCI_BRIDGE_CONTROL, val);
|
|
ssleep(3);
|
|
}
|
|
|
|
/**
|
|
* hl_pci_set_dram_bar_base() - Set DDR BAR to map specific device address.
|
|
* @hdev: Pointer to hl_device structure.
|
|
* @inbound_region: Inbound region number.
|
|
* @bar: PCI BAR number.
|
|
* @addr: Address in DRAM. Must be aligned to DRAM bar size.
|
|
*
|
|
* Configure the iATU so that the DRAM bar will start at the specified address.
|
|
*
|
|
* Return: 0 on success, negative value for failure.
|
|
*/
|
|
int hl_pci_set_dram_bar_base(struct hl_device *hdev, u8 inbound_region, u8 bar,
|
|
u64 addr)
|
|
{
|
|
struct asic_fixed_properties *prop = &hdev->asic_prop;
|
|
u32 offset;
|
|
int rc;
|
|
|
|
switch (inbound_region) {
|
|
case 0:
|
|
offset = 0x100;
|
|
break;
|
|
case 1:
|
|
offset = 0x300;
|
|
break;
|
|
case 2:
|
|
offset = 0x500;
|
|
break;
|
|
default:
|
|
dev_err(hdev->dev, "Invalid inbound region %d\n",
|
|
inbound_region);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (bar != 0 && bar != 2 && bar != 4) {
|
|
dev_err(hdev->dev, "Invalid PCI BAR %d\n", bar);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Point to the specified address */
|
|
rc = hl_pci_iatu_write(hdev, offset + 0x14, lower_32_bits(addr));
|
|
rc |= hl_pci_iatu_write(hdev, offset + 0x18, upper_32_bits(addr));
|
|
rc |= hl_pci_iatu_write(hdev, offset + 0x0, 0);
|
|
/* Enable + BAR match + match enable + BAR number */
|
|
rc |= hl_pci_iatu_write(hdev, offset + 0x4, 0xC0080000 | (bar << 8));
|
|
|
|
/* Return the DBI window to the default location */
|
|
rc |= hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr, 0);
|
|
rc |= hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr + 4, 0);
|
|
|
|
if (rc)
|
|
dev_err(hdev->dev, "failed to map DRAM bar to 0x%08llx\n",
|
|
addr);
|
|
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* hl_pci_init_iatu() - Initialize the iATU unit inside the PCI controller.
|
|
* @hdev: Pointer to hl_device structure.
|
|
* @sram_base_address: SRAM base address.
|
|
* @dram_base_address: DRAM base address.
|
|
* @host_phys_base_address: Base physical address of host memory for device
|
|
* transactions.
|
|
* @host_phys_size: Size of host memory for device transactions.
|
|
*
|
|
* This is needed in case the firmware doesn't initialize the iATU.
|
|
*
|
|
* Return: 0 on success, negative value for failure.
|
|
*/
|
|
int hl_pci_init_iatu(struct hl_device *hdev, u64 sram_base_address,
|
|
u64 dram_base_address, u64 host_phys_base_address,
|
|
u64 host_phys_size)
|
|
{
|
|
struct asic_fixed_properties *prop = &hdev->asic_prop;
|
|
u64 host_phys_end_addr;
|
|
int rc = 0;
|
|
|
|
/* Inbound Region 0 - Bar 0 - Point to SRAM base address */
|
|
rc = hl_pci_iatu_write(hdev, 0x114, lower_32_bits(sram_base_address));
|
|
rc |= hl_pci_iatu_write(hdev, 0x118, upper_32_bits(sram_base_address));
|
|
rc |= hl_pci_iatu_write(hdev, 0x100, 0);
|
|
/* Enable + Bar match + match enable */
|
|
rc |= hl_pci_iatu_write(hdev, 0x104, 0xC0080000);
|
|
|
|
/* Point to DRAM */
|
|
if (!hdev->asic_funcs->set_dram_bar_base)
|
|
return -EINVAL;
|
|
if (hdev->asic_funcs->set_dram_bar_base(hdev, dram_base_address) ==
|
|
U64_MAX)
|
|
return -EIO;
|
|
|
|
|
|
/* Outbound Region 0 - Point to Host */
|
|
host_phys_end_addr = host_phys_base_address + host_phys_size - 1;
|
|
rc |= hl_pci_iatu_write(hdev, 0x008,
|
|
lower_32_bits(host_phys_base_address));
|
|
rc |= hl_pci_iatu_write(hdev, 0x00C,
|
|
upper_32_bits(host_phys_base_address));
|
|
rc |= hl_pci_iatu_write(hdev, 0x010, lower_32_bits(host_phys_end_addr));
|
|
rc |= hl_pci_iatu_write(hdev, 0x014, 0);
|
|
rc |= hl_pci_iatu_write(hdev, 0x018, 0);
|
|
rc |= hl_pci_iatu_write(hdev, 0x020, upper_32_bits(host_phys_end_addr));
|
|
/* Increase region size */
|
|
rc |= hl_pci_iatu_write(hdev, 0x000, 0x00002000);
|
|
/* Enable */
|
|
rc |= hl_pci_iatu_write(hdev, 0x004, 0x80000000);
|
|
|
|
/* Return the DBI window to the default location */
|
|
rc |= hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr, 0);
|
|
rc |= hl_pci_elbi_write(hdev, prop->pcie_aux_dbi_reg_addr + 4, 0);
|
|
|
|
if (rc)
|
|
return -EIO;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* hl_pci_set_dma_mask() - Set DMA masks for the device.
|
|
* @hdev: Pointer to hl_device structure.
|
|
* @dma_mask: number of bits for the requested dma mask.
|
|
*
|
|
* This function sets the DMA masks (regular and consistent) for a specified
|
|
* value. If it doesn't succeed, it tries to set it to a fall-back value
|
|
*
|
|
* Return: 0 on success, non-zero for failure.
|
|
*/
|
|
int hl_pci_set_dma_mask(struct hl_device *hdev, u8 dma_mask)
|
|
{
|
|
struct pci_dev *pdev = hdev->pdev;
|
|
int rc;
|
|
|
|
/* set DMA mask */
|
|
rc = pci_set_dma_mask(pdev, DMA_BIT_MASK(dma_mask));
|
|
if (rc) {
|
|
dev_warn(hdev->dev,
|
|
"Failed to set pci dma mask to %d bits, error %d\n",
|
|
dma_mask, rc);
|
|
|
|
dma_mask = hdev->dma_mask;
|
|
|
|
rc = pci_set_dma_mask(pdev, DMA_BIT_MASK(dma_mask));
|
|
if (rc) {
|
|
dev_err(hdev->dev,
|
|
"Failed to set pci dma mask to %d bits, error %d\n",
|
|
dma_mask, rc);
|
|
return rc;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* We managed to set the dma mask, so update the dma mask field. If
|
|
* the set to the coherent mask will fail with that mask, we will
|
|
* fail the entire function
|
|
*/
|
|
hdev->dma_mask = dma_mask;
|
|
|
|
rc = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(dma_mask));
|
|
if (rc) {
|
|
dev_err(hdev->dev,
|
|
"Failed to set pci consistent dma mask to %d bits, error %d\n",
|
|
dma_mask, rc);
|
|
return rc;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* hl_pci_init() - PCI initialization code.
|
|
* @hdev: Pointer to hl_device structure.
|
|
* @dma_mask: number of bits for the requested dma mask.
|
|
*
|
|
* Set DMA masks, initialize the PCI controller and map the PCI BARs.
|
|
*
|
|
* Return: 0 on success, non-zero for failure.
|
|
*/
|
|
int hl_pci_init(struct hl_device *hdev, u8 dma_mask)
|
|
{
|
|
struct pci_dev *pdev = hdev->pdev;
|
|
int rc;
|
|
|
|
rc = hl_pci_set_dma_mask(hdev, dma_mask);
|
|
if (rc)
|
|
return rc;
|
|
|
|
if (hdev->reset_pcilink)
|
|
hl_pci_reset_link_through_bridge(hdev);
|
|
|
|
rc = pci_enable_device_mem(pdev);
|
|
if (rc) {
|
|
dev_err(hdev->dev, "can't enable PCI device\n");
|
|
return rc;
|
|
}
|
|
|
|
pci_set_master(pdev);
|
|
|
|
rc = hdev->asic_funcs->init_iatu(hdev);
|
|
if (rc) {
|
|
dev_err(hdev->dev, "Failed to initialize iATU\n");
|
|
goto disable_device;
|
|
}
|
|
|
|
rc = hdev->asic_funcs->pci_bars_map(hdev);
|
|
if (rc) {
|
|
dev_err(hdev->dev, "Failed to initialize PCI BARs\n");
|
|
goto disable_device;
|
|
}
|
|
|
|
return 0;
|
|
|
|
disable_device:
|
|
pci_clear_master(pdev);
|
|
pci_disable_device(pdev);
|
|
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* hl_fw_fini() - PCI finalization code.
|
|
* @hdev: Pointer to hl_device structure
|
|
*
|
|
* Unmap PCI bars and disable PCI device.
|
|
*/
|
|
void hl_pci_fini(struct hl_device *hdev)
|
|
{
|
|
hl_pci_bars_unmap(hdev);
|
|
|
|
pci_clear_master(hdev->pdev);
|
|
pci_disable_device(hdev->pdev);
|
|
}
|