mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-21 09:34:51 +07:00
8a25c4be58
Start using device specific parameters instead of module parameters for most things. The module parameters become the immutable initial values for i915 parameters. The device specific parameters in i915->params start life as a copy of i915_modparams. Any later changes are only reflected in the debugfs. The stragglers are: * i915.force_probe and i915.modeset. Needed before dev_priv is available. This is fine because the parameters are read-only and never modified. * i915.verbose_state_checks. Passing dev_priv to I915_STATE_WARN and I915_STATE_WARN_ON would result in massive and ugly churn. This is handled by not exposing the parameter via debugfs, and leaving the parameter writable in sysfs. This may be fixed up in follow-up work. * i915.inject_probe_failure. Only makes sense in terms of the module, not the device. This is handled by not exposing the parameter via debugfs. v2: Fix uc i915 lookup code (Michał Winiarski) Cc: Juha-Pekka Heikkilä <juha-pekka.heikkila@intel.com> Cc: Venkata Sandeep Dhanalakota <venkata.s.dhanalakota@intel.com> Cc: Michał Winiarski <michal.winiarski@intel.com> Reviewed-by: Rodrigo Vivi <rodrigo.vivi@intel.com> Acked-by: Michał Winiarski <michal.winiarski@intel.com> Signed-off-by: Jani Nikula <jani.nikula@intel.com> Link: http://patchwork.freedesktop.org/patch/msgid/20200618150402.14022-1-jani.nikula@intel.com
139 lines
3.3 KiB
C
139 lines
3.3 KiB
C
// SPDX-License-Identifier: MIT
|
|
/*
|
|
* Copyright © 2019 Intel Corporation
|
|
*/
|
|
|
|
#include "i915_drv.h"
|
|
#include "intel_memory_region.h"
|
|
#include "gem/i915_gem_lmem.h"
|
|
#include "gem/i915_gem_region.h"
|
|
#include "intel_region_lmem.h"
|
|
|
|
static int init_fake_lmem_bar(struct intel_memory_region *mem)
|
|
{
|
|
struct drm_i915_private *i915 = mem->i915;
|
|
struct i915_ggtt *ggtt = &i915->ggtt;
|
|
unsigned long n;
|
|
int ret;
|
|
|
|
/* We want to 1:1 map the mappable aperture to our reserved region */
|
|
|
|
mem->fake_mappable.start = 0;
|
|
mem->fake_mappable.size = resource_size(&mem->region);
|
|
mem->fake_mappable.color = I915_COLOR_UNEVICTABLE;
|
|
|
|
ret = drm_mm_reserve_node(&ggtt->vm.mm, &mem->fake_mappable);
|
|
if (ret)
|
|
return ret;
|
|
|
|
mem->remap_addr = dma_map_resource(&i915->drm.pdev->dev,
|
|
mem->region.start,
|
|
mem->fake_mappable.size,
|
|
PCI_DMA_BIDIRECTIONAL,
|
|
DMA_ATTR_FORCE_CONTIGUOUS);
|
|
if (dma_mapping_error(&i915->drm.pdev->dev, mem->remap_addr)) {
|
|
drm_mm_remove_node(&mem->fake_mappable);
|
|
return -EINVAL;
|
|
}
|
|
|
|
for (n = 0; n < mem->fake_mappable.size >> PAGE_SHIFT; ++n) {
|
|
ggtt->vm.insert_page(&ggtt->vm,
|
|
mem->remap_addr + (n << PAGE_SHIFT),
|
|
n << PAGE_SHIFT,
|
|
I915_CACHE_NONE, 0);
|
|
}
|
|
|
|
mem->region = (struct resource)DEFINE_RES_MEM(mem->remap_addr,
|
|
mem->fake_mappable.size);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void release_fake_lmem_bar(struct intel_memory_region *mem)
|
|
{
|
|
if (!drm_mm_node_allocated(&mem->fake_mappable))
|
|
return;
|
|
|
|
drm_mm_remove_node(&mem->fake_mappable);
|
|
|
|
dma_unmap_resource(&mem->i915->drm.pdev->dev,
|
|
mem->remap_addr,
|
|
mem->fake_mappable.size,
|
|
PCI_DMA_BIDIRECTIONAL,
|
|
DMA_ATTR_FORCE_CONTIGUOUS);
|
|
}
|
|
|
|
static void
|
|
region_lmem_release(struct intel_memory_region *mem)
|
|
{
|
|
release_fake_lmem_bar(mem);
|
|
io_mapping_fini(&mem->iomap);
|
|
intel_memory_region_release_buddy(mem);
|
|
}
|
|
|
|
static int
|
|
region_lmem_init(struct intel_memory_region *mem)
|
|
{
|
|
int ret;
|
|
|
|
if (mem->i915->params.fake_lmem_start) {
|
|
ret = init_fake_lmem_bar(mem);
|
|
GEM_BUG_ON(ret);
|
|
}
|
|
|
|
if (!io_mapping_init_wc(&mem->iomap,
|
|
mem->io_start,
|
|
resource_size(&mem->region)))
|
|
return -EIO;
|
|
|
|
ret = intel_memory_region_init_buddy(mem);
|
|
if (ret)
|
|
io_mapping_fini(&mem->iomap);
|
|
|
|
intel_memory_region_set_name(mem, "local");
|
|
|
|
return ret;
|
|
}
|
|
|
|
const struct intel_memory_region_ops intel_region_lmem_ops = {
|
|
.init = region_lmem_init,
|
|
.release = region_lmem_release,
|
|
.create_object = __i915_gem_lmem_object_create,
|
|
};
|
|
|
|
struct intel_memory_region *
|
|
intel_setup_fake_lmem(struct drm_i915_private *i915)
|
|
{
|
|
struct pci_dev *pdev = i915->drm.pdev;
|
|
struct intel_memory_region *mem;
|
|
resource_size_t mappable_end;
|
|
resource_size_t io_start;
|
|
resource_size_t start;
|
|
|
|
GEM_BUG_ON(i915_ggtt_has_aperture(&i915->ggtt));
|
|
GEM_BUG_ON(!i915->params.fake_lmem_start);
|
|
|
|
/* Your mappable aperture belongs to me now! */
|
|
mappable_end = pci_resource_len(pdev, 2);
|
|
io_start = pci_resource_start(pdev, 2),
|
|
start = i915->params.fake_lmem_start;
|
|
|
|
mem = intel_memory_region_create(i915,
|
|
start,
|
|
mappable_end,
|
|
PAGE_SIZE,
|
|
io_start,
|
|
&intel_region_lmem_ops);
|
|
if (!IS_ERR(mem)) {
|
|
drm_info(&i915->drm, "Intel graphics fake LMEM: %pR\n",
|
|
&mem->region);
|
|
drm_info(&i915->drm,
|
|
"Intel graphics fake LMEM IO start: %llx\n",
|
|
(u64)mem->io_start);
|
|
drm_info(&i915->drm, "Intel graphics fake LMEM size: %llx\n",
|
|
(u64)resource_size(&mem->region));
|
|
}
|
|
|
|
return mem;
|
|
}
|