mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-23 12:24:54 +07:00
455d56ce80
Extends the user space ioctl for CL submissions so it can include a request to flush the cache once the CL execution has completed. Fixes memory write violation messages reported by the kernel in workloads involving shader memory writes (SSBOs, shader images, scratch, etc) which sometimes also lead to GPU resets during Piglit and CTS workloads. v2: if v3d_job_init() fails we need to kfree() the job instead of v3d_job_put() it (Eric Anholt). v3 (Eric Anholt): - Drop _FLAG suffix from the new flag name. - Add a new param so userspace can tell whether cache flushing is implemented in the kernel. Signed-off-by: Iago Toral Quiroga <itoral@igalia.com> Reviewed-by: Eric Anholt <eric@anholt.net> Signed-off-by: Eric Anholt <eric@anholt.net> Link: https://patchwork.freedesktop.org/patch/msgid/20190919071016.4578-1-itoral@igalia.com
387 lines
9.4 KiB
C
387 lines
9.4 KiB
C
// SPDX-License-Identifier: GPL-2.0+
|
|
/* Copyright (C) 2014-2018 Broadcom */
|
|
|
|
/**
|
|
* DOC: Broadcom V3D Graphics Driver
|
|
*
|
|
* This driver supports the Broadcom V3D 3.3 and 4.1 OpenGL ES GPUs.
|
|
* For V3D 2.x support, see the VC4 driver.
|
|
*
|
|
* The V3D GPU includes a tiled render (composed of a bin and render
|
|
* pipelines), the TFU (texture formatting unit), and the CSD (compute
|
|
* shader dispatch).
|
|
*/
|
|
|
|
#include <linux/clk.h>
|
|
#include <linux/device.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/io.h>
|
|
#include <linux/module.h>
|
|
#include <linux/of_platform.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/pm_runtime.h>
|
|
#include <linux/reset.h>
|
|
|
|
#include <drm/drm_drv.h>
|
|
#include <drm/drm_fb_cma_helper.h>
|
|
#include <drm/drm_fb_helper.h>
|
|
#include <uapi/drm/v3d_drm.h>
|
|
|
|
#include "v3d_drv.h"
|
|
#include "v3d_regs.h"
|
|
|
|
#define DRIVER_NAME "v3d"
|
|
#define DRIVER_DESC "Broadcom V3D graphics"
|
|
#define DRIVER_DATE "20180419"
|
|
#define DRIVER_MAJOR 1
|
|
#define DRIVER_MINOR 0
|
|
#define DRIVER_PATCHLEVEL 0
|
|
|
|
#ifdef CONFIG_PM
|
|
static int v3d_runtime_suspend(struct device *dev)
|
|
{
|
|
struct drm_device *drm = dev_get_drvdata(dev);
|
|
struct v3d_dev *v3d = to_v3d_dev(drm);
|
|
|
|
v3d_irq_disable(v3d);
|
|
|
|
clk_disable_unprepare(v3d->clk);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int v3d_runtime_resume(struct device *dev)
|
|
{
|
|
struct drm_device *drm = dev_get_drvdata(dev);
|
|
struct v3d_dev *v3d = to_v3d_dev(drm);
|
|
int ret;
|
|
|
|
ret = clk_prepare_enable(v3d->clk);
|
|
if (ret != 0)
|
|
return ret;
|
|
|
|
/* XXX: VPM base */
|
|
|
|
v3d_mmu_set_page_table(v3d);
|
|
v3d_irq_enable(v3d);
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static const struct dev_pm_ops v3d_v3d_pm_ops = {
|
|
SET_RUNTIME_PM_OPS(v3d_runtime_suspend, v3d_runtime_resume, NULL)
|
|
};
|
|
|
|
static int v3d_get_param_ioctl(struct drm_device *dev, void *data,
|
|
struct drm_file *file_priv)
|
|
{
|
|
struct v3d_dev *v3d = to_v3d_dev(dev);
|
|
struct drm_v3d_get_param *args = data;
|
|
int ret;
|
|
static const u32 reg_map[] = {
|
|
[DRM_V3D_PARAM_V3D_UIFCFG] = V3D_HUB_UIFCFG,
|
|
[DRM_V3D_PARAM_V3D_HUB_IDENT1] = V3D_HUB_IDENT1,
|
|
[DRM_V3D_PARAM_V3D_HUB_IDENT2] = V3D_HUB_IDENT2,
|
|
[DRM_V3D_PARAM_V3D_HUB_IDENT3] = V3D_HUB_IDENT3,
|
|
[DRM_V3D_PARAM_V3D_CORE0_IDENT0] = V3D_CTL_IDENT0,
|
|
[DRM_V3D_PARAM_V3D_CORE0_IDENT1] = V3D_CTL_IDENT1,
|
|
[DRM_V3D_PARAM_V3D_CORE0_IDENT2] = V3D_CTL_IDENT2,
|
|
};
|
|
|
|
if (args->pad != 0)
|
|
return -EINVAL;
|
|
|
|
/* Note that DRM_V3D_PARAM_V3D_CORE0_IDENT0 is 0, so we need
|
|
* to explicitly allow it in the "the register in our
|
|
* parameter map" check.
|
|
*/
|
|
if (args->param < ARRAY_SIZE(reg_map) &&
|
|
(reg_map[args->param] ||
|
|
args->param == DRM_V3D_PARAM_V3D_CORE0_IDENT0)) {
|
|
u32 offset = reg_map[args->param];
|
|
|
|
if (args->value != 0)
|
|
return -EINVAL;
|
|
|
|
ret = pm_runtime_get_sync(v3d->dev);
|
|
if (ret < 0)
|
|
return ret;
|
|
if (args->param >= DRM_V3D_PARAM_V3D_CORE0_IDENT0 &&
|
|
args->param <= DRM_V3D_PARAM_V3D_CORE0_IDENT2) {
|
|
args->value = V3D_CORE_READ(0, offset);
|
|
} else {
|
|
args->value = V3D_READ(offset);
|
|
}
|
|
pm_runtime_mark_last_busy(v3d->dev);
|
|
pm_runtime_put_autosuspend(v3d->dev);
|
|
return 0;
|
|
}
|
|
|
|
|
|
switch (args->param) {
|
|
case DRM_V3D_PARAM_SUPPORTS_TFU:
|
|
args->value = 1;
|
|
return 0;
|
|
case DRM_V3D_PARAM_SUPPORTS_CSD:
|
|
args->value = v3d_has_csd(v3d);
|
|
return 0;
|
|
case DRM_V3D_PARAM_SUPPORTS_CACHE_FLUSH:
|
|
args->value = 1;
|
|
return 0;
|
|
default:
|
|
DRM_DEBUG("Unknown parameter %d\n", args->param);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
static int
|
|
v3d_open(struct drm_device *dev, struct drm_file *file)
|
|
{
|
|
struct v3d_dev *v3d = to_v3d_dev(dev);
|
|
struct v3d_file_priv *v3d_priv;
|
|
struct drm_sched_rq *rq;
|
|
int i;
|
|
|
|
v3d_priv = kzalloc(sizeof(*v3d_priv), GFP_KERNEL);
|
|
if (!v3d_priv)
|
|
return -ENOMEM;
|
|
|
|
v3d_priv->v3d = v3d;
|
|
|
|
for (i = 0; i < V3D_MAX_QUEUES; i++) {
|
|
rq = &v3d->queue[i].sched.sched_rq[DRM_SCHED_PRIORITY_NORMAL];
|
|
drm_sched_entity_init(&v3d_priv->sched_entity[i], &rq, 1, NULL);
|
|
}
|
|
|
|
file->driver_priv = v3d_priv;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
v3d_postclose(struct drm_device *dev, struct drm_file *file)
|
|
{
|
|
struct v3d_file_priv *v3d_priv = file->driver_priv;
|
|
enum v3d_queue q;
|
|
|
|
for (q = 0; q < V3D_MAX_QUEUES; q++) {
|
|
drm_sched_entity_destroy(&v3d_priv->sched_entity[q]);
|
|
}
|
|
|
|
kfree(v3d_priv);
|
|
}
|
|
|
|
DEFINE_DRM_GEM_SHMEM_FOPS(v3d_drm_fops);
|
|
|
|
/* DRM_AUTH is required on SUBMIT_CL for now, while we don't have GMP
|
|
* protection between clients. Note that render nodes would be be
|
|
* able to submit CLs that could access BOs from clients authenticated
|
|
* with the master node. The TFU doesn't use the GMP, so it would
|
|
* need to stay DRM_AUTH until we do buffer size/offset validation.
|
|
*/
|
|
static const struct drm_ioctl_desc v3d_drm_ioctls[] = {
|
|
DRM_IOCTL_DEF_DRV(V3D_SUBMIT_CL, v3d_submit_cl_ioctl, DRM_RENDER_ALLOW | DRM_AUTH),
|
|
DRM_IOCTL_DEF_DRV(V3D_WAIT_BO, v3d_wait_bo_ioctl, DRM_RENDER_ALLOW),
|
|
DRM_IOCTL_DEF_DRV(V3D_CREATE_BO, v3d_create_bo_ioctl, DRM_RENDER_ALLOW),
|
|
DRM_IOCTL_DEF_DRV(V3D_MMAP_BO, v3d_mmap_bo_ioctl, DRM_RENDER_ALLOW),
|
|
DRM_IOCTL_DEF_DRV(V3D_GET_PARAM, v3d_get_param_ioctl, DRM_RENDER_ALLOW),
|
|
DRM_IOCTL_DEF_DRV(V3D_GET_BO_OFFSET, v3d_get_bo_offset_ioctl, DRM_RENDER_ALLOW),
|
|
DRM_IOCTL_DEF_DRV(V3D_SUBMIT_TFU, v3d_submit_tfu_ioctl, DRM_RENDER_ALLOW | DRM_AUTH),
|
|
DRM_IOCTL_DEF_DRV(V3D_SUBMIT_CSD, v3d_submit_csd_ioctl, DRM_RENDER_ALLOW | DRM_AUTH),
|
|
};
|
|
|
|
static struct drm_driver v3d_drm_driver = {
|
|
.driver_features = (DRIVER_GEM |
|
|
DRIVER_RENDER |
|
|
DRIVER_SYNCOBJ),
|
|
|
|
.open = v3d_open,
|
|
.postclose = v3d_postclose,
|
|
|
|
#if defined(CONFIG_DEBUG_FS)
|
|
.debugfs_init = v3d_debugfs_init,
|
|
#endif
|
|
|
|
.gem_create_object = v3d_create_object,
|
|
.prime_handle_to_fd = drm_gem_prime_handle_to_fd,
|
|
.prime_fd_to_handle = drm_gem_prime_fd_to_handle,
|
|
.gem_prime_import_sg_table = v3d_prime_import_sg_table,
|
|
.gem_prime_mmap = drm_gem_prime_mmap,
|
|
|
|
.ioctls = v3d_drm_ioctls,
|
|
.num_ioctls = ARRAY_SIZE(v3d_drm_ioctls),
|
|
.fops = &v3d_drm_fops,
|
|
|
|
.name = DRIVER_NAME,
|
|
.desc = DRIVER_DESC,
|
|
.date = DRIVER_DATE,
|
|
.major = DRIVER_MAJOR,
|
|
.minor = DRIVER_MINOR,
|
|
.patchlevel = DRIVER_PATCHLEVEL,
|
|
};
|
|
|
|
static const struct of_device_id v3d_of_match[] = {
|
|
{ .compatible = "brcm,7268-v3d" },
|
|
{ .compatible = "brcm,7278-v3d" },
|
|
{},
|
|
};
|
|
MODULE_DEVICE_TABLE(of, v3d_of_match);
|
|
|
|
static int
|
|
map_regs(struct v3d_dev *v3d, void __iomem **regs, const char *name)
|
|
{
|
|
struct resource *res =
|
|
platform_get_resource_byname(v3d->pdev, IORESOURCE_MEM, name);
|
|
|
|
*regs = devm_ioremap_resource(v3d->dev, res);
|
|
return PTR_ERR_OR_ZERO(*regs);
|
|
}
|
|
|
|
static int v3d_platform_drm_probe(struct platform_device *pdev)
|
|
{
|
|
struct device *dev = &pdev->dev;
|
|
struct drm_device *drm;
|
|
struct v3d_dev *v3d;
|
|
int ret;
|
|
u32 mmu_debug;
|
|
u32 ident1;
|
|
|
|
|
|
v3d = kzalloc(sizeof(*v3d), GFP_KERNEL);
|
|
if (!v3d)
|
|
return -ENOMEM;
|
|
v3d->dev = dev;
|
|
v3d->pdev = pdev;
|
|
drm = &v3d->drm;
|
|
|
|
ret = map_regs(v3d, &v3d->hub_regs, "hub");
|
|
if (ret)
|
|
goto dev_free;
|
|
|
|
ret = map_regs(v3d, &v3d->core_regs[0], "core0");
|
|
if (ret)
|
|
goto dev_free;
|
|
|
|
mmu_debug = V3D_READ(V3D_MMU_DEBUG_INFO);
|
|
dev->coherent_dma_mask =
|
|
DMA_BIT_MASK(30 + V3D_GET_FIELD(mmu_debug, V3D_MMU_PA_WIDTH));
|
|
v3d->va_width = 30 + V3D_GET_FIELD(mmu_debug, V3D_MMU_VA_WIDTH);
|
|
|
|
ident1 = V3D_READ(V3D_HUB_IDENT1);
|
|
v3d->ver = (V3D_GET_FIELD(ident1, V3D_HUB_IDENT1_TVER) * 10 +
|
|
V3D_GET_FIELD(ident1, V3D_HUB_IDENT1_REV));
|
|
v3d->cores = V3D_GET_FIELD(ident1, V3D_HUB_IDENT1_NCORES);
|
|
WARN_ON(v3d->cores > 1); /* multicore not yet implemented */
|
|
|
|
v3d->reset = devm_reset_control_get_exclusive(dev, NULL);
|
|
if (IS_ERR(v3d->reset)) {
|
|
ret = PTR_ERR(v3d->reset);
|
|
|
|
if (ret == -EPROBE_DEFER)
|
|
goto dev_free;
|
|
|
|
v3d->reset = NULL;
|
|
ret = map_regs(v3d, &v3d->bridge_regs, "bridge");
|
|
if (ret) {
|
|
dev_err(dev,
|
|
"Failed to get reset control or bridge regs\n");
|
|
goto dev_free;
|
|
}
|
|
}
|
|
|
|
if (v3d->ver < 41) {
|
|
ret = map_regs(v3d, &v3d->gca_regs, "gca");
|
|
if (ret)
|
|
goto dev_free;
|
|
}
|
|
|
|
v3d->mmu_scratch = dma_alloc_wc(dev, 4096, &v3d->mmu_scratch_paddr,
|
|
GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO);
|
|
if (!v3d->mmu_scratch) {
|
|
dev_err(dev, "Failed to allocate MMU scratch page\n");
|
|
ret = -ENOMEM;
|
|
goto dev_free;
|
|
}
|
|
|
|
pm_runtime_use_autosuspend(dev);
|
|
pm_runtime_set_autosuspend_delay(dev, 50);
|
|
pm_runtime_enable(dev);
|
|
|
|
ret = drm_dev_init(&v3d->drm, &v3d_drm_driver, dev);
|
|
if (ret)
|
|
goto dma_free;
|
|
|
|
platform_set_drvdata(pdev, drm);
|
|
drm->dev_private = v3d;
|
|
|
|
ret = v3d_gem_init(drm);
|
|
if (ret)
|
|
goto dev_destroy;
|
|
|
|
ret = v3d_irq_init(v3d);
|
|
if (ret)
|
|
goto gem_destroy;
|
|
|
|
ret = drm_dev_register(drm, 0);
|
|
if (ret)
|
|
goto irq_disable;
|
|
|
|
return 0;
|
|
|
|
irq_disable:
|
|
v3d_irq_disable(v3d);
|
|
gem_destroy:
|
|
v3d_gem_destroy(drm);
|
|
dev_destroy:
|
|
drm_dev_put(drm);
|
|
dma_free:
|
|
dma_free_wc(dev, 4096, v3d->mmu_scratch, v3d->mmu_scratch_paddr);
|
|
dev_free:
|
|
kfree(v3d);
|
|
return ret;
|
|
}
|
|
|
|
static int v3d_platform_drm_remove(struct platform_device *pdev)
|
|
{
|
|
struct drm_device *drm = platform_get_drvdata(pdev);
|
|
struct v3d_dev *v3d = to_v3d_dev(drm);
|
|
|
|
drm_dev_unregister(drm);
|
|
|
|
v3d_gem_destroy(drm);
|
|
|
|
drm_dev_put(drm);
|
|
|
|
dma_free_wc(v3d->dev, 4096, v3d->mmu_scratch, v3d->mmu_scratch_paddr);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct platform_driver v3d_platform_driver = {
|
|
.probe = v3d_platform_drm_probe,
|
|
.remove = v3d_platform_drm_remove,
|
|
.driver = {
|
|
.name = "v3d",
|
|
.of_match_table = v3d_of_match,
|
|
},
|
|
};
|
|
|
|
static int __init v3d_drm_register(void)
|
|
{
|
|
return platform_driver_register(&v3d_platform_driver);
|
|
}
|
|
|
|
static void __exit v3d_drm_unregister(void)
|
|
{
|
|
platform_driver_unregister(&v3d_platform_driver);
|
|
}
|
|
|
|
module_init(v3d_drm_register);
|
|
module_exit(v3d_drm_unregister);
|
|
|
|
MODULE_ALIAS("platform:v3d-drm");
|
|
MODULE_DESCRIPTION("Broadcom V3D DRM Driver");
|
|
MODULE_AUTHOR("Eric Anholt <eric@anholt.net>");
|
|
MODULE_LICENSE("GPL v2");
|