mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
635430797d
There's a few issues with the runtime PM initialization. The documentation states pm_runtime_set_active() should be called before pm_runtime_enable(). The pm_runtime_put_autosuspend() could suspend the GPU before panfrost_perfcnt_init() is called which touches the h/w. The autosuspend delay keeps things from breaking. There's no need explicitly power off the GPU only to wake back up with pm_runtime_get_sync(). Just delaying pm_runtime_enable to the end of probe is sufficient. Lets move all the runtime PM calls into the probe() function so they are all in one place and are done after all initialization. Cc: Tomeu Vizoso <tomeu.vizoso@collabora.com> Cc: Steven Price <steven.price@arm.com> Cc: David Airlie <airlied@linux.ie> Cc: Daniel Vetter <daniel@ffwll.ch> Acked-by: Alyssa Rosenzweig <alyssa.rosenzweig@collabora.com> Signed-off-by: Rob Herring <robh@kernel.org> Reviewed-by: Robin Murphy <robin.murphy@arm.com> Reviewed-by: Steven Price <steven.price@arm.com> Link: https://patchwork.freedesktop.org/patch/msgid/20190826223317.28509-2-robh@kernel.org
619 lines
16 KiB
C
619 lines
16 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/* Copyright 2018 Marty E. Plummer <hanetzer@startmail.com> */
|
|
/* Copyright 2019 Linaro, Ltd., Rob Herring <robh@kernel.org> */
|
|
/* Copyright 2019 Collabora ltd. */
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/of_platform.h>
|
|
#include <linux/pagemap.h>
|
|
#include <linux/pm_runtime.h>
|
|
#include <drm/panfrost_drm.h>
|
|
#include <drm/drm_drv.h>
|
|
#include <drm/drm_ioctl.h>
|
|
#include <drm/drm_syncobj.h>
|
|
#include <drm/drm_utils.h>
|
|
|
|
#include "panfrost_device.h"
|
|
#include "panfrost_devfreq.h"
|
|
#include "panfrost_gem.h"
|
|
#include "panfrost_mmu.h"
|
|
#include "panfrost_job.h"
|
|
#include "panfrost_gpu.h"
|
|
#include "panfrost_perfcnt.h"
|
|
|
|
static bool unstable_ioctls;
|
|
module_param_unsafe(unstable_ioctls, bool, 0600);
|
|
|
|
static int panfrost_ioctl_get_param(struct drm_device *ddev, void *data, struct drm_file *file)
|
|
{
|
|
struct drm_panfrost_get_param *param = data;
|
|
struct panfrost_device *pfdev = ddev->dev_private;
|
|
|
|
if (param->pad != 0)
|
|
return -EINVAL;
|
|
|
|
#define PANFROST_FEATURE(name, member) \
|
|
case DRM_PANFROST_PARAM_ ## name: \
|
|
param->value = pfdev->features.member; \
|
|
break
|
|
#define PANFROST_FEATURE_ARRAY(name, member, max) \
|
|
case DRM_PANFROST_PARAM_ ## name ## 0 ... \
|
|
DRM_PANFROST_PARAM_ ## name ## max: \
|
|
param->value = pfdev->features.member[param->param - \
|
|
DRM_PANFROST_PARAM_ ## name ## 0]; \
|
|
break
|
|
|
|
switch (param->param) {
|
|
PANFROST_FEATURE(GPU_PROD_ID, id);
|
|
PANFROST_FEATURE(GPU_REVISION, revision);
|
|
PANFROST_FEATURE(SHADER_PRESENT, shader_present);
|
|
PANFROST_FEATURE(TILER_PRESENT, tiler_present);
|
|
PANFROST_FEATURE(L2_PRESENT, l2_present);
|
|
PANFROST_FEATURE(STACK_PRESENT, stack_present);
|
|
PANFROST_FEATURE(AS_PRESENT, as_present);
|
|
PANFROST_FEATURE(JS_PRESENT, js_present);
|
|
PANFROST_FEATURE(L2_FEATURES, l2_features);
|
|
PANFROST_FEATURE(CORE_FEATURES, core_features);
|
|
PANFROST_FEATURE(TILER_FEATURES, tiler_features);
|
|
PANFROST_FEATURE(MEM_FEATURES, mem_features);
|
|
PANFROST_FEATURE(MMU_FEATURES, mmu_features);
|
|
PANFROST_FEATURE(THREAD_FEATURES, thread_features);
|
|
PANFROST_FEATURE(MAX_THREADS, max_threads);
|
|
PANFROST_FEATURE(THREAD_MAX_WORKGROUP_SZ,
|
|
thread_max_workgroup_sz);
|
|
PANFROST_FEATURE(THREAD_MAX_BARRIER_SZ,
|
|
thread_max_barrier_sz);
|
|
PANFROST_FEATURE(COHERENCY_FEATURES, coherency_features);
|
|
PANFROST_FEATURE_ARRAY(TEXTURE_FEATURES, texture_features, 3);
|
|
PANFROST_FEATURE_ARRAY(JS_FEATURES, js_features, 15);
|
|
PANFROST_FEATURE(NR_CORE_GROUPS, nr_core_groups);
|
|
PANFROST_FEATURE(THREAD_TLS_ALLOC, thread_tls_alloc);
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int panfrost_ioctl_create_bo(struct drm_device *dev, void *data,
|
|
struct drm_file *file)
|
|
{
|
|
struct panfrost_gem_object *bo;
|
|
struct drm_panfrost_create_bo *args = data;
|
|
|
|
if (!args->size || args->pad ||
|
|
(args->flags & ~(PANFROST_BO_NOEXEC | PANFROST_BO_HEAP)))
|
|
return -EINVAL;
|
|
|
|
/* Heaps should never be executable */
|
|
if ((args->flags & PANFROST_BO_HEAP) &&
|
|
!(args->flags & PANFROST_BO_NOEXEC))
|
|
return -EINVAL;
|
|
|
|
bo = panfrost_gem_create_with_handle(file, dev, args->size, args->flags,
|
|
&args->handle);
|
|
if (IS_ERR(bo))
|
|
return PTR_ERR(bo);
|
|
|
|
args->offset = bo->node.start << PAGE_SHIFT;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* panfrost_lookup_bos() - Sets up job->bo[] with the GEM objects
|
|
* referenced by the job.
|
|
* @dev: DRM device
|
|
* @file_priv: DRM file for this fd
|
|
* @args: IOCTL args
|
|
* @job: job being set up
|
|
*
|
|
* Resolve handles from userspace to BOs and attach them to job.
|
|
*
|
|
* Note that this function doesn't need to unreference the BOs on
|
|
* failure, because that will happen at panfrost_job_cleanup() time.
|
|
*/
|
|
static int
|
|
panfrost_lookup_bos(struct drm_device *dev,
|
|
struct drm_file *file_priv,
|
|
struct drm_panfrost_submit *args,
|
|
struct panfrost_job *job)
|
|
{
|
|
job->bo_count = args->bo_handle_count;
|
|
|
|
if (!job->bo_count)
|
|
return 0;
|
|
|
|
job->implicit_fences = kvmalloc_array(job->bo_count,
|
|
sizeof(struct dma_fence *),
|
|
GFP_KERNEL | __GFP_ZERO);
|
|
if (!job->implicit_fences)
|
|
return -ENOMEM;
|
|
|
|
return drm_gem_objects_lookup(file_priv,
|
|
(void __user *)(uintptr_t)args->bo_handles,
|
|
job->bo_count, &job->bos);
|
|
}
|
|
|
|
/**
|
|
* panfrost_copy_in_sync() - Sets up job->in_fences[] with the sync objects
|
|
* referenced by the job.
|
|
* @dev: DRM device
|
|
* @file_priv: DRM file for this fd
|
|
* @args: IOCTL args
|
|
* @job: job being set up
|
|
*
|
|
* Resolve syncobjs from userspace to fences and attach them to job.
|
|
*
|
|
* Note that this function doesn't need to unreference the fences on
|
|
* failure, because that will happen at panfrost_job_cleanup() time.
|
|
*/
|
|
static int
|
|
panfrost_copy_in_sync(struct drm_device *dev,
|
|
struct drm_file *file_priv,
|
|
struct drm_panfrost_submit *args,
|
|
struct panfrost_job *job)
|
|
{
|
|
u32 *handles;
|
|
int ret = 0;
|
|
int i;
|
|
|
|
job->in_fence_count = args->in_sync_count;
|
|
|
|
if (!job->in_fence_count)
|
|
return 0;
|
|
|
|
job->in_fences = kvmalloc_array(job->in_fence_count,
|
|
sizeof(struct dma_fence *),
|
|
GFP_KERNEL | __GFP_ZERO);
|
|
if (!job->in_fences) {
|
|
DRM_DEBUG("Failed to allocate job in fences\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
handles = kvmalloc_array(job->in_fence_count, sizeof(u32), GFP_KERNEL);
|
|
if (!handles) {
|
|
ret = -ENOMEM;
|
|
DRM_DEBUG("Failed to allocate incoming syncobj handles\n");
|
|
goto fail;
|
|
}
|
|
|
|
if (copy_from_user(handles,
|
|
(void __user *)(uintptr_t)args->in_syncs,
|
|
job->in_fence_count * sizeof(u32))) {
|
|
ret = -EFAULT;
|
|
DRM_DEBUG("Failed to copy in syncobj handles\n");
|
|
goto fail;
|
|
}
|
|
|
|
for (i = 0; i < job->in_fence_count; i++) {
|
|
ret = drm_syncobj_find_fence(file_priv, handles[i], 0, 0,
|
|
&job->in_fences[i]);
|
|
if (ret == -EINVAL)
|
|
goto fail;
|
|
}
|
|
|
|
fail:
|
|
kvfree(handles);
|
|
return ret;
|
|
}
|
|
|
|
static int panfrost_ioctl_submit(struct drm_device *dev, void *data,
|
|
struct drm_file *file)
|
|
{
|
|
struct panfrost_device *pfdev = dev->dev_private;
|
|
struct drm_panfrost_submit *args = data;
|
|
struct drm_syncobj *sync_out = NULL;
|
|
struct panfrost_job *job;
|
|
int ret = 0;
|
|
|
|
if (!args->jc)
|
|
return -EINVAL;
|
|
|
|
if (args->requirements && args->requirements != PANFROST_JD_REQ_FS)
|
|
return -EINVAL;
|
|
|
|
if (args->out_sync > 0) {
|
|
sync_out = drm_syncobj_find(file, args->out_sync);
|
|
if (!sync_out)
|
|
return -ENODEV;
|
|
}
|
|
|
|
job = kzalloc(sizeof(*job), GFP_KERNEL);
|
|
if (!job) {
|
|
ret = -ENOMEM;
|
|
goto fail_out_sync;
|
|
}
|
|
|
|
kref_init(&job->refcount);
|
|
|
|
job->pfdev = pfdev;
|
|
job->jc = args->jc;
|
|
job->requirements = args->requirements;
|
|
job->flush_id = panfrost_gpu_get_latest_flush_id(pfdev);
|
|
job->file_priv = file->driver_priv;
|
|
|
|
ret = panfrost_copy_in_sync(dev, file, args, job);
|
|
if (ret)
|
|
goto fail_job;
|
|
|
|
ret = panfrost_lookup_bos(dev, file, args, job);
|
|
if (ret)
|
|
goto fail_job;
|
|
|
|
ret = panfrost_job_push(job);
|
|
if (ret)
|
|
goto fail_job;
|
|
|
|
/* Update the return sync object for the job */
|
|
if (sync_out)
|
|
drm_syncobj_replace_fence(sync_out, job->render_done_fence);
|
|
|
|
fail_job:
|
|
panfrost_job_put(job);
|
|
fail_out_sync:
|
|
if (sync_out)
|
|
drm_syncobj_put(sync_out);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int
|
|
panfrost_ioctl_wait_bo(struct drm_device *dev, void *data,
|
|
struct drm_file *file_priv)
|
|
{
|
|
long ret;
|
|
struct drm_panfrost_wait_bo *args = data;
|
|
struct drm_gem_object *gem_obj;
|
|
unsigned long timeout = drm_timeout_abs_to_jiffies(args->timeout_ns);
|
|
|
|
if (args->pad)
|
|
return -EINVAL;
|
|
|
|
gem_obj = drm_gem_object_lookup(file_priv, args->handle);
|
|
if (!gem_obj)
|
|
return -ENOENT;
|
|
|
|
ret = dma_resv_wait_timeout_rcu(gem_obj->resv, true,
|
|
true, timeout);
|
|
if (!ret)
|
|
ret = timeout ? -ETIMEDOUT : -EBUSY;
|
|
|
|
drm_gem_object_put_unlocked(gem_obj);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int panfrost_ioctl_mmap_bo(struct drm_device *dev, void *data,
|
|
struct drm_file *file_priv)
|
|
{
|
|
struct drm_panfrost_mmap_bo *args = data;
|
|
struct drm_gem_object *gem_obj;
|
|
int ret;
|
|
|
|
if (args->flags != 0) {
|
|
DRM_INFO("unknown mmap_bo flags: %d\n", args->flags);
|
|
return -EINVAL;
|
|
}
|
|
|
|
gem_obj = drm_gem_object_lookup(file_priv, args->handle);
|
|
if (!gem_obj) {
|
|
DRM_DEBUG("Failed to look up GEM BO %d\n", args->handle);
|
|
return -ENOENT;
|
|
}
|
|
|
|
/* Don't allow mmapping of heap objects as pages are not pinned. */
|
|
if (to_panfrost_bo(gem_obj)->is_heap)
|
|
return -EINVAL;
|
|
|
|
ret = drm_gem_create_mmap_offset(gem_obj);
|
|
if (ret == 0)
|
|
args->offset = drm_vma_node_offset_addr(&gem_obj->vma_node);
|
|
drm_gem_object_put_unlocked(gem_obj);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int panfrost_ioctl_get_bo_offset(struct drm_device *dev, void *data,
|
|
struct drm_file *file_priv)
|
|
{
|
|
struct drm_panfrost_get_bo_offset *args = data;
|
|
struct drm_gem_object *gem_obj;
|
|
struct panfrost_gem_object *bo;
|
|
|
|
gem_obj = drm_gem_object_lookup(file_priv, args->handle);
|
|
if (!gem_obj) {
|
|
DRM_DEBUG("Failed to look up GEM BO %d\n", args->handle);
|
|
return -ENOENT;
|
|
}
|
|
bo = to_panfrost_bo(gem_obj);
|
|
|
|
args->offset = bo->node.start << PAGE_SHIFT;
|
|
|
|
drm_gem_object_put_unlocked(gem_obj);
|
|
return 0;
|
|
}
|
|
|
|
static int panfrost_ioctl_madvise(struct drm_device *dev, void *data,
|
|
struct drm_file *file_priv)
|
|
{
|
|
struct drm_panfrost_madvise *args = data;
|
|
struct panfrost_device *pfdev = dev->dev_private;
|
|
struct drm_gem_object *gem_obj;
|
|
|
|
gem_obj = drm_gem_object_lookup(file_priv, args->handle);
|
|
if (!gem_obj) {
|
|
DRM_DEBUG("Failed to look up GEM BO %d\n", args->handle);
|
|
return -ENOENT;
|
|
}
|
|
|
|
args->retained = drm_gem_shmem_madvise(gem_obj, args->madv);
|
|
|
|
if (args->retained) {
|
|
struct panfrost_gem_object *bo = to_panfrost_bo(gem_obj);
|
|
|
|
mutex_lock(&pfdev->shrinker_lock);
|
|
|
|
if (args->madv == PANFROST_MADV_DONTNEED)
|
|
list_add_tail(&bo->base.madv_list, &pfdev->shrinker_list);
|
|
else if (args->madv == PANFROST_MADV_WILLNEED)
|
|
list_del_init(&bo->base.madv_list);
|
|
|
|
mutex_unlock(&pfdev->shrinker_lock);
|
|
}
|
|
|
|
drm_gem_object_put_unlocked(gem_obj);
|
|
return 0;
|
|
}
|
|
|
|
int panfrost_unstable_ioctl_check(void)
|
|
{
|
|
if (!unstable_ioctls)
|
|
return -ENOSYS;
|
|
|
|
return 0;
|
|
}
|
|
|
|
#define PFN_4G (SZ_4G >> PAGE_SHIFT)
|
|
#define PFN_4G_MASK (PFN_4G - 1)
|
|
#define PFN_16M (SZ_16M >> PAGE_SHIFT)
|
|
|
|
static void panfrost_drm_mm_color_adjust(const struct drm_mm_node *node,
|
|
unsigned long color,
|
|
u64 *start, u64 *end)
|
|
{
|
|
/* Executable buffers can't start or end on a 4GB boundary */
|
|
if (!(color & PANFROST_BO_NOEXEC)) {
|
|
u64 next_seg;
|
|
|
|
if ((*start & PFN_4G_MASK) == 0)
|
|
(*start)++;
|
|
|
|
if ((*end & PFN_4G_MASK) == 0)
|
|
(*end)--;
|
|
|
|
next_seg = ALIGN(*start, PFN_4G);
|
|
if (next_seg - *start <= PFN_16M)
|
|
*start = next_seg + 1;
|
|
|
|
*end = min(*end, ALIGN(*start, PFN_4G) - 1);
|
|
}
|
|
}
|
|
|
|
static int
|
|
panfrost_open(struct drm_device *dev, struct drm_file *file)
|
|
{
|
|
int ret;
|
|
struct panfrost_device *pfdev = dev->dev_private;
|
|
struct panfrost_file_priv *panfrost_priv;
|
|
|
|
panfrost_priv = kzalloc(sizeof(*panfrost_priv), GFP_KERNEL);
|
|
if (!panfrost_priv)
|
|
return -ENOMEM;
|
|
|
|
panfrost_priv->pfdev = pfdev;
|
|
file->driver_priv = panfrost_priv;
|
|
|
|
spin_lock_init(&panfrost_priv->mm_lock);
|
|
|
|
/* 4G enough for now. can be 48-bit */
|
|
drm_mm_init(&panfrost_priv->mm, SZ_32M >> PAGE_SHIFT, (SZ_4G - SZ_32M) >> PAGE_SHIFT);
|
|
panfrost_priv->mm.color_adjust = panfrost_drm_mm_color_adjust;
|
|
|
|
ret = panfrost_mmu_pgtable_alloc(panfrost_priv);
|
|
if (ret)
|
|
goto err_pgtable;
|
|
|
|
ret = panfrost_job_open(panfrost_priv);
|
|
if (ret)
|
|
goto err_job;
|
|
|
|
return 0;
|
|
|
|
err_job:
|
|
panfrost_mmu_pgtable_free(panfrost_priv);
|
|
err_pgtable:
|
|
drm_mm_takedown(&panfrost_priv->mm);
|
|
kfree(panfrost_priv);
|
|
return ret;
|
|
}
|
|
|
|
static void
|
|
panfrost_postclose(struct drm_device *dev, struct drm_file *file)
|
|
{
|
|
struct panfrost_file_priv *panfrost_priv = file->driver_priv;
|
|
|
|
panfrost_perfcnt_close(panfrost_priv);
|
|
panfrost_job_close(panfrost_priv);
|
|
|
|
panfrost_mmu_pgtable_free(panfrost_priv);
|
|
drm_mm_takedown(&panfrost_priv->mm);
|
|
kfree(panfrost_priv);
|
|
}
|
|
|
|
/* DRM_AUTH is required on SUBMIT for now, while all clients share a single
|
|
* address space. Note that render nodes would be able to submit jobs that
|
|
* could access BOs from clients authenticated with the master node.
|
|
*/
|
|
static const struct drm_ioctl_desc panfrost_drm_driver_ioctls[] = {
|
|
#define PANFROST_IOCTL(n, func, flags) \
|
|
DRM_IOCTL_DEF_DRV(PANFROST_##n, panfrost_ioctl_##func, flags)
|
|
|
|
PANFROST_IOCTL(SUBMIT, submit, DRM_RENDER_ALLOW | DRM_AUTH),
|
|
PANFROST_IOCTL(WAIT_BO, wait_bo, DRM_RENDER_ALLOW),
|
|
PANFROST_IOCTL(CREATE_BO, create_bo, DRM_RENDER_ALLOW),
|
|
PANFROST_IOCTL(MMAP_BO, mmap_bo, DRM_RENDER_ALLOW),
|
|
PANFROST_IOCTL(GET_PARAM, get_param, DRM_RENDER_ALLOW),
|
|
PANFROST_IOCTL(GET_BO_OFFSET, get_bo_offset, DRM_RENDER_ALLOW),
|
|
PANFROST_IOCTL(PERFCNT_ENABLE, perfcnt_enable, DRM_RENDER_ALLOW),
|
|
PANFROST_IOCTL(PERFCNT_DUMP, perfcnt_dump, DRM_RENDER_ALLOW),
|
|
PANFROST_IOCTL(MADVISE, madvise, DRM_RENDER_ALLOW),
|
|
};
|
|
|
|
DEFINE_DRM_GEM_SHMEM_FOPS(panfrost_drm_driver_fops);
|
|
|
|
/*
|
|
* Panfrost driver version:
|
|
* - 1.0 - initial interface
|
|
* - 1.1 - adds HEAP and NOEXEC flags for CREATE_BO
|
|
*/
|
|
static struct drm_driver panfrost_drm_driver = {
|
|
.driver_features = DRIVER_RENDER | DRIVER_GEM | DRIVER_SYNCOBJ,
|
|
.open = panfrost_open,
|
|
.postclose = panfrost_postclose,
|
|
.ioctls = panfrost_drm_driver_ioctls,
|
|
.num_ioctls = ARRAY_SIZE(panfrost_drm_driver_ioctls),
|
|
.fops = &panfrost_drm_driver_fops,
|
|
.name = "panfrost",
|
|
.desc = "panfrost DRM",
|
|
.date = "20180908",
|
|
.major = 1,
|
|
.minor = 1,
|
|
|
|
.gem_create_object = panfrost_gem_create_object,
|
|
.prime_handle_to_fd = drm_gem_prime_handle_to_fd,
|
|
.prime_fd_to_handle = drm_gem_prime_fd_to_handle,
|
|
.gem_prime_import_sg_table = panfrost_gem_prime_import_sg_table,
|
|
.gem_prime_mmap = drm_gem_prime_mmap,
|
|
};
|
|
|
|
static int panfrost_probe(struct platform_device *pdev)
|
|
{
|
|
struct panfrost_device *pfdev;
|
|
struct drm_device *ddev;
|
|
int err;
|
|
|
|
pfdev = devm_kzalloc(&pdev->dev, sizeof(*pfdev), GFP_KERNEL);
|
|
if (!pfdev)
|
|
return -ENOMEM;
|
|
|
|
pfdev->pdev = pdev;
|
|
pfdev->dev = &pdev->dev;
|
|
|
|
platform_set_drvdata(pdev, pfdev);
|
|
|
|
/* Allocate and initialze the DRM device. */
|
|
ddev = drm_dev_alloc(&panfrost_drm_driver, &pdev->dev);
|
|
if (IS_ERR(ddev))
|
|
return PTR_ERR(ddev);
|
|
|
|
ddev->dev_private = pfdev;
|
|
pfdev->ddev = ddev;
|
|
|
|
mutex_init(&pfdev->shrinker_lock);
|
|
INIT_LIST_HEAD(&pfdev->shrinker_list);
|
|
|
|
err = panfrost_device_init(pfdev);
|
|
if (err) {
|
|
if (err != -EPROBE_DEFER)
|
|
dev_err(&pdev->dev, "Fatal error during GPU init\n");
|
|
goto err_out0;
|
|
}
|
|
|
|
err = panfrost_devfreq_init(pfdev);
|
|
if (err) {
|
|
if (err != -EPROBE_DEFER)
|
|
dev_err(&pdev->dev, "Fatal error during devfreq init\n");
|
|
goto err_out1;
|
|
}
|
|
|
|
pm_runtime_set_active(pfdev->dev);
|
|
pm_runtime_mark_last_busy(pfdev->dev);
|
|
pm_runtime_enable(pfdev->dev);
|
|
pm_runtime_set_autosuspend_delay(pfdev->dev, 50); /* ~3 frames */
|
|
pm_runtime_use_autosuspend(pfdev->dev);
|
|
|
|
/*
|
|
* Register the DRM device with the core and the connectors with
|
|
* sysfs
|
|
*/
|
|
err = drm_dev_register(ddev, 0);
|
|
if (err < 0)
|
|
goto err_out2;
|
|
|
|
panfrost_gem_shrinker_init(ddev);
|
|
|
|
return 0;
|
|
|
|
err_out2:
|
|
panfrost_devfreq_fini(pfdev);
|
|
err_out1:
|
|
panfrost_device_fini(pfdev);
|
|
err_out0:
|
|
pm_runtime_disable(pfdev->dev);
|
|
drm_dev_put(ddev);
|
|
return err;
|
|
}
|
|
|
|
static int panfrost_remove(struct platform_device *pdev)
|
|
{
|
|
struct panfrost_device *pfdev = platform_get_drvdata(pdev);
|
|
struct drm_device *ddev = pfdev->ddev;
|
|
|
|
drm_dev_unregister(ddev);
|
|
panfrost_gem_shrinker_cleanup(ddev);
|
|
|
|
pm_runtime_get_sync(pfdev->dev);
|
|
panfrost_devfreq_fini(pfdev);
|
|
panfrost_device_fini(pfdev);
|
|
pm_runtime_put_sync_suspend(pfdev->dev);
|
|
pm_runtime_disable(pfdev->dev);
|
|
|
|
drm_dev_put(ddev);
|
|
return 0;
|
|
}
|
|
|
|
static const struct of_device_id dt_match[] = {
|
|
{ .compatible = "arm,mali-t604" },
|
|
{ .compatible = "arm,mali-t624" },
|
|
{ .compatible = "arm,mali-t628" },
|
|
{ .compatible = "arm,mali-t720" },
|
|
{ .compatible = "arm,mali-t760" },
|
|
{ .compatible = "arm,mali-t820" },
|
|
{ .compatible = "arm,mali-t830" },
|
|
{ .compatible = "arm,mali-t860" },
|
|
{ .compatible = "arm,mali-t880" },
|
|
{}
|
|
};
|
|
MODULE_DEVICE_TABLE(of, dt_match);
|
|
|
|
static const struct dev_pm_ops panfrost_pm_ops = {
|
|
SET_SYSTEM_SLEEP_PM_OPS(pm_runtime_force_suspend, pm_runtime_force_resume)
|
|
SET_RUNTIME_PM_OPS(panfrost_device_suspend, panfrost_device_resume, NULL)
|
|
};
|
|
|
|
static struct platform_driver panfrost_driver = {
|
|
.probe = panfrost_probe,
|
|
.remove = panfrost_remove,
|
|
.driver = {
|
|
.name = "panfrost",
|
|
.pm = &panfrost_pm_ops,
|
|
.of_match_table = dt_match,
|
|
},
|
|
};
|
|
module_platform_driver(panfrost_driver);
|
|
|
|
MODULE_AUTHOR("Panfrost Project Developers");
|
|
MODULE_DESCRIPTION("Panfrost DRM Driver");
|
|
MODULE_LICENSE("GPL v2");
|