mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 08:49:38 +07:00
088880ddc0
With softpin we allow the userspace to take control over the GPU virtual address space. The new capability is relected by a bump of the minor DRM version. There are a few restrictions for userspace to take into account: 1. The kernel reserves a bit of the address space to implement zero page faulting and mapping of the kernel internal ring buffer. Userspace can query the kernel for the first usable GPU VM address via ETNAVIV_PARAM_SOFTPIN_START_ADDR. 2. We only allow softpin on GPUs, which implement proper process separation via PPAS. If softpin is not available the softpin start address will be set to ~0. 3. Softpin is all or nothing. A submit using softpin must not use any address fixups via relocs. Signed-off-by: Lucas Stach <l.stach@pengutronix.de> Reviewed-by: Philipp Zabel <p.zabel@pengutronix.de> Reviewed-by: Guido Günther <agx@sigxcpu.org>
128 lines
4.2 KiB
C
128 lines
4.2 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Copyright (C) 2015-2018 Etnaviv Project
|
|
*/
|
|
|
|
#ifndef __ETNAVIV_DRV_H__
|
|
#define __ETNAVIV_DRV_H__
|
|
|
|
#include <linux/list.h>
|
|
#include <linux/mm_types.h>
|
|
#include <linux/sizes.h>
|
|
#include <linux/time64.h>
|
|
#include <linux/types.h>
|
|
|
|
#include <drm/drm_fb_helper.h>
|
|
#include <drm/drm_gem.h>
|
|
#include <drm/etnaviv_drm.h>
|
|
#include <drm/gpu_scheduler.h>
|
|
|
|
struct etnaviv_cmdbuf;
|
|
struct etnaviv_gpu;
|
|
struct etnaviv_mmu;
|
|
struct etnaviv_gem_object;
|
|
struct etnaviv_gem_submit;
|
|
struct etnaviv_iommu_global;
|
|
|
|
#define ETNAVIV_SOFTPIN_START_ADDRESS SZ_4M /* must be >= SUBALLOC_SIZE */
|
|
|
|
struct etnaviv_file_private {
|
|
struct etnaviv_iommu_context *mmu;
|
|
struct drm_sched_entity sched_entity[ETNA_MAX_PIPES];
|
|
};
|
|
|
|
struct etnaviv_drm_private {
|
|
int num_gpus;
|
|
struct device_dma_parameters dma_parms;
|
|
struct etnaviv_gpu *gpu[ETNA_MAX_PIPES];
|
|
|
|
struct etnaviv_cmdbuf_suballoc *cmdbuf_suballoc;
|
|
struct etnaviv_iommu_global *mmu_global;
|
|
|
|
/* list of GEM objects: */
|
|
struct mutex gem_lock;
|
|
struct list_head gem_list;
|
|
};
|
|
|
|
int etnaviv_ioctl_gem_submit(struct drm_device *dev, void *data,
|
|
struct drm_file *file);
|
|
|
|
int etnaviv_gem_mmap(struct file *filp, struct vm_area_struct *vma);
|
|
vm_fault_t etnaviv_gem_fault(struct vm_fault *vmf);
|
|
int etnaviv_gem_mmap_offset(struct drm_gem_object *obj, u64 *offset);
|
|
struct sg_table *etnaviv_gem_prime_get_sg_table(struct drm_gem_object *obj);
|
|
void *etnaviv_gem_prime_vmap(struct drm_gem_object *obj);
|
|
void etnaviv_gem_prime_vunmap(struct drm_gem_object *obj, void *vaddr);
|
|
int etnaviv_gem_prime_mmap(struct drm_gem_object *obj,
|
|
struct vm_area_struct *vma);
|
|
struct drm_gem_object *etnaviv_gem_prime_import_sg_table(struct drm_device *dev,
|
|
struct dma_buf_attachment *attach, struct sg_table *sg);
|
|
int etnaviv_gem_prime_pin(struct drm_gem_object *obj);
|
|
void etnaviv_gem_prime_unpin(struct drm_gem_object *obj);
|
|
void *etnaviv_gem_vmap(struct drm_gem_object *obj);
|
|
int etnaviv_gem_cpu_prep(struct drm_gem_object *obj, u32 op,
|
|
struct timespec *timeout);
|
|
int etnaviv_gem_cpu_fini(struct drm_gem_object *obj);
|
|
void etnaviv_gem_free_object(struct drm_gem_object *obj);
|
|
int etnaviv_gem_new_handle(struct drm_device *dev, struct drm_file *file,
|
|
u32 size, u32 flags, u32 *handle);
|
|
int etnaviv_gem_new_userptr(struct drm_device *dev, struct drm_file *file,
|
|
uintptr_t ptr, u32 size, u32 flags, u32 *handle);
|
|
u16 etnaviv_buffer_init(struct etnaviv_gpu *gpu);
|
|
u16 etnaviv_buffer_config_mmuv2(struct etnaviv_gpu *gpu, u32 mtlb_addr, u32 safe_addr);
|
|
u16 etnaviv_buffer_config_pta(struct etnaviv_gpu *gpu, unsigned short id);
|
|
void etnaviv_buffer_end(struct etnaviv_gpu *gpu);
|
|
void etnaviv_sync_point_queue(struct etnaviv_gpu *gpu, unsigned int event);
|
|
void etnaviv_buffer_queue(struct etnaviv_gpu *gpu, u32 exec_state,
|
|
struct etnaviv_iommu_context *mmu,
|
|
unsigned int event, struct etnaviv_cmdbuf *cmdbuf);
|
|
void etnaviv_validate_init(void);
|
|
bool etnaviv_cmd_validate_one(struct etnaviv_gpu *gpu,
|
|
u32 *stream, unsigned int size,
|
|
struct drm_etnaviv_gem_submit_reloc *relocs, unsigned int reloc_size);
|
|
|
|
#ifdef CONFIG_DEBUG_FS
|
|
void etnaviv_gem_describe_objects(struct etnaviv_drm_private *priv,
|
|
struct seq_file *m);
|
|
#endif
|
|
|
|
#define DBG(fmt, ...) DRM_DEBUG(fmt"\n", ##__VA_ARGS__)
|
|
#define VERB(fmt, ...) if (0) DRM_DEBUG(fmt"\n", ##__VA_ARGS__)
|
|
|
|
/*
|
|
* Return the storage size of a structure with a variable length array.
|
|
* The array is nelem elements of elem_size, where the base structure
|
|
* is defined by base. If the size overflows size_t, return zero.
|
|
*/
|
|
static inline size_t size_vstruct(size_t nelem, size_t elem_size, size_t base)
|
|
{
|
|
if (elem_size && nelem > (SIZE_MAX - base) / elem_size)
|
|
return 0;
|
|
return base + nelem * elem_size;
|
|
}
|
|
|
|
/*
|
|
* Etnaviv timeouts are specified wrt CLOCK_MONOTONIC, not jiffies.
|
|
* We need to calculate the timeout in terms of number of jiffies
|
|
* between the specified timeout and the current CLOCK_MONOTONIC time.
|
|
*/
|
|
static inline unsigned long etnaviv_timeout_to_jiffies(
|
|
const struct timespec *timeout)
|
|
{
|
|
struct timespec64 ts, to;
|
|
|
|
to = timespec_to_timespec64(*timeout);
|
|
|
|
ktime_get_ts64(&ts);
|
|
|
|
/* timeouts before "now" have already expired */
|
|
if (timespec64_compare(&to, &ts) <= 0)
|
|
return 0;
|
|
|
|
ts = timespec64_sub(to, ts);
|
|
|
|
return timespec64_to_jiffies(&ts);
|
|
}
|
|
|
|
#endif /* __ETNAVIV_DRV_H__ */
|