mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-27 18:45:02 +07:00
41db12df64
Improve the readibility of the function which inserts command buffers and other maintanence commands into the GPUs ring buffer. We do this by splitting the ring buffer reservation in two: one chunk for any commands that need to be issued prior to the command buffer, and a separate chunk for commands issued after the buffer. The result is a much more obvious code flow in this function, and localisation of the conditional maintanence commands prior to the command buffer. Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk> Signed-off-by: Lucas Stach <l.stach@pengutronix.de>
328 lines
9.3 KiB
C
328 lines
9.3 KiB
C
/*
|
|
* Copyright (C) 2014 Etnaviv Project
|
|
* Author: Christian Gmeiner <christian.gmeiner@gmail.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 as published by
|
|
* the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along with
|
|
* this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include "etnaviv_gpu.h"
|
|
#include "etnaviv_gem.h"
|
|
#include "etnaviv_mmu.h"
|
|
|
|
#include "common.xml.h"
|
|
#include "state.xml.h"
|
|
#include "state_3d.xml.h"
|
|
#include "cmdstream.xml.h"
|
|
|
|
/*
|
|
* Command Buffer helper:
|
|
*/
|
|
|
|
|
|
static inline void OUT(struct etnaviv_cmdbuf *buffer, u32 data)
|
|
{
|
|
u32 *vaddr = (u32 *)buffer->vaddr;
|
|
|
|
BUG_ON(buffer->user_size >= buffer->size);
|
|
|
|
vaddr[buffer->user_size / 4] = data;
|
|
buffer->user_size += 4;
|
|
}
|
|
|
|
static inline void CMD_LOAD_STATE(struct etnaviv_cmdbuf *buffer,
|
|
u32 reg, u32 value)
|
|
{
|
|
u32 index = reg >> VIV_FE_LOAD_STATE_HEADER_OFFSET__SHR;
|
|
|
|
buffer->user_size = ALIGN(buffer->user_size, 8);
|
|
|
|
/* write a register via cmd stream */
|
|
OUT(buffer, VIV_FE_LOAD_STATE_HEADER_OP_LOAD_STATE |
|
|
VIV_FE_LOAD_STATE_HEADER_COUNT(1) |
|
|
VIV_FE_LOAD_STATE_HEADER_OFFSET(index));
|
|
OUT(buffer, value);
|
|
}
|
|
|
|
static inline void CMD_END(struct etnaviv_cmdbuf *buffer)
|
|
{
|
|
buffer->user_size = ALIGN(buffer->user_size, 8);
|
|
|
|
OUT(buffer, VIV_FE_END_HEADER_OP_END);
|
|
}
|
|
|
|
static inline void CMD_WAIT(struct etnaviv_cmdbuf *buffer)
|
|
{
|
|
buffer->user_size = ALIGN(buffer->user_size, 8);
|
|
|
|
OUT(buffer, VIV_FE_WAIT_HEADER_OP_WAIT | 200);
|
|
}
|
|
|
|
static inline void CMD_LINK(struct etnaviv_cmdbuf *buffer,
|
|
u16 prefetch, u32 address)
|
|
{
|
|
buffer->user_size = ALIGN(buffer->user_size, 8);
|
|
|
|
OUT(buffer, VIV_FE_LINK_HEADER_OP_LINK |
|
|
VIV_FE_LINK_HEADER_PREFETCH(prefetch));
|
|
OUT(buffer, address);
|
|
}
|
|
|
|
static inline void CMD_STALL(struct etnaviv_cmdbuf *buffer,
|
|
u32 from, u32 to)
|
|
{
|
|
buffer->user_size = ALIGN(buffer->user_size, 8);
|
|
|
|
OUT(buffer, VIV_FE_STALL_HEADER_OP_STALL);
|
|
OUT(buffer, VIV_FE_STALL_TOKEN_FROM(from) | VIV_FE_STALL_TOKEN_TO(to));
|
|
}
|
|
|
|
static inline void CMD_SEM(struct etnaviv_cmdbuf *buffer, u32 from, u32 to)
|
|
{
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_SEMAPHORE_TOKEN,
|
|
VIVS_GL_SEMAPHORE_TOKEN_FROM(from) |
|
|
VIVS_GL_SEMAPHORE_TOKEN_TO(to));
|
|
}
|
|
|
|
static void etnaviv_cmd_select_pipe(struct etnaviv_gpu *gpu,
|
|
struct etnaviv_cmdbuf *buffer, u8 pipe)
|
|
{
|
|
u32 flush = 0;
|
|
|
|
/*
|
|
* This assumes that if we're switching to 2D, we're switching
|
|
* away from 3D, and vice versa. Hence, if we're switching to
|
|
* the 2D core, we need to flush the 3D depth and color caches,
|
|
* otherwise we need to flush the 2D pixel engine cache.
|
|
*/
|
|
if (gpu->exec_state == ETNA_PIPE_2D)
|
|
flush = VIVS_GL_FLUSH_CACHE_PE2D;
|
|
else if (gpu->exec_state == ETNA_PIPE_3D)
|
|
flush = VIVS_GL_FLUSH_CACHE_DEPTH | VIVS_GL_FLUSH_CACHE_COLOR;
|
|
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush);
|
|
CMD_SEM(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
CMD_STALL(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_PIPE_SELECT,
|
|
VIVS_GL_PIPE_SELECT_PIPE(pipe));
|
|
}
|
|
|
|
static u32 gpu_va(struct etnaviv_gpu *gpu, struct etnaviv_cmdbuf *buf)
|
|
{
|
|
return buf->paddr - gpu->memory_base;
|
|
}
|
|
|
|
static void etnaviv_buffer_dump(struct etnaviv_gpu *gpu,
|
|
struct etnaviv_cmdbuf *buf, u32 off, u32 len)
|
|
{
|
|
u32 size = buf->size;
|
|
u32 *ptr = buf->vaddr + off;
|
|
|
|
dev_info(gpu->dev, "virt %p phys 0x%08x free 0x%08x\n",
|
|
ptr, gpu_va(gpu, buf) + off, size - len * 4 - off);
|
|
|
|
print_hex_dump(KERN_INFO, "cmd ", DUMP_PREFIX_OFFSET, 16, 4,
|
|
ptr, len * 4, 0);
|
|
}
|
|
|
|
/*
|
|
* Safely replace the WAIT of a waitlink with a new command and argument.
|
|
* The GPU may be executing this WAIT while we're modifying it, so we have
|
|
* to write it in a specific order to avoid the GPU branching to somewhere
|
|
* else. 'wl_offset' is the offset to the first byte of the WAIT command.
|
|
*/
|
|
static void etnaviv_buffer_replace_wait(struct etnaviv_cmdbuf *buffer,
|
|
unsigned int wl_offset, u32 cmd, u32 arg)
|
|
{
|
|
u32 *lw = buffer->vaddr + wl_offset;
|
|
|
|
lw[1] = arg;
|
|
mb();
|
|
lw[0] = cmd;
|
|
mb();
|
|
}
|
|
|
|
/*
|
|
* Ensure that there is space in the command buffer to contiguously write
|
|
* 'cmd_dwords' 64-bit words into the buffer, wrapping if necessary.
|
|
*/
|
|
static u32 etnaviv_buffer_reserve(struct etnaviv_gpu *gpu,
|
|
struct etnaviv_cmdbuf *buffer, unsigned int cmd_dwords)
|
|
{
|
|
if (buffer->user_size + cmd_dwords * sizeof(u64) > buffer->size)
|
|
buffer->user_size = 0;
|
|
|
|
return gpu_va(gpu, buffer) + buffer->user_size;
|
|
}
|
|
|
|
u16 etnaviv_buffer_init(struct etnaviv_gpu *gpu)
|
|
{
|
|
struct etnaviv_cmdbuf *buffer = gpu->buffer;
|
|
|
|
/* initialize buffer */
|
|
buffer->user_size = 0;
|
|
|
|
CMD_WAIT(buffer);
|
|
CMD_LINK(buffer, 2, gpu_va(gpu, buffer) + buffer->user_size - 4);
|
|
|
|
return buffer->user_size / 8;
|
|
}
|
|
|
|
void etnaviv_buffer_end(struct etnaviv_gpu *gpu)
|
|
{
|
|
struct etnaviv_cmdbuf *buffer = gpu->buffer;
|
|
unsigned int waitlink_offset = buffer->user_size - 16;
|
|
u32 link_target, flush = 0;
|
|
|
|
if (gpu->exec_state == ETNA_PIPE_2D)
|
|
flush = VIVS_GL_FLUSH_CACHE_PE2D;
|
|
else if (gpu->exec_state == ETNA_PIPE_3D)
|
|
flush = VIVS_GL_FLUSH_CACHE_DEPTH |
|
|
VIVS_GL_FLUSH_CACHE_COLOR |
|
|
VIVS_GL_FLUSH_CACHE_TEXTURE |
|
|
VIVS_GL_FLUSH_CACHE_TEXTUREVS |
|
|
VIVS_GL_FLUSH_CACHE_SHADER_L2;
|
|
|
|
if (flush) {
|
|
unsigned int dwords = 7;
|
|
|
|
link_target = etnaviv_buffer_reserve(gpu, buffer, dwords);
|
|
|
|
CMD_SEM(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
CMD_STALL(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush);
|
|
if (gpu->exec_state == ETNA_PIPE_3D)
|
|
CMD_LOAD_STATE(buffer, VIVS_TS_FLUSH_CACHE,
|
|
VIVS_TS_FLUSH_CACHE_FLUSH);
|
|
CMD_SEM(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
CMD_STALL(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
CMD_END(buffer);
|
|
|
|
etnaviv_buffer_replace_wait(buffer, waitlink_offset,
|
|
VIV_FE_LINK_HEADER_OP_LINK |
|
|
VIV_FE_LINK_HEADER_PREFETCH(dwords),
|
|
link_target);
|
|
} else {
|
|
/* Replace the last link-wait with an "END" command */
|
|
etnaviv_buffer_replace_wait(buffer, waitlink_offset,
|
|
VIV_FE_END_HEADER_OP_END, 0);
|
|
}
|
|
}
|
|
|
|
/* Append a command buffer to the ring buffer. */
|
|
void etnaviv_buffer_queue(struct etnaviv_gpu *gpu, unsigned int event,
|
|
struct etnaviv_cmdbuf *cmdbuf)
|
|
{
|
|
struct etnaviv_cmdbuf *buffer = gpu->buffer;
|
|
unsigned int waitlink_offset = buffer->user_size - 16;
|
|
u32 return_target, return_dwords;
|
|
u32 link_target, link_dwords;
|
|
|
|
if (drm_debug & DRM_UT_DRIVER)
|
|
etnaviv_buffer_dump(gpu, buffer, 0, 0x50);
|
|
|
|
link_target = gpu_va(gpu, cmdbuf);
|
|
link_dwords = cmdbuf->size / 8;
|
|
|
|
/*
|
|
* If we need maintanence prior to submitting this buffer, we will
|
|
* need to append a mmu flush load state, followed by a new
|
|
* link to this buffer - a total of four additional words.
|
|
*/
|
|
if (gpu->mmu->need_flush || gpu->switch_context) {
|
|
u32 target, extra_dwords;
|
|
|
|
/* link command */
|
|
extra_dwords = 1;
|
|
|
|
/* flush command */
|
|
if (gpu->mmu->need_flush)
|
|
extra_dwords += 1;
|
|
|
|
/* pipe switch commands */
|
|
if (gpu->switch_context)
|
|
extra_dwords += 4;
|
|
|
|
target = etnaviv_buffer_reserve(gpu, buffer, extra_dwords);
|
|
|
|
if (gpu->mmu->need_flush) {
|
|
/* Add the MMU flush */
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_MMU,
|
|
VIVS_GL_FLUSH_MMU_FLUSH_FEMMU |
|
|
VIVS_GL_FLUSH_MMU_FLUSH_UNK1 |
|
|
VIVS_GL_FLUSH_MMU_FLUSH_UNK2 |
|
|
VIVS_GL_FLUSH_MMU_FLUSH_PEMMU |
|
|
VIVS_GL_FLUSH_MMU_FLUSH_UNK4);
|
|
|
|
gpu->mmu->need_flush = false;
|
|
}
|
|
|
|
if (gpu->switch_context) {
|
|
etnaviv_cmd_select_pipe(gpu, buffer, cmdbuf->exec_state);
|
|
gpu->exec_state = cmdbuf->exec_state;
|
|
gpu->switch_context = false;
|
|
}
|
|
|
|
/* And the link to the submitted buffer */
|
|
CMD_LINK(buffer, link_dwords, link_target);
|
|
|
|
/* Update the link target to point to above instructions */
|
|
link_target = target;
|
|
link_dwords = extra_dwords;
|
|
}
|
|
|
|
/*
|
|
* Append a LINK to the submitted command buffer to return to
|
|
* the ring buffer. return_target is the ring target address.
|
|
* We need three dwords: event, wait, link.
|
|
*/
|
|
return_dwords = 3;
|
|
return_target = etnaviv_buffer_reserve(gpu, buffer, return_dwords);
|
|
CMD_LINK(cmdbuf, return_dwords, return_target);
|
|
|
|
/*
|
|
* Append event, wait and link pointing back to the wait
|
|
* command to the ring buffer.
|
|
*/
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_EVENT, VIVS_GL_EVENT_EVENT_ID(event) |
|
|
VIVS_GL_EVENT_FROM_PE);
|
|
CMD_WAIT(buffer);
|
|
CMD_LINK(buffer, 2, return_target + 8);
|
|
|
|
if (drm_debug & DRM_UT_DRIVER)
|
|
pr_info("stream link to 0x%08x @ 0x%08x %p\n",
|
|
return_target, gpu_va(gpu, cmdbuf), cmdbuf->vaddr);
|
|
|
|
if (drm_debug & DRM_UT_DRIVER) {
|
|
print_hex_dump(KERN_INFO, "cmd ", DUMP_PREFIX_OFFSET, 16, 4,
|
|
cmdbuf->vaddr, cmdbuf->size, 0);
|
|
|
|
pr_info("link op: %p\n", buffer->vaddr + waitlink_offset);
|
|
pr_info("addr: 0x%08x\n", link_target);
|
|
pr_info("back: 0x%08x\n", return_target);
|
|
pr_info("event: %d\n", event);
|
|
}
|
|
|
|
/*
|
|
* Kick off the submitted command by replacing the previous
|
|
* WAIT with a link to the address in the ring buffer.
|
|
*/
|
|
etnaviv_buffer_replace_wait(buffer, waitlink_offset,
|
|
VIV_FE_LINK_HEADER_OP_LINK |
|
|
VIV_FE_LINK_HEADER_PREFETCH(link_dwords),
|
|
link_target);
|
|
|
|
if (drm_debug & DRM_UT_DRIVER)
|
|
etnaviv_buffer_dump(gpu, buffer, 0, 0x50);
|
|
}
|