mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2025-01-10 07:44:23 +08:00
90747b9511
Use the previous GPU pipe state when deciding which GPU caches should be flushed prior to switching the current pipe. This avoids infering what the previously selected pipe was, and potentially flushing the wrong caches. Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk> Signed-off-by: Lucas Stach <l.stach@pengutronix.de>
324 lines
9.2 KiB
C
324 lines
9.2 KiB
C
/*
|
|
* Copyright (C) 2014 Etnaviv Project
|
|
* Author: Christian Gmeiner <christian.gmeiner@gmail.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 as published by
|
|
* the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along with
|
|
* this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include "etnaviv_gpu.h"
|
|
#include "etnaviv_gem.h"
|
|
#include "etnaviv_mmu.h"
|
|
|
|
#include "common.xml.h"
|
|
#include "state.xml.h"
|
|
#include "state_3d.xml.h"
|
|
#include "cmdstream.xml.h"
|
|
|
|
/*
|
|
* Command Buffer helper:
|
|
*/
|
|
|
|
|
|
static inline void OUT(struct etnaviv_cmdbuf *buffer, u32 data)
|
|
{
|
|
u32 *vaddr = (u32 *)buffer->vaddr;
|
|
|
|
BUG_ON(buffer->user_size >= buffer->size);
|
|
|
|
vaddr[buffer->user_size / 4] = data;
|
|
buffer->user_size += 4;
|
|
}
|
|
|
|
static inline void CMD_LOAD_STATE(struct etnaviv_cmdbuf *buffer,
|
|
u32 reg, u32 value)
|
|
{
|
|
u32 index = reg >> VIV_FE_LOAD_STATE_HEADER_OFFSET__SHR;
|
|
|
|
buffer->user_size = ALIGN(buffer->user_size, 8);
|
|
|
|
/* write a register via cmd stream */
|
|
OUT(buffer, VIV_FE_LOAD_STATE_HEADER_OP_LOAD_STATE |
|
|
VIV_FE_LOAD_STATE_HEADER_COUNT(1) |
|
|
VIV_FE_LOAD_STATE_HEADER_OFFSET(index));
|
|
OUT(buffer, value);
|
|
}
|
|
|
|
static inline void CMD_END(struct etnaviv_cmdbuf *buffer)
|
|
{
|
|
buffer->user_size = ALIGN(buffer->user_size, 8);
|
|
|
|
OUT(buffer, VIV_FE_END_HEADER_OP_END);
|
|
}
|
|
|
|
static inline void CMD_WAIT(struct etnaviv_cmdbuf *buffer)
|
|
{
|
|
buffer->user_size = ALIGN(buffer->user_size, 8);
|
|
|
|
OUT(buffer, VIV_FE_WAIT_HEADER_OP_WAIT | 200);
|
|
}
|
|
|
|
static inline void CMD_LINK(struct etnaviv_cmdbuf *buffer,
|
|
u16 prefetch, u32 address)
|
|
{
|
|
buffer->user_size = ALIGN(buffer->user_size, 8);
|
|
|
|
OUT(buffer, VIV_FE_LINK_HEADER_OP_LINK |
|
|
VIV_FE_LINK_HEADER_PREFETCH(prefetch));
|
|
OUT(buffer, address);
|
|
}
|
|
|
|
static inline void CMD_STALL(struct etnaviv_cmdbuf *buffer,
|
|
u32 from, u32 to)
|
|
{
|
|
buffer->user_size = ALIGN(buffer->user_size, 8);
|
|
|
|
OUT(buffer, VIV_FE_STALL_HEADER_OP_STALL);
|
|
OUT(buffer, VIV_FE_STALL_TOKEN_FROM(from) | VIV_FE_STALL_TOKEN_TO(to));
|
|
}
|
|
|
|
static inline void CMD_SEM(struct etnaviv_cmdbuf *buffer, u32 from, u32 to)
|
|
{
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_SEMAPHORE_TOKEN,
|
|
VIVS_GL_SEMAPHORE_TOKEN_FROM(from) |
|
|
VIVS_GL_SEMAPHORE_TOKEN_TO(to));
|
|
}
|
|
|
|
static void etnaviv_cmd_select_pipe(struct etnaviv_gpu *gpu,
|
|
struct etnaviv_cmdbuf *buffer, u8 pipe)
|
|
{
|
|
u32 flush = 0;
|
|
|
|
/*
|
|
* This assumes that if we're switching to 2D, we're switching
|
|
* away from 3D, and vice versa. Hence, if we're switching to
|
|
* the 2D core, we need to flush the 3D depth and color caches,
|
|
* otherwise we need to flush the 2D pixel engine cache.
|
|
*/
|
|
if (gpu->exec_state == ETNA_PIPE_2D)
|
|
flush = VIVS_GL_FLUSH_CACHE_PE2D;
|
|
else if (gpu->exec_state == ETNA_PIPE_3D)
|
|
flush = VIVS_GL_FLUSH_CACHE_DEPTH | VIVS_GL_FLUSH_CACHE_COLOR;
|
|
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush);
|
|
CMD_SEM(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
CMD_STALL(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_PIPE_SELECT,
|
|
VIVS_GL_PIPE_SELECT_PIPE(pipe));
|
|
}
|
|
|
|
static u32 gpu_va(struct etnaviv_gpu *gpu, struct etnaviv_cmdbuf *buf)
|
|
{
|
|
return buf->paddr - gpu->memory_base;
|
|
}
|
|
|
|
static void etnaviv_buffer_dump(struct etnaviv_gpu *gpu,
|
|
struct etnaviv_cmdbuf *buf, u32 off, u32 len)
|
|
{
|
|
u32 size = buf->size;
|
|
u32 *ptr = buf->vaddr + off;
|
|
|
|
dev_info(gpu->dev, "virt %p phys 0x%08x free 0x%08x\n",
|
|
ptr, gpu_va(gpu, buf) + off, size - len * 4 - off);
|
|
|
|
print_hex_dump(KERN_INFO, "cmd ", DUMP_PREFIX_OFFSET, 16, 4,
|
|
ptr, len * 4, 0);
|
|
}
|
|
|
|
/*
|
|
* Safely replace the WAIT of a waitlink with a new command and argument.
|
|
* The GPU may be executing this WAIT while we're modifying it, so we have
|
|
* to write it in a specific order to avoid the GPU branching to somewhere
|
|
* else. 'wl_offset' is the offset to the first byte of the WAIT command.
|
|
*/
|
|
static void etnaviv_buffer_replace_wait(struct etnaviv_cmdbuf *buffer,
|
|
unsigned int wl_offset, u32 cmd, u32 arg)
|
|
{
|
|
u32 *lw = buffer->vaddr + wl_offset;
|
|
|
|
lw[1] = arg;
|
|
mb();
|
|
lw[0] = cmd;
|
|
mb();
|
|
}
|
|
|
|
/*
|
|
* Ensure that there is space in the command buffer to contiguously write
|
|
* 'cmd_dwords' 64-bit words into the buffer, wrapping if necessary.
|
|
*/
|
|
static u32 etnaviv_buffer_reserve(struct etnaviv_gpu *gpu,
|
|
struct etnaviv_cmdbuf *buffer, unsigned int cmd_dwords)
|
|
{
|
|
if (buffer->user_size + cmd_dwords * sizeof(u64) > buffer->size)
|
|
buffer->user_size = 0;
|
|
|
|
return gpu_va(gpu, buffer) + buffer->user_size;
|
|
}
|
|
|
|
u16 etnaviv_buffer_init(struct etnaviv_gpu *gpu)
|
|
{
|
|
struct etnaviv_cmdbuf *buffer = gpu->buffer;
|
|
|
|
/* initialize buffer */
|
|
buffer->user_size = 0;
|
|
|
|
CMD_WAIT(buffer);
|
|
CMD_LINK(buffer, 2, gpu_va(gpu, buffer) + buffer->user_size - 4);
|
|
|
|
return buffer->user_size / 8;
|
|
}
|
|
|
|
void etnaviv_buffer_end(struct etnaviv_gpu *gpu)
|
|
{
|
|
struct etnaviv_cmdbuf *buffer = gpu->buffer;
|
|
unsigned int waitlink_offset = buffer->user_size - 16;
|
|
u32 link_target, flush = 0;
|
|
|
|
if (gpu->exec_state == ETNA_PIPE_2D)
|
|
flush = VIVS_GL_FLUSH_CACHE_PE2D;
|
|
else if (gpu->exec_state == ETNA_PIPE_3D)
|
|
flush = VIVS_GL_FLUSH_CACHE_DEPTH |
|
|
VIVS_GL_FLUSH_CACHE_COLOR |
|
|
VIVS_GL_FLUSH_CACHE_TEXTURE |
|
|
VIVS_GL_FLUSH_CACHE_TEXTUREVS |
|
|
VIVS_GL_FLUSH_CACHE_SHADER_L2;
|
|
|
|
if (flush) {
|
|
unsigned int dwords = 7;
|
|
|
|
link_target = etnaviv_buffer_reserve(gpu, buffer, dwords);
|
|
|
|
CMD_SEM(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
CMD_STALL(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush);
|
|
if (gpu->exec_state == ETNA_PIPE_3D)
|
|
CMD_LOAD_STATE(buffer, VIVS_TS_FLUSH_CACHE,
|
|
VIVS_TS_FLUSH_CACHE_FLUSH);
|
|
CMD_SEM(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
CMD_STALL(buffer, SYNC_RECIPIENT_FE, SYNC_RECIPIENT_PE);
|
|
CMD_END(buffer);
|
|
|
|
etnaviv_buffer_replace_wait(buffer, waitlink_offset,
|
|
VIV_FE_LINK_HEADER_OP_LINK |
|
|
VIV_FE_LINK_HEADER_PREFETCH(dwords),
|
|
link_target);
|
|
} else {
|
|
/* Replace the last link-wait with an "END" command */
|
|
etnaviv_buffer_replace_wait(buffer, waitlink_offset,
|
|
VIV_FE_END_HEADER_OP_END, 0);
|
|
}
|
|
}
|
|
|
|
void etnaviv_buffer_queue(struct etnaviv_gpu *gpu, unsigned int event,
|
|
struct etnaviv_cmdbuf *cmdbuf)
|
|
{
|
|
struct etnaviv_cmdbuf *buffer = gpu->buffer;
|
|
unsigned int waitlink_offset = buffer->user_size - 16;
|
|
u32 back, link_target, link_size, reserve_size, extra_size = 0;
|
|
|
|
if (drm_debug & DRM_UT_DRIVER)
|
|
etnaviv_buffer_dump(gpu, buffer, 0, 0x50);
|
|
|
|
/*
|
|
* If we need to flush the MMU prior to submitting this buffer, we
|
|
* will need to append a mmu flush load state, followed by a new
|
|
* link to this buffer - a total of four additional words.
|
|
*/
|
|
if (gpu->mmu->need_flush || gpu->switch_context) {
|
|
/* link command */
|
|
extra_size += 2;
|
|
/* flush command */
|
|
if (gpu->mmu->need_flush)
|
|
extra_size += 2;
|
|
/* pipe switch commands */
|
|
if (gpu->switch_context)
|
|
extra_size += 8;
|
|
}
|
|
|
|
reserve_size = (6 + extra_size) * 4;
|
|
|
|
link_target = etnaviv_buffer_reserve(gpu, buffer, reserve_size / 8);
|
|
|
|
/* save offset back into main buffer */
|
|
back = buffer->user_size + reserve_size - 6 * 4;
|
|
link_size = 6;
|
|
|
|
/* Skip over any extra instructions */
|
|
link_target += extra_size * sizeof(u32);
|
|
|
|
if (drm_debug & DRM_UT_DRIVER)
|
|
pr_info("stream link to 0x%08x @ 0x%08x %p\n",
|
|
link_target, gpu_va(gpu, cmdbuf), cmdbuf->vaddr);
|
|
|
|
/* jump back from cmd to main buffer */
|
|
CMD_LINK(cmdbuf, link_size, link_target);
|
|
|
|
link_target = gpu_va(gpu, cmdbuf);
|
|
link_size = cmdbuf->size / 8;
|
|
|
|
if (drm_debug & DRM_UT_DRIVER) {
|
|
print_hex_dump(KERN_INFO, "cmd ", DUMP_PREFIX_OFFSET, 16, 4,
|
|
cmdbuf->vaddr, cmdbuf->size, 0);
|
|
|
|
pr_info("link op: %p\n", buffer->vaddr + waitlink_offset);
|
|
pr_info("addr: 0x%08x\n", link_target);
|
|
pr_info("back: 0x%08x\n", gpu_va(gpu, buffer) + back);
|
|
pr_info("event: %d\n", event);
|
|
}
|
|
|
|
if (gpu->mmu->need_flush || gpu->switch_context) {
|
|
u32 new_target = gpu_va(gpu, buffer) + buffer->user_size;
|
|
|
|
if (gpu->mmu->need_flush) {
|
|
/* Add the MMU flush */
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_MMU,
|
|
VIVS_GL_FLUSH_MMU_FLUSH_FEMMU |
|
|
VIVS_GL_FLUSH_MMU_FLUSH_UNK1 |
|
|
VIVS_GL_FLUSH_MMU_FLUSH_UNK2 |
|
|
VIVS_GL_FLUSH_MMU_FLUSH_PEMMU |
|
|
VIVS_GL_FLUSH_MMU_FLUSH_UNK4);
|
|
|
|
gpu->mmu->need_flush = false;
|
|
}
|
|
|
|
if (gpu->switch_context) {
|
|
etnaviv_cmd_select_pipe(gpu, buffer, cmdbuf->exec_state);
|
|
gpu->exec_state = cmdbuf->exec_state;
|
|
gpu->switch_context = false;
|
|
}
|
|
|
|
/* And the link to the first buffer */
|
|
CMD_LINK(buffer, link_size, link_target);
|
|
|
|
/* Update the link target to point to above instructions */
|
|
link_target = new_target;
|
|
link_size = extra_size;
|
|
}
|
|
|
|
/* trigger event */
|
|
CMD_LOAD_STATE(buffer, VIVS_GL_EVENT, VIVS_GL_EVENT_EVENT_ID(event) |
|
|
VIVS_GL_EVENT_FROM_PE);
|
|
|
|
/* append WAIT/LINK to main buffer */
|
|
CMD_WAIT(buffer);
|
|
CMD_LINK(buffer, 2, gpu_va(gpu, buffer) + (buffer->user_size - 4));
|
|
|
|
etnaviv_buffer_replace_wait(buffer, waitlink_offset,
|
|
VIV_FE_LINK_HEADER_OP_LINK |
|
|
VIV_FE_LINK_HEADER_PREFETCH(link_size),
|
|
link_target);
|
|
|
|
if (drm_debug & DRM_UT_DRIVER)
|
|
etnaviv_buffer_dump(gpu, buffer, 0, 0x50);
|
|
}
|