mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-22 20:23:57 +08:00
961e3beae3
Job submission currently relies on the fact that struct drm_tegra_reloc and struct host1x_reloc are the same size and uses a simple call to the copy_from_user() function to copy them to kernel space. This causes the handle to be stored in the buffer object field, which then needs a cast to a 32 bit integer to resolve it to a proper buffer object pointer and store it back in the buffer object field. On 64-bit architectures that will no longer work, since pointers are 64 bits wide whereas handles will remain 32 bits. This causes the sizes of both structures to because different and copying will no longer work. Fix this by adding a new function, host1x_reloc_get_user(), that copies the structures field by field. While at it, use substructures for the command and target buffers in struct host1x_reloc for better readability. Also use unsized types to make it more obvious that this isn't part of userspace ABI. Signed-off-by: Thierry Reding <treding@nvidia.com>
599 lines
13 KiB
C
599 lines
13 KiB
C
/*
|
|
* Tegra host1x Job
|
|
*
|
|
* Copyright (c) 2010-2013, NVIDIA Corporation.
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify it
|
|
* under the terms and conditions of the GNU General Public License,
|
|
* version 2, as published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
* more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/err.h>
|
|
#include <linux/host1x.h>
|
|
#include <linux/kref.h>
|
|
#include <linux/module.h>
|
|
#include <linux/scatterlist.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <trace/events/host1x.h>
|
|
|
|
#include "channel.h"
|
|
#include "dev.h"
|
|
#include "job.h"
|
|
#include "syncpt.h"
|
|
|
|
struct host1x_job *host1x_job_alloc(struct host1x_channel *ch,
|
|
u32 num_cmdbufs, u32 num_relocs,
|
|
u32 num_waitchks)
|
|
{
|
|
struct host1x_job *job = NULL;
|
|
unsigned int num_unpins = num_cmdbufs + num_relocs;
|
|
u64 total;
|
|
void *mem;
|
|
|
|
/* Check that we're not going to overflow */
|
|
total = sizeof(struct host1x_job) +
|
|
(u64)num_relocs * sizeof(struct host1x_reloc) +
|
|
(u64)num_unpins * sizeof(struct host1x_job_unpin_data) +
|
|
(u64)num_waitchks * sizeof(struct host1x_waitchk) +
|
|
(u64)num_cmdbufs * sizeof(struct host1x_job_gather) +
|
|
(u64)num_unpins * sizeof(dma_addr_t) +
|
|
(u64)num_unpins * sizeof(u32 *);
|
|
if (total > ULONG_MAX)
|
|
return NULL;
|
|
|
|
mem = job = kzalloc(total, GFP_KERNEL);
|
|
if (!job)
|
|
return NULL;
|
|
|
|
kref_init(&job->ref);
|
|
job->channel = ch;
|
|
|
|
/* Redistribute memory to the structs */
|
|
mem += sizeof(struct host1x_job);
|
|
job->relocarray = num_relocs ? mem : NULL;
|
|
mem += num_relocs * sizeof(struct host1x_reloc);
|
|
job->unpins = num_unpins ? mem : NULL;
|
|
mem += num_unpins * sizeof(struct host1x_job_unpin_data);
|
|
job->waitchk = num_waitchks ? mem : NULL;
|
|
mem += num_waitchks * sizeof(struct host1x_waitchk);
|
|
job->gathers = num_cmdbufs ? mem : NULL;
|
|
mem += num_cmdbufs * sizeof(struct host1x_job_gather);
|
|
job->addr_phys = num_unpins ? mem : NULL;
|
|
|
|
job->reloc_addr_phys = job->addr_phys;
|
|
job->gather_addr_phys = &job->addr_phys[num_relocs];
|
|
|
|
return job;
|
|
}
|
|
EXPORT_SYMBOL(host1x_job_alloc);
|
|
|
|
struct host1x_job *host1x_job_get(struct host1x_job *job)
|
|
{
|
|
kref_get(&job->ref);
|
|
return job;
|
|
}
|
|
EXPORT_SYMBOL(host1x_job_get);
|
|
|
|
static void job_free(struct kref *ref)
|
|
{
|
|
struct host1x_job *job = container_of(ref, struct host1x_job, ref);
|
|
|
|
kfree(job);
|
|
}
|
|
|
|
void host1x_job_put(struct host1x_job *job)
|
|
{
|
|
kref_put(&job->ref, job_free);
|
|
}
|
|
EXPORT_SYMBOL(host1x_job_put);
|
|
|
|
void host1x_job_add_gather(struct host1x_job *job, struct host1x_bo *bo,
|
|
u32 words, u32 offset)
|
|
{
|
|
struct host1x_job_gather *cur_gather = &job->gathers[job->num_gathers];
|
|
|
|
cur_gather->words = words;
|
|
cur_gather->bo = bo;
|
|
cur_gather->offset = offset;
|
|
job->num_gathers++;
|
|
}
|
|
EXPORT_SYMBOL(host1x_job_add_gather);
|
|
|
|
/*
|
|
* NULL an already satisfied WAIT_SYNCPT host method, by patching its
|
|
* args in the command stream. The method data is changed to reference
|
|
* a reserved (never given out or incr) HOST1X_SYNCPT_RESERVED syncpt
|
|
* with a matching threshold value of 0, so is guaranteed to be popped
|
|
* by the host HW.
|
|
*/
|
|
static void host1x_syncpt_patch_offset(struct host1x_syncpt *sp,
|
|
struct host1x_bo *h, u32 offset)
|
|
{
|
|
void *patch_addr = NULL;
|
|
|
|
/* patch the wait */
|
|
patch_addr = host1x_bo_kmap(h, offset >> PAGE_SHIFT);
|
|
if (patch_addr) {
|
|
host1x_syncpt_patch_wait(sp,
|
|
patch_addr + (offset & ~PAGE_MASK));
|
|
host1x_bo_kunmap(h, offset >> PAGE_SHIFT, patch_addr);
|
|
} else
|
|
pr_err("Could not map cmdbuf for wait check\n");
|
|
}
|
|
|
|
/*
|
|
* Check driver supplied waitchk structs for syncpt thresholds
|
|
* that have already been satisfied and NULL the comparison (to
|
|
* avoid a wrap condition in the HW).
|
|
*/
|
|
static int do_waitchks(struct host1x_job *job, struct host1x *host,
|
|
struct host1x_bo *patch)
|
|
{
|
|
int i;
|
|
|
|
/* compare syncpt vs wait threshold */
|
|
for (i = 0; i < job->num_waitchk; i++) {
|
|
struct host1x_waitchk *wait = &job->waitchk[i];
|
|
struct host1x_syncpt *sp =
|
|
host1x_syncpt_get(host, wait->syncpt_id);
|
|
|
|
/* validate syncpt id */
|
|
if (wait->syncpt_id > host1x_syncpt_nb_pts(host))
|
|
continue;
|
|
|
|
/* skip all other gathers */
|
|
if (patch != wait->bo)
|
|
continue;
|
|
|
|
trace_host1x_syncpt_wait_check(wait->bo, wait->offset,
|
|
wait->syncpt_id, wait->thresh,
|
|
host1x_syncpt_read_min(sp));
|
|
|
|
if (host1x_syncpt_is_expired(sp, wait->thresh)) {
|
|
dev_dbg(host->dev,
|
|
"drop WAIT id %d (%s) thresh 0x%x, min 0x%x\n",
|
|
wait->syncpt_id, sp->name, wait->thresh,
|
|
host1x_syncpt_read_min(sp));
|
|
|
|
host1x_syncpt_patch_offset(sp, patch, wait->offset);
|
|
}
|
|
|
|
wait->bo = NULL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static unsigned int pin_job(struct host1x_job *job)
|
|
{
|
|
unsigned int i;
|
|
|
|
job->num_unpins = 0;
|
|
|
|
for (i = 0; i < job->num_relocs; i++) {
|
|
struct host1x_reloc *reloc = &job->relocarray[i];
|
|
struct sg_table *sgt;
|
|
dma_addr_t phys_addr;
|
|
|
|
reloc->target.bo = host1x_bo_get(reloc->target.bo);
|
|
if (!reloc->target.bo)
|
|
goto unpin;
|
|
|
|
phys_addr = host1x_bo_pin(reloc->target.bo, &sgt);
|
|
if (!phys_addr)
|
|
goto unpin;
|
|
|
|
job->addr_phys[job->num_unpins] = phys_addr;
|
|
job->unpins[job->num_unpins].bo = reloc->target.bo;
|
|
job->unpins[job->num_unpins].sgt = sgt;
|
|
job->num_unpins++;
|
|
}
|
|
|
|
for (i = 0; i < job->num_gathers; i++) {
|
|
struct host1x_job_gather *g = &job->gathers[i];
|
|
struct sg_table *sgt;
|
|
dma_addr_t phys_addr;
|
|
|
|
g->bo = host1x_bo_get(g->bo);
|
|
if (!g->bo)
|
|
goto unpin;
|
|
|
|
phys_addr = host1x_bo_pin(g->bo, &sgt);
|
|
if (!phys_addr)
|
|
goto unpin;
|
|
|
|
job->addr_phys[job->num_unpins] = phys_addr;
|
|
job->unpins[job->num_unpins].bo = g->bo;
|
|
job->unpins[job->num_unpins].sgt = sgt;
|
|
job->num_unpins++;
|
|
}
|
|
|
|
return job->num_unpins;
|
|
|
|
unpin:
|
|
host1x_job_unpin(job);
|
|
return 0;
|
|
}
|
|
|
|
static unsigned int do_relocs(struct host1x_job *job, struct host1x_bo *cmdbuf)
|
|
{
|
|
int i = 0;
|
|
u32 last_page = ~0;
|
|
void *cmdbuf_page_addr = NULL;
|
|
|
|
/* pin & patch the relocs for one gather */
|
|
for (i = 0; i < job->num_relocs; i++) {
|
|
struct host1x_reloc *reloc = &job->relocarray[i];
|
|
u32 reloc_addr = (job->reloc_addr_phys[i] +
|
|
reloc->target.offset) >> reloc->shift;
|
|
u32 *target;
|
|
|
|
/* skip all other gathers */
|
|
if (cmdbuf != reloc->cmdbuf.bo)
|
|
continue;
|
|
|
|
if (last_page != reloc->cmdbuf.offset >> PAGE_SHIFT) {
|
|
if (cmdbuf_page_addr)
|
|
host1x_bo_kunmap(cmdbuf, last_page,
|
|
cmdbuf_page_addr);
|
|
|
|
cmdbuf_page_addr = host1x_bo_kmap(cmdbuf,
|
|
reloc->cmdbuf.offset >> PAGE_SHIFT);
|
|
last_page = reloc->cmdbuf.offset >> PAGE_SHIFT;
|
|
|
|
if (unlikely(!cmdbuf_page_addr)) {
|
|
pr_err("Could not map cmdbuf for relocation\n");
|
|
return -ENOMEM;
|
|
}
|
|
}
|
|
|
|
target = cmdbuf_page_addr + (reloc->cmdbuf.offset & ~PAGE_MASK);
|
|
*target = reloc_addr;
|
|
}
|
|
|
|
if (cmdbuf_page_addr)
|
|
host1x_bo_kunmap(cmdbuf, last_page, cmdbuf_page_addr);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static bool check_reloc(struct host1x_reloc *reloc, struct host1x_bo *cmdbuf,
|
|
unsigned int offset)
|
|
{
|
|
offset *= sizeof(u32);
|
|
|
|
if (reloc->cmdbuf.bo != cmdbuf || reloc->cmdbuf.offset != offset)
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
struct host1x_firewall {
|
|
struct host1x_job *job;
|
|
struct device *dev;
|
|
|
|
unsigned int num_relocs;
|
|
struct host1x_reloc *reloc;
|
|
|
|
struct host1x_bo *cmdbuf;
|
|
unsigned int offset;
|
|
|
|
u32 words;
|
|
u32 class;
|
|
u32 reg;
|
|
u32 mask;
|
|
u32 count;
|
|
};
|
|
|
|
static int check_register(struct host1x_firewall *fw, unsigned long offset)
|
|
{
|
|
if (fw->job->is_addr_reg(fw->dev, fw->class, offset)) {
|
|
if (!fw->num_relocs)
|
|
return -EINVAL;
|
|
|
|
if (!check_reloc(fw->reloc, fw->cmdbuf, fw->offset))
|
|
return -EINVAL;
|
|
|
|
fw->num_relocs--;
|
|
fw->reloc++;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int check_mask(struct host1x_firewall *fw)
|
|
{
|
|
u32 mask = fw->mask;
|
|
u32 reg = fw->reg;
|
|
int ret;
|
|
|
|
while (mask) {
|
|
if (fw->words == 0)
|
|
return -EINVAL;
|
|
|
|
if (mask & 1) {
|
|
ret = check_register(fw, reg);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
fw->words--;
|
|
fw->offset++;
|
|
}
|
|
mask >>= 1;
|
|
reg++;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int check_incr(struct host1x_firewall *fw)
|
|
{
|
|
u32 count = fw->count;
|
|
u32 reg = fw->reg;
|
|
int ret;
|
|
|
|
while (count) {
|
|
if (fw->words == 0)
|
|
return -EINVAL;
|
|
|
|
ret = check_register(fw, reg);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
reg++;
|
|
fw->words--;
|
|
fw->offset++;
|
|
count--;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int check_nonincr(struct host1x_firewall *fw)
|
|
{
|
|
u32 count = fw->count;
|
|
int ret;
|
|
|
|
while (count) {
|
|
if (fw->words == 0)
|
|
return -EINVAL;
|
|
|
|
ret = check_register(fw, fw->reg);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
fw->words--;
|
|
fw->offset++;
|
|
count--;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int validate(struct host1x_firewall *fw, struct host1x_job_gather *g)
|
|
{
|
|
u32 *cmdbuf_base = (u32 *)fw->job->gather_copy_mapped +
|
|
(g->offset / sizeof(u32));
|
|
int err = 0;
|
|
|
|
if (!fw->job->is_addr_reg)
|
|
return 0;
|
|
|
|
fw->words = g->words;
|
|
fw->cmdbuf = g->bo;
|
|
fw->offset = 0;
|
|
|
|
while (fw->words && !err) {
|
|
u32 word = cmdbuf_base[fw->offset];
|
|
u32 opcode = (word & 0xf0000000) >> 28;
|
|
|
|
fw->mask = 0;
|
|
fw->reg = 0;
|
|
fw->count = 0;
|
|
fw->words--;
|
|
fw->offset++;
|
|
|
|
switch (opcode) {
|
|
case 0:
|
|
fw->class = word >> 6 & 0x3ff;
|
|
fw->mask = word & 0x3f;
|
|
fw->reg = word >> 16 & 0xfff;
|
|
err = check_mask(fw);
|
|
if (err)
|
|
goto out;
|
|
break;
|
|
case 1:
|
|
fw->reg = word >> 16 & 0xfff;
|
|
fw->count = word & 0xffff;
|
|
err = check_incr(fw);
|
|
if (err)
|
|
goto out;
|
|
break;
|
|
|
|
case 2:
|
|
fw->reg = word >> 16 & 0xfff;
|
|
fw->count = word & 0xffff;
|
|
err = check_nonincr(fw);
|
|
if (err)
|
|
goto out;
|
|
break;
|
|
|
|
case 3:
|
|
fw->mask = word & 0xffff;
|
|
fw->reg = word >> 16 & 0xfff;
|
|
err = check_mask(fw);
|
|
if (err)
|
|
goto out;
|
|
break;
|
|
case 4:
|
|
case 5:
|
|
case 14:
|
|
break;
|
|
default:
|
|
err = -EINVAL;
|
|
break;
|
|
}
|
|
}
|
|
|
|
out:
|
|
return err;
|
|
}
|
|
|
|
static inline int copy_gathers(struct host1x_job *job, struct device *dev)
|
|
{
|
|
struct host1x_firewall fw;
|
|
size_t size = 0;
|
|
size_t offset = 0;
|
|
int i;
|
|
|
|
fw.job = job;
|
|
fw.dev = dev;
|
|
fw.reloc = job->relocarray;
|
|
fw.num_relocs = job->num_relocs;
|
|
fw.class = 0;
|
|
|
|
for (i = 0; i < job->num_gathers; i++) {
|
|
struct host1x_job_gather *g = &job->gathers[i];
|
|
size += g->words * sizeof(u32);
|
|
}
|
|
|
|
job->gather_copy_mapped = dma_alloc_writecombine(dev, size,
|
|
&job->gather_copy,
|
|
GFP_KERNEL);
|
|
if (!job->gather_copy_mapped) {
|
|
job->gather_copy_mapped = NULL;
|
|
return -ENOMEM;
|
|
}
|
|
|
|
job->gather_copy_size = size;
|
|
|
|
for (i = 0; i < job->num_gathers; i++) {
|
|
struct host1x_job_gather *g = &job->gathers[i];
|
|
void *gather;
|
|
|
|
/* Copy the gather */
|
|
gather = host1x_bo_mmap(g->bo);
|
|
memcpy(job->gather_copy_mapped + offset, gather + g->offset,
|
|
g->words * sizeof(u32));
|
|
host1x_bo_munmap(g->bo, gather);
|
|
|
|
/* Store the location in the buffer */
|
|
g->base = job->gather_copy;
|
|
g->offset = offset;
|
|
|
|
/* Validate the job */
|
|
if (validate(&fw, g))
|
|
return -EINVAL;
|
|
|
|
offset += g->words * sizeof(u32);
|
|
}
|
|
|
|
/* No relocs should remain at this point */
|
|
if (fw.num_relocs)
|
|
return -EINVAL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int host1x_job_pin(struct host1x_job *job, struct device *dev)
|
|
{
|
|
int err;
|
|
unsigned int i, j;
|
|
struct host1x *host = dev_get_drvdata(dev->parent);
|
|
DECLARE_BITMAP(waitchk_mask, host1x_syncpt_nb_pts(host));
|
|
|
|
bitmap_zero(waitchk_mask, host1x_syncpt_nb_pts(host));
|
|
for (i = 0; i < job->num_waitchk; i++) {
|
|
u32 syncpt_id = job->waitchk[i].syncpt_id;
|
|
if (syncpt_id < host1x_syncpt_nb_pts(host))
|
|
set_bit(syncpt_id, waitchk_mask);
|
|
}
|
|
|
|
/* get current syncpt values for waitchk */
|
|
for_each_set_bit(i, waitchk_mask, host1x_syncpt_nb_pts(host))
|
|
host1x_syncpt_load(host->syncpt + i);
|
|
|
|
/* pin memory */
|
|
err = pin_job(job);
|
|
if (!err)
|
|
goto out;
|
|
|
|
/* patch gathers */
|
|
for (i = 0; i < job->num_gathers; i++) {
|
|
struct host1x_job_gather *g = &job->gathers[i];
|
|
|
|
/* process each gather mem only once */
|
|
if (g->handled)
|
|
continue;
|
|
|
|
g->base = job->gather_addr_phys[i];
|
|
|
|
for (j = i + 1; j < job->num_gathers; j++)
|
|
if (job->gathers[j].bo == g->bo)
|
|
job->gathers[j].handled = true;
|
|
|
|
err = do_relocs(job, g->bo);
|
|
if (err)
|
|
break;
|
|
|
|
err = do_waitchks(job, host, g->bo);
|
|
if (err)
|
|
break;
|
|
}
|
|
|
|
if (IS_ENABLED(CONFIG_TEGRA_HOST1X_FIREWALL) && !err) {
|
|
err = copy_gathers(job, dev);
|
|
if (err) {
|
|
host1x_job_unpin(job);
|
|
return err;
|
|
}
|
|
}
|
|
|
|
out:
|
|
wmb();
|
|
|
|
return err;
|
|
}
|
|
EXPORT_SYMBOL(host1x_job_pin);
|
|
|
|
void host1x_job_unpin(struct host1x_job *job)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < job->num_unpins; i++) {
|
|
struct host1x_job_unpin_data *unpin = &job->unpins[i];
|
|
host1x_bo_unpin(unpin->bo, unpin->sgt);
|
|
host1x_bo_put(unpin->bo);
|
|
}
|
|
job->num_unpins = 0;
|
|
|
|
if (job->gather_copy_size)
|
|
dma_free_writecombine(job->channel->dev, job->gather_copy_size,
|
|
job->gather_copy_mapped,
|
|
job->gather_copy);
|
|
}
|
|
EXPORT_SYMBOL(host1x_job_unpin);
|
|
|
|
/*
|
|
* Debug routine used to dump job entries
|
|
*/
|
|
void host1x_job_dump(struct device *dev, struct host1x_job *job)
|
|
{
|
|
dev_dbg(dev, " SYNCPT_ID %d\n", job->syncpt_id);
|
|
dev_dbg(dev, " SYNCPT_VAL %d\n", job->syncpt_end);
|
|
dev_dbg(dev, " FIRST_GET 0x%x\n", job->first_get);
|
|
dev_dbg(dev, " TIMEOUT %d\n", job->timeout);
|
|
dev_dbg(dev, " NUM_SLOTS %d\n", job->num_slots);
|
|
dev_dbg(dev, " NUM_HANDLES %d\n", job->num_unpins);
|
|
}
|