2019-05-29 01:10:04 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2013-03-22 22:34:03 +08:00
|
|
|
/*
|
|
|
|
* Tegra host1x Job
|
|
|
|
*
|
2016-11-09 01:51:32 +08:00
|
|
|
* Copyright (c) 2010-2015, NVIDIA Corporation.
|
2013-03-22 22:34:03 +08:00
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/dma-mapping.h>
|
|
|
|
#include <linux/err.h>
|
2013-09-24 22:30:32 +08:00
|
|
|
#include <linux/host1x.h>
|
2020-02-04 21:59:25 +08:00
|
|
|
#include <linux/iommu.h>
|
2013-03-22 22:34:03 +08:00
|
|
|
#include <linux/kref.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/scatterlist.h>
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/vmalloc.h>
|
|
|
|
#include <trace/events/host1x.h>
|
|
|
|
|
|
|
|
#include "channel.h"
|
|
|
|
#include "dev.h"
|
|
|
|
#include "job.h"
|
|
|
|
#include "syncpt.h"
|
|
|
|
|
2017-06-15 07:18:39 +08:00
|
|
|
#define HOST1X_WAIT_SYNCPT_OFFSET 0x8
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
struct host1x_job *host1x_job_alloc(struct host1x_channel *ch,
|
2018-05-05 14:45:47 +08:00
|
|
|
u32 num_cmdbufs, u32 num_relocs)
|
2013-03-22 22:34:03 +08:00
|
|
|
{
|
|
|
|
struct host1x_job *job = NULL;
|
2020-06-29 11:18:39 +08:00
|
|
|
unsigned int num_unpins = num_relocs;
|
2013-03-22 22:34:03 +08:00
|
|
|
u64 total;
|
|
|
|
void *mem;
|
|
|
|
|
2020-06-29 11:18:39 +08:00
|
|
|
if (!IS_ENABLED(CONFIG_TEGRA_HOST1X_FIREWALL))
|
|
|
|
num_unpins += num_cmdbufs;
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
/* Check that we're not going to overflow */
|
|
|
|
total = sizeof(struct host1x_job) +
|
2013-08-23 18:18:25 +08:00
|
|
|
(u64)num_relocs * sizeof(struct host1x_reloc) +
|
|
|
|
(u64)num_unpins * sizeof(struct host1x_job_unpin_data) +
|
|
|
|
(u64)num_cmdbufs * sizeof(struct host1x_job_gather) +
|
|
|
|
(u64)num_unpins * sizeof(dma_addr_t) +
|
|
|
|
(u64)num_unpins * sizeof(u32 *);
|
2013-03-22 22:34:03 +08:00
|
|
|
if (total > ULONG_MAX)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
mem = job = kzalloc(total, GFP_KERNEL);
|
|
|
|
if (!job)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
kref_init(&job->ref);
|
|
|
|
job->channel = ch;
|
|
|
|
|
|
|
|
/* Redistribute memory to the structs */
|
|
|
|
mem += sizeof(struct host1x_job);
|
2018-05-16 22:58:44 +08:00
|
|
|
job->relocs = num_relocs ? mem : NULL;
|
2013-03-22 22:34:03 +08:00
|
|
|
mem += num_relocs * sizeof(struct host1x_reloc);
|
|
|
|
job->unpins = num_unpins ? mem : NULL;
|
|
|
|
mem += num_unpins * sizeof(struct host1x_job_unpin_data);
|
|
|
|
job->gathers = num_cmdbufs ? mem : NULL;
|
|
|
|
mem += num_cmdbufs * sizeof(struct host1x_job_gather);
|
|
|
|
job->addr_phys = num_unpins ? mem : NULL;
|
|
|
|
|
|
|
|
job->reloc_addr_phys = job->addr_phys;
|
|
|
|
job->gather_addr_phys = &job->addr_phys[num_relocs];
|
|
|
|
|
|
|
|
return job;
|
|
|
|
}
|
2013-11-08 18:41:42 +08:00
|
|
|
EXPORT_SYMBOL(host1x_job_alloc);
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
struct host1x_job *host1x_job_get(struct host1x_job *job)
|
|
|
|
{
|
|
|
|
kref_get(&job->ref);
|
|
|
|
return job;
|
|
|
|
}
|
2013-11-08 18:41:42 +08:00
|
|
|
EXPORT_SYMBOL(host1x_job_get);
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
static void job_free(struct kref *ref)
|
|
|
|
{
|
|
|
|
struct host1x_job *job = container_of(ref, struct host1x_job, ref);
|
|
|
|
|
2021-03-29 21:38:32 +08:00
|
|
|
if (job->syncpt)
|
|
|
|
host1x_syncpt_put(job->syncpt);
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
kfree(job);
|
|
|
|
}
|
|
|
|
|
|
|
|
void host1x_job_put(struct host1x_job *job)
|
|
|
|
{
|
|
|
|
kref_put(&job->ref, job_free);
|
|
|
|
}
|
2013-11-08 18:41:42 +08:00
|
|
|
EXPORT_SYMBOL(host1x_job_put);
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
void host1x_job_add_gather(struct host1x_job *job, struct host1x_bo *bo,
|
2018-05-16 23:01:43 +08:00
|
|
|
unsigned int words, unsigned int offset)
|
2013-03-22 22:34:03 +08:00
|
|
|
{
|
2018-05-16 23:01:43 +08:00
|
|
|
struct host1x_job_gather *gather = &job->gathers[job->num_gathers];
|
|
|
|
|
|
|
|
gather->words = words;
|
|
|
|
gather->bo = bo;
|
|
|
|
gather->offset = offset;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
job->num_gathers++;
|
|
|
|
}
|
2013-11-08 18:41:42 +08:00
|
|
|
EXPORT_SYMBOL(host1x_job_add_gather);
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2016-12-14 19:16:14 +08:00
|
|
|
static unsigned int pin_job(struct host1x *host, struct host1x_job *job)
|
2013-03-22 22:34:03 +08:00
|
|
|
{
|
2019-10-28 20:37:13 +08:00
|
|
|
struct host1x_client *client = job->client;
|
|
|
|
struct device *dev = client->dev;
|
2020-06-29 11:18:40 +08:00
|
|
|
struct host1x_job_gather *g;
|
2020-02-04 21:59:25 +08:00
|
|
|
struct iommu_domain *domain;
|
2013-03-22 22:34:03 +08:00
|
|
|
unsigned int i;
|
2016-12-14 19:16:14 +08:00
|
|
|
int err;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2020-02-04 21:59:25 +08:00
|
|
|
domain = iommu_get_domain_for_dev(dev);
|
2013-03-22 22:34:03 +08:00
|
|
|
job->num_unpins = 0;
|
|
|
|
|
|
|
|
for (i = 0; i < job->num_relocs; i++) {
|
2018-05-16 22:58:44 +08:00
|
|
|
struct host1x_reloc *reloc = &job->relocs[i];
|
2019-10-28 20:37:13 +08:00
|
|
|
dma_addr_t phys_addr, *phys;
|
2013-03-22 22:34:03 +08:00
|
|
|
struct sg_table *sgt;
|
|
|
|
|
2014-06-10 16:25:00 +08:00
|
|
|
reloc->target.bo = host1x_bo_get(reloc->target.bo);
|
2016-12-14 19:16:14 +08:00
|
|
|
if (!reloc->target.bo) {
|
|
|
|
err = -EINVAL;
|
2013-03-22 22:34:03 +08:00
|
|
|
goto unpin;
|
2016-12-14 19:16:14 +08:00
|
|
|
}
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2020-02-04 21:59:25 +08:00
|
|
|
/*
|
|
|
|
* If the client device is not attached to an IOMMU, the
|
|
|
|
* physical address of the buffer object can be used.
|
|
|
|
*
|
|
|
|
* Similarly, when an IOMMU domain is shared between all
|
|
|
|
* host1x clients, the IOVA is already available, so no
|
|
|
|
* need to map the buffer object again.
|
|
|
|
*
|
|
|
|
* XXX Note that this isn't always safe to do because it
|
|
|
|
* relies on an assumption that no cache maintenance is
|
|
|
|
* needed on the buffer objects.
|
|
|
|
*/
|
|
|
|
if (!domain || client->group)
|
2019-10-28 20:37:13 +08:00
|
|
|
phys = &phys_addr;
|
|
|
|
else
|
|
|
|
phys = NULL;
|
|
|
|
|
|
|
|
sgt = host1x_bo_pin(dev, reloc->target.bo, phys);
|
gpu: host1x: Overhaul host1x_bo_{pin,unpin}() API
The host1x_bo_pin() and host1x_bo_unpin() APIs are used to pin and unpin
buffers during host1x job submission. Pinning currently returns the SG
table and the DMA address (an IOVA if an IOMMU is used or a physical
address if no IOMMU is used) of the buffer. The DMA address is only used
for buffers that are relocated, whereas the host1x driver will map
gather buffers into its own IOVA space so that they can be processed by
the CDMA engine.
This approach has a couple of issues. On one hand it's not very useful
to return a DMA address for the buffer if host1x doesn't need it. On the
other hand, returning the SG table of the buffer is suboptimal because a
single SG table cannot be shared for multiple mappings, because the DMA
address is stored within the SG table, and the DMA address may be
different for different devices.
Subsequent patches will move the host1x driver over to the DMA API which
doesn't work with a single shared SG table. Fix this by returning a new
SG table each time a buffer is pinned. This allows the buffer to be
referenced by multiple jobs for different engines.
Change the prototypes of host1x_bo_pin() and host1x_bo_unpin() to take a
struct device *, specifying the device for which the buffer should be
pinned. This is required in order to be able to properly construct the
SG table. While at it, make host1x_bo_pin() return the SG table because
that allows us to return an ERR_PTR()-encoded error code if we need to,
or return NULL to signal that we don't need the SG table to be remapped
and can simply use the DMA address as-is. At the same time, returning
the DMA address is made optional because in the example of command
buffers, host1x doesn't need to know the DMA address since it will have
to create its own mapping anyway.
Signed-off-by: Thierry Reding <treding@nvidia.com>
2019-10-28 20:37:09 +08:00
|
|
|
if (IS_ERR(sgt)) {
|
|
|
|
err = PTR_ERR(sgt);
|
|
|
|
goto unpin;
|
|
|
|
}
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2019-10-28 20:37:13 +08:00
|
|
|
if (sgt) {
|
|
|
|
unsigned long mask = HOST1X_RELOC_READ |
|
|
|
|
HOST1X_RELOC_WRITE;
|
|
|
|
enum dma_data_direction dir;
|
|
|
|
|
|
|
|
switch (reloc->flags & mask) {
|
|
|
|
case HOST1X_RELOC_READ:
|
|
|
|
dir = DMA_TO_DEVICE;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case HOST1X_RELOC_WRITE:
|
|
|
|
dir = DMA_FROM_DEVICE;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case HOST1X_RELOC_READ | HOST1X_RELOC_WRITE:
|
|
|
|
dir = DMA_BIDIRECTIONAL;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
err = -EINVAL;
|
|
|
|
goto unpin;
|
|
|
|
}
|
|
|
|
|
2020-04-28 19:11:16 +08:00
|
|
|
err = dma_map_sgtable(dev, sgt, dir, 0);
|
|
|
|
if (err)
|
2019-10-28 20:37:13 +08:00
|
|
|
goto unpin;
|
|
|
|
|
|
|
|
job->unpins[job->num_unpins].dev = dev;
|
|
|
|
job->unpins[job->num_unpins].dir = dir;
|
|
|
|
phys_addr = sg_dma_address(sgt->sgl);
|
|
|
|
}
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
job->addr_phys[job->num_unpins] = phys_addr;
|
2014-06-10 16:25:00 +08:00
|
|
|
job->unpins[job->num_unpins].bo = reloc->target.bo;
|
2013-03-22 22:34:03 +08:00
|
|
|
job->unpins[job->num_unpins].sgt = sgt;
|
|
|
|
job->num_unpins++;
|
|
|
|
}
|
|
|
|
|
2020-06-29 11:18:39 +08:00
|
|
|
/*
|
|
|
|
* We will copy gathers BO content later, so there is no need to
|
|
|
|
* hold and pin them.
|
|
|
|
*/
|
|
|
|
if (IS_ENABLED(CONFIG_TEGRA_HOST1X_FIREWALL))
|
|
|
|
return 0;
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
for (i = 0; i < job->num_gathers; i++) {
|
2016-12-14 19:16:14 +08:00
|
|
|
size_t gather_size = 0;
|
|
|
|
struct scatterlist *sg;
|
2013-03-22 22:34:03 +08:00
|
|
|
struct sg_table *sgt;
|
|
|
|
dma_addr_t phys_addr;
|
2016-12-14 19:16:14 +08:00
|
|
|
unsigned long shift;
|
|
|
|
struct iova *alloc;
|
2020-02-04 21:59:25 +08:00
|
|
|
dma_addr_t *phys;
|
2016-12-14 19:16:14 +08:00
|
|
|
unsigned int j;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2020-06-29 11:18:40 +08:00
|
|
|
g = &job->gathers[i];
|
2013-03-22 22:34:03 +08:00
|
|
|
g->bo = host1x_bo_get(g->bo);
|
2016-12-14 19:16:14 +08:00
|
|
|
if (!g->bo) {
|
|
|
|
err = -EINVAL;
|
2013-03-22 22:34:03 +08:00
|
|
|
goto unpin;
|
2016-12-14 19:16:14 +08:00
|
|
|
}
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2020-02-04 21:59:25 +08:00
|
|
|
/**
|
|
|
|
* If the host1x is not attached to an IOMMU, there is no need
|
|
|
|
* to map the buffer object for the host1x, since the physical
|
|
|
|
* address can simply be used.
|
|
|
|
*/
|
|
|
|
if (!iommu_get_domain_for_dev(host->dev))
|
|
|
|
phys = &phys_addr;
|
|
|
|
else
|
|
|
|
phys = NULL;
|
|
|
|
|
|
|
|
sgt = host1x_bo_pin(host->dev, g->bo, phys);
|
gpu: host1x: Overhaul host1x_bo_{pin,unpin}() API
The host1x_bo_pin() and host1x_bo_unpin() APIs are used to pin and unpin
buffers during host1x job submission. Pinning currently returns the SG
table and the DMA address (an IOVA if an IOMMU is used or a physical
address if no IOMMU is used) of the buffer. The DMA address is only used
for buffers that are relocated, whereas the host1x driver will map
gather buffers into its own IOVA space so that they can be processed by
the CDMA engine.
This approach has a couple of issues. On one hand it's not very useful
to return a DMA address for the buffer if host1x doesn't need it. On the
other hand, returning the SG table of the buffer is suboptimal because a
single SG table cannot be shared for multiple mappings, because the DMA
address is stored within the SG table, and the DMA address may be
different for different devices.
Subsequent patches will move the host1x driver over to the DMA API which
doesn't work with a single shared SG table. Fix this by returning a new
SG table each time a buffer is pinned. This allows the buffer to be
referenced by multiple jobs for different engines.
Change the prototypes of host1x_bo_pin() and host1x_bo_unpin() to take a
struct device *, specifying the device for which the buffer should be
pinned. This is required in order to be able to properly construct the
SG table. While at it, make host1x_bo_pin() return the SG table because
that allows us to return an ERR_PTR()-encoded error code if we need to,
or return NULL to signal that we don't need the SG table to be remapped
and can simply use the DMA address as-is. At the same time, returning
the DMA address is made optional because in the example of command
buffers, host1x doesn't need to know the DMA address since it will have
to create its own mapping anyway.
Signed-off-by: Thierry Reding <treding@nvidia.com>
2019-10-28 20:37:09 +08:00
|
|
|
if (IS_ERR(sgt)) {
|
|
|
|
err = PTR_ERR(sgt);
|
2020-06-29 11:18:40 +08:00
|
|
|
goto put;
|
gpu: host1x: Overhaul host1x_bo_{pin,unpin}() API
The host1x_bo_pin() and host1x_bo_unpin() APIs are used to pin and unpin
buffers during host1x job submission. Pinning currently returns the SG
table and the DMA address (an IOVA if an IOMMU is used or a physical
address if no IOMMU is used) of the buffer. The DMA address is only used
for buffers that are relocated, whereas the host1x driver will map
gather buffers into its own IOVA space so that they can be processed by
the CDMA engine.
This approach has a couple of issues. On one hand it's not very useful
to return a DMA address for the buffer if host1x doesn't need it. On the
other hand, returning the SG table of the buffer is suboptimal because a
single SG table cannot be shared for multiple mappings, because the DMA
address is stored within the SG table, and the DMA address may be
different for different devices.
Subsequent patches will move the host1x driver over to the DMA API which
doesn't work with a single shared SG table. Fix this by returning a new
SG table each time a buffer is pinned. This allows the buffer to be
referenced by multiple jobs for different engines.
Change the prototypes of host1x_bo_pin() and host1x_bo_unpin() to take a
struct device *, specifying the device for which the buffer should be
pinned. This is required in order to be able to properly construct the
SG table. While at it, make host1x_bo_pin() return the SG table because
that allows us to return an ERR_PTR()-encoded error code if we need to,
or return NULL to signal that we don't need the SG table to be remapped
and can simply use the DMA address as-is. At the same time, returning
the DMA address is made optional because in the example of command
buffers, host1x doesn't need to know the DMA address since it will have
to create its own mapping anyway.
Signed-off-by: Thierry Reding <treding@nvidia.com>
2019-10-28 20:37:09 +08:00
|
|
|
}
|
2016-12-14 19:16:14 +08:00
|
|
|
|
2020-06-29 11:18:39 +08:00
|
|
|
if (host->domain) {
|
2020-04-28 19:11:16 +08:00
|
|
|
for_each_sgtable_sg(sgt, sg, j)
|
2016-12-14 19:16:14 +08:00
|
|
|
gather_size += sg->length;
|
|
|
|
gather_size = iova_align(&host->iova, gather_size);
|
|
|
|
|
|
|
|
shift = iova_shift(&host->iova);
|
|
|
|
alloc = alloc_iova(&host->iova, gather_size >> shift,
|
|
|
|
host->iova_end >> shift, true);
|
|
|
|
if (!alloc) {
|
|
|
|
err = -ENOMEM;
|
2020-06-29 11:18:40 +08:00
|
|
|
goto put;
|
2016-12-14 19:16:14 +08:00
|
|
|
}
|
|
|
|
|
2020-04-28 19:11:16 +08:00
|
|
|
err = iommu_map_sgtable(host->domain,
|
2016-12-14 19:16:14 +08:00
|
|
|
iova_dma_addr(&host->iova, alloc),
|
2020-04-28 19:11:16 +08:00
|
|
|
sgt, IOMMU_READ);
|
2016-12-14 19:16:14 +08:00
|
|
|
if (err == 0) {
|
|
|
|
__free_iova(&host->iova, alloc);
|
|
|
|
err = -EINVAL;
|
2020-06-29 11:18:40 +08:00
|
|
|
goto put;
|
2016-12-14 19:16:14 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
job->unpins[job->num_unpins].size = gather_size;
|
2019-10-28 20:37:13 +08:00
|
|
|
phys_addr = iova_dma_addr(&host->iova, alloc);
|
2020-02-04 21:59:25 +08:00
|
|
|
} else if (sgt) {
|
2020-04-28 19:11:16 +08:00
|
|
|
err = dma_map_sgtable(host->dev, sgt, DMA_TO_DEVICE, 0);
|
|
|
|
if (err)
|
2020-06-29 11:18:40 +08:00
|
|
|
goto put;
|
2019-10-28 20:37:13 +08:00
|
|
|
|
2020-02-04 21:59:26 +08:00
|
|
|
job->unpins[job->num_unpins].dir = DMA_TO_DEVICE;
|
2019-10-28 20:37:13 +08:00
|
|
|
job->unpins[job->num_unpins].dev = host->dev;
|
|
|
|
phys_addr = sg_dma_address(sgt->sgl);
|
2016-12-14 19:16:14 +08:00
|
|
|
}
|
|
|
|
|
2019-10-28 20:37:13 +08:00
|
|
|
job->addr_phys[job->num_unpins] = phys_addr;
|
|
|
|
job->gather_addr_phys[i] = phys_addr;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
job->unpins[job->num_unpins].bo = g->bo;
|
|
|
|
job->unpins[job->num_unpins].sgt = sgt;
|
|
|
|
job->num_unpins++;
|
|
|
|
}
|
|
|
|
|
2016-12-14 19:16:14 +08:00
|
|
|
return 0;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2020-06-29 11:18:40 +08:00
|
|
|
put:
|
|
|
|
host1x_bo_put(g->bo);
|
2013-03-22 22:34:03 +08:00
|
|
|
unpin:
|
|
|
|
host1x_job_unpin(job);
|
2016-12-14 19:16:14 +08:00
|
|
|
return err;
|
2013-03-22 22:34:03 +08:00
|
|
|
}
|
|
|
|
|
2017-06-15 07:18:34 +08:00
|
|
|
static int do_relocs(struct host1x_job *job, struct host1x_job_gather *g)
|
2013-03-22 22:34:03 +08:00
|
|
|
{
|
2019-11-18 18:35:22 +08:00
|
|
|
void *cmdbuf_addr = NULL;
|
2017-06-15 07:18:34 +08:00
|
|
|
struct host1x_bo *cmdbuf = g->bo;
|
2018-03-23 20:31:24 +08:00
|
|
|
unsigned int i;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
/* pin & patch the relocs for one gather */
|
2013-05-29 18:26:05 +08:00
|
|
|
for (i = 0; i < job->num_relocs; i++) {
|
2018-05-16 22:58:44 +08:00
|
|
|
struct host1x_reloc *reloc = &job->relocs[i];
|
2013-03-22 22:34:03 +08:00
|
|
|
u32 reloc_addr = (job->reloc_addr_phys[i] +
|
2014-06-10 16:25:00 +08:00
|
|
|
reloc->target.offset) >> reloc->shift;
|
2013-03-22 22:34:03 +08:00
|
|
|
u32 *target;
|
|
|
|
|
|
|
|
/* skip all other gathers */
|
2014-06-10 16:25:00 +08:00
|
|
|
if (cmdbuf != reloc->cmdbuf.bo)
|
2013-03-22 22:34:03 +08:00
|
|
|
continue;
|
|
|
|
|
2017-06-15 07:18:34 +08:00
|
|
|
if (IS_ENABLED(CONFIG_TEGRA_HOST1X_FIREWALL)) {
|
|
|
|
target = (u32 *)job->gather_copy_mapped +
|
|
|
|
reloc->cmdbuf.offset / sizeof(u32) +
|
|
|
|
g->offset / sizeof(u32);
|
|
|
|
goto patch_reloc;
|
|
|
|
}
|
|
|
|
|
2019-11-18 18:35:22 +08:00
|
|
|
if (!cmdbuf_addr) {
|
|
|
|
cmdbuf_addr = host1x_bo_mmap(cmdbuf);
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2019-11-18 18:35:22 +08:00
|
|
|
if (unlikely(!cmdbuf_addr)) {
|
2013-03-22 22:34:03 +08:00
|
|
|
pr_err("Could not map cmdbuf for relocation\n");
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-11-18 18:35:22 +08:00
|
|
|
target = cmdbuf_addr + reloc->cmdbuf.offset;
|
2017-06-15 07:18:34 +08:00
|
|
|
patch_reloc:
|
2013-03-22 22:34:03 +08:00
|
|
|
*target = reloc_addr;
|
|
|
|
}
|
|
|
|
|
2019-11-18 18:35:22 +08:00
|
|
|
if (cmdbuf_addr)
|
|
|
|
host1x_bo_munmap(cmdbuf, cmdbuf_addr);
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-05-29 18:26:03 +08:00
|
|
|
static bool check_reloc(struct host1x_reloc *reloc, struct host1x_bo *cmdbuf,
|
2013-10-10 16:17:45 +08:00
|
|
|
unsigned int offset)
|
2013-03-22 22:34:03 +08:00
|
|
|
{
|
|
|
|
offset *= sizeof(u32);
|
|
|
|
|
2014-06-10 16:25:00 +08:00
|
|
|
if (reloc->cmdbuf.bo != cmdbuf || reloc->cmdbuf.offset != offset)
|
2013-05-29 18:26:03 +08:00
|
|
|
return false;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2017-06-15 07:18:35 +08:00
|
|
|
/* relocation shift value validation isn't implemented yet */
|
|
|
|
if (reloc->shift)
|
|
|
|
return false;
|
|
|
|
|
2013-05-29 18:26:03 +08:00
|
|
|
return true;
|
2013-03-22 22:34:03 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
struct host1x_firewall {
|
|
|
|
struct host1x_job *job;
|
|
|
|
struct device *dev;
|
|
|
|
|
|
|
|
unsigned int num_relocs;
|
|
|
|
struct host1x_reloc *reloc;
|
|
|
|
|
2013-10-10 16:21:58 +08:00
|
|
|
struct host1x_bo *cmdbuf;
|
2013-03-22 22:34:03 +08:00
|
|
|
unsigned int offset;
|
|
|
|
|
|
|
|
u32 words;
|
|
|
|
u32 class;
|
|
|
|
u32 reg;
|
|
|
|
u32 mask;
|
|
|
|
u32 count;
|
|
|
|
};
|
|
|
|
|
2013-10-10 16:24:04 +08:00
|
|
|
static int check_register(struct host1x_firewall *fw, unsigned long offset)
|
|
|
|
{
|
2017-06-15 07:18:37 +08:00
|
|
|
if (!fw->job->is_addr_reg)
|
|
|
|
return 0;
|
|
|
|
|
2013-10-10 16:24:04 +08:00
|
|
|
if (fw->job->is_addr_reg(fw->dev, fw->class, offset)) {
|
|
|
|
if (!fw->num_relocs)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (!check_reloc(fw->reloc, fw->cmdbuf, fw->offset))
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
fw->num_relocs--;
|
|
|
|
fw->reloc++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-06-15 07:18:37 +08:00
|
|
|
static int check_class(struct host1x_firewall *fw, u32 class)
|
|
|
|
{
|
|
|
|
if (!fw->job->is_valid_class) {
|
|
|
|
if (fw->class != class)
|
|
|
|
return -EINVAL;
|
|
|
|
} else {
|
|
|
|
if (!fw->job->is_valid_class(fw->class))
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
static int check_mask(struct host1x_firewall *fw)
|
|
|
|
{
|
|
|
|
u32 mask = fw->mask;
|
|
|
|
u32 reg = fw->reg;
|
2013-10-10 16:24:04 +08:00
|
|
|
int ret;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
while (mask) {
|
|
|
|
if (fw->words == 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (mask & 1) {
|
2013-10-10 16:24:04 +08:00
|
|
|
ret = check_register(fw, reg);
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
fw->words--;
|
|
|
|
fw->offset++;
|
|
|
|
}
|
|
|
|
mask >>= 1;
|
|
|
|
reg++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int check_incr(struct host1x_firewall *fw)
|
|
|
|
{
|
|
|
|
u32 count = fw->count;
|
|
|
|
u32 reg = fw->reg;
|
2013-10-10 16:24:04 +08:00
|
|
|
int ret;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2013-05-29 18:26:02 +08:00
|
|
|
while (count) {
|
2013-03-22 22:34:03 +08:00
|
|
|
if (fw->words == 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2013-10-10 16:24:04 +08:00
|
|
|
ret = check_register(fw, reg);
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
reg++;
|
|
|
|
fw->words--;
|
|
|
|
fw->offset++;
|
|
|
|
count--;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int check_nonincr(struct host1x_firewall *fw)
|
|
|
|
{
|
|
|
|
u32 count = fw->count;
|
2013-10-10 16:24:04 +08:00
|
|
|
int ret;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
while (count) {
|
|
|
|
if (fw->words == 0)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2013-10-10 16:24:04 +08:00
|
|
|
ret = check_register(fw, fw->reg);
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
fw->words--;
|
|
|
|
fw->offset++;
|
|
|
|
count--;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-05-29 18:26:04 +08:00
|
|
|
static int validate(struct host1x_firewall *fw, struct host1x_job_gather *g)
|
2013-03-22 22:34:03 +08:00
|
|
|
{
|
2013-05-29 18:26:05 +08:00
|
|
|
u32 *cmdbuf_base = (u32 *)fw->job->gather_copy_mapped +
|
|
|
|
(g->offset / sizeof(u32));
|
2017-06-15 07:18:37 +08:00
|
|
|
u32 job_class = fw->class;
|
2013-03-22 22:34:03 +08:00
|
|
|
int err = 0;
|
|
|
|
|
2013-05-29 18:26:04 +08:00
|
|
|
fw->words = g->words;
|
2013-10-10 16:21:58 +08:00
|
|
|
fw->cmdbuf = g->bo;
|
2013-05-29 18:26:04 +08:00
|
|
|
fw->offset = 0;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2013-05-29 18:26:04 +08:00
|
|
|
while (fw->words && !err) {
|
|
|
|
u32 word = cmdbuf_base[fw->offset];
|
2013-03-22 22:34:03 +08:00
|
|
|
u32 opcode = (word & 0xf0000000) >> 28;
|
|
|
|
|
2013-05-29 18:26:04 +08:00
|
|
|
fw->mask = 0;
|
|
|
|
fw->reg = 0;
|
|
|
|
fw->count = 0;
|
|
|
|
fw->words--;
|
|
|
|
fw->offset++;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
switch (opcode) {
|
|
|
|
case 0:
|
2013-05-29 18:26:04 +08:00
|
|
|
fw->class = word >> 6 & 0x3ff;
|
|
|
|
fw->mask = word & 0x3f;
|
|
|
|
fw->reg = word >> 16 & 0xfff;
|
2017-06-15 07:18:37 +08:00
|
|
|
err = check_class(fw, job_class);
|
|
|
|
if (!err)
|
|
|
|
err = check_mask(fw);
|
2013-03-22 22:34:03 +08:00
|
|
|
if (err)
|
|
|
|
goto out;
|
|
|
|
break;
|
|
|
|
case 1:
|
2013-05-29 18:26:04 +08:00
|
|
|
fw->reg = word >> 16 & 0xfff;
|
|
|
|
fw->count = word & 0xffff;
|
|
|
|
err = check_incr(fw);
|
2013-03-22 22:34:03 +08:00
|
|
|
if (err)
|
|
|
|
goto out;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 2:
|
2013-05-29 18:26:04 +08:00
|
|
|
fw->reg = word >> 16 & 0xfff;
|
|
|
|
fw->count = word & 0xffff;
|
|
|
|
err = check_nonincr(fw);
|
2013-03-22 22:34:03 +08:00
|
|
|
if (err)
|
|
|
|
goto out;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 3:
|
2013-05-29 18:26:04 +08:00
|
|
|
fw->mask = word & 0xffff;
|
|
|
|
fw->reg = word >> 16 & 0xfff;
|
|
|
|
err = check_mask(fw);
|
2013-03-22 22:34:03 +08:00
|
|
|
if (err)
|
|
|
|
goto out;
|
|
|
|
break;
|
|
|
|
case 4:
|
|
|
|
case 14:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
err = -EINVAL;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2019-10-28 20:37:12 +08:00
|
|
|
static inline int copy_gathers(struct device *host, struct host1x_job *job,
|
|
|
|
struct device *dev)
|
2013-03-22 22:34:03 +08:00
|
|
|
{
|
2013-05-29 18:26:05 +08:00
|
|
|
struct host1x_firewall fw;
|
2013-03-22 22:34:03 +08:00
|
|
|
size_t size = 0;
|
|
|
|
size_t offset = 0;
|
2018-03-23 20:31:24 +08:00
|
|
|
unsigned int i;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2013-05-29 18:26:05 +08:00
|
|
|
fw.job = job;
|
|
|
|
fw.dev = dev;
|
2018-05-16 22:58:44 +08:00
|
|
|
fw.reloc = job->relocs;
|
2013-05-29 18:26:05 +08:00
|
|
|
fw.num_relocs = job->num_relocs;
|
2017-06-15 07:18:32 +08:00
|
|
|
fw.class = job->class;
|
2013-05-29 18:26:05 +08:00
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
for (i = 0; i < job->num_gathers; i++) {
|
|
|
|
struct host1x_job_gather *g = &job->gathers[i];
|
2016-06-23 17:33:31 +08:00
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
size += g->words * sizeof(u32);
|
|
|
|
}
|
|
|
|
|
2017-06-15 07:18:43 +08:00
|
|
|
/*
|
|
|
|
* Try a non-blocking allocation from a higher priority pools first,
|
|
|
|
* as awaiting for the allocation here is a major performance hit.
|
|
|
|
*/
|
2019-10-28 20:37:12 +08:00
|
|
|
job->gather_copy_mapped = dma_alloc_wc(host, size, &job->gather_copy,
|
2017-06-15 07:18:43 +08:00
|
|
|
GFP_NOWAIT);
|
|
|
|
|
|
|
|
/* the higher priority allocation failed, try the generic-blocking */
|
|
|
|
if (!job->gather_copy_mapped)
|
2019-10-28 20:37:12 +08:00
|
|
|
job->gather_copy_mapped = dma_alloc_wc(host, size,
|
2017-06-15 07:18:43 +08:00
|
|
|
&job->gather_copy,
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (!job->gather_copy_mapped)
|
2013-08-23 18:19:11 +08:00
|
|
|
return -ENOMEM;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
job->gather_copy_size = size;
|
|
|
|
|
|
|
|
for (i = 0; i < job->num_gathers; i++) {
|
|
|
|
struct host1x_job_gather *g = &job->gathers[i];
|
|
|
|
void *gather;
|
|
|
|
|
2013-05-29 18:26:05 +08:00
|
|
|
/* Copy the gather */
|
2013-03-22 22:34:03 +08:00
|
|
|
gather = host1x_bo_mmap(g->bo);
|
|
|
|
memcpy(job->gather_copy_mapped + offset, gather + g->offset,
|
|
|
|
g->words * sizeof(u32));
|
|
|
|
host1x_bo_munmap(g->bo, gather);
|
|
|
|
|
2013-05-29 18:26:05 +08:00
|
|
|
/* Store the location in the buffer */
|
2013-03-22 22:34:03 +08:00
|
|
|
g->base = job->gather_copy;
|
|
|
|
g->offset = offset;
|
2013-05-29 18:26:05 +08:00
|
|
|
|
|
|
|
/* Validate the job */
|
|
|
|
if (validate(&fw, g))
|
|
|
|
return -EINVAL;
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
offset += g->words * sizeof(u32);
|
|
|
|
}
|
|
|
|
|
2018-05-05 14:45:47 +08:00
|
|
|
/* No relocs should remain at this point */
|
|
|
|
if (fw.num_relocs)
|
2013-10-05 04:18:33 +08:00
|
|
|
return -EINVAL;
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int host1x_job_pin(struct host1x_job *job, struct device *dev)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
unsigned int i, j;
|
|
|
|
struct host1x *host = dev_get_drvdata(dev->parent);
|
|
|
|
|
|
|
|
/* pin memory */
|
2016-12-14 19:16:14 +08:00
|
|
|
err = pin_job(host, job);
|
|
|
|
if (err)
|
2013-03-22 22:34:03 +08:00
|
|
|
goto out;
|
|
|
|
|
2017-06-15 07:18:34 +08:00
|
|
|
if (IS_ENABLED(CONFIG_TEGRA_HOST1X_FIREWALL)) {
|
2019-10-28 20:37:12 +08:00
|
|
|
err = copy_gathers(host->dev, job, dev);
|
2017-06-15 07:18:34 +08:00
|
|
|
if (err)
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
/* patch gathers */
|
|
|
|
for (i = 0; i < job->num_gathers; i++) {
|
|
|
|
struct host1x_job_gather *g = &job->gathers[i];
|
|
|
|
|
|
|
|
/* process each gather mem only once */
|
|
|
|
if (g->handled)
|
|
|
|
continue;
|
|
|
|
|
2017-06-15 07:18:34 +08:00
|
|
|
/* copy_gathers() sets gathers base if firewall is enabled */
|
|
|
|
if (!IS_ENABLED(CONFIG_TEGRA_HOST1X_FIREWALL))
|
|
|
|
g->base = job->gather_addr_phys[i];
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2016-11-09 01:51:32 +08:00
|
|
|
for (j = i + 1; j < job->num_gathers; j++) {
|
|
|
|
if (job->gathers[j].bo == g->bo) {
|
2013-03-22 22:34:03 +08:00
|
|
|
job->gathers[j].handled = true;
|
2016-11-09 01:51:32 +08:00
|
|
|
job->gathers[j].base = g->base;
|
|
|
|
}
|
|
|
|
}
|
2013-03-22 22:34:03 +08:00
|
|
|
|
2017-06-15 07:18:34 +08:00
|
|
|
err = do_relocs(job, g);
|
2013-03-22 22:34:03 +08:00
|
|
|
if (err)
|
2017-06-15 07:18:34 +08:00
|
|
|
break;
|
2013-03-22 22:34:03 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
2017-06-15 07:18:33 +08:00
|
|
|
if (err)
|
|
|
|
host1x_job_unpin(job);
|
2013-03-22 22:34:03 +08:00
|
|
|
wmb();
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
2013-11-08 18:41:42 +08:00
|
|
|
EXPORT_SYMBOL(host1x_job_pin);
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
void host1x_job_unpin(struct host1x_job *job)
|
|
|
|
{
|
2016-12-14 19:16:14 +08:00
|
|
|
struct host1x *host = dev_get_drvdata(job->channel->dev->parent);
|
2013-03-22 22:34:03 +08:00
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
for (i = 0; i < job->num_unpins; i++) {
|
|
|
|
struct host1x_job_unpin_data *unpin = &job->unpins[i];
|
2019-10-28 20:37:13 +08:00
|
|
|
struct device *dev = unpin->dev ?: host->dev;
|
|
|
|
struct sg_table *sgt = unpin->sgt;
|
2016-06-23 17:33:31 +08:00
|
|
|
|
2018-07-07 02:02:36 +08:00
|
|
|
if (!IS_ENABLED(CONFIG_TEGRA_HOST1X_FIREWALL) &&
|
|
|
|
unpin->size && host->domain) {
|
2016-12-14 19:16:14 +08:00
|
|
|
iommu_unmap(host->domain, job->addr_phys[i],
|
|
|
|
unpin->size);
|
|
|
|
free_iova(&host->iova,
|
|
|
|
iova_pfn(&host->iova, job->addr_phys[i]));
|
|
|
|
}
|
|
|
|
|
2019-10-28 20:37:13 +08:00
|
|
|
if (unpin->dev && sgt)
|
2020-04-28 19:11:16 +08:00
|
|
|
dma_unmap_sgtable(unpin->dev, sgt, unpin->dir, 0);
|
2019-10-28 20:37:13 +08:00
|
|
|
|
|
|
|
host1x_bo_unpin(dev, unpin->bo, sgt);
|
2013-03-22 22:34:03 +08:00
|
|
|
host1x_bo_put(unpin->bo);
|
|
|
|
}
|
2016-06-23 17:35:50 +08:00
|
|
|
|
2013-03-22 22:34:03 +08:00
|
|
|
job->num_unpins = 0;
|
|
|
|
|
|
|
|
if (job->gather_copy_size)
|
2019-10-28 20:37:12 +08:00
|
|
|
dma_free_wc(host->dev, job->gather_copy_size,
|
2016-06-23 17:35:50 +08:00
|
|
|
job->gather_copy_mapped, job->gather_copy);
|
2013-03-22 22:34:03 +08:00
|
|
|
}
|
2013-11-08 18:41:42 +08:00
|
|
|
EXPORT_SYMBOL(host1x_job_unpin);
|
2013-03-22 22:34:03 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Debug routine used to dump job entries
|
|
|
|
*/
|
|
|
|
void host1x_job_dump(struct device *dev, struct host1x_job *job)
|
|
|
|
{
|
2021-03-29 21:38:32 +08:00
|
|
|
dev_dbg(dev, " SYNCPT_ID %d\n", job->syncpt->id);
|
2013-03-22 22:34:03 +08:00
|
|
|
dev_dbg(dev, " SYNCPT_VAL %d\n", job->syncpt_end);
|
|
|
|
dev_dbg(dev, " FIRST_GET 0x%x\n", job->first_get);
|
|
|
|
dev_dbg(dev, " TIMEOUT %d\n", job->timeout);
|
|
|
|
dev_dbg(dev, " NUM_SLOTS %d\n", job->num_slots);
|
|
|
|
dev_dbg(dev, " NUM_HANDLES %d\n", job->num_unpins);
|
|
|
|
}
|