mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-17 09:14:19 +08:00
bd03b0214a
On systems with coherence issues, packet processed could succeed while it should have failed, e.g. because of an authentication fail. This is because the driver would read stale status information that had all error bits initialised to zero = no error. Since this is potential a security risk, we want to prevent it from being a possibility at all. So initialize all error bits to error state, so that reading stale status information will always result in errors. Signed-off-by: Pascal van Leeuwen <pvanleeuwen@rambus.com> Acked-by: Antoine Tenart <antoine.tenart@bootlin.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
255 lines
6.6 KiB
C
255 lines
6.6 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (C) 2017 Marvell
|
|
*
|
|
* Antoine Tenart <antoine.tenart@free-electrons.com>
|
|
*/
|
|
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/spinlock.h>
|
|
|
|
#include "safexcel.h"
|
|
|
|
int safexcel_init_ring_descriptors(struct safexcel_crypto_priv *priv,
|
|
struct safexcel_desc_ring *cdr,
|
|
struct safexcel_desc_ring *rdr)
|
|
{
|
|
int i;
|
|
struct safexcel_command_desc *cdesc;
|
|
dma_addr_t atok;
|
|
|
|
/* Actual command descriptor ring */
|
|
cdr->offset = priv->config.cd_offset;
|
|
cdr->base = dmam_alloc_coherent(priv->dev,
|
|
cdr->offset * EIP197_DEFAULT_RING_SIZE,
|
|
&cdr->base_dma, GFP_KERNEL);
|
|
if (!cdr->base)
|
|
return -ENOMEM;
|
|
cdr->write = cdr->base;
|
|
cdr->base_end = cdr->base + cdr->offset * (EIP197_DEFAULT_RING_SIZE - 1);
|
|
cdr->read = cdr->base;
|
|
|
|
/* Command descriptor shadow ring for storing additional token data */
|
|
cdr->shoffset = priv->config.cdsh_offset;
|
|
cdr->shbase = dmam_alloc_coherent(priv->dev,
|
|
cdr->shoffset *
|
|
EIP197_DEFAULT_RING_SIZE,
|
|
&cdr->shbase_dma, GFP_KERNEL);
|
|
if (!cdr->shbase)
|
|
return -ENOMEM;
|
|
cdr->shwrite = cdr->shbase;
|
|
cdr->shbase_end = cdr->shbase + cdr->shoffset *
|
|
(EIP197_DEFAULT_RING_SIZE - 1);
|
|
|
|
/*
|
|
* Populate command descriptors with physical pointers to shadow descs.
|
|
* Note that we only need to do this once if we don't overwrite them.
|
|
*/
|
|
cdesc = cdr->base;
|
|
atok = cdr->shbase_dma;
|
|
for (i = 0; i < EIP197_DEFAULT_RING_SIZE; i++) {
|
|
cdesc->atok_lo = lower_32_bits(atok);
|
|
cdesc->atok_hi = upper_32_bits(atok);
|
|
cdesc = (void *)cdesc + cdr->offset;
|
|
atok += cdr->shoffset;
|
|
}
|
|
|
|
rdr->offset = priv->config.rd_offset;
|
|
/* Use shoffset for result token offset here */
|
|
rdr->shoffset = priv->config.res_offset;
|
|
rdr->base = dmam_alloc_coherent(priv->dev,
|
|
rdr->offset * EIP197_DEFAULT_RING_SIZE,
|
|
&rdr->base_dma, GFP_KERNEL);
|
|
if (!rdr->base)
|
|
return -ENOMEM;
|
|
rdr->write = rdr->base;
|
|
rdr->base_end = rdr->base + rdr->offset * (EIP197_DEFAULT_RING_SIZE - 1);
|
|
rdr->read = rdr->base;
|
|
|
|
return 0;
|
|
}
|
|
|
|
inline int safexcel_select_ring(struct safexcel_crypto_priv *priv)
|
|
{
|
|
return (atomic_inc_return(&priv->ring_used) % priv->config.rings);
|
|
}
|
|
|
|
static void *safexcel_ring_next_cwptr(struct safexcel_crypto_priv *priv,
|
|
struct safexcel_desc_ring *ring,
|
|
bool first,
|
|
struct safexcel_token **atoken)
|
|
{
|
|
void *ptr = ring->write;
|
|
|
|
if (first)
|
|
*atoken = ring->shwrite;
|
|
|
|
if ((ring->write == ring->read - ring->offset) ||
|
|
(ring->read == ring->base && ring->write == ring->base_end))
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
if (ring->write == ring->base_end) {
|
|
ring->write = ring->base;
|
|
ring->shwrite = ring->shbase;
|
|
} else {
|
|
ring->write += ring->offset;
|
|
ring->shwrite += ring->shoffset;
|
|
}
|
|
|
|
return ptr;
|
|
}
|
|
|
|
static void *safexcel_ring_next_rwptr(struct safexcel_crypto_priv *priv,
|
|
struct safexcel_desc_ring *ring,
|
|
struct result_data_desc **rtoken)
|
|
{
|
|
void *ptr = ring->write;
|
|
|
|
/* Result token at relative offset shoffset */
|
|
*rtoken = ring->write + ring->shoffset;
|
|
|
|
if ((ring->write == ring->read - ring->offset) ||
|
|
(ring->read == ring->base && ring->write == ring->base_end))
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
if (ring->write == ring->base_end)
|
|
ring->write = ring->base;
|
|
else
|
|
ring->write += ring->offset;
|
|
|
|
return ptr;
|
|
}
|
|
|
|
void *safexcel_ring_next_rptr(struct safexcel_crypto_priv *priv,
|
|
struct safexcel_desc_ring *ring)
|
|
{
|
|
void *ptr = ring->read;
|
|
|
|
if (ring->write == ring->read)
|
|
return ERR_PTR(-ENOENT);
|
|
|
|
if (ring->read == ring->base_end)
|
|
ring->read = ring->base;
|
|
else
|
|
ring->read += ring->offset;
|
|
|
|
return ptr;
|
|
}
|
|
|
|
inline void *safexcel_ring_curr_rptr(struct safexcel_crypto_priv *priv,
|
|
int ring)
|
|
{
|
|
struct safexcel_desc_ring *rdr = &priv->ring[ring].rdr;
|
|
|
|
return rdr->read;
|
|
}
|
|
|
|
inline int safexcel_ring_first_rdr_index(struct safexcel_crypto_priv *priv,
|
|
int ring)
|
|
{
|
|
struct safexcel_desc_ring *rdr = &priv->ring[ring].rdr;
|
|
|
|
return (rdr->read - rdr->base) / rdr->offset;
|
|
}
|
|
|
|
inline int safexcel_ring_rdr_rdesc_index(struct safexcel_crypto_priv *priv,
|
|
int ring,
|
|
struct safexcel_result_desc *rdesc)
|
|
{
|
|
struct safexcel_desc_ring *rdr = &priv->ring[ring].rdr;
|
|
|
|
return ((void *)rdesc - rdr->base) / rdr->offset;
|
|
}
|
|
|
|
void safexcel_ring_rollback_wptr(struct safexcel_crypto_priv *priv,
|
|
struct safexcel_desc_ring *ring)
|
|
{
|
|
if (ring->write == ring->read)
|
|
return;
|
|
|
|
if (ring->write == ring->base) {
|
|
ring->write = ring->base_end;
|
|
ring->shwrite = ring->shbase_end;
|
|
} else {
|
|
ring->write -= ring->offset;
|
|
ring->shwrite -= ring->shoffset;
|
|
}
|
|
}
|
|
|
|
struct safexcel_command_desc *safexcel_add_cdesc(struct safexcel_crypto_priv *priv,
|
|
int ring_id,
|
|
bool first, bool last,
|
|
dma_addr_t data, u32 data_len,
|
|
u32 full_data_len,
|
|
dma_addr_t context,
|
|
struct safexcel_token **atoken)
|
|
{
|
|
struct safexcel_command_desc *cdesc;
|
|
|
|
cdesc = safexcel_ring_next_cwptr(priv, &priv->ring[ring_id].cdr,
|
|
first, atoken);
|
|
if (IS_ERR(cdesc))
|
|
return cdesc;
|
|
|
|
cdesc->particle_size = data_len;
|
|
cdesc->rsvd0 = 0;
|
|
cdesc->last_seg = last;
|
|
cdesc->first_seg = first;
|
|
cdesc->additional_cdata_size = 0;
|
|
cdesc->rsvd1 = 0;
|
|
cdesc->data_lo = lower_32_bits(data);
|
|
cdesc->data_hi = upper_32_bits(data);
|
|
|
|
if (first) {
|
|
/*
|
|
* Note that the length here MUST be >0 or else the EIP(1)97
|
|
* may hang. Newer EIP197 firmware actually incorporates this
|
|
* fix already, but that doesn't help the EIP97 and we may
|
|
* also be running older firmware.
|
|
*/
|
|
cdesc->control_data.packet_length = full_data_len ?: 1;
|
|
cdesc->control_data.options = EIP197_OPTION_MAGIC_VALUE |
|
|
EIP197_OPTION_64BIT_CTX |
|
|
EIP197_OPTION_CTX_CTRL_IN_CMD |
|
|
EIP197_OPTION_RC_AUTO;
|
|
cdesc->control_data.type = EIP197_TYPE_BCLA;
|
|
cdesc->control_data.context_lo = lower_32_bits(context) |
|
|
EIP197_CONTEXT_SMALL;
|
|
cdesc->control_data.context_hi = upper_32_bits(context);
|
|
}
|
|
|
|
return cdesc;
|
|
}
|
|
|
|
struct safexcel_result_desc *safexcel_add_rdesc(struct safexcel_crypto_priv *priv,
|
|
int ring_id,
|
|
bool first, bool last,
|
|
dma_addr_t data, u32 len)
|
|
{
|
|
struct safexcel_result_desc *rdesc;
|
|
struct result_data_desc *rtoken;
|
|
|
|
rdesc = safexcel_ring_next_rwptr(priv, &priv->ring[ring_id].rdr,
|
|
&rtoken);
|
|
if (IS_ERR(rdesc))
|
|
return rdesc;
|
|
|
|
rdesc->particle_size = len;
|
|
rdesc->rsvd0 = 0;
|
|
rdesc->descriptor_overflow = 1; /* assume error */
|
|
rdesc->buffer_overflow = 1; /* assume error */
|
|
rdesc->last_seg = last;
|
|
rdesc->first_seg = first;
|
|
rdesc->result_size = EIP197_RD64_RESULT_SIZE;
|
|
rdesc->rsvd1 = 0;
|
|
rdesc->data_lo = lower_32_bits(data);
|
|
rdesc->data_hi = upper_32_bits(data);
|
|
|
|
/* Clear length in result token */
|
|
rtoken->packet_length = 0;
|
|
/* Assume errors - HW will clear if not the case */
|
|
rtoken->error_code = 0x7fff;
|
|
|
|
return rdesc;
|
|
}
|