mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-30 16:13:54 +08:00
crypto: marvell - Move SRAM I/O operations to step functions
Currently the crypto requests were sent to engines sequentially. This commit moves the SRAM I/O operations from the prepare to the step functions. It provides flexibility for future works and allow to prepare a request while the engine is running. Signed-off-by: Romain Perier <romain.perier@free-electrons.com> Acked-by: Boris Brezillon <boris.brezillon@free-electrons.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
1bf6682cb3
commit
2786cee8e5
@ -89,6 +89,9 @@ static void mv_cesa_ablkcipher_std_step(struct ablkcipher_request *req)
|
||||
size_t len = min_t(size_t, req->nbytes - sreq->offset,
|
||||
CESA_SA_SRAM_PAYLOAD_SIZE);
|
||||
|
||||
mv_cesa_adjust_op(engine, &sreq->op);
|
||||
memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op));
|
||||
|
||||
len = sg_pcopy_to_buffer(req->src, creq->src_nents,
|
||||
engine->sram + CESA_SA_DATA_SRAM_OFFSET,
|
||||
len, sreq->offset);
|
||||
@ -177,12 +180,9 @@ mv_cesa_ablkcipher_std_prepare(struct ablkcipher_request *req)
|
||||
{
|
||||
struct mv_cesa_ablkcipher_req *creq = ablkcipher_request_ctx(req);
|
||||
struct mv_cesa_ablkcipher_std_req *sreq = &creq->std;
|
||||
struct mv_cesa_engine *engine = creq->base.engine;
|
||||
|
||||
sreq->size = 0;
|
||||
sreq->offset = 0;
|
||||
mv_cesa_adjust_op(engine, &sreq->op);
|
||||
memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op));
|
||||
}
|
||||
|
||||
static inline void mv_cesa_ablkcipher_prepare(struct crypto_async_request *req,
|
||||
|
@ -162,6 +162,15 @@ static void mv_cesa_ahash_std_step(struct ahash_request *req)
|
||||
unsigned int new_cache_ptr = 0;
|
||||
u32 frag_mode;
|
||||
size_t len;
|
||||
unsigned int digsize;
|
||||
int i;
|
||||
|
||||
mv_cesa_adjust_op(engine, &creq->op_tmpl);
|
||||
memcpy_toio(engine->sram, &creq->op_tmpl, sizeof(creq->op_tmpl));
|
||||
|
||||
digsize = crypto_ahash_digestsize(crypto_ahash_reqtfm(req));
|
||||
for (i = 0; i < digsize / 4; i++)
|
||||
writel_relaxed(creq->state[i], engine->regs + CESA_IVDIG(i));
|
||||
|
||||
if (creq->cache_ptr)
|
||||
memcpy_toio(engine->sram + CESA_SA_DATA_SRAM_OFFSET,
|
||||
@ -265,11 +274,8 @@ static void mv_cesa_ahash_std_prepare(struct ahash_request *req)
|
||||
{
|
||||
struct mv_cesa_ahash_req *creq = ahash_request_ctx(req);
|
||||
struct mv_cesa_ahash_std_req *sreq = &creq->req.std;
|
||||
struct mv_cesa_engine *engine = creq->base.engine;
|
||||
|
||||
sreq->offset = 0;
|
||||
mv_cesa_adjust_op(engine, &creq->op_tmpl);
|
||||
memcpy_toio(engine->sram, &creq->op_tmpl, sizeof(creq->op_tmpl));
|
||||
}
|
||||
|
||||
static void mv_cesa_ahash_step(struct crypto_async_request *req)
|
||||
@ -336,8 +342,6 @@ static void mv_cesa_ahash_prepare(struct crypto_async_request *req,
|
||||
{
|
||||
struct ahash_request *ahashreq = ahash_request_cast(req);
|
||||
struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq);
|
||||
unsigned int digsize;
|
||||
int i;
|
||||
|
||||
creq->base.engine = engine;
|
||||
|
||||
@ -345,10 +349,6 @@ static void mv_cesa_ahash_prepare(struct crypto_async_request *req,
|
||||
mv_cesa_ahash_dma_prepare(ahashreq);
|
||||
else
|
||||
mv_cesa_ahash_std_prepare(ahashreq);
|
||||
|
||||
digsize = crypto_ahash_digestsize(crypto_ahash_reqtfm(ahashreq));
|
||||
for (i = 0; i < digsize / 4; i++)
|
||||
writel_relaxed(creq->state[i], engine->regs + CESA_IVDIG(i));
|
||||
}
|
||||
|
||||
static void mv_cesa_ahash_req_cleanup(struct crypto_async_request *req)
|
||||
|
Loading…
Reference in New Issue
Block a user