2
0
mirror of https://github.com/edk2-porting/linux-next.git synced 2024-12-20 03:04:01 +08:00

crypto: hisilicon - Adjust some inner logic

1.Adjust call back function.
2.Adjust parameter checking function.

Signed-off-by: Zaibo Xu <xuzaibo@huawei.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
Zaibo Xu 2020-01-11 10:41:52 +08:00 committed by Herbert Xu
parent 7c7d902aa4
commit d6de2a5943

View File

@ -35,6 +35,8 @@
#define SEC_TOTAL_IV_SZ (SEC_IV_SIZE * QM_Q_DEPTH)
#define SEC_SGL_SGE_NR 128
#define SEC_CTX_DEV(ctx) (&(ctx)->sec->qm.pdev->dev)
#define SEC_SQE_CFLAG 2
#define SEC_SQE_DONE 0x1
static DEFINE_MUTEX(sec_algs_lock);
static unsigned int sec_active_devs;
@ -99,32 +101,34 @@ static void sec_req_cb(struct hisi_qp *qp, void *resp)
{
struct sec_qp_ctx *qp_ctx = qp->qp_ctx;
struct sec_sqe *bd = resp;
struct sec_ctx *ctx;
struct sec_req *req;
u16 done, flag;
u8 type;
struct sec_req *req;
type = bd->type_cipher_auth & SEC_TYPE_MASK;
if (type == SEC_BD_TYPE2) {
req = qp_ctx->req_list[le16_to_cpu(bd->type2.tag)];
req->err_type = bd->type2.error_type;
done = le16_to_cpu(bd->type2.done_flag) & SEC_DONE_MASK;
flag = (le16_to_cpu(bd->type2.done_flag) &
SEC_FLAG_MASK) >> SEC_FLAG_OFFSET;
if (req->err_type || done != 0x1 || flag != 0x2)
dev_err(SEC_CTX_DEV(req->ctx),
"err_type[%d],done[%d],flag[%d]\n",
req->err_type, done, flag);
} else {
if (type != SEC_BD_TYPE2) {
pr_err("err bd type [%d]\n", type);
return;
}
atomic64_inc(&req->ctx->sec->debug.dfx.recv_cnt);
req = qp_ctx->req_list[le16_to_cpu(bd->type2.tag)];
req->err_type = bd->type2.error_type;
ctx = req->ctx;
done = le16_to_cpu(bd->type2.done_flag) & SEC_DONE_MASK;
flag = (le16_to_cpu(bd->type2.done_flag) &
SEC_FLAG_MASK) >> SEC_FLAG_OFFSET;
if (req->err_type || done != SEC_SQE_DONE ||
flag != SEC_SQE_CFLAG)
dev_err(SEC_CTX_DEV(ctx),
"err_type[%d],done[%d],flag[%d]\n",
req->err_type, done, flag);
req->ctx->req_op->buf_unmap(req->ctx, req);
atomic64_inc(&ctx->sec->debug.dfx.recv_cnt);
req->ctx->req_op->callback(req->ctx, req);
ctx->req_op->buf_unmap(ctx, req);
ctx->req_op->callback(ctx, req);
}
static int sec_bd_send(struct sec_ctx *ctx, struct sec_req *req)
@ -545,9 +549,7 @@ static void sec_skcipher_copy_iv(struct sec_ctx *ctx, struct sec_req *req)
{
struct skcipher_request *sk_req = req->c_req.sk_req;
u8 *c_ivin = req->qp_ctx->res[req->req_id].c_ivin;
struct sec_cipher_req *c_req = &req->c_req;
c_req->c_len = sk_req->cryptlen;
memcpy(c_ivin, sk_req->iv, ctx->c_ctx.ivsize);
}
@ -728,17 +730,17 @@ static void sec_skcipher_ctx_exit(struct crypto_skcipher *tfm)
sec_skcipher_uninit(tfm);
}
static int sec_skcipher_param_check(struct sec_ctx *ctx,
struct skcipher_request *sk_req)
static int sec_skcipher_param_check(struct sec_ctx *ctx, struct sec_req *sreq)
{
u8 c_alg = ctx->c_ctx.c_alg;
struct skcipher_request *sk_req = sreq->c_req.sk_req;
struct device *dev = SEC_CTX_DEV(ctx);
u8 c_alg = ctx->c_ctx.c_alg;
if (!sk_req->src || !sk_req->dst) {
dev_err(dev, "skcipher input param error!\n");
return -EINVAL;
}
sreq->c_req.c_len = sk_req->cryptlen;
if (c_alg == SEC_CALG_3DES) {
if (sk_req->cryptlen & (DES3_EDE_BLOCK_SIZE - 1)) {
dev_err(dev, "skcipher 3des input length error!\n");
@ -767,14 +769,14 @@ static int sec_skcipher_crypto(struct skcipher_request *sk_req, bool encrypt)
if (!sk_req->cryptlen)
return 0;
ret = sec_skcipher_param_check(ctx, sk_req);
if (ret)
return ret;
req->c_req.sk_req = sk_req;
req->c_req.encrypt = encrypt;
req->ctx = ctx;
ret = sec_skcipher_param_check(ctx, req);
if (unlikely(ret))
return -EINVAL;
return ctx->req_op->process(ctx, req);
}