mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-21 11:44:01 +08:00
crypto: caam - add support for givencrypt cbc(aes) and rfc3686(ctr(aes))
Add support for one-shot givencrypt algorithms. Givencrypt algorithms will generate their IV and encrypt data within the same shared job descriptors. Current algorithms merged from ablkcipher to givencrypt are: - AES Cipher Block Chaining (CBC) - AES Counter Mode (CTR) compliant with RFC3686 Signed-off-by: Catalin Vasile <catalin.vasile@freescale.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
daebc46585
commit
7222d1a341
@ -1835,6 +1835,7 @@ static int ablkcipher_setkey(struct crypto_ablkcipher *ablkcipher,
|
|||||||
u32 *key_jump_cmd;
|
u32 *key_jump_cmd;
|
||||||
u32 *desc;
|
u32 *desc;
|
||||||
u32 *nonce;
|
u32 *nonce;
|
||||||
|
u32 geniv;
|
||||||
u32 ctx1_iv_off = 0;
|
u32 ctx1_iv_off = 0;
|
||||||
const bool ctr_mode = ((ctx->class1_alg_type & OP_ALG_AAI_MASK) ==
|
const bool ctr_mode = ((ctx->class1_alg_type & OP_ALG_AAI_MASK) ==
|
||||||
OP_ALG_AAI_CTR_MOD128);
|
OP_ALG_AAI_CTR_MOD128);
|
||||||
@ -1993,6 +1994,83 @@ static int ablkcipher_setkey(struct crypto_ablkcipher *ablkcipher,
|
|||||||
DUMP_PREFIX_ADDRESS, 16, 4, desc,
|
DUMP_PREFIX_ADDRESS, 16, 4, desc,
|
||||||
desc_bytes(desc), 1);
|
desc_bytes(desc), 1);
|
||||||
#endif
|
#endif
|
||||||
|
/* ablkcipher_givencrypt shared descriptor */
|
||||||
|
desc = ctx->sh_desc_givenc;
|
||||||
|
|
||||||
|
init_sh_desc(desc, HDR_SHARE_SERIAL | HDR_SAVECTX);
|
||||||
|
/* Skip if already shared */
|
||||||
|
key_jump_cmd = append_jump(desc, JUMP_JSL | JUMP_TEST_ALL |
|
||||||
|
JUMP_COND_SHRD);
|
||||||
|
|
||||||
|
/* Load class1 key only */
|
||||||
|
append_key_as_imm(desc, (void *)ctx->key, ctx->enckeylen,
|
||||||
|
ctx->enckeylen, CLASS_1 |
|
||||||
|
KEY_DEST_CLASS_REG);
|
||||||
|
|
||||||
|
/* Load Nonce into CONTEXT1 reg */
|
||||||
|
if (is_rfc3686) {
|
||||||
|
nonce = (u32 *)(key + keylen);
|
||||||
|
append_load_imm_u32(desc, *nonce, LDST_CLASS_IND_CCB |
|
||||||
|
LDST_SRCDST_BYTE_OUTFIFO | LDST_IMM);
|
||||||
|
append_move(desc, MOVE_WAITCOMP |
|
||||||
|
MOVE_SRC_OUTFIFO |
|
||||||
|
MOVE_DEST_CLASS1CTX |
|
||||||
|
(16 << MOVE_OFFSET_SHIFT) |
|
||||||
|
(CTR_RFC3686_NONCE_SIZE << MOVE_LEN_SHIFT));
|
||||||
|
}
|
||||||
|
set_jump_tgt_here(desc, key_jump_cmd);
|
||||||
|
|
||||||
|
/* Generate IV */
|
||||||
|
geniv = NFIFOENTRY_STYPE_PAD | NFIFOENTRY_DEST_DECO |
|
||||||
|
NFIFOENTRY_DTYPE_MSG | NFIFOENTRY_LC1 |
|
||||||
|
NFIFOENTRY_PTYPE_RND | (crt->ivsize << NFIFOENTRY_DLEN_SHIFT);
|
||||||
|
append_load_imm_u32(desc, geniv, LDST_CLASS_IND_CCB |
|
||||||
|
LDST_SRCDST_WORD_INFO_FIFO | LDST_IMM);
|
||||||
|
append_cmd(desc, CMD_LOAD | DISABLE_AUTO_INFO_FIFO);
|
||||||
|
append_move(desc, MOVE_WAITCOMP |
|
||||||
|
MOVE_SRC_INFIFO |
|
||||||
|
MOVE_DEST_CLASS1CTX |
|
||||||
|
(crt->ivsize << MOVE_LEN_SHIFT) |
|
||||||
|
(ctx1_iv_off << MOVE_OFFSET_SHIFT));
|
||||||
|
append_cmd(desc, CMD_LOAD | ENABLE_AUTO_INFO_FIFO);
|
||||||
|
|
||||||
|
/* Copy generated IV to memory */
|
||||||
|
append_seq_store(desc, crt->ivsize,
|
||||||
|
LDST_SRCDST_BYTE_CONTEXT | LDST_CLASS_1_CCB |
|
||||||
|
(ctx1_iv_off << LDST_OFFSET_SHIFT));
|
||||||
|
|
||||||
|
/* Load Counter into CONTEXT1 reg */
|
||||||
|
if (is_rfc3686)
|
||||||
|
append_load_imm_u32(desc, (u32)1, LDST_IMM |
|
||||||
|
LDST_CLASS_1_CCB |
|
||||||
|
LDST_SRCDST_BYTE_CONTEXT |
|
||||||
|
((ctx1_iv_off + CTR_RFC3686_IV_SIZE) <<
|
||||||
|
LDST_OFFSET_SHIFT));
|
||||||
|
|
||||||
|
if (ctx1_iv_off)
|
||||||
|
append_jump(desc, JUMP_JSL | JUMP_TEST_ALL | JUMP_COND_NCP |
|
||||||
|
(1 << JUMP_OFFSET_SHIFT));
|
||||||
|
|
||||||
|
/* Load operation */
|
||||||
|
append_operation(desc, ctx->class1_alg_type |
|
||||||
|
OP_ALG_AS_INITFINAL | OP_ALG_ENCRYPT);
|
||||||
|
|
||||||
|
/* Perform operation */
|
||||||
|
ablkcipher_append_src_dst(desc);
|
||||||
|
|
||||||
|
ctx->sh_desc_givenc_dma = dma_map_single(jrdev, desc,
|
||||||
|
desc_bytes(desc),
|
||||||
|
DMA_TO_DEVICE);
|
||||||
|
if (dma_mapping_error(jrdev, ctx->sh_desc_givenc_dma)) {
|
||||||
|
dev_err(jrdev, "unable to map shared descriptor\n");
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
#ifdef DEBUG
|
||||||
|
print_hex_dump(KERN_ERR,
|
||||||
|
"ablkcipher givenc shdesc@" __stringify(__LINE__) ": ",
|
||||||
|
DUMP_PREFIX_ADDRESS, 16, 4, desc,
|
||||||
|
desc_bytes(desc), 1);
|
||||||
|
#endif
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
@ -2479,6 +2557,54 @@ static void init_ablkcipher_job(u32 *sh_desc, dma_addr_t ptr,
|
|||||||
append_seq_out_ptr(desc, dst_dma, req->nbytes, out_options);
|
append_seq_out_ptr(desc, dst_dma, req->nbytes, out_options);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Fill in ablkcipher givencrypt job descriptor
|
||||||
|
*/
|
||||||
|
static void init_ablkcipher_giv_job(u32 *sh_desc, dma_addr_t ptr,
|
||||||
|
struct ablkcipher_edesc *edesc,
|
||||||
|
struct ablkcipher_request *req,
|
||||||
|
bool iv_contig)
|
||||||
|
{
|
||||||
|
struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
|
||||||
|
int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
|
||||||
|
u32 *desc = edesc->hw_desc;
|
||||||
|
u32 out_options, in_options;
|
||||||
|
dma_addr_t dst_dma, src_dma;
|
||||||
|
int len, sec4_sg_index = 0;
|
||||||
|
|
||||||
|
#ifdef DEBUG
|
||||||
|
print_hex_dump(KERN_ERR, "presciv@" __stringify(__LINE__) ": ",
|
||||||
|
DUMP_PREFIX_ADDRESS, 16, 4, req->info,
|
||||||
|
ivsize, 1);
|
||||||
|
print_hex_dump(KERN_ERR, "src @" __stringify(__LINE__) ": ",
|
||||||
|
DUMP_PREFIX_ADDRESS, 16, 4, sg_virt(req->src),
|
||||||
|
edesc->src_nents ? 100 : req->nbytes, 1);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
len = desc_len(sh_desc);
|
||||||
|
init_job_desc_shared(desc, ptr, len, HDR_SHARE_DEFER | HDR_REVERSE);
|
||||||
|
|
||||||
|
if (!edesc->src_nents) {
|
||||||
|
src_dma = sg_dma_address(req->src);
|
||||||
|
in_options = 0;
|
||||||
|
} else {
|
||||||
|
src_dma = edesc->sec4_sg_dma;
|
||||||
|
sec4_sg_index += edesc->src_nents;
|
||||||
|
in_options = LDST_SGF;
|
||||||
|
}
|
||||||
|
append_seq_in_ptr(desc, src_dma, req->nbytes, in_options);
|
||||||
|
|
||||||
|
if (iv_contig) {
|
||||||
|
dst_dma = edesc->iv_dma;
|
||||||
|
out_options = 0;
|
||||||
|
} else {
|
||||||
|
dst_dma = edesc->sec4_sg_dma +
|
||||||
|
sec4_sg_index * sizeof(struct sec4_sg_entry);
|
||||||
|
out_options = LDST_SGF;
|
||||||
|
}
|
||||||
|
append_seq_out_ptr(desc, dst_dma, req->nbytes + ivsize, out_options);
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* allocate and map the aead extended descriptor
|
* allocate and map the aead extended descriptor
|
||||||
*/
|
*/
|
||||||
@ -3099,6 +3225,151 @@ static int ablkcipher_decrypt(struct ablkcipher_request *req)
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* allocate and map the ablkcipher extended descriptor
|
||||||
|
* for ablkcipher givencrypt
|
||||||
|
*/
|
||||||
|
static struct ablkcipher_edesc *ablkcipher_giv_edesc_alloc(
|
||||||
|
struct skcipher_givcrypt_request *greq,
|
||||||
|
int desc_bytes,
|
||||||
|
bool *iv_contig_out)
|
||||||
|
{
|
||||||
|
struct ablkcipher_request *req = &greq->creq;
|
||||||
|
struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
|
||||||
|
struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
|
||||||
|
struct device *jrdev = ctx->jrdev;
|
||||||
|
gfp_t flags = (req->base.flags & (CRYPTO_TFM_REQ_MAY_BACKLOG |
|
||||||
|
CRYPTO_TFM_REQ_MAY_SLEEP)) ?
|
||||||
|
GFP_KERNEL : GFP_ATOMIC;
|
||||||
|
int src_nents, dst_nents = 0, sec4_sg_bytes;
|
||||||
|
struct ablkcipher_edesc *edesc;
|
||||||
|
dma_addr_t iv_dma = 0;
|
||||||
|
bool iv_contig = false;
|
||||||
|
int sgc;
|
||||||
|
int ivsize = crypto_ablkcipher_ivsize(ablkcipher);
|
||||||
|
bool src_chained = false, dst_chained = false;
|
||||||
|
int sec4_sg_index;
|
||||||
|
|
||||||
|
src_nents = sg_count(req->src, req->nbytes, &src_chained);
|
||||||
|
|
||||||
|
if (unlikely(req->dst != req->src))
|
||||||
|
dst_nents = sg_count(req->dst, req->nbytes, &dst_chained);
|
||||||
|
|
||||||
|
if (likely(req->src == req->dst)) {
|
||||||
|
sgc = dma_map_sg_chained(jrdev, req->src, src_nents ? : 1,
|
||||||
|
DMA_BIDIRECTIONAL, src_chained);
|
||||||
|
} else {
|
||||||
|
sgc = dma_map_sg_chained(jrdev, req->src, src_nents ? : 1,
|
||||||
|
DMA_TO_DEVICE, src_chained);
|
||||||
|
sgc = dma_map_sg_chained(jrdev, req->dst, dst_nents ? : 1,
|
||||||
|
DMA_FROM_DEVICE, dst_chained);
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Check if iv can be contiguous with source and destination.
|
||||||
|
* If so, include it. If not, create scatterlist.
|
||||||
|
*/
|
||||||
|
iv_dma = dma_map_single(jrdev, greq->giv, ivsize, DMA_TO_DEVICE);
|
||||||
|
if (dma_mapping_error(jrdev, iv_dma)) {
|
||||||
|
dev_err(jrdev, "unable to map IV\n");
|
||||||
|
return ERR_PTR(-ENOMEM);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!dst_nents && iv_dma + ivsize == sg_dma_address(req->dst))
|
||||||
|
iv_contig = true;
|
||||||
|
else
|
||||||
|
dst_nents = dst_nents ? : 1;
|
||||||
|
sec4_sg_bytes = ((iv_contig ? 0 : 1) + src_nents + dst_nents) *
|
||||||
|
sizeof(struct sec4_sg_entry);
|
||||||
|
|
||||||
|
/* allocate space for base edesc and hw desc commands, link tables */
|
||||||
|
edesc = kmalloc(sizeof(*edesc) + desc_bytes +
|
||||||
|
sec4_sg_bytes, GFP_DMA | flags);
|
||||||
|
if (!edesc) {
|
||||||
|
dev_err(jrdev, "could not allocate extended descriptor\n");
|
||||||
|
return ERR_PTR(-ENOMEM);
|
||||||
|
}
|
||||||
|
|
||||||
|
edesc->src_nents = src_nents;
|
||||||
|
edesc->src_chained = src_chained;
|
||||||
|
edesc->dst_nents = dst_nents;
|
||||||
|
edesc->dst_chained = dst_chained;
|
||||||
|
edesc->sec4_sg_bytes = sec4_sg_bytes;
|
||||||
|
edesc->sec4_sg = (void *)edesc + sizeof(struct ablkcipher_edesc) +
|
||||||
|
desc_bytes;
|
||||||
|
|
||||||
|
sec4_sg_index = 0;
|
||||||
|
if (src_nents) {
|
||||||
|
sg_to_sec4_sg_last(req->src, src_nents, edesc->sec4_sg, 0);
|
||||||
|
sec4_sg_index += src_nents;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!iv_contig) {
|
||||||
|
dma_to_sec4_sg_one(edesc->sec4_sg + sec4_sg_index,
|
||||||
|
iv_dma, ivsize, 0);
|
||||||
|
sec4_sg_index += 1;
|
||||||
|
sg_to_sec4_sg_last(req->dst, dst_nents,
|
||||||
|
edesc->sec4_sg + sec4_sg_index, 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
edesc->sec4_sg_dma = dma_map_single(jrdev, edesc->sec4_sg,
|
||||||
|
sec4_sg_bytes, DMA_TO_DEVICE);
|
||||||
|
if (dma_mapping_error(jrdev, edesc->sec4_sg_dma)) {
|
||||||
|
dev_err(jrdev, "unable to map S/G table\n");
|
||||||
|
return ERR_PTR(-ENOMEM);
|
||||||
|
}
|
||||||
|
edesc->iv_dma = iv_dma;
|
||||||
|
|
||||||
|
#ifdef DEBUG
|
||||||
|
print_hex_dump(KERN_ERR,
|
||||||
|
"ablkcipher sec4_sg@" __stringify(__LINE__) ": ",
|
||||||
|
DUMP_PREFIX_ADDRESS, 16, 4, edesc->sec4_sg,
|
||||||
|
sec4_sg_bytes, 1);
|
||||||
|
#endif
|
||||||
|
|
||||||
|
*iv_contig_out = iv_contig;
|
||||||
|
return edesc;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int ablkcipher_givencrypt(struct skcipher_givcrypt_request *creq)
|
||||||
|
{
|
||||||
|
struct ablkcipher_request *req = &creq->creq;
|
||||||
|
struct ablkcipher_edesc *edesc;
|
||||||
|
struct crypto_ablkcipher *ablkcipher = crypto_ablkcipher_reqtfm(req);
|
||||||
|
struct caam_ctx *ctx = crypto_ablkcipher_ctx(ablkcipher);
|
||||||
|
struct device *jrdev = ctx->jrdev;
|
||||||
|
bool iv_contig;
|
||||||
|
u32 *desc;
|
||||||
|
int ret = 0;
|
||||||
|
|
||||||
|
/* allocate extended descriptor */
|
||||||
|
edesc = ablkcipher_giv_edesc_alloc(creq, DESC_JOB_IO_LEN *
|
||||||
|
CAAM_CMD_SZ, &iv_contig);
|
||||||
|
if (IS_ERR(edesc))
|
||||||
|
return PTR_ERR(edesc);
|
||||||
|
|
||||||
|
/* Create and submit job descriptor*/
|
||||||
|
init_ablkcipher_giv_job(ctx->sh_desc_givenc, ctx->sh_desc_givenc_dma,
|
||||||
|
edesc, req, iv_contig);
|
||||||
|
#ifdef DEBUG
|
||||||
|
print_hex_dump(KERN_ERR,
|
||||||
|
"ablkcipher jobdesc@" __stringify(__LINE__) ": ",
|
||||||
|
DUMP_PREFIX_ADDRESS, 16, 4, edesc->hw_desc,
|
||||||
|
desc_bytes(edesc->hw_desc), 1);
|
||||||
|
#endif
|
||||||
|
desc = edesc->hw_desc;
|
||||||
|
ret = caam_jr_enqueue(jrdev, desc, ablkcipher_encrypt_done, req);
|
||||||
|
|
||||||
|
if (!ret) {
|
||||||
|
ret = -EINPROGRESS;
|
||||||
|
} else {
|
||||||
|
ablkcipher_unmap(jrdev, edesc, req);
|
||||||
|
kfree(edesc);
|
||||||
|
}
|
||||||
|
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
#define template_aead template_u.aead
|
#define template_aead template_u.aead
|
||||||
#define template_ablkcipher template_u.ablkcipher
|
#define template_ablkcipher template_u.ablkcipher
|
||||||
struct caam_alg_template {
|
struct caam_alg_template {
|
||||||
@ -3769,12 +4040,13 @@ static struct caam_alg_template driver_algs[] = {
|
|||||||
.name = "cbc(aes)",
|
.name = "cbc(aes)",
|
||||||
.driver_name = "cbc-aes-caam",
|
.driver_name = "cbc-aes-caam",
|
||||||
.blocksize = AES_BLOCK_SIZE,
|
.blocksize = AES_BLOCK_SIZE,
|
||||||
.type = CRYPTO_ALG_TYPE_ABLKCIPHER,
|
.type = CRYPTO_ALG_TYPE_GIVCIPHER,
|
||||||
.template_ablkcipher = {
|
.template_ablkcipher = {
|
||||||
.setkey = ablkcipher_setkey,
|
.setkey = ablkcipher_setkey,
|
||||||
.encrypt = ablkcipher_encrypt,
|
.encrypt = ablkcipher_encrypt,
|
||||||
.decrypt = ablkcipher_decrypt,
|
.decrypt = ablkcipher_decrypt,
|
||||||
.geniv = "eseqiv",
|
.givencrypt = ablkcipher_givencrypt,
|
||||||
|
.geniv = "<built-in>",
|
||||||
.min_keysize = AES_MIN_KEY_SIZE,
|
.min_keysize = AES_MIN_KEY_SIZE,
|
||||||
.max_keysize = AES_MAX_KEY_SIZE,
|
.max_keysize = AES_MAX_KEY_SIZE,
|
||||||
.ivsize = AES_BLOCK_SIZE,
|
.ivsize = AES_BLOCK_SIZE,
|
||||||
@ -3833,12 +4105,13 @@ static struct caam_alg_template driver_algs[] = {
|
|||||||
.name = "rfc3686(ctr(aes))",
|
.name = "rfc3686(ctr(aes))",
|
||||||
.driver_name = "rfc3686-ctr-aes-caam",
|
.driver_name = "rfc3686-ctr-aes-caam",
|
||||||
.blocksize = 1,
|
.blocksize = 1,
|
||||||
.type = CRYPTO_ALG_TYPE_ABLKCIPHER,
|
.type = CRYPTO_ALG_TYPE_GIVCIPHER,
|
||||||
.template_ablkcipher = {
|
.template_ablkcipher = {
|
||||||
.setkey = ablkcipher_setkey,
|
.setkey = ablkcipher_setkey,
|
||||||
.encrypt = ablkcipher_encrypt,
|
.encrypt = ablkcipher_encrypt,
|
||||||
.decrypt = ablkcipher_decrypt,
|
.decrypt = ablkcipher_decrypt,
|
||||||
.geniv = "seqiv",
|
.givencrypt = ablkcipher_givencrypt,
|
||||||
|
.geniv = "<built-in>",
|
||||||
.min_keysize = AES_MIN_KEY_SIZE +
|
.min_keysize = AES_MIN_KEY_SIZE +
|
||||||
CTR_RFC3686_NONCE_SIZE,
|
CTR_RFC3686_NONCE_SIZE,
|
||||||
.max_keysize = AES_MAX_KEY_SIZE +
|
.max_keysize = AES_MAX_KEY_SIZE +
|
||||||
@ -3946,6 +4219,10 @@ static struct caam_crypto_alg *caam_alg_alloc(struct caam_alg_template
|
|||||||
alg->cra_flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_KERN_DRIVER_ONLY |
|
alg->cra_flags = CRYPTO_ALG_ASYNC | CRYPTO_ALG_KERN_DRIVER_ONLY |
|
||||||
template->type;
|
template->type;
|
||||||
switch (template->type) {
|
switch (template->type) {
|
||||||
|
case CRYPTO_ALG_TYPE_GIVCIPHER:
|
||||||
|
alg->cra_type = &crypto_givcipher_type;
|
||||||
|
alg->cra_ablkcipher = template->template_ablkcipher;
|
||||||
|
break;
|
||||||
case CRYPTO_ALG_TYPE_ABLKCIPHER:
|
case CRYPTO_ALG_TYPE_ABLKCIPHER:
|
||||||
alg->cra_type = &crypto_ablkcipher_type;
|
alg->cra_type = &crypto_ablkcipher_type;
|
||||||
alg->cra_ablkcipher = template->template_ablkcipher;
|
alg->cra_ablkcipher = template->template_ablkcipher;
|
||||||
|
Loading…
Reference in New Issue
Block a user