mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-14 15:54:15 +08:00
7d709af180
Use the frame_push and frame_pop macros to create the stack frames in the AES chaining mode wrappers so that they will get PAC and/or shadow call stack protection when configured. Signed-off-by: Ard Biesheuvel <ardb@kernel.org> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
867 lines
21 KiB
ArmAsm
867 lines
21 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* linux/arch/arm64/crypto/aes-modes.S - chaining mode wrappers for AES
|
|
*
|
|
* Copyright (C) 2013 - 2017 Linaro Ltd <ard.biesheuvel@linaro.org>
|
|
*/
|
|
|
|
/* included by aes-ce.S and aes-neon.S */
|
|
|
|
.text
|
|
.align 4
|
|
|
|
#ifndef MAX_STRIDE
|
|
#define MAX_STRIDE 4
|
|
#endif
|
|
|
|
#if MAX_STRIDE == 4
|
|
#define ST4(x...) x
|
|
#define ST5(x...)
|
|
#else
|
|
#define ST4(x...)
|
|
#define ST5(x...) x
|
|
#endif
|
|
|
|
SYM_FUNC_START_LOCAL(aes_encrypt_block4x)
|
|
encrypt_block4x v0, v1, v2, v3, w3, x2, x8, w7
|
|
ret
|
|
SYM_FUNC_END(aes_encrypt_block4x)
|
|
|
|
SYM_FUNC_START_LOCAL(aes_decrypt_block4x)
|
|
decrypt_block4x v0, v1, v2, v3, w3, x2, x8, w7
|
|
ret
|
|
SYM_FUNC_END(aes_decrypt_block4x)
|
|
|
|
#if MAX_STRIDE == 5
|
|
SYM_FUNC_START_LOCAL(aes_encrypt_block5x)
|
|
encrypt_block5x v0, v1, v2, v3, v4, w3, x2, x8, w7
|
|
ret
|
|
SYM_FUNC_END(aes_encrypt_block5x)
|
|
|
|
SYM_FUNC_START_LOCAL(aes_decrypt_block5x)
|
|
decrypt_block5x v0, v1, v2, v3, v4, w3, x2, x8, w7
|
|
ret
|
|
SYM_FUNC_END(aes_decrypt_block5x)
|
|
#endif
|
|
|
|
/*
|
|
* aes_ecb_encrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks)
|
|
* aes_ecb_decrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks)
|
|
*/
|
|
|
|
AES_FUNC_START(aes_ecb_encrypt)
|
|
frame_push 0
|
|
|
|
enc_prepare w3, x2, x5
|
|
|
|
.LecbencloopNx:
|
|
subs w4, w4, #MAX_STRIDE
|
|
bmi .Lecbenc1x
|
|
ld1 {v0.16b-v3.16b}, [x1], #64 /* get 4 pt blocks */
|
|
ST4( bl aes_encrypt_block4x )
|
|
ST5( ld1 {v4.16b}, [x1], #16 )
|
|
ST5( bl aes_encrypt_block5x )
|
|
st1 {v0.16b-v3.16b}, [x0], #64
|
|
ST5( st1 {v4.16b}, [x0], #16 )
|
|
b .LecbencloopNx
|
|
.Lecbenc1x:
|
|
adds w4, w4, #MAX_STRIDE
|
|
beq .Lecbencout
|
|
.Lecbencloop:
|
|
ld1 {v0.16b}, [x1], #16 /* get next pt block */
|
|
encrypt_block v0, w3, x2, x5, w6
|
|
st1 {v0.16b}, [x0], #16
|
|
subs w4, w4, #1
|
|
bne .Lecbencloop
|
|
.Lecbencout:
|
|
frame_pop
|
|
ret
|
|
AES_FUNC_END(aes_ecb_encrypt)
|
|
|
|
|
|
AES_FUNC_START(aes_ecb_decrypt)
|
|
frame_push 0
|
|
|
|
dec_prepare w3, x2, x5
|
|
|
|
.LecbdecloopNx:
|
|
subs w4, w4, #MAX_STRIDE
|
|
bmi .Lecbdec1x
|
|
ld1 {v0.16b-v3.16b}, [x1], #64 /* get 4 ct blocks */
|
|
ST4( bl aes_decrypt_block4x )
|
|
ST5( ld1 {v4.16b}, [x1], #16 )
|
|
ST5( bl aes_decrypt_block5x )
|
|
st1 {v0.16b-v3.16b}, [x0], #64
|
|
ST5( st1 {v4.16b}, [x0], #16 )
|
|
b .LecbdecloopNx
|
|
.Lecbdec1x:
|
|
adds w4, w4, #MAX_STRIDE
|
|
beq .Lecbdecout
|
|
.Lecbdecloop:
|
|
ld1 {v0.16b}, [x1], #16 /* get next ct block */
|
|
decrypt_block v0, w3, x2, x5, w6
|
|
st1 {v0.16b}, [x0], #16
|
|
subs w4, w4, #1
|
|
bne .Lecbdecloop
|
|
.Lecbdecout:
|
|
frame_pop
|
|
ret
|
|
AES_FUNC_END(aes_ecb_decrypt)
|
|
|
|
|
|
/*
|
|
* aes_cbc_encrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks, u8 iv[])
|
|
* aes_cbc_decrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int blocks, u8 iv[])
|
|
* aes_essiv_cbc_encrypt(u8 out[], u8 const in[], u32 const rk1[],
|
|
* int rounds, int blocks, u8 iv[],
|
|
* u32 const rk2[]);
|
|
* aes_essiv_cbc_decrypt(u8 out[], u8 const in[], u32 const rk1[],
|
|
* int rounds, int blocks, u8 iv[],
|
|
* u32 const rk2[]);
|
|
*/
|
|
|
|
AES_FUNC_START(aes_essiv_cbc_encrypt)
|
|
ld1 {v4.16b}, [x5] /* get iv */
|
|
|
|
mov w8, #14 /* AES-256: 14 rounds */
|
|
enc_prepare w8, x6, x7
|
|
encrypt_block v4, w8, x6, x7, w9
|
|
enc_switch_key w3, x2, x6
|
|
b .Lcbcencloop4x
|
|
|
|
AES_FUNC_START(aes_cbc_encrypt)
|
|
ld1 {v4.16b}, [x5] /* get iv */
|
|
enc_prepare w3, x2, x6
|
|
|
|
.Lcbcencloop4x:
|
|
subs w4, w4, #4
|
|
bmi .Lcbcenc1x
|
|
ld1 {v0.16b-v3.16b}, [x1], #64 /* get 4 pt blocks */
|
|
eor v0.16b, v0.16b, v4.16b /* ..and xor with iv */
|
|
encrypt_block v0, w3, x2, x6, w7
|
|
eor v1.16b, v1.16b, v0.16b
|
|
encrypt_block v1, w3, x2, x6, w7
|
|
eor v2.16b, v2.16b, v1.16b
|
|
encrypt_block v2, w3, x2, x6, w7
|
|
eor v3.16b, v3.16b, v2.16b
|
|
encrypt_block v3, w3, x2, x6, w7
|
|
st1 {v0.16b-v3.16b}, [x0], #64
|
|
mov v4.16b, v3.16b
|
|
b .Lcbcencloop4x
|
|
.Lcbcenc1x:
|
|
adds w4, w4, #4
|
|
beq .Lcbcencout
|
|
.Lcbcencloop:
|
|
ld1 {v0.16b}, [x1], #16 /* get next pt block */
|
|
eor v4.16b, v4.16b, v0.16b /* ..and xor with iv */
|
|
encrypt_block v4, w3, x2, x6, w7
|
|
st1 {v4.16b}, [x0], #16
|
|
subs w4, w4, #1
|
|
bne .Lcbcencloop
|
|
.Lcbcencout:
|
|
st1 {v4.16b}, [x5] /* return iv */
|
|
ret
|
|
AES_FUNC_END(aes_cbc_encrypt)
|
|
AES_FUNC_END(aes_essiv_cbc_encrypt)
|
|
|
|
AES_FUNC_START(aes_essiv_cbc_decrypt)
|
|
ld1 {cbciv.16b}, [x5] /* get iv */
|
|
|
|
mov w8, #14 /* AES-256: 14 rounds */
|
|
enc_prepare w8, x6, x7
|
|
encrypt_block cbciv, w8, x6, x7, w9
|
|
b .Lessivcbcdecstart
|
|
|
|
AES_FUNC_START(aes_cbc_decrypt)
|
|
ld1 {cbciv.16b}, [x5] /* get iv */
|
|
.Lessivcbcdecstart:
|
|
frame_push 0
|
|
dec_prepare w3, x2, x6
|
|
|
|
.LcbcdecloopNx:
|
|
subs w4, w4, #MAX_STRIDE
|
|
bmi .Lcbcdec1x
|
|
ld1 {v0.16b-v3.16b}, [x1], #64 /* get 4 ct blocks */
|
|
#if MAX_STRIDE == 5
|
|
ld1 {v4.16b}, [x1], #16 /* get 1 ct block */
|
|
mov v5.16b, v0.16b
|
|
mov v6.16b, v1.16b
|
|
mov v7.16b, v2.16b
|
|
bl aes_decrypt_block5x
|
|
sub x1, x1, #32
|
|
eor v0.16b, v0.16b, cbciv.16b
|
|
eor v1.16b, v1.16b, v5.16b
|
|
ld1 {v5.16b}, [x1], #16 /* reload 1 ct block */
|
|
ld1 {cbciv.16b}, [x1], #16 /* reload 1 ct block */
|
|
eor v2.16b, v2.16b, v6.16b
|
|
eor v3.16b, v3.16b, v7.16b
|
|
eor v4.16b, v4.16b, v5.16b
|
|
#else
|
|
mov v4.16b, v0.16b
|
|
mov v5.16b, v1.16b
|
|
mov v6.16b, v2.16b
|
|
bl aes_decrypt_block4x
|
|
sub x1, x1, #16
|
|
eor v0.16b, v0.16b, cbciv.16b
|
|
eor v1.16b, v1.16b, v4.16b
|
|
ld1 {cbciv.16b}, [x1], #16 /* reload 1 ct block */
|
|
eor v2.16b, v2.16b, v5.16b
|
|
eor v3.16b, v3.16b, v6.16b
|
|
#endif
|
|
st1 {v0.16b-v3.16b}, [x0], #64
|
|
ST5( st1 {v4.16b}, [x0], #16 )
|
|
b .LcbcdecloopNx
|
|
.Lcbcdec1x:
|
|
adds w4, w4, #MAX_STRIDE
|
|
beq .Lcbcdecout
|
|
.Lcbcdecloop:
|
|
ld1 {v1.16b}, [x1], #16 /* get next ct block */
|
|
mov v0.16b, v1.16b /* ...and copy to v0 */
|
|
decrypt_block v0, w3, x2, x6, w7
|
|
eor v0.16b, v0.16b, cbciv.16b /* xor with iv => pt */
|
|
mov cbciv.16b, v1.16b /* ct is next iv */
|
|
st1 {v0.16b}, [x0], #16
|
|
subs w4, w4, #1
|
|
bne .Lcbcdecloop
|
|
.Lcbcdecout:
|
|
st1 {cbciv.16b}, [x5] /* return iv */
|
|
frame_pop
|
|
ret
|
|
AES_FUNC_END(aes_cbc_decrypt)
|
|
AES_FUNC_END(aes_essiv_cbc_decrypt)
|
|
|
|
|
|
/*
|
|
* aes_cbc_cts_encrypt(u8 out[], u8 const in[], u32 const rk[],
|
|
* int rounds, int bytes, u8 const iv[])
|
|
* aes_cbc_cts_decrypt(u8 out[], u8 const in[], u32 const rk[],
|
|
* int rounds, int bytes, u8 const iv[])
|
|
*/
|
|
|
|
AES_FUNC_START(aes_cbc_cts_encrypt)
|
|
adr_l x8, .Lcts_permute_table
|
|
sub x4, x4, #16
|
|
add x9, x8, #32
|
|
add x8, x8, x4
|
|
sub x9, x9, x4
|
|
ld1 {v3.16b}, [x8]
|
|
ld1 {v4.16b}, [x9]
|
|
|
|
ld1 {v0.16b}, [x1], x4 /* overlapping loads */
|
|
ld1 {v1.16b}, [x1]
|
|
|
|
ld1 {v5.16b}, [x5] /* get iv */
|
|
enc_prepare w3, x2, x6
|
|
|
|
eor v0.16b, v0.16b, v5.16b /* xor with iv */
|
|
tbl v1.16b, {v1.16b}, v4.16b
|
|
encrypt_block v0, w3, x2, x6, w7
|
|
|
|
eor v1.16b, v1.16b, v0.16b
|
|
tbl v0.16b, {v0.16b}, v3.16b
|
|
encrypt_block v1, w3, x2, x6, w7
|
|
|
|
add x4, x0, x4
|
|
st1 {v0.16b}, [x4] /* overlapping stores */
|
|
st1 {v1.16b}, [x0]
|
|
ret
|
|
AES_FUNC_END(aes_cbc_cts_encrypt)
|
|
|
|
AES_FUNC_START(aes_cbc_cts_decrypt)
|
|
adr_l x8, .Lcts_permute_table
|
|
sub x4, x4, #16
|
|
add x9, x8, #32
|
|
add x8, x8, x4
|
|
sub x9, x9, x4
|
|
ld1 {v3.16b}, [x8]
|
|
ld1 {v4.16b}, [x9]
|
|
|
|
ld1 {v0.16b}, [x1], x4 /* overlapping loads */
|
|
ld1 {v1.16b}, [x1]
|
|
|
|
ld1 {v5.16b}, [x5] /* get iv */
|
|
dec_prepare w3, x2, x6
|
|
|
|
decrypt_block v0, w3, x2, x6, w7
|
|
tbl v2.16b, {v0.16b}, v3.16b
|
|
eor v2.16b, v2.16b, v1.16b
|
|
|
|
tbx v0.16b, {v1.16b}, v4.16b
|
|
decrypt_block v0, w3, x2, x6, w7
|
|
eor v0.16b, v0.16b, v5.16b /* xor with iv */
|
|
|
|
add x4, x0, x4
|
|
st1 {v2.16b}, [x4] /* overlapping stores */
|
|
st1 {v0.16b}, [x0]
|
|
ret
|
|
AES_FUNC_END(aes_cbc_cts_decrypt)
|
|
|
|
.section ".rodata", "a"
|
|
.align 6
|
|
.Lcts_permute_table:
|
|
.byte 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff
|
|
.byte 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff
|
|
.byte 0x0, 0x1, 0x2, 0x3, 0x4, 0x5, 0x6, 0x7
|
|
.byte 0x8, 0x9, 0xa, 0xb, 0xc, 0xd, 0xe, 0xf
|
|
.byte 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff
|
|
.byte 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff, 0xff
|
|
.previous
|
|
|
|
/*
|
|
* This macro generates the code for CTR and XCTR mode.
|
|
*/
|
|
.macro ctr_encrypt xctr
|
|
// Arguments
|
|
OUT .req x0
|
|
IN .req x1
|
|
KEY .req x2
|
|
ROUNDS_W .req w3
|
|
BYTES_W .req w4
|
|
IV .req x5
|
|
BYTE_CTR_W .req w6 // XCTR only
|
|
// Intermediate values
|
|
CTR_W .req w11 // XCTR only
|
|
CTR .req x11 // XCTR only
|
|
IV_PART .req x12
|
|
BLOCKS .req x13
|
|
BLOCKS_W .req w13
|
|
|
|
frame_push 0
|
|
|
|
enc_prepare ROUNDS_W, KEY, IV_PART
|
|
ld1 {vctr.16b}, [IV]
|
|
|
|
/*
|
|
* Keep 64 bits of the IV in a register. For CTR mode this lets us
|
|
* easily increment the IV. For XCTR mode this lets us efficiently XOR
|
|
* the 64-bit counter with the IV.
|
|
*/
|
|
.if \xctr
|
|
umov IV_PART, vctr.d[0]
|
|
lsr CTR_W, BYTE_CTR_W, #4
|
|
.else
|
|
umov IV_PART, vctr.d[1]
|
|
rev IV_PART, IV_PART
|
|
.endif
|
|
|
|
.LctrloopNx\xctr:
|
|
add BLOCKS_W, BYTES_W, #15
|
|
sub BYTES_W, BYTES_W, #MAX_STRIDE << 4
|
|
lsr BLOCKS_W, BLOCKS_W, #4
|
|
mov w8, #MAX_STRIDE
|
|
cmp BLOCKS_W, w8
|
|
csel BLOCKS_W, BLOCKS_W, w8, lt
|
|
|
|
/*
|
|
* Set up the counter values in v0-v{MAX_STRIDE-1}.
|
|
*
|
|
* If we are encrypting less than MAX_STRIDE blocks, the tail block
|
|
* handling code expects the last keystream block to be in
|
|
* v{MAX_STRIDE-1}. For example: if encrypting two blocks with
|
|
* MAX_STRIDE=5, then v3 and v4 should have the next two counter blocks.
|
|
*/
|
|
.if \xctr
|
|
add CTR, CTR, BLOCKS
|
|
.else
|
|
adds IV_PART, IV_PART, BLOCKS
|
|
.endif
|
|
mov v0.16b, vctr.16b
|
|
mov v1.16b, vctr.16b
|
|
mov v2.16b, vctr.16b
|
|
mov v3.16b, vctr.16b
|
|
ST5( mov v4.16b, vctr.16b )
|
|
.if \xctr
|
|
sub x6, CTR, #MAX_STRIDE - 1
|
|
sub x7, CTR, #MAX_STRIDE - 2
|
|
sub x8, CTR, #MAX_STRIDE - 3
|
|
sub x9, CTR, #MAX_STRIDE - 4
|
|
ST5( sub x10, CTR, #MAX_STRIDE - 5 )
|
|
eor x6, x6, IV_PART
|
|
eor x7, x7, IV_PART
|
|
eor x8, x8, IV_PART
|
|
eor x9, x9, IV_PART
|
|
ST5( eor x10, x10, IV_PART )
|
|
mov v0.d[0], x6
|
|
mov v1.d[0], x7
|
|
mov v2.d[0], x8
|
|
mov v3.d[0], x9
|
|
ST5( mov v4.d[0], x10 )
|
|
.else
|
|
bcs 0f
|
|
.subsection 1
|
|
/*
|
|
* This subsection handles carries.
|
|
*
|
|
* Conditional branching here is allowed with respect to time
|
|
* invariance since the branches are dependent on the IV instead
|
|
* of the plaintext or key. This code is rarely executed in
|
|
* practice anyway.
|
|
*/
|
|
|
|
/* Apply carry to outgoing counter. */
|
|
0: umov x8, vctr.d[0]
|
|
rev x8, x8
|
|
add x8, x8, #1
|
|
rev x8, x8
|
|
ins vctr.d[0], x8
|
|
|
|
/*
|
|
* Apply carry to counter blocks if needed.
|
|
*
|
|
* Since the carry flag was set, we know 0 <= IV_PART <
|
|
* MAX_STRIDE. Using the value of IV_PART we can determine how
|
|
* many counter blocks need to be updated.
|
|
*/
|
|
cbz IV_PART, 2f
|
|
adr x16, 1f
|
|
sub x16, x16, IV_PART, lsl #3
|
|
br x16
|
|
bti c
|
|
mov v0.d[0], vctr.d[0]
|
|
bti c
|
|
mov v1.d[0], vctr.d[0]
|
|
bti c
|
|
mov v2.d[0], vctr.d[0]
|
|
bti c
|
|
mov v3.d[0], vctr.d[0]
|
|
ST5( bti c )
|
|
ST5( mov v4.d[0], vctr.d[0] )
|
|
1: b 2f
|
|
.previous
|
|
|
|
2: rev x7, IV_PART
|
|
ins vctr.d[1], x7
|
|
sub x7, IV_PART, #MAX_STRIDE - 1
|
|
sub x8, IV_PART, #MAX_STRIDE - 2
|
|
sub x9, IV_PART, #MAX_STRIDE - 3
|
|
rev x7, x7
|
|
rev x8, x8
|
|
mov v1.d[1], x7
|
|
rev x9, x9
|
|
ST5( sub x10, IV_PART, #MAX_STRIDE - 4 )
|
|
mov v2.d[1], x8
|
|
ST5( rev x10, x10 )
|
|
mov v3.d[1], x9
|
|
ST5( mov v4.d[1], x10 )
|
|
.endif
|
|
|
|
/*
|
|
* If there are at least MAX_STRIDE blocks left, XOR the data with
|
|
* keystream and store. Otherwise jump to tail handling.
|
|
*/
|
|
tbnz BYTES_W, #31, .Lctrtail\xctr
|
|
ld1 {v5.16b-v7.16b}, [IN], #48
|
|
ST4( bl aes_encrypt_block4x )
|
|
ST5( bl aes_encrypt_block5x )
|
|
eor v0.16b, v5.16b, v0.16b
|
|
ST4( ld1 {v5.16b}, [IN], #16 )
|
|
eor v1.16b, v6.16b, v1.16b
|
|
ST5( ld1 {v5.16b-v6.16b}, [IN], #32 )
|
|
eor v2.16b, v7.16b, v2.16b
|
|
eor v3.16b, v5.16b, v3.16b
|
|
ST5( eor v4.16b, v6.16b, v4.16b )
|
|
st1 {v0.16b-v3.16b}, [OUT], #64
|
|
ST5( st1 {v4.16b}, [OUT], #16 )
|
|
cbz BYTES_W, .Lctrout\xctr
|
|
b .LctrloopNx\xctr
|
|
|
|
.Lctrout\xctr:
|
|
.if !\xctr
|
|
st1 {vctr.16b}, [IV] /* return next CTR value */
|
|
.endif
|
|
frame_pop
|
|
ret
|
|
|
|
.Lctrtail\xctr:
|
|
/*
|
|
* Handle up to MAX_STRIDE * 16 - 1 bytes of plaintext
|
|
*
|
|
* This code expects the last keystream block to be in v{MAX_STRIDE-1}.
|
|
* For example: if encrypting two blocks with MAX_STRIDE=5, then v3 and
|
|
* v4 should have the next two counter blocks.
|
|
*
|
|
* This allows us to store the ciphertext by writing to overlapping
|
|
* regions of memory. Any invalid ciphertext blocks get overwritten by
|
|
* correctly computed blocks. This approach greatly simplifies the
|
|
* logic for storing the ciphertext.
|
|
*/
|
|
mov x16, #16
|
|
ands w7, BYTES_W, #0xf
|
|
csel x13, x7, x16, ne
|
|
|
|
ST5( cmp BYTES_W, #64 - (MAX_STRIDE << 4))
|
|
ST5( csel x14, x16, xzr, gt )
|
|
cmp BYTES_W, #48 - (MAX_STRIDE << 4)
|
|
csel x15, x16, xzr, gt
|
|
cmp BYTES_W, #32 - (MAX_STRIDE << 4)
|
|
csel x16, x16, xzr, gt
|
|
cmp BYTES_W, #16 - (MAX_STRIDE << 4)
|
|
|
|
adr_l x9, .Lcts_permute_table
|
|
add x9, x9, x13
|
|
ble .Lctrtail1x\xctr
|
|
|
|
ST5( ld1 {v5.16b}, [IN], x14 )
|
|
ld1 {v6.16b}, [IN], x15
|
|
ld1 {v7.16b}, [IN], x16
|
|
|
|
ST4( bl aes_encrypt_block4x )
|
|
ST5( bl aes_encrypt_block5x )
|
|
|
|
ld1 {v8.16b}, [IN], x13
|
|
ld1 {v9.16b}, [IN]
|
|
ld1 {v10.16b}, [x9]
|
|
|
|
ST4( eor v6.16b, v6.16b, v0.16b )
|
|
ST4( eor v7.16b, v7.16b, v1.16b )
|
|
ST4( tbl v3.16b, {v3.16b}, v10.16b )
|
|
ST4( eor v8.16b, v8.16b, v2.16b )
|
|
ST4( eor v9.16b, v9.16b, v3.16b )
|
|
|
|
ST5( eor v5.16b, v5.16b, v0.16b )
|
|
ST5( eor v6.16b, v6.16b, v1.16b )
|
|
ST5( tbl v4.16b, {v4.16b}, v10.16b )
|
|
ST5( eor v7.16b, v7.16b, v2.16b )
|
|
ST5( eor v8.16b, v8.16b, v3.16b )
|
|
ST5( eor v9.16b, v9.16b, v4.16b )
|
|
|
|
ST5( st1 {v5.16b}, [OUT], x14 )
|
|
st1 {v6.16b}, [OUT], x15
|
|
st1 {v7.16b}, [OUT], x16
|
|
add x13, x13, OUT
|
|
st1 {v9.16b}, [x13] // overlapping stores
|
|
st1 {v8.16b}, [OUT]
|
|
b .Lctrout\xctr
|
|
|
|
.Lctrtail1x\xctr:
|
|
/*
|
|
* Handle <= 16 bytes of plaintext
|
|
*
|
|
* This code always reads and writes 16 bytes. To avoid out of bounds
|
|
* accesses, XCTR and CTR modes must use a temporary buffer when
|
|
* encrypting/decrypting less than 16 bytes.
|
|
*
|
|
* This code is unusual in that it loads the input and stores the output
|
|
* relative to the end of the buffers rather than relative to the start.
|
|
* This causes unusual behaviour when encrypting/decrypting less than 16
|
|
* bytes; the end of the data is expected to be at the end of the
|
|
* temporary buffer rather than the start of the data being at the start
|
|
* of the temporary buffer.
|
|
*/
|
|
sub x8, x7, #16
|
|
csel x7, x7, x8, eq
|
|
add IN, IN, x7
|
|
add OUT, OUT, x7
|
|
ld1 {v5.16b}, [IN]
|
|
ld1 {v6.16b}, [OUT]
|
|
ST5( mov v3.16b, v4.16b )
|
|
encrypt_block v3, ROUNDS_W, KEY, x8, w7
|
|
ld1 {v10.16b-v11.16b}, [x9]
|
|
tbl v3.16b, {v3.16b}, v10.16b
|
|
sshr v11.16b, v11.16b, #7
|
|
eor v5.16b, v5.16b, v3.16b
|
|
bif v5.16b, v6.16b, v11.16b
|
|
st1 {v5.16b}, [OUT]
|
|
b .Lctrout\xctr
|
|
|
|
// Arguments
|
|
.unreq OUT
|
|
.unreq IN
|
|
.unreq KEY
|
|
.unreq ROUNDS_W
|
|
.unreq BYTES_W
|
|
.unreq IV
|
|
.unreq BYTE_CTR_W // XCTR only
|
|
// Intermediate values
|
|
.unreq CTR_W // XCTR only
|
|
.unreq CTR // XCTR only
|
|
.unreq IV_PART
|
|
.unreq BLOCKS
|
|
.unreq BLOCKS_W
|
|
.endm
|
|
|
|
/*
|
|
* aes_ctr_encrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int bytes, u8 ctr[])
|
|
*
|
|
* The input and output buffers must always be at least 16 bytes even if
|
|
* encrypting/decrypting less than 16 bytes. Otherwise out of bounds
|
|
* accesses will occur. The data to be encrypted/decrypted is expected
|
|
* to be at the end of this 16-byte temporary buffer rather than the
|
|
* start.
|
|
*/
|
|
|
|
AES_FUNC_START(aes_ctr_encrypt)
|
|
ctr_encrypt 0
|
|
AES_FUNC_END(aes_ctr_encrypt)
|
|
|
|
/*
|
|
* aes_xctr_encrypt(u8 out[], u8 const in[], u8 const rk[], int rounds,
|
|
* int bytes, u8 const iv[], int byte_ctr)
|
|
*
|
|
* The input and output buffers must always be at least 16 bytes even if
|
|
* encrypting/decrypting less than 16 bytes. Otherwise out of bounds
|
|
* accesses will occur. The data to be encrypted/decrypted is expected
|
|
* to be at the end of this 16-byte temporary buffer rather than the
|
|
* start.
|
|
*/
|
|
|
|
AES_FUNC_START(aes_xctr_encrypt)
|
|
ctr_encrypt 1
|
|
AES_FUNC_END(aes_xctr_encrypt)
|
|
|
|
|
|
/*
|
|
* aes_xts_encrypt(u8 out[], u8 const in[], u8 const rk1[], int rounds,
|
|
* int bytes, u8 const rk2[], u8 iv[], int first)
|
|
* aes_xts_decrypt(u8 out[], u8 const in[], u8 const rk1[], int rounds,
|
|
* int bytes, u8 const rk2[], u8 iv[], int first)
|
|
*/
|
|
|
|
.macro next_tweak, out, in, tmp
|
|
sshr \tmp\().2d, \in\().2d, #63
|
|
and \tmp\().16b, \tmp\().16b, xtsmask.16b
|
|
add \out\().2d, \in\().2d, \in\().2d
|
|
ext \tmp\().16b, \tmp\().16b, \tmp\().16b, #8
|
|
eor \out\().16b, \out\().16b, \tmp\().16b
|
|
.endm
|
|
|
|
.macro xts_load_mask, tmp
|
|
movi xtsmask.2s, #0x1
|
|
movi \tmp\().2s, #0x87
|
|
uzp1 xtsmask.4s, xtsmask.4s, \tmp\().4s
|
|
.endm
|
|
|
|
AES_FUNC_START(aes_xts_encrypt)
|
|
frame_push 0
|
|
|
|
ld1 {v4.16b}, [x6]
|
|
xts_load_mask v8
|
|
cbz w7, .Lxtsencnotfirst
|
|
|
|
enc_prepare w3, x5, x8
|
|
xts_cts_skip_tw w7, .LxtsencNx
|
|
encrypt_block v4, w3, x5, x8, w7 /* first tweak */
|
|
enc_switch_key w3, x2, x8
|
|
b .LxtsencNx
|
|
|
|
.Lxtsencnotfirst:
|
|
enc_prepare w3, x2, x8
|
|
.LxtsencloopNx:
|
|
next_tweak v4, v4, v8
|
|
.LxtsencNx:
|
|
subs w4, w4, #64
|
|
bmi .Lxtsenc1x
|
|
ld1 {v0.16b-v3.16b}, [x1], #64 /* get 4 pt blocks */
|
|
next_tweak v5, v4, v8
|
|
eor v0.16b, v0.16b, v4.16b
|
|
next_tweak v6, v5, v8
|
|
eor v1.16b, v1.16b, v5.16b
|
|
eor v2.16b, v2.16b, v6.16b
|
|
next_tweak v7, v6, v8
|
|
eor v3.16b, v3.16b, v7.16b
|
|
bl aes_encrypt_block4x
|
|
eor v3.16b, v3.16b, v7.16b
|
|
eor v0.16b, v0.16b, v4.16b
|
|
eor v1.16b, v1.16b, v5.16b
|
|
eor v2.16b, v2.16b, v6.16b
|
|
st1 {v0.16b-v3.16b}, [x0], #64
|
|
mov v4.16b, v7.16b
|
|
cbz w4, .Lxtsencret
|
|
xts_reload_mask v8
|
|
b .LxtsencloopNx
|
|
.Lxtsenc1x:
|
|
adds w4, w4, #64
|
|
beq .Lxtsencout
|
|
subs w4, w4, #16
|
|
bmi .LxtsencctsNx
|
|
.Lxtsencloop:
|
|
ld1 {v0.16b}, [x1], #16
|
|
.Lxtsencctsout:
|
|
eor v0.16b, v0.16b, v4.16b
|
|
encrypt_block v0, w3, x2, x8, w7
|
|
eor v0.16b, v0.16b, v4.16b
|
|
cbz w4, .Lxtsencout
|
|
subs w4, w4, #16
|
|
next_tweak v4, v4, v8
|
|
bmi .Lxtsenccts
|
|
st1 {v0.16b}, [x0], #16
|
|
b .Lxtsencloop
|
|
.Lxtsencout:
|
|
st1 {v0.16b}, [x0]
|
|
.Lxtsencret:
|
|
st1 {v4.16b}, [x6]
|
|
frame_pop
|
|
ret
|
|
|
|
.LxtsencctsNx:
|
|
mov v0.16b, v3.16b
|
|
sub x0, x0, #16
|
|
.Lxtsenccts:
|
|
adr_l x8, .Lcts_permute_table
|
|
|
|
add x1, x1, w4, sxtw /* rewind input pointer */
|
|
add w4, w4, #16 /* # bytes in final block */
|
|
add x9, x8, #32
|
|
add x8, x8, x4
|
|
sub x9, x9, x4
|
|
add x4, x0, x4 /* output address of final block */
|
|
|
|
ld1 {v1.16b}, [x1] /* load final block */
|
|
ld1 {v2.16b}, [x8]
|
|
ld1 {v3.16b}, [x9]
|
|
|
|
tbl v2.16b, {v0.16b}, v2.16b
|
|
tbx v0.16b, {v1.16b}, v3.16b
|
|
st1 {v2.16b}, [x4] /* overlapping stores */
|
|
mov w4, wzr
|
|
b .Lxtsencctsout
|
|
AES_FUNC_END(aes_xts_encrypt)
|
|
|
|
AES_FUNC_START(aes_xts_decrypt)
|
|
frame_push 0
|
|
|
|
/* subtract 16 bytes if we are doing CTS */
|
|
sub w8, w4, #0x10
|
|
tst w4, #0xf
|
|
csel w4, w4, w8, eq
|
|
|
|
ld1 {v4.16b}, [x6]
|
|
xts_load_mask v8
|
|
xts_cts_skip_tw w7, .Lxtsdecskiptw
|
|
cbz w7, .Lxtsdecnotfirst
|
|
|
|
enc_prepare w3, x5, x8
|
|
encrypt_block v4, w3, x5, x8, w7 /* first tweak */
|
|
.Lxtsdecskiptw:
|
|
dec_prepare w3, x2, x8
|
|
b .LxtsdecNx
|
|
|
|
.Lxtsdecnotfirst:
|
|
dec_prepare w3, x2, x8
|
|
.LxtsdecloopNx:
|
|
next_tweak v4, v4, v8
|
|
.LxtsdecNx:
|
|
subs w4, w4, #64
|
|
bmi .Lxtsdec1x
|
|
ld1 {v0.16b-v3.16b}, [x1], #64 /* get 4 ct blocks */
|
|
next_tweak v5, v4, v8
|
|
eor v0.16b, v0.16b, v4.16b
|
|
next_tweak v6, v5, v8
|
|
eor v1.16b, v1.16b, v5.16b
|
|
eor v2.16b, v2.16b, v6.16b
|
|
next_tweak v7, v6, v8
|
|
eor v3.16b, v3.16b, v7.16b
|
|
bl aes_decrypt_block4x
|
|
eor v3.16b, v3.16b, v7.16b
|
|
eor v0.16b, v0.16b, v4.16b
|
|
eor v1.16b, v1.16b, v5.16b
|
|
eor v2.16b, v2.16b, v6.16b
|
|
st1 {v0.16b-v3.16b}, [x0], #64
|
|
mov v4.16b, v7.16b
|
|
cbz w4, .Lxtsdecout
|
|
xts_reload_mask v8
|
|
b .LxtsdecloopNx
|
|
.Lxtsdec1x:
|
|
adds w4, w4, #64
|
|
beq .Lxtsdecout
|
|
subs w4, w4, #16
|
|
.Lxtsdecloop:
|
|
ld1 {v0.16b}, [x1], #16
|
|
bmi .Lxtsdeccts
|
|
.Lxtsdecctsout:
|
|
eor v0.16b, v0.16b, v4.16b
|
|
decrypt_block v0, w3, x2, x8, w7
|
|
eor v0.16b, v0.16b, v4.16b
|
|
st1 {v0.16b}, [x0], #16
|
|
cbz w4, .Lxtsdecout
|
|
subs w4, w4, #16
|
|
next_tweak v4, v4, v8
|
|
b .Lxtsdecloop
|
|
.Lxtsdecout:
|
|
st1 {v4.16b}, [x6]
|
|
frame_pop
|
|
ret
|
|
|
|
.Lxtsdeccts:
|
|
adr_l x8, .Lcts_permute_table
|
|
|
|
add x1, x1, w4, sxtw /* rewind input pointer */
|
|
add w4, w4, #16 /* # bytes in final block */
|
|
add x9, x8, #32
|
|
add x8, x8, x4
|
|
sub x9, x9, x4
|
|
add x4, x0, x4 /* output address of final block */
|
|
|
|
next_tweak v5, v4, v8
|
|
|
|
ld1 {v1.16b}, [x1] /* load final block */
|
|
ld1 {v2.16b}, [x8]
|
|
ld1 {v3.16b}, [x9]
|
|
|
|
eor v0.16b, v0.16b, v5.16b
|
|
decrypt_block v0, w3, x2, x8, w7
|
|
eor v0.16b, v0.16b, v5.16b
|
|
|
|
tbl v2.16b, {v0.16b}, v2.16b
|
|
tbx v0.16b, {v1.16b}, v3.16b
|
|
|
|
st1 {v2.16b}, [x4] /* overlapping stores */
|
|
mov w4, wzr
|
|
b .Lxtsdecctsout
|
|
AES_FUNC_END(aes_xts_decrypt)
|
|
|
|
/*
|
|
* aes_mac_update(u8 const in[], u32 const rk[], int rounds,
|
|
* int blocks, u8 dg[], int enc_before, int enc_after)
|
|
*/
|
|
AES_FUNC_START(aes_mac_update)
|
|
ld1 {v0.16b}, [x4] /* get dg */
|
|
enc_prepare w2, x1, x7
|
|
cbz w5, .Lmacloop4x
|
|
|
|
encrypt_block v0, w2, x1, x7, w8
|
|
|
|
.Lmacloop4x:
|
|
subs w3, w3, #4
|
|
bmi .Lmac1x
|
|
ld1 {v1.16b-v4.16b}, [x0], #64 /* get next pt block */
|
|
eor v0.16b, v0.16b, v1.16b /* ..and xor with dg */
|
|
encrypt_block v0, w2, x1, x7, w8
|
|
eor v0.16b, v0.16b, v2.16b
|
|
encrypt_block v0, w2, x1, x7, w8
|
|
eor v0.16b, v0.16b, v3.16b
|
|
encrypt_block v0, w2, x1, x7, w8
|
|
eor v0.16b, v0.16b, v4.16b
|
|
cmp w3, wzr
|
|
csinv x5, x6, xzr, eq
|
|
cbz w5, .Lmacout
|
|
encrypt_block v0, w2, x1, x7, w8
|
|
st1 {v0.16b}, [x4] /* return dg */
|
|
cond_yield .Lmacout, x7, x8
|
|
b .Lmacloop4x
|
|
.Lmac1x:
|
|
add w3, w3, #4
|
|
.Lmacloop:
|
|
cbz w3, .Lmacout
|
|
ld1 {v1.16b}, [x0], #16 /* get next pt block */
|
|
eor v0.16b, v0.16b, v1.16b /* ..and xor with dg */
|
|
|
|
subs w3, w3, #1
|
|
csinv x5, x6, xzr, eq
|
|
cbz w5, .Lmacout
|
|
|
|
.Lmacenc:
|
|
encrypt_block v0, w2, x1, x7, w8
|
|
b .Lmacloop
|
|
|
|
.Lmacout:
|
|
st1 {v0.16b}, [x4] /* return dg */
|
|
mov w0, w3
|
|
ret
|
|
AES_FUNC_END(aes_mac_update)
|