mirror of
https://github.com/qemu/qemu.git
synced 2024-12-05 09:43:44 +08:00
target/riscv: Add Zvksed ISA extension support
This commit adds support for the Zvksed vector-crypto extension, which consists of the following instructions: * vsm4k.vi * vsm4r.[vv,vs] Translation functions are defined in `target/riscv/insn_trans/trans_rvvk.c.inc` and helpers are defined in `target/riscv/vcrypto_helper.c`. Signed-off-by: Max Chou <max.chou@sifive.com> Reviewed-by: Frank Chang <frank.chang@sifive.com> [lawrence.hunter@codethink.co.uk: Moved SM4 functions from crypto_helper.c to vcrypto_helper.c] [nazar.kazakov@codethink.co.uk: Added alignment checks, refactored code to use macros, and minor style changes] Signed-off-by: Max Chou <max.chou@sifive.com> Message-ID: <20230711165917.2629866-16-max.chou@sifive.com> Signed-off-by: Alistair Francis <alistair.francis@wdc.com>
This commit is contained in:
parent
f5f3a9152a
commit
8b045ff454
@ -133,6 +133,7 @@ static const struct isa_ext_data isa_edata_arr[] = {
|
||||
ISA_EXT_DATA_ENTRY(zvkned, PRIV_VERSION_1_12_0, ext_zvkned),
|
||||
ISA_EXT_DATA_ENTRY(zvknha, PRIV_VERSION_1_12_0, ext_zvknha),
|
||||
ISA_EXT_DATA_ENTRY(zvknhb, PRIV_VERSION_1_12_0, ext_zvknhb),
|
||||
ISA_EXT_DATA_ENTRY(zvksed, PRIV_VERSION_1_12_0, ext_zvksed),
|
||||
ISA_EXT_DATA_ENTRY(zvksh, PRIV_VERSION_1_12_0, ext_zvksh),
|
||||
ISA_EXT_DATA_ENTRY(zhinx, PRIV_VERSION_1_12_0, ext_zhinx),
|
||||
ISA_EXT_DATA_ENTRY(zhinxmin, PRIV_VERSION_1_12_0, ext_zhinxmin),
|
||||
@ -1283,7 +1284,8 @@ void riscv_cpu_validate_set_extensions(RISCVCPU *cpu, Error **errp)
|
||||
* in qemu
|
||||
*/
|
||||
if ((cpu->cfg.ext_zvbb || cpu->cfg.ext_zvkg || cpu->cfg.ext_zvkned ||
|
||||
cpu->cfg.ext_zvknha || cpu->cfg.ext_zvksh) && !cpu->cfg.ext_zve32f) {
|
||||
cpu->cfg.ext_zvknha || cpu->cfg.ext_zvksed || cpu->cfg.ext_zvksh) &&
|
||||
!cpu->cfg.ext_zve32f) {
|
||||
error_setg(errp,
|
||||
"Vector crypto extensions require V or Zve* extensions");
|
||||
return;
|
||||
@ -1885,6 +1887,7 @@ static Property riscv_cpu_extensions[] = {
|
||||
DEFINE_PROP_BOOL("x-zvkned", RISCVCPU, cfg.ext_zvkned, false),
|
||||
DEFINE_PROP_BOOL("x-zvknha", RISCVCPU, cfg.ext_zvknha, false),
|
||||
DEFINE_PROP_BOOL("x-zvknhb", RISCVCPU, cfg.ext_zvknhb, false),
|
||||
DEFINE_PROP_BOOL("x-zvksed", RISCVCPU, cfg.ext_zvksed, false),
|
||||
DEFINE_PROP_BOOL("x-zvksh", RISCVCPU, cfg.ext_zvksh, false),
|
||||
|
||||
DEFINE_PROP_END_OF_LIST(),
|
||||
|
@ -91,6 +91,7 @@ struct RISCVCPUConfig {
|
||||
bool ext_zvkned;
|
||||
bool ext_zvknha;
|
||||
bool ext_zvknhb;
|
||||
bool ext_zvksed;
|
||||
bool ext_zvksh;
|
||||
bool ext_zmmul;
|
||||
bool ext_zvfbfmin;
|
||||
|
@ -1276,3 +1276,7 @@ DEF_HELPER_5(vsm3c_vi, void, ptr, ptr, i32, env, i32)
|
||||
|
||||
DEF_HELPER_5(vghsh_vv, void, ptr, ptr, ptr, env, i32)
|
||||
DEF_HELPER_4(vgmul_vv, void, ptr, ptr, env, i32)
|
||||
|
||||
DEF_HELPER_5(vsm4k_vi, void, ptr, ptr, i32, env, i32)
|
||||
DEF_HELPER_4(vsm4r_vv, void, ptr, ptr, env, i32)
|
||||
DEF_HELPER_4(vsm4r_vs, void, ptr, ptr, env, i32)
|
||||
|
@ -999,3 +999,8 @@ vsm3c_vi 101011 1 ..... ..... 010 ..... 1110111 @r_vm_1
|
||||
# *** Zvkg vector crypto extension ***
|
||||
vghsh_vv 101100 1 ..... ..... 010 ..... 1110111 @r_vm_1
|
||||
vgmul_vv 101000 1 ..... 10001 010 ..... 1110111 @r2_vm_1
|
||||
|
||||
# *** Zvksed vector crypto extension ***
|
||||
vsm4k_vi 100001 1 ..... ..... 010 ..... 1110111 @r_vm_1
|
||||
vsm4r_vv 101000 1 ..... 10000 010 ..... 1110111 @r2_vm_1
|
||||
vsm4r_vs 101001 1 ..... 10000 010 ..... 1110111 @r2_vm_1
|
||||
|
@ -561,3 +561,46 @@ static bool vghsh_check(DisasContext *s, arg_rmrr *a)
|
||||
}
|
||||
|
||||
GEN_VV_UNMASKED_TRANS(vghsh_vv, vghsh_check, ZVKG_EGS)
|
||||
|
||||
/*
|
||||
* Zvksed
|
||||
*/
|
||||
|
||||
#define ZVKSED_EGS 4
|
||||
|
||||
static bool zvksed_check(DisasContext *s)
|
||||
{
|
||||
int egw_bytes = ZVKSED_EGS << s->sew;
|
||||
return s->cfg_ptr->ext_zvksed == true &&
|
||||
require_rvv(s) &&
|
||||
vext_check_isa_ill(s) &&
|
||||
MAXSZ(s) >= egw_bytes &&
|
||||
s->sew == MO_32;
|
||||
}
|
||||
|
||||
static bool vsm4k_vi_check(DisasContext *s, arg_rmrr *a)
|
||||
{
|
||||
return zvksed_check(s) &&
|
||||
require_align(a->rd, s->lmul) &&
|
||||
require_align(a->rs2, s->lmul);
|
||||
}
|
||||
|
||||
GEN_VI_UNMASKED_TRANS(vsm4k_vi, vsm4k_vi_check, ZVKSED_EGS)
|
||||
|
||||
static bool vsm4r_vv_check(DisasContext *s, arg_rmr *a)
|
||||
{
|
||||
return zvksed_check(s) &&
|
||||
require_align(a->rd, s->lmul) &&
|
||||
require_align(a->rs2, s->lmul);
|
||||
}
|
||||
|
||||
GEN_V_UNMASKED_TRANS(vsm4r_vv, vsm4r_vv_check, ZVKSED_EGS)
|
||||
|
||||
static bool vsm4r_vs_check(DisasContext *s, arg_rmr *a)
|
||||
{
|
||||
return zvksed_check(s) &&
|
||||
!is_overlapped(a->rd, 1 << MAX(s->lmul, 0), a->rs2, 1) &&
|
||||
require_align(a->rd, s->lmul);
|
||||
}
|
||||
|
||||
GEN_V_UNMASKED_TRANS(vsm4r_vs, vsm4r_vs_check, ZVKSED_EGS)
|
||||
|
@ -24,6 +24,7 @@
|
||||
#include "cpu.h"
|
||||
#include "crypto/aes.h"
|
||||
#include "crypto/aes-round.h"
|
||||
#include "crypto/sm4.h"
|
||||
#include "exec/memop.h"
|
||||
#include "exec/exec-all.h"
|
||||
#include "exec/helper-proto.h"
|
||||
@ -841,3 +842,129 @@ void HELPER(vgmul_vv)(void *vd_vptr, void *vs2_vptr, CPURISCVState *env,
|
||||
vext_set_elems_1s(vd, vta, env->vl * 4, total_elems * 4);
|
||||
env->vstart = 0;
|
||||
}
|
||||
|
||||
void HELPER(vsm4k_vi)(void *vd, void *vs2, uint32_t uimm5, CPURISCVState *env,
|
||||
uint32_t desc)
|
||||
{
|
||||
const uint32_t egs = 4;
|
||||
uint32_t rnd = uimm5 & 0x7;
|
||||
uint32_t group_start = env->vstart / egs;
|
||||
uint32_t group_end = env->vl / egs;
|
||||
uint32_t esz = sizeof(uint32_t);
|
||||
uint32_t total_elems = vext_get_total_elems(env, desc, esz);
|
||||
|
||||
for (uint32_t i = group_start; i < group_end; ++i) {
|
||||
uint32_t vstart = i * egs;
|
||||
uint32_t vend = (i + 1) * egs;
|
||||
uint32_t rk[4] = {0};
|
||||
uint32_t tmp[8] = {0};
|
||||
|
||||
for (uint32_t j = vstart; j < vend; ++j) {
|
||||
rk[j - vstart] = *((uint32_t *)vs2 + H4(j));
|
||||
}
|
||||
|
||||
for (uint32_t j = 0; j < egs; ++j) {
|
||||
tmp[j] = rk[j];
|
||||
}
|
||||
|
||||
for (uint32_t j = 0; j < egs; ++j) {
|
||||
uint32_t b, s;
|
||||
b = tmp[j + 1] ^ tmp[j + 2] ^ tmp[j + 3] ^ sm4_ck[rnd * 4 + j];
|
||||
|
||||
s = sm4_subword(b);
|
||||
|
||||
tmp[j + 4] = tmp[j] ^ (s ^ rol32(s, 13) ^ rol32(s, 23));
|
||||
}
|
||||
|
||||
for (uint32_t j = vstart; j < vend; ++j) {
|
||||
*((uint32_t *)vd + H4(j)) = tmp[egs + (j - vstart)];
|
||||
}
|
||||
}
|
||||
|
||||
env->vstart = 0;
|
||||
/* set tail elements to 1s */
|
||||
vext_set_elems_1s(vd, vext_vta(desc), env->vl * esz, total_elems * esz);
|
||||
}
|
||||
|
||||
static void do_sm4_round(uint32_t *rk, uint32_t *buf)
|
||||
{
|
||||
const uint32_t egs = 4;
|
||||
uint32_t s, b;
|
||||
|
||||
for (uint32_t j = egs; j < egs * 2; ++j) {
|
||||
b = buf[j - 3] ^ buf[j - 2] ^ buf[j - 1] ^ rk[j - 4];
|
||||
|
||||
s = sm4_subword(b);
|
||||
|
||||
buf[j] = buf[j - 4] ^ (s ^ rol32(s, 2) ^ rol32(s, 10) ^ rol32(s, 18) ^
|
||||
rol32(s, 24));
|
||||
}
|
||||
}
|
||||
|
||||
void HELPER(vsm4r_vv)(void *vd, void *vs2, CPURISCVState *env, uint32_t desc)
|
||||
{
|
||||
const uint32_t egs = 4;
|
||||
uint32_t group_start = env->vstart / egs;
|
||||
uint32_t group_end = env->vl / egs;
|
||||
uint32_t esz = sizeof(uint32_t);
|
||||
uint32_t total_elems = vext_get_total_elems(env, desc, esz);
|
||||
|
||||
for (uint32_t i = group_start; i < group_end; ++i) {
|
||||
uint32_t vstart = i * egs;
|
||||
uint32_t vend = (i + 1) * egs;
|
||||
uint32_t rk[4] = {0};
|
||||
uint32_t tmp[8] = {0};
|
||||
|
||||
for (uint32_t j = vstart; j < vend; ++j) {
|
||||
rk[j - vstart] = *((uint32_t *)vs2 + H4(j));
|
||||
}
|
||||
|
||||
for (uint32_t j = vstart; j < vend; ++j) {
|
||||
tmp[j - vstart] = *((uint32_t *)vd + H4(j));
|
||||
}
|
||||
|
||||
do_sm4_round(rk, tmp);
|
||||
|
||||
for (uint32_t j = vstart; j < vend; ++j) {
|
||||
*((uint32_t *)vd + H4(j)) = tmp[egs + (j - vstart)];
|
||||
}
|
||||
}
|
||||
|
||||
env->vstart = 0;
|
||||
/* set tail elements to 1s */
|
||||
vext_set_elems_1s(vd, vext_vta(desc), env->vl * esz, total_elems * esz);
|
||||
}
|
||||
|
||||
void HELPER(vsm4r_vs)(void *vd, void *vs2, CPURISCVState *env, uint32_t desc)
|
||||
{
|
||||
const uint32_t egs = 4;
|
||||
uint32_t group_start = env->vstart / egs;
|
||||
uint32_t group_end = env->vl / egs;
|
||||
uint32_t esz = sizeof(uint32_t);
|
||||
uint32_t total_elems = vext_get_total_elems(env, desc, esz);
|
||||
|
||||
for (uint32_t i = group_start; i < group_end; ++i) {
|
||||
uint32_t vstart = i * egs;
|
||||
uint32_t vend = (i + 1) * egs;
|
||||
uint32_t rk[4] = {0};
|
||||
uint32_t tmp[8] = {0};
|
||||
|
||||
for (uint32_t j = 0; j < egs; ++j) {
|
||||
rk[j] = *((uint32_t *)vs2 + H4(j));
|
||||
}
|
||||
|
||||
for (uint32_t j = vstart; j < vend; ++j) {
|
||||
tmp[j - vstart] = *((uint32_t *)vd + H4(j));
|
||||
}
|
||||
|
||||
do_sm4_round(rk, tmp);
|
||||
|
||||
for (uint32_t j = vstart; j < vend; ++j) {
|
||||
*((uint32_t *)vd + H4(j)) = tmp[egs + (j - vstart)];
|
||||
}
|
||||
}
|
||||
|
||||
env->vstart = 0;
|
||||
/* set tail elements to 1s */
|
||||
vext_set_elems_1s(vd, vext_vta(desc), env->vl * esz, total_elems * esz);
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user