mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-17 09:43:59 +08:00
7d209d13e7
Intel IBT requires the target of any indirect CALL or JMP instruction to be the ENDBR instruction; optionally it allows those two instructions to have a NOTRACK prefix in order to avoid this requirement. The kernel will not enable the use of NOTRACK, as such any occurence of it in compiler generated code should be flagged. Teach objtool to Decode ENDBR instructions and WARN about NOTRACK prefixes. Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Acked-by: Josh Poimboeuf <jpoimboe@redhat.com> Link: https://lore.kernel.org/r/20220308154319.645963517@infradead.org
790 lines
14 KiB
C
790 lines
14 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Copyright (C) 2015 Josh Poimboeuf <jpoimboe@redhat.com>
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
|
|
#define unlikely(cond) (cond)
|
|
#include <asm/insn.h>
|
|
#include "../../../arch/x86/lib/inat.c"
|
|
#include "../../../arch/x86/lib/insn.c"
|
|
|
|
#define CONFIG_64BIT 1
|
|
#include <asm/nops.h>
|
|
|
|
#include <asm/orc_types.h>
|
|
#include <objtool/check.h>
|
|
#include <objtool/elf.h>
|
|
#include <objtool/arch.h>
|
|
#include <objtool/warn.h>
|
|
#include <objtool/endianness.h>
|
|
#include <objtool/builtin.h>
|
|
#include <arch/elf.h>
|
|
|
|
static int is_x86_64(const struct elf *elf)
|
|
{
|
|
switch (elf->ehdr.e_machine) {
|
|
case EM_X86_64:
|
|
return 1;
|
|
case EM_386:
|
|
return 0;
|
|
default:
|
|
WARN("unexpected ELF machine type %d", elf->ehdr.e_machine);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
bool arch_callee_saved_reg(unsigned char reg)
|
|
{
|
|
switch (reg) {
|
|
case CFI_BP:
|
|
case CFI_BX:
|
|
case CFI_R12:
|
|
case CFI_R13:
|
|
case CFI_R14:
|
|
case CFI_R15:
|
|
return true;
|
|
|
|
case CFI_AX:
|
|
case CFI_CX:
|
|
case CFI_DX:
|
|
case CFI_SI:
|
|
case CFI_DI:
|
|
case CFI_SP:
|
|
case CFI_R8:
|
|
case CFI_R9:
|
|
case CFI_R10:
|
|
case CFI_R11:
|
|
case CFI_RA:
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
unsigned long arch_dest_reloc_offset(int addend)
|
|
{
|
|
return addend + 4;
|
|
}
|
|
|
|
unsigned long arch_jump_destination(struct instruction *insn)
|
|
{
|
|
return insn->offset + insn->len + insn->immediate;
|
|
}
|
|
|
|
#define ADD_OP(op) \
|
|
if (!(op = calloc(1, sizeof(*op)))) \
|
|
return -1; \
|
|
else for (list_add_tail(&op->list, ops_list); op; op = NULL)
|
|
|
|
/*
|
|
* Helpers to decode ModRM/SIB:
|
|
*
|
|
* r/m| AX CX DX BX | SP | BP | SI DI |
|
|
* | R8 R9 R10 R11 | R12 | R13 | R14 R15 |
|
|
* Mod+----------------+-----+-----+---------+
|
|
* 00 | [r/m] |[SIB]|[IP+]| [r/m] |
|
|
* 01 | [r/m + d8] |[S+d]| [r/m + d8] |
|
|
* 10 | [r/m + d32] |[S+D]| [r/m + d32] |
|
|
* 11 | r/ m |
|
|
*/
|
|
|
|
#define mod_is_mem() (modrm_mod != 3)
|
|
#define mod_is_reg() (modrm_mod == 3)
|
|
|
|
#define is_RIP() ((modrm_rm & 7) == CFI_BP && modrm_mod == 0)
|
|
#define have_SIB() ((modrm_rm & 7) == CFI_SP && mod_is_mem())
|
|
|
|
#define rm_is(reg) (have_SIB() ? \
|
|
sib_base == (reg) && sib_index == CFI_SP : \
|
|
modrm_rm == (reg))
|
|
|
|
#define rm_is_mem(reg) (mod_is_mem() && !is_RIP() && rm_is(reg))
|
|
#define rm_is_reg(reg) (mod_is_reg() && modrm_rm == (reg))
|
|
|
|
static bool has_notrack_prefix(struct insn *insn)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < insn->prefixes.nbytes; i++) {
|
|
if (insn->prefixes.bytes[i] == 0x3e)
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
int arch_decode_instruction(struct objtool_file *file, const struct section *sec,
|
|
unsigned long offset, unsigned int maxlen,
|
|
unsigned int *len, enum insn_type *type,
|
|
unsigned long *immediate,
|
|
struct list_head *ops_list)
|
|
{
|
|
const struct elf *elf = file->elf;
|
|
struct insn insn;
|
|
int x86_64, ret;
|
|
unsigned char op1, op2, op3, prefix,
|
|
rex = 0, rex_b = 0, rex_r = 0, rex_w = 0, rex_x = 0,
|
|
modrm = 0, modrm_mod = 0, modrm_rm = 0, modrm_reg = 0,
|
|
sib = 0, /* sib_scale = 0, */ sib_index = 0, sib_base = 0;
|
|
struct stack_op *op = NULL;
|
|
struct symbol *sym;
|
|
u64 imm;
|
|
|
|
x86_64 = is_x86_64(elf);
|
|
if (x86_64 == -1)
|
|
return -1;
|
|
|
|
ret = insn_decode(&insn, sec->data->d_buf + offset, maxlen,
|
|
x86_64 ? INSN_MODE_64 : INSN_MODE_32);
|
|
if (ret < 0) {
|
|
WARN("can't decode instruction at %s:0x%lx", sec->name, offset);
|
|
return -1;
|
|
}
|
|
|
|
*len = insn.length;
|
|
*type = INSN_OTHER;
|
|
|
|
if (insn.vex_prefix.nbytes)
|
|
return 0;
|
|
|
|
prefix = insn.prefixes.bytes[0];
|
|
|
|
op1 = insn.opcode.bytes[0];
|
|
op2 = insn.opcode.bytes[1];
|
|
op3 = insn.opcode.bytes[2];
|
|
|
|
if (insn.rex_prefix.nbytes) {
|
|
rex = insn.rex_prefix.bytes[0];
|
|
rex_w = X86_REX_W(rex) >> 3;
|
|
rex_r = X86_REX_R(rex) >> 2;
|
|
rex_x = X86_REX_X(rex) >> 1;
|
|
rex_b = X86_REX_B(rex);
|
|
}
|
|
|
|
if (insn.modrm.nbytes) {
|
|
modrm = insn.modrm.bytes[0];
|
|
modrm_mod = X86_MODRM_MOD(modrm);
|
|
modrm_reg = X86_MODRM_REG(modrm) + 8*rex_r;
|
|
modrm_rm = X86_MODRM_RM(modrm) + 8*rex_b;
|
|
}
|
|
|
|
if (insn.sib.nbytes) {
|
|
sib = insn.sib.bytes[0];
|
|
/* sib_scale = X86_SIB_SCALE(sib); */
|
|
sib_index = X86_SIB_INDEX(sib) + 8*rex_x;
|
|
sib_base = X86_SIB_BASE(sib) + 8*rex_b;
|
|
}
|
|
|
|
switch (op1) {
|
|
|
|
case 0x1:
|
|
case 0x29:
|
|
if (rex_w && rm_is_reg(CFI_SP)) {
|
|
|
|
/* add/sub reg, %rsp */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_ADD;
|
|
op->src.reg = modrm_reg;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
}
|
|
break;
|
|
|
|
case 0x50 ... 0x57:
|
|
|
|
/* push reg */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = (op1 & 0x7) + 8*rex_b;
|
|
op->dest.type = OP_DEST_PUSH;
|
|
}
|
|
|
|
break;
|
|
|
|
case 0x58 ... 0x5f:
|
|
|
|
/* pop reg */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_POP;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = (op1 & 0x7) + 8*rex_b;
|
|
}
|
|
|
|
break;
|
|
|
|
case 0x68:
|
|
case 0x6a:
|
|
/* push immediate */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_CONST;
|
|
op->dest.type = OP_DEST_PUSH;
|
|
}
|
|
break;
|
|
|
|
case 0x70 ... 0x7f:
|
|
*type = INSN_JUMP_CONDITIONAL;
|
|
break;
|
|
|
|
case 0x80 ... 0x83:
|
|
/*
|
|
* 1000 00sw : mod OP r/m : immediate
|
|
*
|
|
* s - sign extend immediate
|
|
* w - imm8 / imm32
|
|
*
|
|
* OP: 000 ADD 100 AND
|
|
* 001 OR 101 SUB
|
|
* 010 ADC 110 XOR
|
|
* 011 SBB 111 CMP
|
|
*/
|
|
|
|
/* 64bit only */
|
|
if (!rex_w)
|
|
break;
|
|
|
|
/* %rsp target only */
|
|
if (!rm_is_reg(CFI_SP))
|
|
break;
|
|
|
|
imm = insn.immediate.value;
|
|
if (op1 & 2) { /* sign extend */
|
|
if (op1 & 1) { /* imm32 */
|
|
imm <<= 32;
|
|
imm = (s64)imm >> 32;
|
|
} else { /* imm8 */
|
|
imm <<= 56;
|
|
imm = (s64)imm >> 56;
|
|
}
|
|
}
|
|
|
|
switch (modrm_reg & 7) {
|
|
case 5:
|
|
imm = -imm;
|
|
/* fallthrough */
|
|
case 0:
|
|
/* add/sub imm, %rsp */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_ADD;
|
|
op->src.reg = CFI_SP;
|
|
op->src.offset = imm;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
break;
|
|
|
|
case 4:
|
|
/* and imm, %rsp */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_AND;
|
|
op->src.reg = CFI_SP;
|
|
op->src.offset = insn.immediate.value;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
/* WARN ? */
|
|
break;
|
|
}
|
|
|
|
break;
|
|
|
|
case 0x89:
|
|
if (!rex_w)
|
|
break;
|
|
|
|
if (modrm_reg == CFI_SP) {
|
|
|
|
if (mod_is_reg()) {
|
|
/* mov %rsp, reg */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = CFI_SP;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = modrm_rm;
|
|
}
|
|
break;
|
|
|
|
} else {
|
|
/* skip RIP relative displacement */
|
|
if (is_RIP())
|
|
break;
|
|
|
|
/* skip nontrivial SIB */
|
|
if (have_SIB()) {
|
|
modrm_rm = sib_base;
|
|
if (sib_index != CFI_SP)
|
|
break;
|
|
}
|
|
|
|
/* mov %rsp, disp(%reg) */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = CFI_SP;
|
|
op->dest.type = OP_DEST_REG_INDIRECT;
|
|
op->dest.reg = modrm_rm;
|
|
op->dest.offset = insn.displacement.value;
|
|
}
|
|
break;
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
if (rm_is_reg(CFI_SP)) {
|
|
|
|
/* mov reg, %rsp */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = modrm_reg;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
break;
|
|
}
|
|
|
|
/* fallthrough */
|
|
case 0x88:
|
|
if (!rex_w)
|
|
break;
|
|
|
|
if (rm_is_mem(CFI_BP)) {
|
|
|
|
/* mov reg, disp(%rbp) */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = modrm_reg;
|
|
op->dest.type = OP_DEST_REG_INDIRECT;
|
|
op->dest.reg = CFI_BP;
|
|
op->dest.offset = insn.displacement.value;
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (rm_is_mem(CFI_SP)) {
|
|
|
|
/* mov reg, disp(%rsp) */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = modrm_reg;
|
|
op->dest.type = OP_DEST_REG_INDIRECT;
|
|
op->dest.reg = CFI_SP;
|
|
op->dest.offset = insn.displacement.value;
|
|
}
|
|
break;
|
|
}
|
|
|
|
break;
|
|
|
|
case 0x8b:
|
|
if (!rex_w)
|
|
break;
|
|
|
|
if (rm_is_mem(CFI_BP)) {
|
|
|
|
/* mov disp(%rbp), reg */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG_INDIRECT;
|
|
op->src.reg = CFI_BP;
|
|
op->src.offset = insn.displacement.value;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = modrm_reg;
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (rm_is_mem(CFI_SP)) {
|
|
|
|
/* mov disp(%rsp), reg */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG_INDIRECT;
|
|
op->src.reg = CFI_SP;
|
|
op->src.offset = insn.displacement.value;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = modrm_reg;
|
|
}
|
|
break;
|
|
}
|
|
|
|
break;
|
|
|
|
case 0x8d:
|
|
if (mod_is_reg()) {
|
|
WARN("invalid LEA encoding at %s:0x%lx", sec->name, offset);
|
|
break;
|
|
}
|
|
|
|
/* skip non 64bit ops */
|
|
if (!rex_w)
|
|
break;
|
|
|
|
/* skip RIP relative displacement */
|
|
if (is_RIP())
|
|
break;
|
|
|
|
/* skip nontrivial SIB */
|
|
if (have_SIB()) {
|
|
modrm_rm = sib_base;
|
|
if (sib_index != CFI_SP)
|
|
break;
|
|
}
|
|
|
|
/* lea disp(%src), %dst */
|
|
ADD_OP(op) {
|
|
op->src.offset = insn.displacement.value;
|
|
if (!op->src.offset) {
|
|
/* lea (%src), %dst */
|
|
op->src.type = OP_SRC_REG;
|
|
} else {
|
|
/* lea disp(%src), %dst */
|
|
op->src.type = OP_SRC_ADD;
|
|
}
|
|
op->src.reg = modrm_rm;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = modrm_reg;
|
|
}
|
|
break;
|
|
|
|
case 0x8f:
|
|
/* pop to mem */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_POP;
|
|
op->dest.type = OP_DEST_MEM;
|
|
}
|
|
break;
|
|
|
|
case 0x90:
|
|
*type = INSN_NOP;
|
|
break;
|
|
|
|
case 0x9c:
|
|
/* pushf */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_CONST;
|
|
op->dest.type = OP_DEST_PUSHF;
|
|
}
|
|
break;
|
|
|
|
case 0x9d:
|
|
/* popf */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_POPF;
|
|
op->dest.type = OP_DEST_MEM;
|
|
}
|
|
break;
|
|
|
|
case 0x0f:
|
|
|
|
if (op2 == 0x01) {
|
|
|
|
if (modrm == 0xca)
|
|
*type = INSN_CLAC;
|
|
else if (modrm == 0xcb)
|
|
*type = INSN_STAC;
|
|
|
|
} else if (op2 >= 0x80 && op2 <= 0x8f) {
|
|
|
|
*type = INSN_JUMP_CONDITIONAL;
|
|
|
|
} else if (op2 == 0x05 || op2 == 0x07 || op2 == 0x34 ||
|
|
op2 == 0x35) {
|
|
|
|
/* sysenter, sysret */
|
|
*type = INSN_CONTEXT_SWITCH;
|
|
|
|
} else if (op2 == 0x0b || op2 == 0xb9) {
|
|
|
|
/* ud2 */
|
|
*type = INSN_BUG;
|
|
|
|
} else if (op2 == 0x0d || op2 == 0x1f) {
|
|
|
|
/* nopl/nopw */
|
|
*type = INSN_NOP;
|
|
|
|
} else if (op2 == 0x1e) {
|
|
|
|
if (prefix == 0xf3 && (modrm == 0xfa || modrm == 0xfb))
|
|
*type = INSN_ENDBR;
|
|
|
|
|
|
} else if (op2 == 0x38 && op3 == 0xf8) {
|
|
if (insn.prefixes.nbytes == 1 &&
|
|
insn.prefixes.bytes[0] == 0xf2) {
|
|
/* ENQCMD cannot be used in the kernel. */
|
|
WARN("ENQCMD instruction at %s:%lx", sec->name,
|
|
offset);
|
|
}
|
|
|
|
} else if (op2 == 0xa0 || op2 == 0xa8) {
|
|
|
|
/* push fs/gs */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_CONST;
|
|
op->dest.type = OP_DEST_PUSH;
|
|
}
|
|
|
|
} else if (op2 == 0xa1 || op2 == 0xa9) {
|
|
|
|
/* pop fs/gs */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_POP;
|
|
op->dest.type = OP_DEST_MEM;
|
|
}
|
|
}
|
|
|
|
break;
|
|
|
|
case 0xc9:
|
|
/*
|
|
* leave
|
|
*
|
|
* equivalent to:
|
|
* mov bp, sp
|
|
* pop bp
|
|
*/
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = CFI_BP;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_POP;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_BP;
|
|
}
|
|
break;
|
|
|
|
case 0xcc:
|
|
/* int3 */
|
|
*type = INSN_TRAP;
|
|
break;
|
|
|
|
case 0xe3:
|
|
/* jecxz/jrcxz */
|
|
*type = INSN_JUMP_CONDITIONAL;
|
|
break;
|
|
|
|
case 0xe9:
|
|
case 0xeb:
|
|
*type = INSN_JUMP_UNCONDITIONAL;
|
|
break;
|
|
|
|
case 0xc2:
|
|
case 0xc3:
|
|
*type = INSN_RETURN;
|
|
break;
|
|
|
|
case 0xc7: /* mov imm, r/m */
|
|
if (!noinstr)
|
|
break;
|
|
|
|
if (insn.length == 3+4+4 && !strncmp(sec->name, ".init.text", 10)) {
|
|
struct reloc *immr, *disp;
|
|
struct symbol *func;
|
|
int idx;
|
|
|
|
immr = find_reloc_by_dest(elf, (void *)sec, offset+3);
|
|
disp = find_reloc_by_dest(elf, (void *)sec, offset+7);
|
|
|
|
if (!immr || strcmp(immr->sym->name, "pv_ops"))
|
|
break;
|
|
|
|
idx = (immr->addend + 8) / sizeof(void *);
|
|
|
|
func = disp->sym;
|
|
if (disp->sym->type == STT_SECTION)
|
|
func = find_symbol_by_offset(disp->sym->sec, disp->addend);
|
|
if (!func) {
|
|
WARN("no func for pv_ops[]");
|
|
return -1;
|
|
}
|
|
|
|
objtool_pv_add(file, idx, func);
|
|
}
|
|
|
|
break;
|
|
|
|
case 0xcf: /* iret */
|
|
/*
|
|
* Handle sync_core(), which has an IRET to self.
|
|
* All other IRET are in STT_NONE entry code.
|
|
*/
|
|
sym = find_symbol_containing(sec, offset);
|
|
if (sym && sym->type == STT_FUNC) {
|
|
ADD_OP(op) {
|
|
/* add $40, %rsp */
|
|
op->src.type = OP_SRC_ADD;
|
|
op->src.reg = CFI_SP;
|
|
op->src.offset = 5*8;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
break;
|
|
}
|
|
|
|
/* fallthrough */
|
|
|
|
case 0xca: /* retf */
|
|
case 0xcb: /* retf */
|
|
*type = INSN_CONTEXT_SWITCH;
|
|
break;
|
|
|
|
case 0xe8:
|
|
*type = INSN_CALL;
|
|
/*
|
|
* For the impact on the stack, a CALL behaves like
|
|
* a PUSH of an immediate value (the return address).
|
|
*/
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_CONST;
|
|
op->dest.type = OP_DEST_PUSH;
|
|
}
|
|
break;
|
|
|
|
case 0xfc:
|
|
*type = INSN_CLD;
|
|
break;
|
|
|
|
case 0xfd:
|
|
*type = INSN_STD;
|
|
break;
|
|
|
|
case 0xff:
|
|
if (modrm_reg == 2 || modrm_reg == 3) {
|
|
|
|
*type = INSN_CALL_DYNAMIC;
|
|
if (has_notrack_prefix(&insn))
|
|
WARN("notrack prefix found at %s:0x%lx", sec->name, offset);
|
|
|
|
} else if (modrm_reg == 4) {
|
|
|
|
*type = INSN_JUMP_DYNAMIC;
|
|
if (has_notrack_prefix(&insn))
|
|
WARN("notrack prefix found at %s:0x%lx", sec->name, offset);
|
|
|
|
} else if (modrm_reg == 5) {
|
|
|
|
/* jmpf */
|
|
*type = INSN_CONTEXT_SWITCH;
|
|
|
|
} else if (modrm_reg == 6) {
|
|
|
|
/* push from mem */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_CONST;
|
|
op->dest.type = OP_DEST_PUSH;
|
|
}
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
*immediate = insn.immediate.nbytes ? insn.immediate.value : 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void arch_initial_func_cfi_state(struct cfi_init_state *state)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < CFI_NUM_REGS; i++) {
|
|
state->regs[i].base = CFI_UNDEFINED;
|
|
state->regs[i].offset = 0;
|
|
}
|
|
|
|
/* initial CFA (call frame address) */
|
|
state->cfa.base = CFI_SP;
|
|
state->cfa.offset = 8;
|
|
|
|
/* initial RA (return address) */
|
|
state->regs[CFI_RA].base = CFI_CFA;
|
|
state->regs[CFI_RA].offset = -8;
|
|
}
|
|
|
|
const char *arch_nop_insn(int len)
|
|
{
|
|
static const char nops[5][5] = {
|
|
{ BYTES_NOP1 },
|
|
{ BYTES_NOP2 },
|
|
{ BYTES_NOP3 },
|
|
{ BYTES_NOP4 },
|
|
{ BYTES_NOP5 },
|
|
};
|
|
|
|
if (len < 1 || len > 5) {
|
|
WARN("invalid NOP size: %d\n", len);
|
|
return NULL;
|
|
}
|
|
|
|
return nops[len-1];
|
|
}
|
|
|
|
#define BYTE_RET 0xC3
|
|
|
|
const char *arch_ret_insn(int len)
|
|
{
|
|
static const char ret[5][5] = {
|
|
{ BYTE_RET },
|
|
{ BYTE_RET, 0xcc },
|
|
{ BYTE_RET, 0xcc, BYTES_NOP1 },
|
|
{ BYTE_RET, 0xcc, BYTES_NOP2 },
|
|
{ BYTE_RET, 0xcc, BYTES_NOP3 },
|
|
};
|
|
|
|
if (len < 1 || len > 5) {
|
|
WARN("invalid RET size: %d\n", len);
|
|
return NULL;
|
|
}
|
|
|
|
return ret[len-1];
|
|
}
|
|
|
|
int arch_decode_hint_reg(u8 sp_reg, int *base)
|
|
{
|
|
switch (sp_reg) {
|
|
case ORC_REG_UNDEFINED:
|
|
*base = CFI_UNDEFINED;
|
|
break;
|
|
case ORC_REG_SP:
|
|
*base = CFI_SP;
|
|
break;
|
|
case ORC_REG_BP:
|
|
*base = CFI_BP;
|
|
break;
|
|
case ORC_REG_SP_INDIRECT:
|
|
*base = CFI_SP_INDIRECT;
|
|
break;
|
|
case ORC_REG_R10:
|
|
*base = CFI_R10;
|
|
break;
|
|
case ORC_REG_R13:
|
|
*base = CFI_R13;
|
|
break;
|
|
case ORC_REG_DI:
|
|
*base = CFI_DI;
|
|
break;
|
|
case ORC_REG_DX:
|
|
*base = CFI_DX;
|
|
break;
|
|
default:
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
bool arch_is_retpoline(struct symbol *sym)
|
|
{
|
|
return !strncmp(sym->name, "__x86_indirect_", 15);
|
|
}
|