// SPDX-License-Identifier: GPL-2.0-or-later /* * Copyright (C) 2015 Josh Poimboeuf <jpoimboe@redhat.com> */ #include <stdio.h> #include <stdlib.h> #define unlikely(cond) (cond) #include <asm/insn.h> #include "../../../arch/x86/lib/inat.c" #include "../../../arch/x86/lib/insn.c" #define CONFIG_64BIT 1 #include <asm/nops.h> #include <asm/orc_types.h> #include <objtool/check.h> #include <objtool/elf.h> #include <objtool/arch.h> #include <objtool/warn.h> #include <objtool/endianness.h> #include <objtool/builtin.h> #include <arch/elf.h> int arch_ftrace_match(char *name) { return !strcmp(name, "__fentry__"); } static int is_x86_64(const struct elf *elf) { switch (elf->ehdr.e_machine) { case EM_X86_64: return 1; case EM_386: return 0; default: WARN("unexpected ELF machine type %d", elf->ehdr.e_machine); return -1; } } bool arch_callee_saved_reg(unsigned char reg) { switch (reg) { case CFI_BP: case CFI_BX: case CFI_R12: case CFI_R13: case CFI_R14: case CFI_R15: return true; case CFI_AX: case CFI_CX: case CFI_DX: case CFI_SI: case CFI_DI: case CFI_SP: case CFI_R8: case CFI_R9: case CFI_R10: case CFI_R11: case CFI_RA: default: return false; } } unsigned long arch_dest_reloc_offset(int addend) { return addend + 4; } unsigned long arch_jump_destination(struct instruction *insn) { return insn->offset + insn->len + insn->immediate; } bool arch_pc_relative_reloc(struct reloc *reloc) { /* * All relocation types where P (the address of the target) * is included in the computation. */ switch (reloc_type(reloc)) { case R_X86_64_PC8: case R_X86_64_PC16: case R_X86_64_PC32: case R_X86_64_PC64: case R_X86_64_PLT32: case R_X86_64_GOTPC32: case R_X86_64_GOTPCREL: return true; default: break; } return false; } #define ADD_OP(op) \ if (!(op = calloc(1, sizeof(*op)))) \ return -1; \ else for (*ops_list = op, ops_list = &op->next; op; op = NULL) /* * Helpers to decode ModRM/SIB: * * r/m| AX CX DX BX | SP | BP | SI DI | * | R8 R9 R10 R11 | R12 | R13 | R14 R15 | * Mod+----------------+-----+-----+---------+ * 00 | [r/m] |[SIB]|[IP+]| [r/m] | * 01 | [r/m + d8] |[S+d]| [r/m + d8] | * 10 | [r/m + d32] |[S+D]| [r/m + d32] | * 11 | r/ m | */ #define mod_is_mem() (modrm_mod != 3) #define mod_is_reg() (modrm_mod == 3) #define is_RIP() ((modrm_rm & 7) == CFI_BP && modrm_mod == 0) #define have_SIB() ((modrm_rm & 7) == CFI_SP && mod_is_mem()) #define rm_is(reg) (have_SIB() ? \ sib_base == (reg) && sib_index == CFI_SP : \ modrm_rm == (reg)) #define rm_is_mem(reg) (mod_is_mem() && !is_RIP() && rm_is(reg)) #define rm_is_reg(reg) (mod_is_reg() && modrm_rm == (reg)) static bool has_notrack_prefix(struct insn *insn) { int i; for (i = 0; i < insn->prefixes.nbytes; i++) { if (insn->prefixes.bytes[i] == 0x3e) return true; } return false; } int arch_decode_instruction(struct objtool_file *file, const struct section *sec, unsigned long offset, unsigned int maxlen, struct instruction *insn) { struct stack_op **ops_list = &insn->stack_ops; const struct elf *elf = file->elf; struct insn ins; int x86_64, ret; unsigned char op1, op2, op3, prefix, rex = 0, rex_b = 0, rex_r = 0, rex_w = 0, rex_x = 0, modrm = 0, modrm_mod = 0, modrm_rm = 0, modrm_reg = 0, sib = 0, /* sib_scale = 0, */ sib_index = 0, sib_base = 0; struct stack_op *op = NULL; struct symbol *sym; u64 imm; x86_64 = is_x86_64(elf); if (x86_64 == -1) return -1; ret = insn_decode(&ins, sec->data->d_buf + offset, maxlen, x86_64 ? INSN_MODE_64 : INSN_MODE_32); if (ret < 0) { WARN("can't decode instruction at %s:0x%lx", sec->name, offset); return -1; } insn->len = ins.length; insn->type = INSN_OTHER; if (ins.vex_prefix.nbytes) return 0; prefix = ins.prefixes.bytes[0]; op1 = ins.opcode.bytes[0]; op2 = ins.opcode.bytes[1]; op3 = ins.opcode.bytes[2]; if (ins.rex_prefix.nbytes) { rex = ins.rex_prefix.bytes[0]; rex_w = X86_REX_W(rex) >> 3; rex_r = X86_REX_R(rex) >> 2; rex_x = X86_REX_X(rex) >> 1; rex_b = X86_REX_B(rex); } if (ins.modrm.nbytes) { modrm = ins.modrm.bytes[0]; modrm_mod = X86_MODRM_MOD(modrm); modrm_reg = X86_MODRM_REG(modrm) + 8*rex_r; modrm_rm = X86_MODRM_RM(modrm) + 8*rex_b; } if (ins.sib.nbytes) { sib = ins.sib.bytes[0]; /* sib_scale = X86_SIB_SCALE(sib); */ sib_index = X86_SIB_INDEX(sib) + 8*rex_x; sib_base = X86_SIB_BASE(sib) + 8*rex_b; } switch (op1) { case 0x1: case 0x29: if (rex_w && rm_is_reg(CFI_SP)) { /* add/sub reg, %rsp */ ADD_OP(op) { op->src.type = OP_SRC_ADD; op->src.reg = modrm_reg; op->dest.type = OP_DEST_REG; op->dest.reg = CFI_SP; } } break; case 0x50 ... 0x57: /* push reg */ ADD_OP(op) { op->src.type = OP_SRC_REG; op->src.reg = (op1 & 0x7) + 8*rex_b; op->dest.type = OP_DEST_PUSH; } break; case 0x58 ... 0x5f: /* pop reg */ ADD_OP(op) { op->src.type = OP_SRC_POP; op->dest.type = OP_DEST_REG; op->dest.reg = (op1 & 0x7) + 8*rex_b; } break; case 0x68: case 0x6a: /* push immediate */ ADD_OP(op) { op->src.type = OP_SRC_CONST; op->dest.type = OP_DEST_PUSH; } break; case 0x70 ... 0x7f: insn->type = INSN_JUMP_CONDITIONAL; break; case 0x80 ... 0x83: /* * 1000 00sw : mod OP r/m : immediate * * s - sign extend immediate * w - imm8 / imm32 * * OP: 000 ADD 100 AND * 001 OR 101 SUB * 010 ADC 110 XOR * 011 SBB 111 CMP */ /* 64bit only */ if (!rex_w) break; /* %rsp target only */ if (!rm_is_reg(CFI_SP)) break; imm = ins.immediate.value; if (op1 & 2) { /* sign extend */ if (op1 & 1) { /* imm32 */ imm <<= 32; imm = (s64)imm >> 32; } else { /* imm8 */ imm <<= 56; imm = (s64)imm >> 56; } } switch (modrm_reg & 7) { case 5: imm = -imm; /* fallthrough */ case 0: /* add/sub imm, %rsp */ ADD_OP(op) { op->src.type = OP_SRC_ADD; op->src.reg = CFI_SP; op->src.offset = imm; op->dest.type = OP_DEST_REG; op->dest.reg = CFI_SP; } break; case 4: /* and imm, %rsp */ ADD_OP(op) { op->src.type = OP_SRC_AND; op->src.reg = CFI_SP; op->src.offset = ins.immediate.value; op->dest.type = OP_DEST_REG; op->dest.reg = CFI_SP; } break; default: /* WARN ? */ break; } break; case 0x89: if (!rex_w) break; if (modrm_reg == CFI_SP) { if (mod_is_reg()) { /* mov %rsp, reg */ ADD_OP(op) { op->src.type = OP_SRC_REG; op->src.reg = CFI_SP; op->dest.type = OP_DEST_REG; op->dest.reg = modrm_rm; } break; } else { /* skip RIP relative displacement */ if (is_RIP()) break; /* skip nontrivial SIB */ if (have_SIB()) { modrm_rm = sib_base; if (sib_index != CFI_SP) break; } /* mov %rsp, disp(%reg) */ ADD_OP(op) { op->src.type = OP_SRC_REG; op->src.reg = CFI_SP; op->dest.type = OP_DEST_REG_INDIRECT; op->dest.reg = modrm_rm; op->dest.offset = ins.displacement.value; } break; } break; } if (rm_is_reg(CFI_SP)) { /* mov reg, %rsp */ ADD_OP(op) { op->src.type = OP_SRC_REG; op->src.reg = modrm_reg; op->dest.type = OP_DEST_REG; op->dest.reg = CFI_SP; } break; } /* fallthrough */ case 0x88: if (!rex_w) break; if (rm_is_mem(CFI_BP)) { /* mov reg, disp(%rbp) */ ADD_OP(op) { op->src.type = OP_SRC_REG; op->src.reg = modrm_reg; op->dest.type = OP_DEST_REG_INDIRECT; op->dest.reg = CFI_BP; op->dest.offset = ins.displacement.value; } break; } if (rm_is_mem(CFI_SP)) { /* mov reg, disp(%rsp) */ ADD_OP(op) { op->src.type = OP_SRC_REG; op->src.reg = modrm_reg; op->dest.type = OP_DEST_REG_INDIRECT; op->dest.reg = CFI_SP; op->dest.offset = ins.displacement.value; } break; } break; case 0x8b: if (!rex_w) break; if (rm_is_mem(CFI_BP)) { /* mov disp(%rbp), reg */ ADD_OP(op) { op->src.type = OP_SRC_REG_INDIRECT; op->src.reg = CFI_BP; op->src.offset = ins.displacement.value; op->dest.type = OP_DEST_REG; op->dest.reg = modrm_reg; } break; } if (rm_is_mem(CFI_SP)) { /* mov disp(%rsp), reg */ ADD_OP(op) { op->src.type = OP_SRC_REG_INDIRECT; op->src.reg = CFI_SP; op->src.offset = ins.displacement.value; op->dest.type = OP_DEST_REG; op->dest.reg = modrm_reg; } break; } break; case 0x8d: if (mod_is_reg()) { WARN("invalid LEA encoding at %s:0x%lx", sec->name, offset); break; } /* skip non 64bit ops */ if (!rex_w) break; /* skip RIP relative displacement */ if (is_RIP()) break; /* skip nontrivial SIB */ if (have_SIB()) { modrm_rm = sib_base; if (sib_index != CFI_SP) break; } /* lea disp(%src), %dst */ ADD_OP(op) { op->src.offset = ins.displacement.value; if (!op->src.offset) { /* lea (%src), %dst */ op->src.type = OP_SRC_REG; } else { /* lea disp(%src), %dst */ op->src.type = OP_SRC_ADD; } op->src.reg = modrm_rm; op->dest.type = OP_DEST_REG; op->dest.reg = modrm_reg; } break; case 0x8f: /* pop to mem */ ADD_OP(op) { op->src.type = OP_SRC_POP; op->dest.type = OP_DEST_MEM; } break; case 0x90: insn->type = INSN_NOP; break; case 0x9c: /* pushf */ ADD_OP(op) { op->src.type = OP_SRC_CONST; op->dest.type = OP_DEST_PUSHF; } break; case 0x9d: /* popf */ ADD_OP(op) { op->src.type = OP_SRC_POPF; op->dest.type = OP_DEST_MEM; } break; case 0x0f: if (op2 == 0x01) { if (modrm == 0xca) insn->type = INSN_CLAC; else if (modrm == 0xcb) insn->type = INSN_STAC; } else if (op2 >= 0x80 && op2 <= 0x8f) { insn->type = INSN_JUMP_CONDITIONAL; } else if (op2 == 0x05 || op2 == 0x07 || op2 == 0x34 || op2 == 0x35) { /* sysenter, sysret */ insn->type = INSN_CONTEXT_SWITCH; } else if (op2 == 0x0b || op2 == 0xb9) { /* ud2 */ insn->type = INSN_BUG; } else if (op2 == 0x0d || op2 == 0x1f) { /* nopl/nopw */ insn->type = INSN_NOP; } else if (op2 == 0x1e) { if (prefix == 0xf3 && (modrm == 0xfa || modrm == 0xfb)) insn->type = INSN_ENDBR; } else if (op2 == 0x38 && op3 == 0xf8) { if (ins.prefixes.nbytes == 1 && ins.prefixes.bytes[0] == 0xf2) { /* ENQCMD cannot be used in the kernel. */ WARN("ENQCMD instruction at %s:%lx", sec->name, offset); } } else if (op2 == 0xa0 || op2 == 0xa8) { /* push fs/gs */ ADD_OP(op) { op->src.type = OP_SRC_CONST; op->dest.type = OP_DEST_PUSH; } } else if (op2 == 0xa1 || op2 == 0xa9) { /* pop fs/gs */ ADD_OP(op) { op->src.type = OP_SRC_POP; op->dest.type = OP_DEST_MEM; } } break; case 0xc9: /* * leave * * equivalent to: * mov bp, sp * pop bp */ ADD_OP(op) { op->src.type = OP_SRC_REG; op->src.reg = CFI_BP; op->dest.type = OP_DEST_REG; op->dest.reg = CFI_SP; } ADD_OP(op) { op->src.type = OP_SRC_POP; op->dest.type = OP_DEST_REG; op->dest.reg = CFI_BP; } break; case 0xcc: /* int3 */ insn->type = INSN_TRAP; break; case 0xe3: /* jecxz/jrcxz */ insn->type = INSN_JUMP_CONDITIONAL; break; case 0xe9: case 0xeb: insn->type = INSN_JUMP_UNCONDITIONAL; break; case 0xc2: case 0xc3: insn->type = INSN_RETURN; break; case 0xc7: /* mov imm, r/m */ if (!opts.noinstr) break; if (ins.length == 3+4+4 && !strncmp(sec->name, ".init.text", 10)) { struct reloc *immr, *disp; struct symbol *func; int idx; immr = find_reloc_by_dest(elf, (void *)sec, offset+3); disp = find_reloc_by_dest(elf, (void *)sec, offset+7); if (!immr || strcmp(immr->sym->name, "pv_ops")) break; idx = (reloc_addend(immr) + 8) / sizeof(void *); func = disp->sym; if (disp->sym->type == STT_SECTION) func = find_symbol_by_offset(disp->sym->sec, reloc_addend(disp)); if (!func) { WARN("no func for pv_ops[]"); return -1; } objtool_pv_add(file, idx, func); } break; case 0xcf: /* iret */ /* * Handle sync_core(), which has an IRET to self. * All other IRET are in STT_NONE entry code. */ sym = find_symbol_containing(sec, offset); if (sym && sym->type == STT_FUNC) { ADD_OP(op) { /* add $40, %rsp */ op->src.type = OP_SRC_ADD; op->src.reg = CFI_SP; op->src.offset = 5*8; op->dest.type = OP_DEST_REG; op->dest.reg = CFI_SP; } break; } /* fallthrough */ case 0xca: /* retf */ case 0xcb: /* retf */ insn->type = INSN_CONTEXT_SWITCH; break; case 0xe0: /* loopne */ case 0xe1: /* loope */ case 0xe2: /* loop */ insn->type = INSN_JUMP_CONDITIONAL; break; case 0xe8: insn->type = INSN_CALL; /* * For the impact on the stack, a CALL behaves like * a PUSH of an immediate value (the return address). */ ADD_OP(op) { op->src.type = OP_SRC_CONST; op->dest.type = OP_DEST_PUSH; } break; case 0xfc: insn->type = INSN_CLD; break; case 0xfd: insn->type = INSN_STD; break; case 0xff: if (modrm_reg == 2 || modrm_reg == 3) { insn->type = INSN_CALL_DYNAMIC; if (has_notrack_prefix(&ins)) WARN("notrack prefix found at %s:0x%lx", sec->name, offset); } else if (modrm_reg == 4) { insn->type = INSN_JUMP_DYNAMIC; if (has_notrack_prefix(&ins)) WARN("notrack prefix found at %s:0x%lx", sec->name, offset); } else if (modrm_reg == 5) { /* jmpf */ insn->type = INSN_CONTEXT_SWITCH; } else if (modrm_reg == 6) { /* push from mem */ ADD_OP(op) { op->src.type = OP_SRC_CONST; op->dest.type = OP_DEST_PUSH; } } break; default: break; } insn->immediate = ins.immediate.nbytes ? ins.immediate.value : 0; return 0; } void arch_initial_func_cfi_state(struct cfi_init_state *state) { int i; for (i = 0; i < CFI_NUM_REGS; i++) { state->regs[i].base = CFI_UNDEFINED; state->regs[i].offset = 0; } /* initial CFA (call frame address) */ state->cfa.base = CFI_SP; state->cfa.offset = 8; /* initial RA (return address) */ state->regs[CFI_RA].base = CFI_CFA; state->regs[CFI_RA].offset = -8; } const char *arch_nop_insn(int len) { static const char nops[5][5] = { { BYTES_NOP1 }, { BYTES_NOP2 }, { BYTES_NOP3 }, { BYTES_NOP4 }, { BYTES_NOP5 }, }; if (len < 1 || len > 5) { WARN("invalid NOP size: %d\n", len); return NULL; } return nops[len-1]; } #define BYTE_RET 0xC3 const char *arch_ret_insn(int len) { static const char ret[5][5] = { { BYTE_RET }, { BYTE_RET, 0xcc }, { BYTE_RET, 0xcc, BYTES_NOP1 }, { BYTE_RET, 0xcc, BYTES_NOP2 }, { BYTE_RET, 0xcc, BYTES_NOP3 }, }; if (len < 1 || len > 5) { WARN("invalid RET size: %d\n", len); return NULL; } return ret[len-1]; } int arch_decode_hint_reg(u8 sp_reg, int *base) { switch (sp_reg) { case ORC_REG_UNDEFINED: *base = CFI_UNDEFINED; break; case ORC_REG_SP: *base = CFI_SP; break; case ORC_REG_BP: *base = CFI_BP; break; case ORC_REG_SP_INDIRECT: *base = CFI_SP_INDIRECT; break; case ORC_REG_R10: *base = CFI_R10; break; case ORC_REG_R13: *base = CFI_R13; break; case ORC_REG_DI: *base = CFI_DI; break; case ORC_REG_DX: *base = CFI_DX; break; default: return -1; } return 0; } bool arch_is_retpoline(struct symbol *sym) { return !strncmp(sym->name, "__x86_indirect_", 15); } bool arch_is_rethunk(struct symbol *sym) { return !strcmp(sym->name, "__x86_return_thunk"); } bool arch_is_embedded_insn(struct symbol *sym) { return !strcmp(sym->name, "retbleed_return_thunk") || !strcmp(sym->name, "srso_safe_ret"); }