- Add powerpc qspinlock implementation optimised for large system scalability and paravirt. See the merge message for more details. - Enable objtool to be built on powerpc to generate mcount locations. - Use a temporary mm for code patching with the Radix MMU, so the writable mapping is restricted to the patching CPU. - Add an option to build the 64-bit big-endian kernel with the ELFv2 ABI. - Sanitise user registers on interrupt entry on 64-bit Book3S. - Many other small features and fixes. Thanks to: Aboorva Devarajan, Angel Iglesias, Benjamin Gray, Bjorn Helgaas, Bo Liu, Chen Lifu, Christoph Hellwig, Christophe JAILLET, Christophe Leroy, Christopher M. Riedl, Colin Ian King, Deming Wang, Disha Goel, Dmitry Torokhov, Finn Thain, Geert Uytterhoeven, Gustavo A. R. Silva, Haowen Bai, Joel Stanley, Jordan Niethe, Julia Lawall, Kajol Jain, Laurent Dufour, Li zeming, Miaoqian Lin, Michael Jeanson, Nathan Lynch, Naveen N. Rao, Nayna Jain, Nicholas Miehlbradt, Nicholas Piggin, Pali Rohár, Randy Dunlap, Rohan McLure, Russell Currey, Sathvika Vasireddy, Shaomin Deng, Stephen Kitt, Stephen Rothwell, Thomas Weißschuh, Tiezhu Yang, Uwe Kleine-König, Xie Shaowen, Xiu Jianfeng, XueBing Chen, Yang Yingliang, Zhang Jiaming, ruanjinjie, Jessica Yu, Wolfram Sang. -----BEGIN PGP SIGNATURE----- iQJHBAABCAAxFiEEJFGtCPCthwEv2Y/bUevqPMjhpYAFAmOfrj8THG1wZUBlbGxl cm1hbi5pZC5hdQAKCRBR6+o8yOGlgIWtD/9mGF/ze2k+qFTo+30fb7bO8WJIDgsR dIASnZjXV7q/45elvymhUdkQv4R7xL3pzC40P1+ZKtWzGTNe+zWUQLoALNwRK85j 8CsxZbqefGNKE5Z6ZHo9s37wsu3+jJu9yEQpGFo1LINyzeclCn5St5oqfRam+Hd/ cPF+VfvREwZ0+YOKGBhJ2EgC+Gc9xsFY7DLQsoYlu71iZZr6Z6rgZW/EY5h3RMGS YKBoVwDsWaU0FpFWrr/rYTI6DqSr3AHr1+ftDg7ncCZMD6vQva6aMCCt94aLB1aE vC+DNdhZlA558bXGa5yA7Wr//7aUBUIwyC60DogOeZ6vw3kD9tdEd1fbH5hmqNKY K5bfqm28XU2959CTE8RDgsYYZvwDcfrjBIML14WZGdCQOTcGKpgOGp22o6yNb1Pq JKpHHnVpvu2PZ/p2XdKSm9+etr2yI6lXZAEVTS7ehdtMukButjSHEVbSCEZ8tlWz KokQt2J23BMHuSrXK6+67wWQBtdsLEk+LBOQmweiwarMocqvL/Zjz/5J7DR2DtH8 wlY3wOtB1+E5j7xZ+RgK3c3jNg5dH39ZwvFsSATWTI3P+iq6OK/bbk4q4LmZt2l9 ZIfH/CXPf9BvGCHzHa3AAd3UBbJLFwj17btMEv1wFVPS0T4LPUzkgTNTNUYeP6zL h1e5QfgUxvKPuQ== =7k3p -----END PGP SIGNATURE----- Merge tag 'powerpc-6.2-1' of git://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux Pull powerpc updates from Michael Ellerman: - Add powerpc qspinlock implementation optimised for large system scalability and paravirt. See the merge message for more details - Enable objtool to be built on powerpc to generate mcount locations - Use a temporary mm for code patching with the Radix MMU, so the writable mapping is restricted to the patching CPU - Add an option to build the 64-bit big-endian kernel with the ELFv2 ABI - Sanitise user registers on interrupt entry on 64-bit Book3S - Many other small features and fixes Thanks to Aboorva Devarajan, Angel Iglesias, Benjamin Gray, Bjorn Helgaas, Bo Liu, Chen Lifu, Christoph Hellwig, Christophe JAILLET, Christophe Leroy, Christopher M. Riedl, Colin Ian King, Deming Wang, Disha Goel, Dmitry Torokhov, Finn Thain, Geert Uytterhoeven, Gustavo A. R. Silva, Haowen Bai, Joel Stanley, Jordan Niethe, Julia Lawall, Kajol Jain, Laurent Dufour, Li zeming, Miaoqian Lin, Michael Jeanson, Nathan Lynch, Naveen N. Rao, Nayna Jain, Nicholas Miehlbradt, Nicholas Piggin, Pali Rohár, Randy Dunlap, Rohan McLure, Russell Currey, Sathvika Vasireddy, Shaomin Deng, Stephen Kitt, Stephen Rothwell, Thomas Weißschuh, Tiezhu Yang, Uwe Kleine-König, Xie Shaowen, Xiu Jianfeng, XueBing Chen, Yang Yingliang, Zhang Jiaming, ruanjinjie, Jessica Yu, and Wolfram Sang. * tag 'powerpc-6.2-1' of git://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux: (181 commits) powerpc/code-patching: Fix oops with DEBUG_VM enabled powerpc/qspinlock: Fix 32-bit build powerpc/prom: Fix 32-bit build powerpc/rtas: mandate RTAS syscall filtering powerpc/rtas: define pr_fmt and convert printk call sites powerpc/rtas: clean up includes powerpc/rtas: clean up rtas_error_log_max initialization powerpc/pseries/eeh: use correct API for error log size powerpc/rtas: avoid scheduling in rtas_os_term() powerpc/rtas: avoid device tree lookups in rtas_os_term() powerpc/rtasd: use correct OF API for event scan rate powerpc/rtas: document rtas_call() powerpc/pseries: unregister VPA when hot unplugging a CPU powerpc/pseries: reset the RCU watchdogs after a LPM powerpc: Take in account addition CPU node when building kexec FDT powerpc: export the CPU node count powerpc/cpuidle: Set CPUIDLE_FLAG_POLLING for snooze state powerpc/dts/fsl: Fix pca954x i2c-mux node names cxl: Remove unnecessary cxl_pci_window_alignment() selftests/powerpc: Fix resource leaks ...
829 lines
15 KiB
C
829 lines
15 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Copyright (C) 2015 Josh Poimboeuf <jpoimboe@redhat.com>
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
|
|
#define unlikely(cond) (cond)
|
|
#include <asm/insn.h>
|
|
#include "../../../arch/x86/lib/inat.c"
|
|
#include "../../../arch/x86/lib/insn.c"
|
|
|
|
#define CONFIG_64BIT 1
|
|
#include <asm/nops.h>
|
|
|
|
#include <asm/orc_types.h>
|
|
#include <objtool/check.h>
|
|
#include <objtool/elf.h>
|
|
#include <objtool/arch.h>
|
|
#include <objtool/warn.h>
|
|
#include <objtool/endianness.h>
|
|
#include <objtool/builtin.h>
|
|
#include <arch/elf.h>
|
|
|
|
int arch_ftrace_match(char *name)
|
|
{
|
|
return !strcmp(name, "__fentry__");
|
|
}
|
|
|
|
static int is_x86_64(const struct elf *elf)
|
|
{
|
|
switch (elf->ehdr.e_machine) {
|
|
case EM_X86_64:
|
|
return 1;
|
|
case EM_386:
|
|
return 0;
|
|
default:
|
|
WARN("unexpected ELF machine type %d", elf->ehdr.e_machine);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
bool arch_callee_saved_reg(unsigned char reg)
|
|
{
|
|
switch (reg) {
|
|
case CFI_BP:
|
|
case CFI_BX:
|
|
case CFI_R12:
|
|
case CFI_R13:
|
|
case CFI_R14:
|
|
case CFI_R15:
|
|
return true;
|
|
|
|
case CFI_AX:
|
|
case CFI_CX:
|
|
case CFI_DX:
|
|
case CFI_SI:
|
|
case CFI_DI:
|
|
case CFI_SP:
|
|
case CFI_R8:
|
|
case CFI_R9:
|
|
case CFI_R10:
|
|
case CFI_R11:
|
|
case CFI_RA:
|
|
default:
|
|
return false;
|
|
}
|
|
}
|
|
|
|
unsigned long arch_dest_reloc_offset(int addend)
|
|
{
|
|
return addend + 4;
|
|
}
|
|
|
|
unsigned long arch_jump_destination(struct instruction *insn)
|
|
{
|
|
return insn->offset + insn->len + insn->immediate;
|
|
}
|
|
|
|
bool arch_pc_relative_reloc(struct reloc *reloc)
|
|
{
|
|
/*
|
|
* All relocation types where P (the address of the target)
|
|
* is included in the computation.
|
|
*/
|
|
switch (reloc->type) {
|
|
case R_X86_64_PC8:
|
|
case R_X86_64_PC16:
|
|
case R_X86_64_PC32:
|
|
case R_X86_64_PC64:
|
|
|
|
case R_X86_64_PLT32:
|
|
case R_X86_64_GOTPC32:
|
|
case R_X86_64_GOTPCREL:
|
|
return true;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
#define ADD_OP(op) \
|
|
if (!(op = calloc(1, sizeof(*op)))) \
|
|
return -1; \
|
|
else for (list_add_tail(&op->list, ops_list); op; op = NULL)
|
|
|
|
/*
|
|
* Helpers to decode ModRM/SIB:
|
|
*
|
|
* r/m| AX CX DX BX | SP | BP | SI DI |
|
|
* | R8 R9 R10 R11 | R12 | R13 | R14 R15 |
|
|
* Mod+----------------+-----+-----+---------+
|
|
* 00 | [r/m] |[SIB]|[IP+]| [r/m] |
|
|
* 01 | [r/m + d8] |[S+d]| [r/m + d8] |
|
|
* 10 | [r/m + d32] |[S+D]| [r/m + d32] |
|
|
* 11 | r/ m |
|
|
*/
|
|
|
|
#define mod_is_mem() (modrm_mod != 3)
|
|
#define mod_is_reg() (modrm_mod == 3)
|
|
|
|
#define is_RIP() ((modrm_rm & 7) == CFI_BP && modrm_mod == 0)
|
|
#define have_SIB() ((modrm_rm & 7) == CFI_SP && mod_is_mem())
|
|
|
|
#define rm_is(reg) (have_SIB() ? \
|
|
sib_base == (reg) && sib_index == CFI_SP : \
|
|
modrm_rm == (reg))
|
|
|
|
#define rm_is_mem(reg) (mod_is_mem() && !is_RIP() && rm_is(reg))
|
|
#define rm_is_reg(reg) (mod_is_reg() && modrm_rm == (reg))
|
|
|
|
static bool has_notrack_prefix(struct insn *insn)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < insn->prefixes.nbytes; i++) {
|
|
if (insn->prefixes.bytes[i] == 0x3e)
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
int arch_decode_instruction(struct objtool_file *file, const struct section *sec,
|
|
unsigned long offset, unsigned int maxlen,
|
|
unsigned int *len, enum insn_type *type,
|
|
unsigned long *immediate,
|
|
struct list_head *ops_list)
|
|
{
|
|
const struct elf *elf = file->elf;
|
|
struct insn insn;
|
|
int x86_64, ret;
|
|
unsigned char op1, op2, op3, prefix,
|
|
rex = 0, rex_b = 0, rex_r = 0, rex_w = 0, rex_x = 0,
|
|
modrm = 0, modrm_mod = 0, modrm_rm = 0, modrm_reg = 0,
|
|
sib = 0, /* sib_scale = 0, */ sib_index = 0, sib_base = 0;
|
|
struct stack_op *op = NULL;
|
|
struct symbol *sym;
|
|
u64 imm;
|
|
|
|
x86_64 = is_x86_64(elf);
|
|
if (x86_64 == -1)
|
|
return -1;
|
|
|
|
ret = insn_decode(&insn, sec->data->d_buf + offset, maxlen,
|
|
x86_64 ? INSN_MODE_64 : INSN_MODE_32);
|
|
if (ret < 0) {
|
|
WARN("can't decode instruction at %s:0x%lx", sec->name, offset);
|
|
return -1;
|
|
}
|
|
|
|
*len = insn.length;
|
|
*type = INSN_OTHER;
|
|
|
|
if (insn.vex_prefix.nbytes)
|
|
return 0;
|
|
|
|
prefix = insn.prefixes.bytes[0];
|
|
|
|
op1 = insn.opcode.bytes[0];
|
|
op2 = insn.opcode.bytes[1];
|
|
op3 = insn.opcode.bytes[2];
|
|
|
|
if (insn.rex_prefix.nbytes) {
|
|
rex = insn.rex_prefix.bytes[0];
|
|
rex_w = X86_REX_W(rex) >> 3;
|
|
rex_r = X86_REX_R(rex) >> 2;
|
|
rex_x = X86_REX_X(rex) >> 1;
|
|
rex_b = X86_REX_B(rex);
|
|
}
|
|
|
|
if (insn.modrm.nbytes) {
|
|
modrm = insn.modrm.bytes[0];
|
|
modrm_mod = X86_MODRM_MOD(modrm);
|
|
modrm_reg = X86_MODRM_REG(modrm) + 8*rex_r;
|
|
modrm_rm = X86_MODRM_RM(modrm) + 8*rex_b;
|
|
}
|
|
|
|
if (insn.sib.nbytes) {
|
|
sib = insn.sib.bytes[0];
|
|
/* sib_scale = X86_SIB_SCALE(sib); */
|
|
sib_index = X86_SIB_INDEX(sib) + 8*rex_x;
|
|
sib_base = X86_SIB_BASE(sib) + 8*rex_b;
|
|
}
|
|
|
|
switch (op1) {
|
|
|
|
case 0x1:
|
|
case 0x29:
|
|
if (rex_w && rm_is_reg(CFI_SP)) {
|
|
|
|
/* add/sub reg, %rsp */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_ADD;
|
|
op->src.reg = modrm_reg;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
}
|
|
break;
|
|
|
|
case 0x50 ... 0x57:
|
|
|
|
/* push reg */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = (op1 & 0x7) + 8*rex_b;
|
|
op->dest.type = OP_DEST_PUSH;
|
|
}
|
|
|
|
break;
|
|
|
|
case 0x58 ... 0x5f:
|
|
|
|
/* pop reg */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_POP;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = (op1 & 0x7) + 8*rex_b;
|
|
}
|
|
|
|
break;
|
|
|
|
case 0x68:
|
|
case 0x6a:
|
|
/* push immediate */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_CONST;
|
|
op->dest.type = OP_DEST_PUSH;
|
|
}
|
|
break;
|
|
|
|
case 0x70 ... 0x7f:
|
|
*type = INSN_JUMP_CONDITIONAL;
|
|
break;
|
|
|
|
case 0x80 ... 0x83:
|
|
/*
|
|
* 1000 00sw : mod OP r/m : immediate
|
|
*
|
|
* s - sign extend immediate
|
|
* w - imm8 / imm32
|
|
*
|
|
* OP: 000 ADD 100 AND
|
|
* 001 OR 101 SUB
|
|
* 010 ADC 110 XOR
|
|
* 011 SBB 111 CMP
|
|
*/
|
|
|
|
/* 64bit only */
|
|
if (!rex_w)
|
|
break;
|
|
|
|
/* %rsp target only */
|
|
if (!rm_is_reg(CFI_SP))
|
|
break;
|
|
|
|
imm = insn.immediate.value;
|
|
if (op1 & 2) { /* sign extend */
|
|
if (op1 & 1) { /* imm32 */
|
|
imm <<= 32;
|
|
imm = (s64)imm >> 32;
|
|
} else { /* imm8 */
|
|
imm <<= 56;
|
|
imm = (s64)imm >> 56;
|
|
}
|
|
}
|
|
|
|
switch (modrm_reg & 7) {
|
|
case 5:
|
|
imm = -imm;
|
|
/* fallthrough */
|
|
case 0:
|
|
/* add/sub imm, %rsp */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_ADD;
|
|
op->src.reg = CFI_SP;
|
|
op->src.offset = imm;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
break;
|
|
|
|
case 4:
|
|
/* and imm, %rsp */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_AND;
|
|
op->src.reg = CFI_SP;
|
|
op->src.offset = insn.immediate.value;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
/* WARN ? */
|
|
break;
|
|
}
|
|
|
|
break;
|
|
|
|
case 0x89:
|
|
if (!rex_w)
|
|
break;
|
|
|
|
if (modrm_reg == CFI_SP) {
|
|
|
|
if (mod_is_reg()) {
|
|
/* mov %rsp, reg */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = CFI_SP;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = modrm_rm;
|
|
}
|
|
break;
|
|
|
|
} else {
|
|
/* skip RIP relative displacement */
|
|
if (is_RIP())
|
|
break;
|
|
|
|
/* skip nontrivial SIB */
|
|
if (have_SIB()) {
|
|
modrm_rm = sib_base;
|
|
if (sib_index != CFI_SP)
|
|
break;
|
|
}
|
|
|
|
/* mov %rsp, disp(%reg) */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = CFI_SP;
|
|
op->dest.type = OP_DEST_REG_INDIRECT;
|
|
op->dest.reg = modrm_rm;
|
|
op->dest.offset = insn.displacement.value;
|
|
}
|
|
break;
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
if (rm_is_reg(CFI_SP)) {
|
|
|
|
/* mov reg, %rsp */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = modrm_reg;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
break;
|
|
}
|
|
|
|
/* fallthrough */
|
|
case 0x88:
|
|
if (!rex_w)
|
|
break;
|
|
|
|
if (rm_is_mem(CFI_BP)) {
|
|
|
|
/* mov reg, disp(%rbp) */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = modrm_reg;
|
|
op->dest.type = OP_DEST_REG_INDIRECT;
|
|
op->dest.reg = CFI_BP;
|
|
op->dest.offset = insn.displacement.value;
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (rm_is_mem(CFI_SP)) {
|
|
|
|
/* mov reg, disp(%rsp) */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = modrm_reg;
|
|
op->dest.type = OP_DEST_REG_INDIRECT;
|
|
op->dest.reg = CFI_SP;
|
|
op->dest.offset = insn.displacement.value;
|
|
}
|
|
break;
|
|
}
|
|
|
|
break;
|
|
|
|
case 0x8b:
|
|
if (!rex_w)
|
|
break;
|
|
|
|
if (rm_is_mem(CFI_BP)) {
|
|
|
|
/* mov disp(%rbp), reg */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG_INDIRECT;
|
|
op->src.reg = CFI_BP;
|
|
op->src.offset = insn.displacement.value;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = modrm_reg;
|
|
}
|
|
break;
|
|
}
|
|
|
|
if (rm_is_mem(CFI_SP)) {
|
|
|
|
/* mov disp(%rsp), reg */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG_INDIRECT;
|
|
op->src.reg = CFI_SP;
|
|
op->src.offset = insn.displacement.value;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = modrm_reg;
|
|
}
|
|
break;
|
|
}
|
|
|
|
break;
|
|
|
|
case 0x8d:
|
|
if (mod_is_reg()) {
|
|
WARN("invalid LEA encoding at %s:0x%lx", sec->name, offset);
|
|
break;
|
|
}
|
|
|
|
/* skip non 64bit ops */
|
|
if (!rex_w)
|
|
break;
|
|
|
|
/* skip RIP relative displacement */
|
|
if (is_RIP())
|
|
break;
|
|
|
|
/* skip nontrivial SIB */
|
|
if (have_SIB()) {
|
|
modrm_rm = sib_base;
|
|
if (sib_index != CFI_SP)
|
|
break;
|
|
}
|
|
|
|
/* lea disp(%src), %dst */
|
|
ADD_OP(op) {
|
|
op->src.offset = insn.displacement.value;
|
|
if (!op->src.offset) {
|
|
/* lea (%src), %dst */
|
|
op->src.type = OP_SRC_REG;
|
|
} else {
|
|
/* lea disp(%src), %dst */
|
|
op->src.type = OP_SRC_ADD;
|
|
}
|
|
op->src.reg = modrm_rm;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = modrm_reg;
|
|
}
|
|
break;
|
|
|
|
case 0x8f:
|
|
/* pop to mem */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_POP;
|
|
op->dest.type = OP_DEST_MEM;
|
|
}
|
|
break;
|
|
|
|
case 0x90:
|
|
*type = INSN_NOP;
|
|
break;
|
|
|
|
case 0x9c:
|
|
/* pushf */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_CONST;
|
|
op->dest.type = OP_DEST_PUSHF;
|
|
}
|
|
break;
|
|
|
|
case 0x9d:
|
|
/* popf */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_POPF;
|
|
op->dest.type = OP_DEST_MEM;
|
|
}
|
|
break;
|
|
|
|
case 0x0f:
|
|
|
|
if (op2 == 0x01) {
|
|
|
|
if (modrm == 0xca)
|
|
*type = INSN_CLAC;
|
|
else if (modrm == 0xcb)
|
|
*type = INSN_STAC;
|
|
|
|
} else if (op2 >= 0x80 && op2 <= 0x8f) {
|
|
|
|
*type = INSN_JUMP_CONDITIONAL;
|
|
|
|
} else if (op2 == 0x05 || op2 == 0x07 || op2 == 0x34 ||
|
|
op2 == 0x35) {
|
|
|
|
/* sysenter, sysret */
|
|
*type = INSN_CONTEXT_SWITCH;
|
|
|
|
} else if (op2 == 0x0b || op2 == 0xb9) {
|
|
|
|
/* ud2 */
|
|
*type = INSN_BUG;
|
|
|
|
} else if (op2 == 0x0d || op2 == 0x1f) {
|
|
|
|
/* nopl/nopw */
|
|
*type = INSN_NOP;
|
|
|
|
} else if (op2 == 0x1e) {
|
|
|
|
if (prefix == 0xf3 && (modrm == 0xfa || modrm == 0xfb))
|
|
*type = INSN_ENDBR;
|
|
|
|
|
|
} else if (op2 == 0x38 && op3 == 0xf8) {
|
|
if (insn.prefixes.nbytes == 1 &&
|
|
insn.prefixes.bytes[0] == 0xf2) {
|
|
/* ENQCMD cannot be used in the kernel. */
|
|
WARN("ENQCMD instruction at %s:%lx", sec->name,
|
|
offset);
|
|
}
|
|
|
|
} else if (op2 == 0xa0 || op2 == 0xa8) {
|
|
|
|
/* push fs/gs */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_CONST;
|
|
op->dest.type = OP_DEST_PUSH;
|
|
}
|
|
|
|
} else if (op2 == 0xa1 || op2 == 0xa9) {
|
|
|
|
/* pop fs/gs */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_POP;
|
|
op->dest.type = OP_DEST_MEM;
|
|
}
|
|
}
|
|
|
|
break;
|
|
|
|
case 0xc9:
|
|
/*
|
|
* leave
|
|
*
|
|
* equivalent to:
|
|
* mov bp, sp
|
|
* pop bp
|
|
*/
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_REG;
|
|
op->src.reg = CFI_BP;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_POP;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_BP;
|
|
}
|
|
break;
|
|
|
|
case 0xcc:
|
|
/* int3 */
|
|
*type = INSN_TRAP;
|
|
break;
|
|
|
|
case 0xe3:
|
|
/* jecxz/jrcxz */
|
|
*type = INSN_JUMP_CONDITIONAL;
|
|
break;
|
|
|
|
case 0xe9:
|
|
case 0xeb:
|
|
*type = INSN_JUMP_UNCONDITIONAL;
|
|
break;
|
|
|
|
case 0xc2:
|
|
case 0xc3:
|
|
*type = INSN_RETURN;
|
|
break;
|
|
|
|
case 0xc7: /* mov imm, r/m */
|
|
if (!opts.noinstr)
|
|
break;
|
|
|
|
if (insn.length == 3+4+4 && !strncmp(sec->name, ".init.text", 10)) {
|
|
struct reloc *immr, *disp;
|
|
struct symbol *func;
|
|
int idx;
|
|
|
|
immr = find_reloc_by_dest(elf, (void *)sec, offset+3);
|
|
disp = find_reloc_by_dest(elf, (void *)sec, offset+7);
|
|
|
|
if (!immr || strcmp(immr->sym->name, "pv_ops"))
|
|
break;
|
|
|
|
idx = (immr->addend + 8) / sizeof(void *);
|
|
|
|
func = disp->sym;
|
|
if (disp->sym->type == STT_SECTION)
|
|
func = find_symbol_by_offset(disp->sym->sec, disp->addend);
|
|
if (!func) {
|
|
WARN("no func for pv_ops[]");
|
|
return -1;
|
|
}
|
|
|
|
objtool_pv_add(file, idx, func);
|
|
}
|
|
|
|
break;
|
|
|
|
case 0xcf: /* iret */
|
|
/*
|
|
* Handle sync_core(), which has an IRET to self.
|
|
* All other IRET are in STT_NONE entry code.
|
|
*/
|
|
sym = find_symbol_containing(sec, offset);
|
|
if (sym && sym->type == STT_FUNC) {
|
|
ADD_OP(op) {
|
|
/* add $40, %rsp */
|
|
op->src.type = OP_SRC_ADD;
|
|
op->src.reg = CFI_SP;
|
|
op->src.offset = 5*8;
|
|
op->dest.type = OP_DEST_REG;
|
|
op->dest.reg = CFI_SP;
|
|
}
|
|
break;
|
|
}
|
|
|
|
/* fallthrough */
|
|
|
|
case 0xca: /* retf */
|
|
case 0xcb: /* retf */
|
|
*type = INSN_CONTEXT_SWITCH;
|
|
break;
|
|
|
|
case 0xe0: /* loopne */
|
|
case 0xe1: /* loope */
|
|
case 0xe2: /* loop */
|
|
*type = INSN_JUMP_CONDITIONAL;
|
|
break;
|
|
|
|
case 0xe8:
|
|
*type = INSN_CALL;
|
|
/*
|
|
* For the impact on the stack, a CALL behaves like
|
|
* a PUSH of an immediate value (the return address).
|
|
*/
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_CONST;
|
|
op->dest.type = OP_DEST_PUSH;
|
|
}
|
|
break;
|
|
|
|
case 0xfc:
|
|
*type = INSN_CLD;
|
|
break;
|
|
|
|
case 0xfd:
|
|
*type = INSN_STD;
|
|
break;
|
|
|
|
case 0xff:
|
|
if (modrm_reg == 2 || modrm_reg == 3) {
|
|
|
|
*type = INSN_CALL_DYNAMIC;
|
|
if (has_notrack_prefix(&insn))
|
|
WARN("notrack prefix found at %s:0x%lx", sec->name, offset);
|
|
|
|
} else if (modrm_reg == 4) {
|
|
|
|
*type = INSN_JUMP_DYNAMIC;
|
|
if (has_notrack_prefix(&insn))
|
|
WARN("notrack prefix found at %s:0x%lx", sec->name, offset);
|
|
|
|
} else if (modrm_reg == 5) {
|
|
|
|
/* jmpf */
|
|
*type = INSN_CONTEXT_SWITCH;
|
|
|
|
} else if (modrm_reg == 6) {
|
|
|
|
/* push from mem */
|
|
ADD_OP(op) {
|
|
op->src.type = OP_SRC_CONST;
|
|
op->dest.type = OP_DEST_PUSH;
|
|
}
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
*immediate = insn.immediate.nbytes ? insn.immediate.value : 0;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void arch_initial_func_cfi_state(struct cfi_init_state *state)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < CFI_NUM_REGS; i++) {
|
|
state->regs[i].base = CFI_UNDEFINED;
|
|
state->regs[i].offset = 0;
|
|
}
|
|
|
|
/* initial CFA (call frame address) */
|
|
state->cfa.base = CFI_SP;
|
|
state->cfa.offset = 8;
|
|
|
|
/* initial RA (return address) */
|
|
state->regs[CFI_RA].base = CFI_CFA;
|
|
state->regs[CFI_RA].offset = -8;
|
|
}
|
|
|
|
const char *arch_nop_insn(int len)
|
|
{
|
|
static const char nops[5][5] = {
|
|
{ BYTES_NOP1 },
|
|
{ BYTES_NOP2 },
|
|
{ BYTES_NOP3 },
|
|
{ BYTES_NOP4 },
|
|
{ BYTES_NOP5 },
|
|
};
|
|
|
|
if (len < 1 || len > 5) {
|
|
WARN("invalid NOP size: %d\n", len);
|
|
return NULL;
|
|
}
|
|
|
|
return nops[len-1];
|
|
}
|
|
|
|
#define BYTE_RET 0xC3
|
|
|
|
const char *arch_ret_insn(int len)
|
|
{
|
|
static const char ret[5][5] = {
|
|
{ BYTE_RET },
|
|
{ BYTE_RET, 0xcc },
|
|
{ BYTE_RET, 0xcc, BYTES_NOP1 },
|
|
{ BYTE_RET, 0xcc, BYTES_NOP2 },
|
|
{ BYTE_RET, 0xcc, BYTES_NOP3 },
|
|
};
|
|
|
|
if (len < 1 || len > 5) {
|
|
WARN("invalid RET size: %d\n", len);
|
|
return NULL;
|
|
}
|
|
|
|
return ret[len-1];
|
|
}
|
|
|
|
int arch_decode_hint_reg(u8 sp_reg, int *base)
|
|
{
|
|
switch (sp_reg) {
|
|
case ORC_REG_UNDEFINED:
|
|
*base = CFI_UNDEFINED;
|
|
break;
|
|
case ORC_REG_SP:
|
|
*base = CFI_SP;
|
|
break;
|
|
case ORC_REG_BP:
|
|
*base = CFI_BP;
|
|
break;
|
|
case ORC_REG_SP_INDIRECT:
|
|
*base = CFI_SP_INDIRECT;
|
|
break;
|
|
case ORC_REG_R10:
|
|
*base = CFI_R10;
|
|
break;
|
|
case ORC_REG_R13:
|
|
*base = CFI_R13;
|
|
break;
|
|
case ORC_REG_DI:
|
|
*base = CFI_DI;
|
|
break;
|
|
case ORC_REG_DX:
|
|
*base = CFI_DX;
|
|
break;
|
|
default:
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
bool arch_is_retpoline(struct symbol *sym)
|
|
{
|
|
return !strncmp(sym->name, "__x86_indirect_", 15);
|
|
}
|
|
|
|
bool arch_is_rethunk(struct symbol *sym)
|
|
{
|
|
return !strcmp(sym->name, "__x86_return_thunk");
|
|
}
|