2021-12-22 02:21:41 +08:00
|
|
|
//===- bolt/Target/X86/X86MCPlusBuilder.cpp -------------------------------===//
|
2018-03-10 01:45:13 +08:00
|
|
|
//
|
2021-03-16 09:04:18 +08:00
|
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
2018-03-10 01:45:13 +08:00
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
//
|
2021-12-22 02:21:41 +08:00
|
|
|
// This file provides X86-specific MCPlus builder.
|
2018-03-10 01:45:13 +08:00
|
|
|
//
|
|
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
|
2021-05-01 04:54:02 +08:00
|
|
|
#include "MCTargetDesc/X86BaseInfo.h"
|
2022-03-11 04:21:39 +08:00
|
|
|
#include "MCTargetDesc/X86InstrRelaxTables.h"
|
2021-05-01 04:54:02 +08:00
|
|
|
#include "MCTargetDesc/X86MCTargetDesc.h"
|
2022-02-23 11:06:25 +08:00
|
|
|
#include "X86MCSymbolizer.h"
|
[BOLT] CMOVConversion pass
Convert simple hammocks into cmov based on misprediction rate.
Test Plan:
- Assembly test: `cmov-conversion.s`
- Testing on a binary:
# Bootstrap clang with `-x86-cmov-converter-force-all` and `-Wl,--emit-relocs`
(Release build)
# Collect perf.data:
- `clang++ <opts> bolt/lib/Core/BinaryFunction.cpp -E > bf.cpp`
- `perf record -e cycles:u -j any,u -- clang-15 bf.cpp -O2 -std=c++14 -c -o bf.o`
# Optimize clang-15 with and w/o -cmov-conversion:
- `llvm-bolt clang-15 -p perf.data -o clang-15.bolt`
- `llvm-bolt clang-15 -p perf.data -cmov-conversion -o clang-15.bolt.cmovconv`
# Run perf experiment:
- test: `clang-15.bolt.cmovconv`,
- control: `clang-15.bolt`,
- workload (clang options): `bf.cpp -O2 -std=c++14 -c -o bf.o`
Results:
```
task-clock [delta: -360.21 ± 356.75, delta(%): -1.7760 ± 1.7589, p-value: 0.047951, balance: -6]
instructions [delta: 44061118 ± 13246382, delta(%): 0.0690 ± 0.0207, p-value: 0.000001, balance: 50]
icache-misses [delta: -5534468 ± 2779620, delta(%): -0.4331 ± 0.2175, p-value: 0.028014, balance: -28]
branch-misses [delta: -1624270 ± 1113244, delta(%): -0.3456 ± 0.2368, p-value: 0.030300, balance: -22]
```
Reviewed By: rafauler
Differential Revision: https://reviews.llvm.org/D120177
2022-02-08 12:16:13 +08:00
|
|
|
#include "bolt/Core/MCPlus.h"
|
2021-10-09 02:47:10 +08:00
|
|
|
#include "bolt/Core/MCPlusBuilder.h"
|
2022-02-09 21:26:30 +08:00
|
|
|
#include "llvm/BinaryFormat/ELF.h"
|
2018-03-10 01:45:13 +08:00
|
|
|
#include "llvm/MC/MCContext.h"
|
[BOLT] Support for lite mode with relocations
Summary:
Add '-lite' support for relocations for improved processing time,
memory consumption, and more resilient processing of binaries with
embedded assembly code.
In lite relocation mode, BOLT will skip full processing of functions
without a profile. It will run scanExternalRefs() on such functions
to discover external references and to create internal relocations
to update references to optimized functions.
Note that we could have relied on the compiler/linker to provide
relocations for function references. However, there's no assurance
that all such references are reported. E.g., the compiler can resolve
inter-procedural references internally, leaving no relocations
for the linker.
The scan process takes about <10 seconds per 100MB of code on modern
hardware. It's a reasonable overhead to live with considering the
flexibility it provides.
If BOLT fails to scan or disassemble a function, .e.g., due to a data
object embedded in code, or an unsupported instruction, it enables a
patching mode to guarantee that the failed function will call
optimized/moved versions of functions. The patching happens at original
function entry points.
'-skip=<func1,func2,...>' option now can be used to skip processing of
arbitrary functions in the relocation mode.
With '-use-old-text' or '-strict' we require all functions to be
processed. As such, it is incompatible with '-lite' option,
and '-skip' option will only disable optimizations of listed
functions, not their disassembly and emission.
(cherry picked from FBD22040717)
2020-06-15 15:15:47 +08:00
|
|
|
#include "llvm/MC/MCFixupKindInfo.h"
|
[BOLT] CMOVConversion pass
Convert simple hammocks into cmov based on misprediction rate.
Test Plan:
- Assembly test: `cmov-conversion.s`
- Testing on a binary:
# Bootstrap clang with `-x86-cmov-converter-force-all` and `-Wl,--emit-relocs`
(Release build)
# Collect perf.data:
- `clang++ <opts> bolt/lib/Core/BinaryFunction.cpp -E > bf.cpp`
- `perf record -e cycles:u -j any,u -- clang-15 bf.cpp -O2 -std=c++14 -c -o bf.o`
# Optimize clang-15 with and w/o -cmov-conversion:
- `llvm-bolt clang-15 -p perf.data -o clang-15.bolt`
- `llvm-bolt clang-15 -p perf.data -cmov-conversion -o clang-15.bolt.cmovconv`
# Run perf experiment:
- test: `clang-15.bolt.cmovconv`,
- control: `clang-15.bolt`,
- workload (clang options): `bf.cpp -O2 -std=c++14 -c -o bf.o`
Results:
```
task-clock [delta: -360.21 ± 356.75, delta(%): -1.7760 ± 1.7589, p-value: 0.047951, balance: -6]
instructions [delta: 44061118 ± 13246382, delta(%): 0.0690 ± 0.0207, p-value: 0.000001, balance: 50]
icache-misses [delta: -5534468 ± 2779620, delta(%): -0.4331 ± 0.2175, p-value: 0.028014, balance: -28]
branch-misses [delta: -1624270 ± 1113244, delta(%): -0.3456 ± 0.2368, p-value: 0.030300, balance: -22]
```
Reviewed By: rafauler
Differential Revision: https://reviews.llvm.org/D120177
2022-02-08 12:16:13 +08:00
|
|
|
#include "llvm/MC/MCInst.h"
|
2018-05-27 03:40:51 +08:00
|
|
|
#include "llvm/MC/MCInstBuilder.h"
|
2018-03-10 01:45:13 +08:00
|
|
|
#include "llvm/MC/MCInstrInfo.h"
|
[BOLT] CMOVConversion pass
Convert simple hammocks into cmov based on misprediction rate.
Test Plan:
- Assembly test: `cmov-conversion.s`
- Testing on a binary:
# Bootstrap clang with `-x86-cmov-converter-force-all` and `-Wl,--emit-relocs`
(Release build)
# Collect perf.data:
- `clang++ <opts> bolt/lib/Core/BinaryFunction.cpp -E > bf.cpp`
- `perf record -e cycles:u -j any,u -- clang-15 bf.cpp -O2 -std=c++14 -c -o bf.o`
# Optimize clang-15 with and w/o -cmov-conversion:
- `llvm-bolt clang-15 -p perf.data -o clang-15.bolt`
- `llvm-bolt clang-15 -p perf.data -cmov-conversion -o clang-15.bolt.cmovconv`
# Run perf experiment:
- test: `clang-15.bolt.cmovconv`,
- control: `clang-15.bolt`,
- workload (clang options): `bf.cpp -O2 -std=c++14 -c -o bf.o`
Results:
```
task-clock [delta: -360.21 ± 356.75, delta(%): -1.7760 ± 1.7589, p-value: 0.047951, balance: -6]
instructions [delta: 44061118 ± 13246382, delta(%): 0.0690 ± 0.0207, p-value: 0.000001, balance: 50]
icache-misses [delta: -5534468 ± 2779620, delta(%): -0.4331 ± 0.2175, p-value: 0.028014, balance: -28]
branch-misses [delta: -1624270 ± 1113244, delta(%): -0.3456 ± 0.2368, p-value: 0.030300, balance: -22]
```
Reviewed By: rafauler
Differential Revision: https://reviews.llvm.org/D120177
2022-02-08 12:16:13 +08:00
|
|
|
#include "llvm/MC/MCRegister.h"
|
2018-03-10 01:45:13 +08:00
|
|
|
#include "llvm/MC/MCRegisterInfo.h"
|
2022-03-17 00:37:54 +08:00
|
|
|
#include "llvm/Support/CommandLine.h"
|
2018-03-10 01:45:13 +08:00
|
|
|
#include "llvm/Support/DataExtractor.h"
|
|
|
|
#include "llvm/Support/Debug.h"
|
|
|
|
#include "llvm/Support/Errc.h"
|
|
|
|
#include "llvm/Support/ErrorHandling.h"
|
|
|
|
#include "llvm/Support/ErrorOr.h"
|
2021-05-01 04:54:02 +08:00
|
|
|
#include <set>
|
2018-03-10 01:45:13 +08:00
|
|
|
|
2022-01-06 04:02:54 +08:00
|
|
|
#define DEBUG_TYPE "mcplus"
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
using namespace llvm;
|
|
|
|
using namespace bolt;
|
|
|
|
|
2022-03-17 00:37:54 +08:00
|
|
|
namespace opts {
|
|
|
|
|
|
|
|
extern cl::OptionCategory BoltOptCategory;
|
|
|
|
|
|
|
|
static cl::opt<bool> X86StripRedundantAddressSize(
|
|
|
|
"x86-strip-redundant-address-size",
|
|
|
|
cl::desc("Remove redundant Address-Size override prefix"), cl::init(true),
|
|
|
|
cl::ZeroOrMore, cl::cat(BoltOptCategory));
|
|
|
|
|
|
|
|
} // namespace opts
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
namespace {
|
|
|
|
|
|
|
|
unsigned getShortBranchOpcode(unsigned Opcode) {
|
|
|
|
switch (Opcode) {
|
|
|
|
default:
|
|
|
|
return Opcode;
|
|
|
|
case X86::JMP_2: return X86::JMP_1;
|
|
|
|
case X86::JMP_4: return X86::JMP_1;
|
2020-12-02 08:29:39 +08:00
|
|
|
case X86::JCC_2: return X86::JCC_1;
|
|
|
|
case X86::JCC_4: return X86::JCC_1;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned getShortArithOpcode(unsigned Opcode) {
|
2022-03-11 04:21:39 +08:00
|
|
|
return X86::getShortOpcodeArith(Opcode);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
2022-03-16 23:14:50 +08:00
|
|
|
bool isMOVSX64rm32(const MCInst &Inst) {
|
|
|
|
return Inst.getOpcode() == X86::MOVSX64rm32;
|
|
|
|
}
|
|
|
|
|
2022-04-06 05:30:44 +08:00
|
|
|
bool isADD64rr(const MCInst &Inst) { return Inst.getOpcode() == X86::ADD64rr; }
|
|
|
|
|
|
|
|
bool isADDri(const MCInst &Inst) {
|
|
|
|
return Inst.getOpcode() == X86::ADD64ri32 ||
|
|
|
|
Inst.getOpcode() == X86::ADD64ri8;
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
class X86MCPlusBuilder : public MCPlusBuilder {
|
|
|
|
public:
|
|
|
|
X86MCPlusBuilder(const MCInstrAnalysis *Analysis, const MCInstrInfo *Info,
|
2022-01-09 09:48:33 +08:00
|
|
|
const MCRegisterInfo *RegInfo)
|
|
|
|
: MCPlusBuilder(Analysis, Info, RegInfo) {}
|
2018-03-10 01:45:13 +08:00
|
|
|
|
2022-02-23 11:06:25 +08:00
|
|
|
std::unique_ptr<MCSymbolizer>
|
|
|
|
createTargetSymbolizer(BinaryFunction &Function) const override {
|
|
|
|
return std::make_unique<X86MCSymbolizer>(Function);
|
|
|
|
}
|
|
|
|
|
2020-12-02 08:29:39 +08:00
|
|
|
bool isBranch(const MCInst &Inst) const override {
|
|
|
|
return Analysis->isBranch(Inst) && !isTailCall(Inst);
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
bool isNoop(const MCInst &Inst) const override {
|
2022-03-18 16:48:00 +08:00
|
|
|
return X86::isNOP(Inst.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
2020-12-02 08:29:39 +08:00
|
|
|
unsigned getCondCode(const MCInst &Inst) const override {
|
2022-04-05 05:04:39 +08:00
|
|
|
unsigned Opcode = Inst.getOpcode();
|
|
|
|
if (X86::isJCC(Opcode))
|
|
|
|
return Inst.getOperand(Info->get(Opcode).NumOperands - 1).getImm();
|
|
|
|
return X86::COND_INVALID;
|
2020-12-02 08:29:39 +08:00
|
|
|
}
|
|
|
|
|
2021-08-18 01:15:21 +08:00
|
|
|
unsigned getInvertedCondCode(unsigned CC) const override {
|
|
|
|
switch (CC) {
|
|
|
|
default: return X86::COND_INVALID;
|
|
|
|
case X86::COND_E: return X86::COND_NE;
|
|
|
|
case X86::COND_NE: return X86::COND_E;
|
|
|
|
case X86::COND_L: return X86::COND_GE;
|
|
|
|
case X86::COND_LE: return X86::COND_G;
|
|
|
|
case X86::COND_G: return X86::COND_LE;
|
|
|
|
case X86::COND_GE: return X86::COND_L;
|
|
|
|
case X86::COND_B: return X86::COND_AE;
|
|
|
|
case X86::COND_BE: return X86::COND_A;
|
|
|
|
case X86::COND_A: return X86::COND_BE;
|
|
|
|
case X86::COND_AE: return X86::COND_B;
|
|
|
|
case X86::COND_S: return X86::COND_NS;
|
|
|
|
case X86::COND_NS: return X86::COND_S;
|
|
|
|
case X86::COND_P: return X86::COND_NP;
|
|
|
|
case X86::COND_NP: return X86::COND_P;
|
|
|
|
case X86::COND_O: return X86::COND_NO;
|
|
|
|
case X86::COND_NO: return X86::COND_O;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned getCondCodesLogicalOr(unsigned CC1, unsigned CC2) const override {
|
|
|
|
enum DecodedCondCode : uint8_t {
|
|
|
|
DCC_EQUAL = 0x1,
|
|
|
|
DCC_GREATER = 0x2,
|
|
|
|
DCC_LESSER = 0x4,
|
|
|
|
DCC_GREATER_OR_LESSER = 0x6,
|
|
|
|
DCC_UNSIGNED = 0x8,
|
|
|
|
DCC_SIGNED = 0x10,
|
|
|
|
DCC_INVALID = 0x20,
|
|
|
|
};
|
|
|
|
|
|
|
|
auto decodeCondCode = [&](unsigned CC) -> uint8_t {
|
|
|
|
switch (CC) {
|
|
|
|
default: return DCC_INVALID;
|
|
|
|
case X86::COND_E: return DCC_EQUAL;
|
|
|
|
case X86::COND_NE: return DCC_GREATER | DCC_LESSER;
|
|
|
|
case X86::COND_L: return DCC_LESSER | DCC_SIGNED;
|
|
|
|
case X86::COND_LE: return DCC_EQUAL | DCC_LESSER | DCC_SIGNED;
|
|
|
|
case X86::COND_G: return DCC_GREATER | DCC_SIGNED;
|
|
|
|
case X86::COND_GE: return DCC_GREATER | DCC_EQUAL | DCC_SIGNED;
|
|
|
|
case X86::COND_B: return DCC_LESSER | DCC_UNSIGNED;
|
|
|
|
case X86::COND_BE: return DCC_EQUAL | DCC_LESSER | DCC_UNSIGNED;
|
|
|
|
case X86::COND_A: return DCC_GREATER | DCC_UNSIGNED;
|
|
|
|
case X86::COND_AE: return DCC_GREATER | DCC_EQUAL | DCC_UNSIGNED;
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
uint8_t DCC = decodeCondCode(CC1) | decodeCondCode(CC2);
|
|
|
|
|
|
|
|
if (DCC & DCC_INVALID)
|
|
|
|
return X86::COND_INVALID;
|
|
|
|
|
|
|
|
if (DCC & DCC_SIGNED && DCC & DCC_UNSIGNED)
|
|
|
|
return X86::COND_INVALID;
|
|
|
|
|
|
|
|
switch (DCC) {
|
2022-01-09 09:48:33 +08:00
|
|
|
default: return X86::COND_INVALID;
|
|
|
|
case DCC_EQUAL | DCC_LESSER | DCC_SIGNED: return X86::COND_LE;
|
|
|
|
case DCC_EQUAL | DCC_LESSER | DCC_UNSIGNED: return X86::COND_BE;
|
|
|
|
case DCC_EQUAL | DCC_GREATER | DCC_SIGNED: return X86::COND_GE;
|
|
|
|
case DCC_EQUAL | DCC_GREATER | DCC_UNSIGNED: return X86::COND_AE;
|
|
|
|
case DCC_GREATER | DCC_LESSER | DCC_SIGNED: return X86::COND_NE;
|
|
|
|
case DCC_GREATER | DCC_LESSER | DCC_UNSIGNED: return X86::COND_NE;
|
|
|
|
case DCC_GREATER | DCC_LESSER: return X86::COND_NE;
|
|
|
|
case DCC_EQUAL | DCC_SIGNED: return X86::COND_E;
|
|
|
|
case DCC_EQUAL | DCC_UNSIGNED: return X86::COND_E;
|
|
|
|
case DCC_EQUAL: return X86::COND_E;
|
|
|
|
case DCC_LESSER | DCC_SIGNED: return X86::COND_L;
|
|
|
|
case DCC_LESSER | DCC_UNSIGNED: return X86::COND_B;
|
|
|
|
case DCC_GREATER | DCC_SIGNED: return X86::COND_G;
|
|
|
|
case DCC_GREATER | DCC_UNSIGNED: return X86::COND_A;
|
2021-08-18 01:15:21 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isValidCondCode(unsigned CC) const override {
|
|
|
|
return (CC != X86::COND_INVALID);
|
|
|
|
}
|
|
|
|
|
[BOLT] Add code padding verification
Summary:
In non-relocation mode, we allow data objects to be embedded in the
code. Such objects could be unmarked, and could occupy an area between
functions, the area which is considered to be code padding.
When we disassemble code, we detect references into the padding area
and adjust it, so that it is not overwritten during the code emission.
We assume the reference to be pointing to the beginning of the object.
However, assembly-written functions may reference the middle of an
object and use negative offsets to reference data fields. Thus,
conservatively, we reduce the possibly-overwritten padding area to
a minimum if the object reference was detected.
Since we also allow functions with unknown code in non-relocation mode,
it is possible that we miss references to some objects in code.
To cover such cases, we need to verify the padding area before we
allow to overwrite it.
(cherry picked from FBD16477787)
2019-07-24 11:48:41 +08:00
|
|
|
bool isBreakpoint(const MCInst &Inst) const override {
|
|
|
|
return Inst.getOpcode() == X86::INT3;
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
bool isPrefix(const MCInst &Inst) const override {
|
2022-04-05 05:04:39 +08:00
|
|
|
const MCInstrDesc &Desc = Info->get(Inst.getOpcode());
|
|
|
|
return X86II::isPrefix(Desc.TSFlags);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
2021-08-11 01:02:32 +08:00
|
|
|
bool isRep(const MCInst &Inst) const override {
|
|
|
|
return Inst.getFlags() == X86::IP_HAS_REPEAT;
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
bool deleteREPPrefix(MCInst &Inst) const override {
|
|
|
|
if (Inst.getFlags() == X86::IP_HAS_REPEAT) {
|
|
|
|
Inst.setFlags(0);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// FIXME: For compatibility with old LLVM only!
|
|
|
|
bool isTerminator(const MCInst &Inst) const override {
|
2022-04-05 05:04:39 +08:00
|
|
|
unsigned Opcode = Inst.getOpcode();
|
|
|
|
return Info->get(Opcode).isTerminator() || X86::isUD1(Opcode) ||
|
|
|
|
X86::isUD2(Opcode);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
bool isIndirectCall(const MCInst &Inst) const override {
|
|
|
|
return isCall(Inst) &&
|
|
|
|
((getMemoryOperandNo(Inst) != -1) || Inst.getOperand(0).isReg());
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isPop(const MCInst &Inst) const override {
|
|
|
|
return getPopSize(Inst) == 0 ? false : true;
|
|
|
|
}
|
|
|
|
|
2021-07-14 16:35:34 +08:00
|
|
|
bool isTerminateBranch(const MCInst &Inst) const override {
|
|
|
|
return Inst.getOpcode() == X86::ENDBR32 || Inst.getOpcode() == X86::ENDBR64;
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
int getPopSize(const MCInst &Inst) const override {
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
case X86::POP16r:
|
|
|
|
case X86::POP16rmm:
|
|
|
|
case X86::POP16rmr:
|
|
|
|
case X86::POPF16:
|
|
|
|
case X86::POPA16:
|
|
|
|
case X86::POPDS16:
|
|
|
|
case X86::POPES16:
|
|
|
|
case X86::POPFS16:
|
|
|
|
case X86::POPGS16:
|
|
|
|
case X86::POPSS16:
|
|
|
|
return 2;
|
|
|
|
case X86::POP32r:
|
|
|
|
case X86::POP32rmm:
|
|
|
|
case X86::POP32rmr:
|
|
|
|
case X86::POPA32:
|
|
|
|
case X86::POPDS32:
|
|
|
|
case X86::POPES32:
|
|
|
|
case X86::POPF32:
|
|
|
|
case X86::POPFS32:
|
|
|
|
case X86::POPGS32:
|
|
|
|
case X86::POPSS32:
|
|
|
|
return 4;
|
|
|
|
case X86::POP64r:
|
|
|
|
case X86::POP64rmm:
|
|
|
|
case X86::POP64rmr:
|
|
|
|
case X86::POPF64:
|
|
|
|
case X86::POPFS64:
|
|
|
|
case X86::POPGS64:
|
|
|
|
return 8;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isPush(const MCInst &Inst) const override {
|
|
|
|
return getPushSize(Inst) == 0 ? false : true;
|
|
|
|
}
|
|
|
|
|
|
|
|
int getPushSize(const MCInst &Inst) const override {
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
case X86::PUSH16i8:
|
|
|
|
case X86::PUSH16r:
|
|
|
|
case X86::PUSH16rmm:
|
|
|
|
case X86::PUSH16rmr:
|
|
|
|
case X86::PUSHA16:
|
|
|
|
case X86::PUSHCS16:
|
|
|
|
case X86::PUSHDS16:
|
|
|
|
case X86::PUSHES16:
|
|
|
|
case X86::PUSHF16:
|
|
|
|
case X86::PUSHFS16:
|
|
|
|
case X86::PUSHGS16:
|
|
|
|
case X86::PUSHSS16:
|
|
|
|
case X86::PUSHi16:
|
|
|
|
return 2;
|
|
|
|
case X86::PUSH32i8:
|
|
|
|
case X86::PUSH32r:
|
|
|
|
case X86::PUSH32rmm:
|
|
|
|
case X86::PUSH32rmr:
|
|
|
|
case X86::PUSHA32:
|
|
|
|
case X86::PUSHCS32:
|
|
|
|
case X86::PUSHDS32:
|
|
|
|
case X86::PUSHES32:
|
|
|
|
case X86::PUSHF32:
|
|
|
|
case X86::PUSHFS32:
|
|
|
|
case X86::PUSHGS32:
|
|
|
|
case X86::PUSHSS32:
|
|
|
|
case X86::PUSHi32:
|
|
|
|
return 4;
|
|
|
|
case X86::PUSH64i32:
|
|
|
|
case X86::PUSH64i8:
|
|
|
|
case X86::PUSH64r:
|
|
|
|
case X86::PUSH64rmm:
|
|
|
|
case X86::PUSH64rmr:
|
|
|
|
case X86::PUSHF64:
|
|
|
|
case X86::PUSHFS64:
|
|
|
|
case X86::PUSHGS64:
|
|
|
|
return 8;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isSUB(const MCInst &Inst) const override {
|
2022-03-18 16:48:00 +08:00
|
|
|
return X86::isSUB(Inst.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
bool isLEA64r(const MCInst &Inst) const override {
|
|
|
|
return Inst.getOpcode() == X86::LEA64r;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isLeave(const MCInst &Inst) const override {
|
2022-01-09 09:48:33 +08:00
|
|
|
return Inst.getOpcode() == X86::LEAVE || Inst.getOpcode() == X86::LEAVE64;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
bool isMoveMem2Reg(const MCInst &Inst) const override {
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
case X86::MOV16rm:
|
|
|
|
case X86::MOV32rm:
|
|
|
|
case X86::MOV64rm:
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2019-06-20 11:10:49 +08:00
|
|
|
bool isUnsupportedBranch(unsigned Opcode) const override {
|
|
|
|
switch (Opcode) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case X86::LOOP:
|
|
|
|
case X86::LOOPE:
|
|
|
|
case X86::LOOPNE:
|
|
|
|
case X86::JECXZ:
|
|
|
|
case X86::JRCXZ:
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
bool isLoad(const MCInst &Inst) const override {
|
|
|
|
if (isPop(Inst))
|
|
|
|
return true;
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
int MemOpNo = getMemoryOperandNo(Inst);
|
2021-04-18 12:48:46 +08:00
|
|
|
const MCInstrDesc &MCII = Info->get(Inst.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
if (MemOpNo == -1)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return MCII.mayLoad();
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isStore(const MCInst &Inst) const override {
|
|
|
|
if (isPush(Inst))
|
|
|
|
return true;
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
int MemOpNo = getMemoryOperandNo(Inst);
|
2021-04-18 12:48:46 +08:00
|
|
|
const MCInstrDesc &MCII = Info->get(Inst.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
if (MemOpNo == -1)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return MCII.mayStore();
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isCleanRegXOR(const MCInst &Inst) const override {
|
|
|
|
switch (Inst.getOpcode()) {
|
2022-01-09 09:48:33 +08:00
|
|
|
case X86::XOR16rr:
|
|
|
|
case X86::XOR32rr:
|
|
|
|
case X86::XOR64rr:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return false;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
2022-01-09 09:48:33 +08:00
|
|
|
return (Inst.getOperand(0).getReg() == Inst.getOperand(2).getReg());
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
2021-08-18 01:15:21 +08:00
|
|
|
bool isPacked(const MCInst &Inst) const override {
|
|
|
|
const MCInstrDesc &Desc = Info->get(Inst.getOpcode());
|
|
|
|
return (Desc.TSFlags & X86II::OpPrefixMask) == X86II::PD;
|
|
|
|
}
|
|
|
|
|
2020-12-02 08:29:39 +08:00
|
|
|
unsigned getTrapFillValue() const override { return 0xCC; }
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
struct IndJmpMatcherFrag1 : MCInstMatcher {
|
|
|
|
std::unique_ptr<MCInstMatcher> Base;
|
|
|
|
std::unique_ptr<MCInstMatcher> Scale;
|
|
|
|
std::unique_ptr<MCInstMatcher> Index;
|
|
|
|
std::unique_ptr<MCInstMatcher> Offset;
|
|
|
|
|
|
|
|
IndJmpMatcherFrag1(std::unique_ptr<MCInstMatcher> Base,
|
|
|
|
std::unique_ptr<MCInstMatcher> Scale,
|
|
|
|
std::unique_ptr<MCInstMatcher> Index,
|
|
|
|
std::unique_ptr<MCInstMatcher> Offset)
|
|
|
|
: Base(std::move(Base)), Scale(std::move(Scale)),
|
|
|
|
Index(std::move(Index)), Offset(std::move(Offset)) {}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
bool match(const MCRegisterInfo &MRI, MCPlusBuilder &MIB,
|
2018-03-10 01:45:13 +08:00
|
|
|
MutableArrayRef<MCInst> InInstrWindow, int OpNum) override {
|
|
|
|
if (!MCInstMatcher::match(MRI, MIB, InInstrWindow, OpNum))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (CurInst->getOpcode() != X86::JMP64m)
|
|
|
|
return false;
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
int MemOpNo = MIB.getMemoryOperandNo(*CurInst);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (MemOpNo == -1)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (!Base->match(MRI, MIB, this->InstrWindow, MemOpNo + X86::AddrBaseReg))
|
|
|
|
return false;
|
|
|
|
if (!Scale->match(MRI, MIB, this->InstrWindow,
|
|
|
|
MemOpNo + X86::AddrScaleAmt))
|
|
|
|
return false;
|
|
|
|
if (!Index->match(MRI, MIB, this->InstrWindow,
|
|
|
|
MemOpNo + X86::AddrIndexReg))
|
|
|
|
return false;
|
|
|
|
if (!Offset->match(MRI, MIB, this->InstrWindow, MemOpNo + X86::AddrDisp))
|
|
|
|
return false;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
void annotate(MCPlusBuilder &MIB, StringRef Annotation) override {
|
|
|
|
MIB.addAnnotation(*CurInst, Annotation, true);
|
|
|
|
Base->annotate(MIB, Annotation);
|
|
|
|
Scale->annotate(MIB, Annotation);
|
|
|
|
Index->annotate(MIB, Annotation);
|
|
|
|
Offset->annotate(MIB, Annotation);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
std::unique_ptr<MCInstMatcher>
|
|
|
|
matchIndJmp(std::unique_ptr<MCInstMatcher> Base,
|
|
|
|
std::unique_ptr<MCInstMatcher> Scale,
|
|
|
|
std::unique_ptr<MCInstMatcher> Index,
|
|
|
|
std::unique_ptr<MCInstMatcher> Offset) const override {
|
|
|
|
return std::unique_ptr<MCInstMatcher>(
|
|
|
|
new IndJmpMatcherFrag1(std::move(Base), std::move(Scale),
|
|
|
|
std::move(Index), std::move(Offset)));
|
|
|
|
}
|
|
|
|
|
|
|
|
struct IndJmpMatcherFrag2 : MCInstMatcher {
|
|
|
|
std::unique_ptr<MCInstMatcher> Reg;
|
|
|
|
|
|
|
|
IndJmpMatcherFrag2(std::unique_ptr<MCInstMatcher> Reg)
|
|
|
|
: Reg(std::move(Reg)) {}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
bool match(const MCRegisterInfo &MRI, MCPlusBuilder &MIB,
|
2018-03-10 01:45:13 +08:00
|
|
|
MutableArrayRef<MCInst> InInstrWindow, int OpNum) override {
|
|
|
|
if (!MCInstMatcher::match(MRI, MIB, InInstrWindow, OpNum))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (CurInst->getOpcode() != X86::JMP64r)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return Reg->match(MRI, MIB, this->InstrWindow, 0);
|
|
|
|
}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
void annotate(MCPlusBuilder &MIB, StringRef Annotation) override {
|
|
|
|
MIB.addAnnotation(*CurInst, Annotation, true);
|
|
|
|
Reg->annotate(MIB, Annotation);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
std::unique_ptr<MCInstMatcher>
|
|
|
|
matchIndJmp(std::unique_ptr<MCInstMatcher> Target) const override {
|
|
|
|
return std::unique_ptr<MCInstMatcher>(
|
|
|
|
new IndJmpMatcherFrag2(std::move(Target)));
|
|
|
|
}
|
|
|
|
|
|
|
|
struct LoadMatcherFrag1 : MCInstMatcher {
|
|
|
|
std::unique_ptr<MCInstMatcher> Base;
|
|
|
|
std::unique_ptr<MCInstMatcher> Scale;
|
|
|
|
std::unique_ptr<MCInstMatcher> Index;
|
|
|
|
std::unique_ptr<MCInstMatcher> Offset;
|
|
|
|
|
|
|
|
LoadMatcherFrag1(std::unique_ptr<MCInstMatcher> Base,
|
|
|
|
std::unique_ptr<MCInstMatcher> Scale,
|
|
|
|
std::unique_ptr<MCInstMatcher> Index,
|
|
|
|
std::unique_ptr<MCInstMatcher> Offset)
|
|
|
|
: Base(std::move(Base)), Scale(std::move(Scale)),
|
|
|
|
Index(std::move(Index)), Offset(std::move(Offset)) {}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
bool match(const MCRegisterInfo &MRI, MCPlusBuilder &MIB,
|
2018-03-10 01:45:13 +08:00
|
|
|
MutableArrayRef<MCInst> InInstrWindow, int OpNum) override {
|
|
|
|
if (!MCInstMatcher::match(MRI, MIB, InInstrWindow, OpNum))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (CurInst->getOpcode() != X86::MOV64rm &&
|
|
|
|
CurInst->getOpcode() != X86::MOVSX64rm32)
|
|
|
|
return false;
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
int MemOpNo = MIB.getMemoryOperandNo(*CurInst);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (MemOpNo == -1)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (!Base->match(MRI, MIB, this->InstrWindow, MemOpNo + X86::AddrBaseReg))
|
|
|
|
return false;
|
|
|
|
if (!Scale->match(MRI, MIB, this->InstrWindow,
|
|
|
|
MemOpNo + X86::AddrScaleAmt))
|
|
|
|
return false;
|
|
|
|
if (!Index->match(MRI, MIB, this->InstrWindow,
|
|
|
|
MemOpNo + X86::AddrIndexReg))
|
|
|
|
return false;
|
|
|
|
if (!Offset->match(MRI, MIB, this->InstrWindow, MemOpNo + X86::AddrDisp))
|
|
|
|
return false;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
void annotate(MCPlusBuilder &MIB, StringRef Annotation) override {
|
|
|
|
MIB.addAnnotation(*CurInst, Annotation, true);
|
|
|
|
Base->annotate(MIB, Annotation);
|
|
|
|
Scale->annotate(MIB, Annotation);
|
|
|
|
Index->annotate(MIB, Annotation);
|
|
|
|
Offset->annotate(MIB, Annotation);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
std::unique_ptr<MCInstMatcher>
|
|
|
|
matchLoad(std::unique_ptr<MCInstMatcher> Base,
|
|
|
|
std::unique_ptr<MCInstMatcher> Scale,
|
|
|
|
std::unique_ptr<MCInstMatcher> Index,
|
|
|
|
std::unique_ptr<MCInstMatcher> Offset) const override {
|
|
|
|
return std::unique_ptr<MCInstMatcher>(
|
|
|
|
new LoadMatcherFrag1(std::move(Base), std::move(Scale),
|
2022-01-09 09:48:33 +08:00
|
|
|
std::move(Index), std::move(Offset)));
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
struct AddMatcher : MCInstMatcher {
|
|
|
|
std::unique_ptr<MCInstMatcher> A;
|
|
|
|
std::unique_ptr<MCInstMatcher> B;
|
|
|
|
|
|
|
|
AddMatcher(std::unique_ptr<MCInstMatcher> A,
|
|
|
|
std::unique_ptr<MCInstMatcher> B)
|
|
|
|
: A(std::move(A)), B(std::move(B)) {}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
bool match(const MCRegisterInfo &MRI, MCPlusBuilder &MIB,
|
2018-03-10 01:45:13 +08:00
|
|
|
MutableArrayRef<MCInst> InInstrWindow, int OpNum) override {
|
|
|
|
if (!MCInstMatcher::match(MRI, MIB, InInstrWindow, OpNum))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (CurInst->getOpcode() == X86::ADD64rr ||
|
|
|
|
CurInst->getOpcode() == X86::ADD64rr_DB ||
|
|
|
|
CurInst->getOpcode() == X86::ADD64rr_REV) {
|
|
|
|
if (!A->match(MRI, MIB, this->InstrWindow, 1)) {
|
|
|
|
if (!B->match(MRI, MIB, this->InstrWindow, 1))
|
|
|
|
return false;
|
|
|
|
return A->match(MRI, MIB, this->InstrWindow, 2);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (B->match(MRI, MIB, this->InstrWindow, 2))
|
|
|
|
return true;
|
|
|
|
|
|
|
|
if (!B->match(MRI, MIB, this->InstrWindow, 1))
|
|
|
|
return false;
|
|
|
|
return A->match(MRI, MIB, this->InstrWindow, 2);
|
|
|
|
}
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
void annotate(MCPlusBuilder &MIB, StringRef Annotation) override {
|
|
|
|
MIB.addAnnotation(*CurInst, Annotation, true);
|
|
|
|
A->annotate(MIB, Annotation);
|
|
|
|
B->annotate(MIB, Annotation);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
virtual std::unique_ptr<MCInstMatcher>
|
|
|
|
matchAdd(std::unique_ptr<MCInstMatcher> A,
|
|
|
|
std::unique_ptr<MCInstMatcher> B) const override {
|
|
|
|
return std::unique_ptr<MCInstMatcher>(
|
|
|
|
new AddMatcher(std::move(A), std::move(B)));
|
|
|
|
}
|
|
|
|
|
|
|
|
struct LEAMatcher : MCInstMatcher {
|
|
|
|
std::unique_ptr<MCInstMatcher> Target;
|
|
|
|
|
|
|
|
LEAMatcher(std::unique_ptr<MCInstMatcher> Target)
|
|
|
|
: Target(std::move(Target)) {}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
bool match(const MCRegisterInfo &MRI, MCPlusBuilder &MIB,
|
2018-03-10 01:45:13 +08:00
|
|
|
MutableArrayRef<MCInst> InInstrWindow, int OpNum) override {
|
|
|
|
if (!MCInstMatcher::match(MRI, MIB, InInstrWindow, OpNum))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (CurInst->getOpcode() != X86::LEA64r)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (CurInst->getOperand(1 + X86::AddrScaleAmt).getImm() != 1 ||
|
|
|
|
CurInst->getOperand(1 + X86::AddrIndexReg).getReg() !=
|
|
|
|
X86::NoRegister ||
|
|
|
|
(CurInst->getOperand(1 + X86::AddrBaseReg).getReg() !=
|
|
|
|
X86::NoRegister &&
|
|
|
|
CurInst->getOperand(1 + X86::AddrBaseReg).getReg() != X86::RIP))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return Target->match(MRI, MIB, this->InstrWindow, 1 + X86::AddrDisp);
|
|
|
|
}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
void annotate(MCPlusBuilder &MIB, StringRef Annotation) override {
|
|
|
|
MIB.addAnnotation(*CurInst, Annotation, true);
|
|
|
|
Target->annotate(MIB, Annotation);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
virtual std::unique_ptr<MCInstMatcher>
|
|
|
|
matchLoadAddr(std::unique_ptr<MCInstMatcher> Target) const override {
|
|
|
|
return std::unique_ptr<MCInstMatcher>(new LEAMatcher(std::move(Target)));
|
|
|
|
}
|
|
|
|
|
|
|
|
bool hasPCRelOperand(const MCInst &Inst) const override {
|
2021-12-29 09:52:08 +08:00
|
|
|
for (const MCOperand &Operand : Inst)
|
2018-03-10 01:45:13 +08:00
|
|
|
if (Operand.isReg() && Operand.getReg() == X86::RIP)
|
|
|
|
return true;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
int getMemoryOperandNo(const MCInst &Inst) const override {
|
2021-04-08 15:19:26 +08:00
|
|
|
unsigned Opcode = Inst.getOpcode();
|
|
|
|
const MCInstrDesc &Desc = Info->get(Opcode);
|
|
|
|
int MemOpNo = X86II::getMemoryOperandNo(Desc.TSFlags);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (MemOpNo >= 0)
|
|
|
|
MemOpNo += X86II::getOperandBias(Desc);
|
|
|
|
return MemOpNo;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool hasEVEXEncoding(const MCInst &Inst) const override {
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCInstrDesc &Desc = Info->get(Inst.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
return (Desc.TSFlags & X86II::EncodingMask) == X86II::EVEX;
|
|
|
|
}
|
|
|
|
|
2018-04-14 06:46:19 +08:00
|
|
|
bool isMacroOpFusionPair(ArrayRef<MCInst> Insts) const override {
|
2021-04-08 15:19:26 +08:00
|
|
|
const auto *I = Insts.begin();
|
2018-04-14 06:46:19 +08:00
|
|
|
while (I != Insts.end() && isPrefix(*I))
|
|
|
|
++I;
|
|
|
|
if (I == Insts.end())
|
|
|
|
return false;
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCInst &FirstInst = *I;
|
2018-04-14 06:46:19 +08:00
|
|
|
++I;
|
|
|
|
while (I != Insts.end() && isPrefix(*I))
|
|
|
|
++I;
|
|
|
|
if (I == Insts.end())
|
|
|
|
return false;
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCInst &SecondInst = *I;
|
2018-04-14 06:46:19 +08:00
|
|
|
|
|
|
|
if (!isConditionalBranch(SecondInst))
|
|
|
|
return false;
|
|
|
|
// Cannot fuse if the first instruction uses RIP-relative memory.
|
|
|
|
if (hasPCRelOperand(FirstInst))
|
|
|
|
return false;
|
|
|
|
|
2022-01-06 07:58:01 +08:00
|
|
|
const X86::FirstMacroFusionInstKind CmpKind =
|
|
|
|
X86::classifyFirstOpcodeInMacroFusion(FirstInst.getOpcode());
|
|
|
|
if (CmpKind == X86::FirstMacroFusionInstKind::Invalid)
|
2018-04-14 06:46:19 +08:00
|
|
|
return false;
|
|
|
|
|
2022-01-06 07:58:01 +08:00
|
|
|
X86::CondCode CC = static_cast<X86::CondCode>(getCondCode(SecondInst));
|
|
|
|
X86::SecondMacroFusionInstKind BranchKind =
|
|
|
|
X86::classifySecondCondCodeInMacroFusion(CC);
|
|
|
|
if (BranchKind == X86::SecondMacroFusionInstKind::Invalid)
|
2018-04-14 06:46:19 +08:00
|
|
|
return false;
|
2022-01-06 07:58:01 +08:00
|
|
|
return X86::isMacroFused(CmpKind, BranchKind);
|
2018-04-14 06:46:19 +08:00
|
|
|
}
|
2018-05-05 01:10:41 +08:00
|
|
|
|
2022-01-09 09:48:33 +08:00
|
|
|
bool
|
|
|
|
evaluateX86MemoryOperand(const MCInst &Inst, unsigned *BaseRegNum,
|
|
|
|
int64_t *ScaleImm, unsigned *IndexRegNum,
|
|
|
|
int64_t *DispImm, unsigned *SegmentRegNum,
|
|
|
|
const MCExpr **DispExpr = nullptr) const override {
|
2018-03-10 01:45:13 +08:00
|
|
|
assert(BaseRegNum && ScaleImm && IndexRegNum && SegmentRegNum &&
|
|
|
|
"one of the input pointers is null");
|
2021-04-08 15:19:26 +08:00
|
|
|
int MemOpNo = getMemoryOperandNo(Inst);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (MemOpNo < 0)
|
|
|
|
return false;
|
|
|
|
unsigned MemOpOffset = static_cast<unsigned>(MemOpNo);
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
if (MemOpOffset + X86::AddrSegmentReg >= MCPlus::getNumPrimeOperands(Inst))
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCOperand &Base = Inst.getOperand(MemOpOffset + X86::AddrBaseReg);
|
|
|
|
const MCOperand &Scale = Inst.getOperand(MemOpOffset + X86::AddrScaleAmt);
|
|
|
|
const MCOperand &Index = Inst.getOperand(MemOpOffset + X86::AddrIndexReg);
|
|
|
|
const MCOperand &Disp = Inst.getOperand(MemOpOffset + X86::AddrDisp);
|
|
|
|
const MCOperand &Segment =
|
|
|
|
Inst.getOperand(MemOpOffset + X86::AddrSegmentReg);
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// Make sure it is a well-formed memory operand.
|
|
|
|
if (!Base.isReg() || !Scale.isImm() || !Index.isReg() ||
|
|
|
|
(!Disp.isImm() && !Disp.isExpr()) || !Segment.isReg())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
*BaseRegNum = Base.getReg();
|
|
|
|
*ScaleImm = Scale.getImm();
|
|
|
|
*IndexRegNum = Index.getReg();
|
|
|
|
if (Disp.isImm()) {
|
|
|
|
assert(DispImm && "DispImm needs to be set");
|
|
|
|
*DispImm = Disp.getImm();
|
2021-12-29 09:52:08 +08:00
|
|
|
if (DispExpr)
|
2018-03-10 01:45:13 +08:00
|
|
|
*DispExpr = nullptr;
|
|
|
|
} else {
|
|
|
|
assert(DispExpr && "DispExpr needs to be set");
|
|
|
|
*DispExpr = Disp.getExpr();
|
2021-12-29 09:52:08 +08:00
|
|
|
if (DispImm)
|
2018-08-04 07:36:06 +08:00
|
|
|
*DispImm = 0;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
*SegmentRegNum = Segment.getReg();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2022-01-09 09:48:33 +08:00
|
|
|
bool evaluateMemOperandTarget(const MCInst &Inst, uint64_t &Target,
|
2018-03-10 01:45:13 +08:00
|
|
|
uint64_t Address,
|
|
|
|
uint64_t Size) const override {
|
|
|
|
unsigned BaseRegNum;
|
|
|
|
int64_t ScaleValue;
|
|
|
|
unsigned IndexRegNum;
|
|
|
|
int64_t DispValue;
|
|
|
|
unsigned SegRegNum;
|
2021-05-14 01:50:47 +08:00
|
|
|
const MCExpr *DispExpr = nullptr;
|
2018-03-10 01:45:13 +08:00
|
|
|
if (!evaluateX86MemoryOperand(Inst, &BaseRegNum, &ScaleValue, &IndexRegNum,
|
2021-12-29 09:52:08 +08:00
|
|
|
&DispValue, &SegRegNum, &DispExpr))
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
// Make sure it's a well-formed addressing we can statically evaluate.
|
|
|
|
if ((BaseRegNum != X86::RIP && BaseRegNum != X86::NoRegister) ||
|
|
|
|
IndexRegNum != X86::NoRegister || SegRegNum != X86::NoRegister ||
|
2021-12-29 09:52:08 +08:00
|
|
|
DispExpr)
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
Target = DispValue;
|
|
|
|
if (BaseRegNum == X86::RIP) {
|
|
|
|
assert(Size != 0 && "instruction size required in order to statically "
|
|
|
|
"evaluate RIP-relative address");
|
|
|
|
Target += Address + Size;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
MCInst::iterator getMemOperandDisp(MCInst &Inst) const override {
|
2021-04-08 15:19:26 +08:00
|
|
|
int MemOpNo = getMemoryOperandNo(Inst);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (MemOpNo < 0)
|
|
|
|
return Inst.end();
|
|
|
|
return Inst.begin() + (MemOpNo + X86::AddrDisp);
|
|
|
|
}
|
|
|
|
|
|
|
|
bool replaceMemOperandDisp(MCInst &Inst, MCOperand Operand) const override {
|
2021-04-08 15:19:26 +08:00
|
|
|
MCOperand *OI = getMemOperandDisp(Inst);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (OI == Inst.end())
|
|
|
|
return false;
|
2020-12-02 08:29:39 +08:00
|
|
|
*OI = Operand;
|
2018-03-10 01:45:13 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Get the registers used as function parameters.
|
|
|
|
/// This function is specific to the x86_64 abi on Linux.
|
|
|
|
BitVector getRegsUsedAsParams() const override {
|
|
|
|
BitVector Regs = BitVector(RegInfo->getNumRegs(), false);
|
|
|
|
Regs |= getAliases(X86::RSI);
|
|
|
|
Regs |= getAliases(X86::RDI);
|
|
|
|
Regs |= getAliases(X86::RDX);
|
|
|
|
Regs |= getAliases(X86::RCX);
|
|
|
|
Regs |= getAliases(X86::R8);
|
|
|
|
Regs |= getAliases(X86::R9);
|
|
|
|
return Regs;
|
|
|
|
}
|
|
|
|
|
|
|
|
void getCalleeSavedRegs(BitVector &Regs) const override {
|
|
|
|
Regs |= getAliases(X86::RBX);
|
|
|
|
Regs |= getAliases(X86::RBP);
|
|
|
|
Regs |= getAliases(X86::R12);
|
|
|
|
Regs |= getAliases(X86::R13);
|
|
|
|
Regs |= getAliases(X86::R14);
|
|
|
|
Regs |= getAliases(X86::R15);
|
|
|
|
}
|
|
|
|
|
|
|
|
void getDefaultDefIn(BitVector &Regs) const override {
|
|
|
|
assert(Regs.size() >= RegInfo->getNumRegs() &&
|
2022-01-09 09:48:33 +08:00
|
|
|
"The size of BitVector is less than RegInfo->getNumRegs().");
|
2018-03-10 01:45:13 +08:00
|
|
|
Regs.set(X86::RAX);
|
|
|
|
Regs.set(X86::RCX);
|
|
|
|
Regs.set(X86::RDX);
|
|
|
|
Regs.set(X86::RSI);
|
|
|
|
Regs.set(X86::RDI);
|
|
|
|
Regs.set(X86::R8);
|
|
|
|
Regs.set(X86::R9);
|
|
|
|
Regs.set(X86::XMM0);
|
|
|
|
Regs.set(X86::XMM1);
|
|
|
|
Regs.set(X86::XMM2);
|
|
|
|
Regs.set(X86::XMM3);
|
|
|
|
Regs.set(X86::XMM4);
|
|
|
|
Regs.set(X86::XMM5);
|
|
|
|
Regs.set(X86::XMM6);
|
|
|
|
Regs.set(X86::XMM7);
|
|
|
|
}
|
|
|
|
|
|
|
|
void getDefaultLiveOut(BitVector &Regs) const override {
|
|
|
|
assert(Regs.size() >= RegInfo->getNumRegs() &&
|
2022-01-09 09:48:33 +08:00
|
|
|
"The size of BitVector is less than RegInfo->getNumRegs().");
|
2018-03-10 01:45:13 +08:00
|
|
|
Regs |= getAliases(X86::RAX);
|
|
|
|
Regs |= getAliases(X86::RDX);
|
|
|
|
Regs |= getAliases(X86::RCX);
|
|
|
|
Regs |= getAliases(X86::XMM0);
|
|
|
|
Regs |= getAliases(X86::XMM1);
|
|
|
|
}
|
|
|
|
|
|
|
|
void getGPRegs(BitVector &Regs, bool IncludeAlias) const override {
|
|
|
|
if (IncludeAlias) {
|
|
|
|
Regs |= getAliases(X86::RAX);
|
|
|
|
Regs |= getAliases(X86::RBX);
|
|
|
|
Regs |= getAliases(X86::RBP);
|
|
|
|
Regs |= getAliases(X86::RSI);
|
|
|
|
Regs |= getAliases(X86::RDI);
|
|
|
|
Regs |= getAliases(X86::RDX);
|
|
|
|
Regs |= getAliases(X86::RCX);
|
|
|
|
Regs |= getAliases(X86::R8);
|
|
|
|
Regs |= getAliases(X86::R9);
|
|
|
|
Regs |= getAliases(X86::R10);
|
|
|
|
Regs |= getAliases(X86::R11);
|
|
|
|
Regs |= getAliases(X86::R12);
|
|
|
|
Regs |= getAliases(X86::R13);
|
|
|
|
Regs |= getAliases(X86::R14);
|
|
|
|
Regs |= getAliases(X86::R15);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
Regs.set(X86::RAX);
|
|
|
|
Regs.set(X86::RBX);
|
|
|
|
Regs.set(X86::RBP);
|
|
|
|
Regs.set(X86::RSI);
|
|
|
|
Regs.set(X86::RDI);
|
|
|
|
Regs.set(X86::RDX);
|
|
|
|
Regs.set(X86::RCX);
|
|
|
|
Regs.set(X86::R8);
|
|
|
|
Regs.set(X86::R9);
|
|
|
|
Regs.set(X86::R10);
|
|
|
|
Regs.set(X86::R11);
|
|
|
|
Regs.set(X86::R12);
|
|
|
|
Regs.set(X86::R13);
|
|
|
|
Regs.set(X86::R14);
|
|
|
|
Regs.set(X86::R15);
|
|
|
|
}
|
|
|
|
|
|
|
|
void getClassicGPRegs(BitVector &Regs) const override {
|
|
|
|
Regs |= getAliases(X86::RAX);
|
|
|
|
Regs |= getAliases(X86::RBX);
|
|
|
|
Regs |= getAliases(X86::RBP);
|
|
|
|
Regs |= getAliases(X86::RSI);
|
|
|
|
Regs |= getAliases(X86::RDI);
|
|
|
|
Regs |= getAliases(X86::RDX);
|
|
|
|
Regs |= getAliases(X86::RCX);
|
|
|
|
}
|
|
|
|
|
2021-08-11 01:02:32 +08:00
|
|
|
void getRepRegs(BitVector &Regs) const override {
|
|
|
|
Regs |= getAliases(X86::RCX);
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
MCPhysReg getAliasSized(MCPhysReg Reg, uint8_t Size) const override {
|
|
|
|
switch (Reg) {
|
2022-01-09 09:48:33 +08:00
|
|
|
case X86::RAX: case X86::EAX: case X86::AX: case X86::AL: case X86::AH:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::RAX; case 4: return X86::EAX;
|
|
|
|
case 2: return X86::AX; case 1: return X86::AL;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::RBX: case X86::EBX: case X86::BX: case X86::BL: case X86::BH:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::RBX; case 4: return X86::EBX;
|
|
|
|
case 2: return X86::BX; case 1: return X86::BL;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::RDX: case X86::EDX: case X86::DX: case X86::DL: case X86::DH:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::RDX; case 4: return X86::EDX;
|
|
|
|
case 2: return X86::DX; case 1: return X86::DL;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::RDI: case X86::EDI: case X86::DI: case X86::DIL:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::RDI; case 4: return X86::EDI;
|
|
|
|
case 2: return X86::DI; case 1: return X86::DIL;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::RSI: case X86::ESI: case X86::SI: case X86::SIL:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::RSI; case 4: return X86::ESI;
|
|
|
|
case 2: return X86::SI; case 1: return X86::SIL;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::RCX: case X86::ECX: case X86::CX: case X86::CL: case X86::CH:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::RCX; case 4: return X86::ECX;
|
|
|
|
case 2: return X86::CX; case 1: return X86::CL;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::RSP: case X86::ESP: case X86::SP: case X86::SPL:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::RSP; case 4: return X86::ESP;
|
|
|
|
case 2: return X86::SP; case 1: return X86::SPL;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::RBP: case X86::EBP: case X86::BP: case X86::BPL:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::RBP; case 4: return X86::EBP;
|
|
|
|
case 2: return X86::BP; case 1: return X86::BPL;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::R8: case X86::R8D: case X86::R8W: case X86::R8B:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::R8; case 4: return X86::R8D;
|
|
|
|
case 2: return X86::R8W; case 1: return X86::R8B;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::R9: case X86::R9D: case X86::R9W: case X86::R9B:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::R9; case 4: return X86::R9D;
|
|
|
|
case 2: return X86::R9W; case 1: return X86::R9B;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::R10: case X86::R10D: case X86::R10W: case X86::R10B:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::R10; case 4: return X86::R10D;
|
|
|
|
case 2: return X86::R10W; case 1: return X86::R10B;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::R11: case X86::R11D: case X86::R11W: case X86::R11B:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::R11; case 4: return X86::R11D;
|
|
|
|
case 2: return X86::R11W; case 1: return X86::R11B;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::R12: case X86::R12D: case X86::R12W: case X86::R12B:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::R12; case 4: return X86::R12D;
|
|
|
|
case 2: return X86::R12W; case 1: return X86::R12B;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::R13: case X86::R13D: case X86::R13W: case X86::R13B:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::R13; case 4: return X86::R13D;
|
|
|
|
case 2: return X86::R13W; case 1: return X86::R13B;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::R14: case X86::R14D: case X86::R14W: case X86::R14B:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::R14; case 4: return X86::R14D;
|
|
|
|
case 2: return X86::R14W; case 1: return X86::R14B;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
case X86::R15: case X86::R15D: case X86::R15W: case X86::R15B:
|
|
|
|
switch (Size) {
|
|
|
|
case 8: return X86::R15; case 4: return X86::R15D;
|
|
|
|
case 2: return X86::R15W; case 1: return X86::R15B;
|
|
|
|
default: llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
default:
|
|
|
|
dbgs() << Reg << " (get alias sized)\n";
|
|
|
|
llvm_unreachable("Unexpected reg number");
|
|
|
|
break;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isUpper8BitReg(MCPhysReg Reg) const override {
|
|
|
|
switch (Reg) {
|
2022-01-09 09:48:33 +08:00
|
|
|
case X86::AH:
|
|
|
|
case X86::BH:
|
|
|
|
case X86::CH:
|
|
|
|
case X86::DH:
|
|
|
|
return true;
|
|
|
|
default:
|
|
|
|
return false;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool cannotUseREX(const MCInst &Inst) const override {
|
2022-01-09 09:48:33 +08:00
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
case X86::MOV8mr_NOREX:
|
|
|
|
case X86::MOV8rm_NOREX:
|
|
|
|
case X86::MOV8rr_NOREX:
|
|
|
|
case X86::MOVSX32rm8_NOREX:
|
|
|
|
case X86::MOVSX32rr8_NOREX:
|
|
|
|
case X86::MOVZX32rm8_NOREX:
|
|
|
|
case X86::MOVZX32rr8_NOREX:
|
|
|
|
case X86::MOV8mr:
|
|
|
|
case X86::MOV8rm:
|
|
|
|
case X86::MOV8rr:
|
|
|
|
case X86::MOVSX32rm8:
|
|
|
|
case X86::MOVSX32rr8:
|
|
|
|
case X86::MOVZX32rm8:
|
|
|
|
case X86::MOVZX32rr8:
|
|
|
|
case X86::TEST8ri:
|
2022-05-12 00:34:10 +08:00
|
|
|
for (const MCOperand &Operand : MCPlus::primeOperands(Inst)) {
|
2022-01-09 09:48:33 +08:00
|
|
|
if (!Operand.isReg())
|
|
|
|
continue;
|
|
|
|
if (isUpper8BitReg(Operand.getReg()))
|
|
|
|
return true;
|
|
|
|
}
|
2020-12-02 08:29:39 +08:00
|
|
|
LLVM_FALLTHROUGH;
|
2022-01-09 09:48:33 +08:00
|
|
|
default:
|
|
|
|
return false;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isStackAccess(const MCInst &Inst, bool &IsLoad, bool &IsStore,
|
|
|
|
bool &IsStoreFromReg, MCPhysReg &Reg, int32_t &SrcImm,
|
2022-01-09 09:48:33 +08:00
|
|
|
uint16_t &StackPtrReg, int64_t &StackOffset, uint8_t &Size,
|
|
|
|
bool &IsSimple, bool &IsIndexed) const override {
|
2018-03-10 01:45:13 +08:00
|
|
|
// Detect simple push/pop cases first
|
2021-04-08 15:19:26 +08:00
|
|
|
if (int Sz = getPushSize(Inst)) {
|
2018-03-10 01:45:13 +08:00
|
|
|
IsLoad = false;
|
|
|
|
IsStore = true;
|
|
|
|
IsStoreFromReg = true;
|
|
|
|
StackPtrReg = X86::RSP;
|
|
|
|
StackOffset = -Sz;
|
|
|
|
Size = Sz;
|
|
|
|
IsSimple = true;
|
2021-12-29 09:52:08 +08:00
|
|
|
if (Inst.getOperand(0).isImm())
|
2018-03-10 01:45:13 +08:00
|
|
|
SrcImm = Inst.getOperand(0).getImm();
|
2021-12-29 09:52:08 +08:00
|
|
|
else if (Inst.getOperand(0).isReg())
|
2018-03-10 01:45:13 +08:00
|
|
|
Reg = Inst.getOperand(0).getReg();
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-03-10 01:45:13 +08:00
|
|
|
IsSimple = false;
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
return true;
|
|
|
|
}
|
2021-04-08 15:19:26 +08:00
|
|
|
if (int Sz = getPopSize(Inst)) {
|
2018-03-10 01:45:13 +08:00
|
|
|
IsLoad = true;
|
|
|
|
IsStore = false;
|
2021-12-16 05:29:46 +08:00
|
|
|
if (Inst.getNumOperands() == 0 || !Inst.getOperand(0).isReg()) {
|
|
|
|
IsSimple = false;
|
|
|
|
} else {
|
|
|
|
Reg = Inst.getOperand(0).getReg();
|
|
|
|
IsSimple = true;
|
|
|
|
}
|
2018-03-10 01:45:13 +08:00
|
|
|
StackPtrReg = X86::RSP;
|
|
|
|
StackOffset = 0;
|
|
|
|
Size = Sz;
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct InstInfo {
|
|
|
|
// Size in bytes that Inst loads from memory.
|
|
|
|
uint8_t DataSize;
|
|
|
|
bool IsLoad;
|
|
|
|
bool IsStore;
|
|
|
|
bool StoreFromReg;
|
|
|
|
bool Simple;
|
|
|
|
};
|
|
|
|
|
|
|
|
InstInfo I;
|
2021-04-08 15:19:26 +08:00
|
|
|
int MemOpNo = getMemoryOperandNo(Inst);
|
2021-04-18 12:48:46 +08:00
|
|
|
const MCInstrDesc &MCII = Info->get(Inst.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
// If it is not dealing with a memory operand, we discard it
|
|
|
|
if (MemOpNo == -1 || MCII.isCall())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default: {
|
|
|
|
uint8_t Sz = 0;
|
|
|
|
bool IsLoad = MCII.mayLoad();
|
|
|
|
bool IsStore = MCII.mayStore();
|
|
|
|
// Is it LEA? (deals with memory but is not loading nor storing)
|
|
|
|
if (!IsLoad && !IsStore)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Try to guess data size involved in the load/store by looking at the
|
|
|
|
// register size. If there's no reg involved, return 0 as size, meaning
|
|
|
|
// we don't know.
|
|
|
|
for (unsigned I = 0, E = MCII.getNumOperands(); I != E; ++I) {
|
|
|
|
if (MCII.OpInfo[I].OperandType != MCOI::OPERAND_REGISTER)
|
|
|
|
continue;
|
|
|
|
if (static_cast<int>(I) >= MemOpNo && I < X86::AddrNumOperands)
|
|
|
|
continue;
|
2020-12-02 08:29:39 +08:00
|
|
|
Sz = RegInfo->getRegClass(MCII.OpInfo[I].RegClass).getSizeInBits() / 8;
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
I = {Sz, IsLoad, IsStore, false, false};
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case X86::MOV16rm: I = {2, true, false, false, true}; break;
|
|
|
|
case X86::MOV32rm: I = {4, true, false, false, true}; break;
|
|
|
|
case X86::MOV64rm: I = {8, true, false, false, true}; break;
|
|
|
|
case X86::MOV16mr: I = {2, false, true, true, true}; break;
|
|
|
|
case X86::MOV32mr: I = {4, false, true, true, true}; break;
|
|
|
|
case X86::MOV64mr: I = {8, false, true, true, true}; break;
|
|
|
|
case X86::MOV16mi: I = {2, false, true, false, true}; break;
|
|
|
|
case X86::MOV32mi: I = {4, false, true, false, true}; break;
|
|
|
|
} // end switch (Inst.getOpcode())
|
|
|
|
|
|
|
|
unsigned BaseRegNum;
|
|
|
|
int64_t ScaleValue;
|
|
|
|
unsigned IndexRegNum;
|
|
|
|
int64_t DispValue;
|
|
|
|
unsigned SegRegNum;
|
|
|
|
const MCExpr *DispExpr;
|
|
|
|
if (!evaluateX86MemoryOperand(Inst, &BaseRegNum, &ScaleValue, &IndexRegNum,
|
|
|
|
&DispValue, &SegRegNum, &DispExpr)) {
|
2020-12-02 08:29:39 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "Evaluate failed on ");
|
|
|
|
LLVM_DEBUG(Inst.dump());
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure it's a stack access
|
2021-12-29 09:52:08 +08:00
|
|
|
if (BaseRegNum != X86::RBP && BaseRegNum != X86::RSP)
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
IsLoad = I.IsLoad;
|
|
|
|
IsStore = I.IsStore;
|
|
|
|
IsStoreFromReg = I.StoreFromReg;
|
|
|
|
Size = I.DataSize;
|
|
|
|
IsSimple = I.Simple;
|
|
|
|
StackPtrReg = BaseRegNum;
|
|
|
|
StackOffset = DispValue;
|
|
|
|
IsIndexed = IndexRegNum != X86::NoRegister || SegRegNum != X86::NoRegister;
|
|
|
|
|
|
|
|
if (!I.Simple)
|
|
|
|
return true;
|
|
|
|
|
|
|
|
// Retrieve related register in simple MOV from/to stack operations.
|
|
|
|
unsigned MemOpOffset = static_cast<unsigned>(MemOpNo);
|
|
|
|
if (I.IsLoad) {
|
2021-04-08 15:19:26 +08:00
|
|
|
MCOperand RegOpnd = Inst.getOperand(0);
|
2018-03-10 01:45:13 +08:00
|
|
|
assert(RegOpnd.isReg() && "unexpected destination operand");
|
|
|
|
Reg = RegOpnd.getReg();
|
|
|
|
} else if (I.IsStore) {
|
2021-04-08 15:19:26 +08:00
|
|
|
MCOperand SrcOpnd =
|
|
|
|
Inst.getOperand(MemOpOffset + X86::AddrSegmentReg + 1);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (I.StoreFromReg) {
|
|
|
|
assert(SrcOpnd.isReg() && "unexpected source operand");
|
|
|
|
Reg = SrcOpnd.getReg();
|
|
|
|
} else {
|
|
|
|
assert(SrcOpnd.isImm() && "unexpected source operand");
|
|
|
|
SrcImm = SrcOpnd.getImm();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
void changeToPushOrPop(MCInst &Inst) const override {
|
|
|
|
assert(!isPush(Inst) && !isPop(Inst));
|
|
|
|
|
|
|
|
struct InstInfo {
|
|
|
|
// Size in bytes that Inst loads from memory.
|
|
|
|
uint8_t DataSize;
|
|
|
|
bool IsLoad;
|
|
|
|
bool StoreFromReg;
|
|
|
|
};
|
|
|
|
|
|
|
|
InstInfo I;
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default: {
|
|
|
|
llvm_unreachable("Unhandled opcode");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
case X86::MOV16rm: I = {2, true, false}; break;
|
|
|
|
case X86::MOV32rm: I = {4, true, false}; break;
|
|
|
|
case X86::MOV64rm: I = {8, true, false}; break;
|
|
|
|
case X86::MOV16mr: I = {2, false, true}; break;
|
|
|
|
case X86::MOV32mr: I = {4, false, true}; break;
|
|
|
|
case X86::MOV64mr: I = {8, false, true}; break;
|
|
|
|
case X86::MOV16mi: I = {2, false, false}; break;
|
|
|
|
case X86::MOV32mi: I = {4, false, false}; break;
|
|
|
|
} // end switch (Inst.getOpcode())
|
|
|
|
|
|
|
|
unsigned BaseRegNum;
|
|
|
|
int64_t ScaleValue;
|
|
|
|
unsigned IndexRegNum;
|
|
|
|
int64_t DispValue;
|
|
|
|
unsigned SegRegNum;
|
|
|
|
const MCExpr *DispExpr;
|
|
|
|
if (!evaluateX86MemoryOperand(Inst, &BaseRegNum, &ScaleValue, &IndexRegNum,
|
|
|
|
&DispValue, &SegRegNum, &DispExpr)) {
|
|
|
|
llvm_unreachable("Evaluate failed");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
// Make sure it's a stack access
|
|
|
|
if (BaseRegNum != X86::RBP && BaseRegNum != X86::RSP) {
|
|
|
|
llvm_unreachable("Not a stack access");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned MemOpOffset = getMemoryOperandNo(Inst);
|
|
|
|
unsigned NewOpcode = 0;
|
|
|
|
if (I.IsLoad) {
|
|
|
|
switch (I.DataSize) {
|
2022-01-09 09:48:33 +08:00
|
|
|
case 2: NewOpcode = X86::POP16r; break;
|
|
|
|
case 4: NewOpcode = X86::POP32r; break;
|
|
|
|
case 8: NewOpcode = X86::POP64r; break;
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Unexpected size");
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
2021-04-08 15:19:26 +08:00
|
|
|
unsigned RegOpndNum = Inst.getOperand(0).getReg();
|
2018-03-10 01:45:13 +08:00
|
|
|
Inst.clear();
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
Inst.addOperand(MCOperand::createReg(RegOpndNum));
|
|
|
|
} else {
|
2021-04-08 15:19:26 +08:00
|
|
|
MCOperand SrcOpnd =
|
|
|
|
Inst.getOperand(MemOpOffset + X86::AddrSegmentReg + 1);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (I.StoreFromReg) {
|
|
|
|
switch (I.DataSize) {
|
2022-01-09 09:48:33 +08:00
|
|
|
case 2: NewOpcode = X86::PUSH16r; break;
|
|
|
|
case 4: NewOpcode = X86::PUSH32r; break;
|
|
|
|
case 8: NewOpcode = X86::PUSH64r; break;
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Unexpected size");
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
2022-01-09 09:48:33 +08:00
|
|
|
assert(SrcOpnd.isReg() && "Unexpected source operand");
|
2021-04-08 15:19:26 +08:00
|
|
|
unsigned RegOpndNum = SrcOpnd.getReg();
|
2018-03-10 01:45:13 +08:00
|
|
|
Inst.clear();
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
Inst.addOperand(MCOperand::createReg(RegOpndNum));
|
|
|
|
} else {
|
|
|
|
switch (I.DataSize) {
|
2022-01-09 09:48:33 +08:00
|
|
|
case 2: NewOpcode = X86::PUSH16i8; break;
|
|
|
|
case 4: NewOpcode = X86::PUSH32i8; break;
|
|
|
|
case 8: NewOpcode = X86::PUSH64i32; break;
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Unexpected size");
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
2022-01-09 09:48:33 +08:00
|
|
|
assert(SrcOpnd.isImm() && "Unexpected source operand");
|
2021-04-08 15:19:26 +08:00
|
|
|
int64_t SrcImm = SrcOpnd.getImm();
|
2018-03-10 01:45:13 +08:00
|
|
|
Inst.clear();
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
Inst.addOperand(MCOperand::createImm(SrcImm));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isStackAdjustment(const MCInst &Inst) const override {
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case X86::SUB64ri32:
|
|
|
|
case X86::SUB64ri8:
|
|
|
|
case X86::ADD64ri32:
|
|
|
|
case X86::ADD64ri8:
|
|
|
|
case X86::LEA64r:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2021-04-18 12:48:46 +08:00
|
|
|
const MCInstrDesc &MCII = Info->get(Inst.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
for (int I = 0, E = MCII.getNumDefs(); I != E; ++I) {
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCOperand &Operand = Inst.getOperand(I);
|
2021-12-29 09:52:08 +08:00
|
|
|
if (Operand.isReg() && Operand.getReg() == X86::RSP)
|
2018-03-10 01:45:13 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2022-05-21 08:42:58 +08:00
|
|
|
bool
|
|
|
|
evaluateStackOffsetExpr(const MCInst &Inst, int64_t &Output,
|
|
|
|
std::pair<MCPhysReg, int64_t> Input1,
|
|
|
|
std::pair<MCPhysReg, int64_t> Input2) const override {
|
2018-03-10 01:45:13 +08:00
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
auto getOperandVal = [&](MCPhysReg Reg) -> ErrorOr<int64_t> {
|
2018-03-10 01:45:13 +08:00
|
|
|
if (Reg == Input1.first)
|
|
|
|
return Input1.second;
|
|
|
|
if (Reg == Input2.first)
|
|
|
|
return Input2.second;
|
|
|
|
return make_error_code(errc::result_out_of_range);
|
|
|
|
};
|
|
|
|
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
|
|
|
|
case X86::SUB64ri32:
|
|
|
|
case X86::SUB64ri8:
|
2019-06-29 00:21:27 +08:00
|
|
|
if (!Inst.getOperand(2).isImm())
|
|
|
|
return false;
|
2021-04-08 15:19:26 +08:00
|
|
|
if (ErrorOr<int64_t> InputVal =
|
2021-12-29 09:52:08 +08:00
|
|
|
getOperandVal(Inst.getOperand(1).getReg()))
|
2018-03-10 01:45:13 +08:00
|
|
|
Output = *InputVal - Inst.getOperand(2).getImm();
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
|
|
|
break;
|
|
|
|
case X86::ADD64ri32:
|
|
|
|
case X86::ADD64ri8:
|
2019-06-29 00:21:27 +08:00
|
|
|
if (!Inst.getOperand(2).isImm())
|
|
|
|
return false;
|
2021-04-08 15:19:26 +08:00
|
|
|
if (ErrorOr<int64_t> InputVal =
|
2021-12-29 09:52:08 +08:00
|
|
|
getOperandVal(Inst.getOperand(1).getReg()))
|
2018-03-10 01:45:13 +08:00
|
|
|
Output = *InputVal + Inst.getOperand(2).getImm();
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
|
|
|
break;
|
2018-06-12 04:18:44 +08:00
|
|
|
case X86::ADD64i32:
|
2019-06-29 00:21:27 +08:00
|
|
|
if (!Inst.getOperand(0).isImm())
|
|
|
|
return false;
|
2021-12-29 09:52:08 +08:00
|
|
|
if (ErrorOr<int64_t> InputVal = getOperandVal(X86::RAX))
|
2018-06-12 04:18:44 +08:00
|
|
|
Output = *InputVal + Inst.getOperand(0).getImm();
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-06-12 04:18:44 +08:00
|
|
|
return false;
|
|
|
|
break;
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
case X86::LEA64r: {
|
|
|
|
unsigned BaseRegNum;
|
|
|
|
int64_t ScaleValue;
|
|
|
|
unsigned IndexRegNum;
|
|
|
|
int64_t DispValue;
|
|
|
|
unsigned SegRegNum;
|
2021-05-14 01:50:47 +08:00
|
|
|
const MCExpr *DispExpr = nullptr;
|
2018-03-10 01:45:13 +08:00
|
|
|
if (!evaluateX86MemoryOperand(Inst, &BaseRegNum, &ScaleValue,
|
|
|
|
&IndexRegNum, &DispValue, &SegRegNum,
|
2021-12-29 09:52:08 +08:00
|
|
|
&DispExpr))
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
if (BaseRegNum == X86::NoRegister || IndexRegNum != X86::NoRegister ||
|
2021-12-29 09:52:08 +08:00
|
|
|
SegRegNum != X86::NoRegister || DispExpr)
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
|
|
|
|
2021-12-29 09:52:08 +08:00
|
|
|
if (ErrorOr<int64_t> InputVal = getOperandVal(BaseRegNum))
|
2018-03-10 01:45:13 +08:00
|
|
|
Output = *InputVal + DispValue;
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isRegToRegMove(const MCInst &Inst, MCPhysReg &From,
|
|
|
|
MCPhysReg &To) const override {
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case X86::LEAVE:
|
|
|
|
case X86::LEAVE64:
|
|
|
|
To = getStackPointer();
|
|
|
|
From = getFramePointer();
|
|
|
|
return true;
|
|
|
|
case X86::MOV64rr:
|
|
|
|
To = Inst.getOperand(0).getReg();
|
|
|
|
From = Inst.getOperand(1).getReg();
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
MCPhysReg getStackPointer() const override { return X86::RSP; }
|
|
|
|
MCPhysReg getFramePointer() const override { return X86::RBP; }
|
|
|
|
MCPhysReg getFlagsReg() const override { return X86::EFLAGS; }
|
|
|
|
|
2022-01-09 09:48:33 +08:00
|
|
|
bool escapesVariable(const MCInst &Inst,
|
|
|
|
bool HasFramePointer) const override {
|
2021-04-08 15:19:26 +08:00
|
|
|
int MemOpNo = getMemoryOperandNo(Inst);
|
2021-04-18 12:48:46 +08:00
|
|
|
const MCInstrDesc &MCII = Info->get(Inst.getOpcode());
|
2021-04-08 15:19:26 +08:00
|
|
|
const unsigned NumDefs = MCII.getNumDefs();
|
2020-12-02 08:29:39 +08:00
|
|
|
static BitVector SPBPAliases(BitVector(getAliases(X86::RSP)) |=
|
2018-03-10 01:45:13 +08:00
|
|
|
getAliases(X86::RBP));
|
|
|
|
static BitVector SPAliases(getAliases(X86::RSP));
|
|
|
|
|
|
|
|
// FIXME: PUSH can be technically a leak, but let's ignore this for now
|
2021-11-09 11:54:05 +08:00
|
|
|
// because a lot of harmless prologue code will spill SP to the stack.
|
|
|
|
// Unless push is clearly pushing an object address to the stack as
|
|
|
|
// demonstrated by having a MemOp.
|
2018-03-10 01:45:13 +08:00
|
|
|
bool IsPush = isPush(Inst);
|
|
|
|
if (IsPush && MemOpNo == -1)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// We use this to detect LEA (has memop but does not access mem)
|
|
|
|
bool AccessMem = MCII.mayLoad() || MCII.mayStore();
|
|
|
|
bool DoesLeak = false;
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
for (int I = 0, E = MCPlus::getNumPrimeOperands(Inst); I != E; ++I) {
|
2021-11-09 11:54:05 +08:00
|
|
|
// Ignore if SP/BP is used to dereference memory -- that's fine
|
2018-03-10 01:45:13 +08:00
|
|
|
if (MemOpNo != -1 && !IsPush && AccessMem && I >= MemOpNo &&
|
|
|
|
I <= MemOpNo + 5)
|
|
|
|
continue;
|
|
|
|
// Ignore if someone is writing to SP/BP
|
|
|
|
if (I < static_cast<int>(NumDefs))
|
|
|
|
continue;
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCOperand &Operand = Inst.getOperand(I);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (HasFramePointer && Operand.isReg() && SPBPAliases[Operand.getReg()]) {
|
|
|
|
DoesLeak = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (!HasFramePointer && Operand.isReg() && SPAliases[Operand.getReg()]) {
|
|
|
|
DoesLeak = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// If potential leak, check if it is not just writing to itself/sp/bp
|
|
|
|
if (DoesLeak) {
|
|
|
|
for (int I = 0, E = NumDefs; I != E; ++I) {
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCOperand &Operand = Inst.getOperand(I);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (HasFramePointer && Operand.isReg() &&
|
|
|
|
SPBPAliases[Operand.getReg()]) {
|
|
|
|
DoesLeak = false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (!HasFramePointer && Operand.isReg() &&
|
|
|
|
SPAliases[Operand.getReg()]) {
|
|
|
|
DoesLeak = false;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return DoesLeak;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool addToImm(MCInst &Inst, int64_t &Amt, MCContext *Ctx) const override {
|
|
|
|
unsigned ImmOpNo = -1U;
|
2021-04-08 15:19:26 +08:00
|
|
|
int MemOpNo = getMemoryOperandNo(Inst);
|
2021-12-29 09:52:08 +08:00
|
|
|
if (MemOpNo != -1)
|
2018-03-10 01:45:13 +08:00
|
|
|
ImmOpNo = MemOpNo + X86::AddrDisp;
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2022-01-09 09:48:33 +08:00
|
|
|
for (unsigned Index = 0; Index < MCPlus::getNumPrimeOperands(Inst);
|
|
|
|
++Index)
|
2021-12-29 09:52:08 +08:00
|
|
|
if (Inst.getOperand(Index).isImm())
|
2018-03-10 01:45:13 +08:00
|
|
|
ImmOpNo = Index;
|
|
|
|
if (ImmOpNo == -1U)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
MCOperand &Operand = Inst.getOperand(ImmOpNo);
|
|
|
|
Amt += Operand.getImm();
|
|
|
|
Operand.setImm(Amt);
|
|
|
|
// Check for the need for relaxation
|
|
|
|
if (int64_t(Amt) == int64_t(int8_t(Amt)))
|
|
|
|
return true;
|
|
|
|
|
|
|
|
// Relax instruction
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
case X86::SUB64ri8:
|
|
|
|
Inst.setOpcode(X86::SUB64ri32);
|
|
|
|
break;
|
|
|
|
case X86::ADD64ri8:
|
|
|
|
Inst.setOpcode(X86::ADD64ri32);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
// No need for relaxation
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// TODO: this implementation currently works for the most common opcodes that
|
|
|
|
/// load from memory. It can be extended to work with memory store opcodes as
|
|
|
|
/// well as more memory load opcodes.
|
|
|
|
bool replaceMemOperandWithImm(MCInst &Inst, StringRef ConstantData,
|
2020-12-02 08:29:39 +08:00
|
|
|
uint64_t Offset) const override {
|
2018-03-10 01:45:13 +08:00
|
|
|
enum CheckSignExt : uint8_t {
|
|
|
|
NOCHECK = 0,
|
|
|
|
CHECK8,
|
|
|
|
CHECK32,
|
|
|
|
};
|
|
|
|
|
|
|
|
using CheckList = std::vector<std::pair<CheckSignExt, unsigned>>;
|
|
|
|
struct InstInfo {
|
|
|
|
// Size in bytes that Inst loads from memory.
|
|
|
|
uint8_t DataSize;
|
|
|
|
|
|
|
|
// True when the target operand has to be duplicated because the opcode
|
|
|
|
// expects a LHS operand.
|
|
|
|
bool HasLHS;
|
|
|
|
|
|
|
|
// List of checks and corresponding opcodes to be used. We try to use the
|
|
|
|
// smallest possible immediate value when various sizes are available,
|
|
|
|
// hence we may need to check whether a larger constant fits in a smaller
|
|
|
|
// immediate.
|
|
|
|
CheckList Checks;
|
|
|
|
};
|
|
|
|
|
|
|
|
InstInfo I;
|
|
|
|
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default: {
|
|
|
|
switch (getPopSize(Inst)) {
|
|
|
|
case 2: I = {2, false, {{NOCHECK, X86::MOV16ri}}}; break;
|
|
|
|
case 4: I = {4, false, {{NOCHECK, X86::MOV32ri}}}; break;
|
|
|
|
case 8: I = {8, false, {{CHECK32, X86::MOV64ri32},
|
|
|
|
{NOCHECK, X86::MOV64rm}}}; break;
|
|
|
|
default: return false;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// MOV
|
|
|
|
case X86::MOV8rm: I = {1, false, {{NOCHECK, X86::MOV8ri}}}; break;
|
|
|
|
case X86::MOV16rm: I = {2, false, {{NOCHECK, X86::MOV16ri}}}; break;
|
|
|
|
case X86::MOV32rm: I = {4, false, {{NOCHECK, X86::MOV32ri}}}; break;
|
|
|
|
case X86::MOV64rm: I = {8, false, {{CHECK32, X86::MOV64ri32},
|
|
|
|
{NOCHECK, X86::MOV64rm}}}; break;
|
|
|
|
|
|
|
|
// MOVZX
|
|
|
|
case X86::MOVZX16rm8: I = {1, false, {{NOCHECK, X86::MOV16ri}}}; break;
|
|
|
|
case X86::MOVZX32rm8: I = {1, false, {{NOCHECK, X86::MOV32ri}}}; break;
|
|
|
|
case X86::MOVZX32rm16: I = {2, false, {{NOCHECK, X86::MOV32ri}}}; break;
|
|
|
|
|
|
|
|
// CMP
|
|
|
|
case X86::CMP8rm: I = {1, false, {{NOCHECK, X86::CMP8ri}}}; break;
|
|
|
|
case X86::CMP16rm: I = {2, false, {{CHECK8, X86::CMP16ri8},
|
|
|
|
{NOCHECK, X86::CMP16ri}}}; break;
|
|
|
|
case X86::CMP32rm: I = {4, false, {{CHECK8, X86::CMP32ri8},
|
|
|
|
{NOCHECK, X86::CMP32ri}}}; break;
|
|
|
|
case X86::CMP64rm: I = {8, false, {{CHECK8, X86::CMP64ri8},
|
|
|
|
{CHECK32, X86::CMP64ri32},
|
|
|
|
{NOCHECK, X86::CMP64rm}}}; break;
|
|
|
|
|
|
|
|
// TEST
|
|
|
|
case X86::TEST8mr: I = {1, false, {{NOCHECK, X86::TEST8ri}}}; break;
|
|
|
|
case X86::TEST16mr: I = {2, false, {{NOCHECK, X86::TEST16ri}}}; break;
|
|
|
|
case X86::TEST32mr: I = {4, false, {{NOCHECK, X86::TEST32ri}}}; break;
|
|
|
|
case X86::TEST64mr: I = {8, false, {{CHECK32, X86::TEST64ri32},
|
|
|
|
{NOCHECK, X86::TEST64mr}}}; break;
|
|
|
|
|
|
|
|
// ADD
|
|
|
|
case X86::ADD8rm: I = {1, true, {{NOCHECK, X86::ADD8ri}}}; break;
|
|
|
|
case X86::ADD16rm: I = {2, true, {{CHECK8, X86::ADD16ri8},
|
|
|
|
{NOCHECK, X86::ADD16ri}}}; break;
|
|
|
|
case X86::ADD32rm: I = {4, true, {{CHECK8, X86::ADD32ri8},
|
|
|
|
{NOCHECK, X86::ADD32ri}}}; break;
|
|
|
|
case X86::ADD64rm: I = {8, true, {{CHECK8, X86::ADD64ri8},
|
|
|
|
{CHECK32, X86::ADD64ri32},
|
|
|
|
{NOCHECK, X86::ADD64rm}}}; break;
|
|
|
|
|
|
|
|
// SUB
|
|
|
|
case X86::SUB8rm: I = {1, true, {{NOCHECK, X86::SUB8ri}}}; break;
|
|
|
|
case X86::SUB16rm: I = {2, true, {{CHECK8, X86::SUB16ri8},
|
|
|
|
{NOCHECK, X86::SUB16ri}}}; break;
|
|
|
|
case X86::SUB32rm: I = {4, true, {{CHECK8, X86::SUB32ri8},
|
|
|
|
{NOCHECK, X86::SUB32ri}}}; break;
|
|
|
|
case X86::SUB64rm: I = {8, true, {{CHECK8, X86::SUB64ri8},
|
|
|
|
{CHECK32, X86::SUB64ri32},
|
|
|
|
{NOCHECK, X86::SUB64rm}}}; break;
|
|
|
|
|
|
|
|
// AND
|
|
|
|
case X86::AND8rm: I = {1, true, {{NOCHECK, X86::AND8ri}}}; break;
|
|
|
|
case X86::AND16rm: I = {2, true, {{CHECK8, X86::AND16ri8},
|
|
|
|
{NOCHECK, X86::AND16ri}}}; break;
|
|
|
|
case X86::AND32rm: I = {4, true, {{CHECK8, X86::AND32ri8},
|
|
|
|
{NOCHECK, X86::AND32ri}}}; break;
|
|
|
|
case X86::AND64rm: I = {8, true, {{CHECK8, X86::AND64ri8},
|
|
|
|
{CHECK32, X86::AND64ri32},
|
|
|
|
{NOCHECK, X86::AND64rm}}}; break;
|
|
|
|
|
|
|
|
// OR
|
|
|
|
case X86::OR8rm: I = {1, true, {{NOCHECK, X86::OR8ri}}}; break;
|
|
|
|
case X86::OR16rm: I = {2, true, {{CHECK8, X86::OR16ri8},
|
|
|
|
{NOCHECK, X86::OR16ri}}}; break;
|
|
|
|
case X86::OR32rm: I = {4, true, {{CHECK8, X86::OR32ri8},
|
|
|
|
{NOCHECK, X86::OR32ri}}}; break;
|
|
|
|
case X86::OR64rm: I = {8, true, {{CHECK8, X86::OR64ri8},
|
|
|
|
{CHECK32, X86::OR64ri32},
|
|
|
|
{NOCHECK, X86::OR64rm}}}; break;
|
|
|
|
|
|
|
|
// XOR
|
|
|
|
case X86::XOR8rm: I = {1, true, {{NOCHECK, X86::XOR8ri}}}; break;
|
|
|
|
case X86::XOR16rm: I = {2, true, {{CHECK8, X86::XOR16ri8},
|
|
|
|
{NOCHECK, X86::XOR16ri}}}; break;
|
|
|
|
case X86::XOR32rm: I = {4, true, {{CHECK8, X86::XOR32ri8},
|
|
|
|
{NOCHECK, X86::XOR32ri}}}; break;
|
|
|
|
case X86::XOR64rm: I = {8, true, {{CHECK8, X86::XOR64ri8},
|
|
|
|
{CHECK32, X86::XOR64ri32},
|
|
|
|
{NOCHECK, X86::XOR64rm}}}; break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Compute the immediate value.
|
|
|
|
assert(Offset + I.DataSize <= ConstantData.size() &&
|
|
|
|
"invalid offset for given constant data");
|
|
|
|
int64_t ImmVal =
|
2022-01-09 09:48:33 +08:00
|
|
|
DataExtractor(ConstantData, true, 8).getSigned(&Offset, I.DataSize);
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// Compute the new opcode.
|
|
|
|
unsigned NewOpcode = 0;
|
2021-04-08 15:19:26 +08:00
|
|
|
for (const std::pair<CheckSignExt, unsigned> &Check : I.Checks) {
|
2018-03-10 01:45:13 +08:00
|
|
|
NewOpcode = Check.second;
|
|
|
|
if (Check.first == NOCHECK)
|
|
|
|
break;
|
2022-02-03 01:04:10 +08:00
|
|
|
if (Check.first == CHECK8 && isInt<8>(ImmVal))
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
2022-02-03 01:04:10 +08:00
|
|
|
if (Check.first == CHECK32 && isInt<32>(ImmVal))
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (NewOpcode == Inst.getOpcode())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// Modify the instruction.
|
|
|
|
MCOperand ImmOp = MCOperand::createImm(ImmVal);
|
2021-05-14 01:50:47 +08:00
|
|
|
uint32_t TargetOpNum = 0;
|
2018-03-10 01:45:13 +08:00
|
|
|
// Test instruction does not follow the regular pattern of putting the
|
|
|
|
// memory reference of a load (5 MCOperands) last in the list of operands.
|
|
|
|
// Since it is not modifying the register operand, it is not treated as
|
|
|
|
// a destination operand and it is not the first operand as it is in the
|
|
|
|
// other instructions we treat here.
|
2022-01-09 09:48:33 +08:00
|
|
|
if (NewOpcode == X86::TEST8ri || NewOpcode == X86::TEST16ri ||
|
|
|
|
NewOpcode == X86::TEST32ri || NewOpcode == X86::TEST64ri32)
|
2018-03-10 01:45:13 +08:00
|
|
|
TargetOpNum = getMemoryOperandNo(Inst) + X86::AddrNumOperands;
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
MCOperand TargetOp = Inst.getOperand(TargetOpNum);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
Inst.addOperand(TargetOp);
|
|
|
|
if (I.HasLHS)
|
|
|
|
Inst.addOperand(TargetOp);
|
|
|
|
Inst.addOperand(ImmOp);
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// TODO: this implementation currently works for the most common opcodes that
|
|
|
|
/// load from memory. It can be extended to work with memory store opcodes as
|
|
|
|
/// well as more memory load opcodes.
|
|
|
|
bool replaceMemOperandWithReg(MCInst &Inst, MCPhysReg RegNum) const override {
|
|
|
|
unsigned NewOpcode;
|
|
|
|
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default: {
|
|
|
|
switch (getPopSize(Inst)) {
|
|
|
|
case 2: NewOpcode = X86::MOV16rr; break;
|
|
|
|
case 4: NewOpcode = X86::MOV32rr; break;
|
|
|
|
case 8: NewOpcode = X86::MOV64rr; break;
|
|
|
|
default: return false;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// MOV
|
|
|
|
case X86::MOV8rm: NewOpcode = X86::MOV8rr; break;
|
|
|
|
case X86::MOV16rm: NewOpcode = X86::MOV16rr; break;
|
|
|
|
case X86::MOV32rm: NewOpcode = X86::MOV32rr; break;
|
|
|
|
case X86::MOV64rm: NewOpcode = X86::MOV64rr; break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Modify the instruction.
|
|
|
|
MCOperand RegOp = MCOperand::createReg(RegNum);
|
|
|
|
MCOperand TargetOp = Inst.getOperand(0);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
Inst.addOperand(TargetOp);
|
|
|
|
Inst.addOperand(RegOp);
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isRedundantMove(const MCInst &Inst) const override {
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// MOV
|
|
|
|
case X86::MOV8rr:
|
|
|
|
case X86::MOV16rr:
|
|
|
|
case X86::MOV32rr:
|
|
|
|
case X86::MOV64rr:
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
assert(Inst.getOperand(0).isReg() && Inst.getOperand(1).isReg());
|
|
|
|
return Inst.getOperand(0).getReg() == Inst.getOperand(1).getReg();
|
|
|
|
}
|
|
|
|
|
|
|
|
bool requiresAlignedAddress(const MCInst &Inst) const override {
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCInstrDesc &Desc = Info->get(Inst.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
for (unsigned int I = 0; I < Desc.getNumOperands(); ++I) {
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCOperandInfo &Op = Desc.OpInfo[I];
|
2018-03-10 01:45:13 +08:00
|
|
|
if (Op.OperandType != MCOI::OPERAND_REGISTER)
|
|
|
|
continue;
|
|
|
|
if (Op.RegClass == X86::VR128RegClassID)
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2019-06-29 00:21:27 +08:00
|
|
|
bool convertJmpToTailCall(MCInst &Inst) override {
|
2020-12-02 08:29:39 +08:00
|
|
|
if (isTailCall(Inst))
|
|
|
|
return false;
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
int NewOpcode;
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case X86::JMP_1:
|
|
|
|
case X86::JMP_2:
|
|
|
|
case X86::JMP_4:
|
2020-12-02 08:29:39 +08:00
|
|
|
NewOpcode = X86::JMP_4;
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
|
|
|
case X86::JMP16m:
|
|
|
|
case X86::JMP32m:
|
|
|
|
case X86::JMP64m:
|
2020-12-02 08:29:39 +08:00
|
|
|
NewOpcode = X86::JMP32m;
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
|
|
|
case X86::JMP16r:
|
|
|
|
case X86::JMP32r:
|
|
|
|
case X86::JMP64r:
|
2020-12-02 08:29:39 +08:00
|
|
|
NewOpcode = X86::JMP32r;
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
Inst.setOpcode(NewOpcode);
|
2021-07-30 08:28:51 +08:00
|
|
|
setTailCall(Inst);
|
2018-03-10 01:45:13 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-06-29 00:21:27 +08:00
|
|
|
bool convertTailCallToJmp(MCInst &Inst) override {
|
|
|
|
int NewOpcode;
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default:
|
|
|
|
return false;
|
2020-12-02 08:29:39 +08:00
|
|
|
case X86::JMP_4:
|
2019-06-29 00:21:27 +08:00
|
|
|
NewOpcode = X86::JMP_1;
|
|
|
|
break;
|
2020-12-02 08:29:39 +08:00
|
|
|
case X86::JMP32m:
|
2019-06-29 00:21:27 +08:00
|
|
|
NewOpcode = X86::JMP64m;
|
|
|
|
break;
|
2020-12-02 08:29:39 +08:00
|
|
|
case X86::JMP32r:
|
2019-06-29 00:21:27 +08:00
|
|
|
NewOpcode = X86::JMP64r;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
Inst.setOpcode(NewOpcode);
|
2021-07-30 08:28:51 +08:00
|
|
|
removeAnnotation(Inst, MCPlus::MCAnnotation::kTailCall);
|
2021-08-04 08:53:32 +08:00
|
|
|
clearOffset(Inst);
|
2019-06-29 00:21:27 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-12-02 08:29:39 +08:00
|
|
|
bool convertTailCallToCall(MCInst &Inst) override {
|
2018-03-10 01:45:13 +08:00
|
|
|
int NewOpcode;
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default:
|
|
|
|
return false;
|
2020-12-02 08:29:39 +08:00
|
|
|
case X86::JMP_4:
|
2018-03-10 01:45:13 +08:00
|
|
|
NewOpcode = X86::CALL64pcrel32;
|
|
|
|
break;
|
2020-12-02 08:29:39 +08:00
|
|
|
case X86::JMP32m:
|
2018-03-10 01:45:13 +08:00
|
|
|
NewOpcode = X86::CALL64m;
|
|
|
|
break;
|
2020-12-02 08:29:39 +08:00
|
|
|
case X86::JMP32r:
|
2018-03-10 01:45:13 +08:00
|
|
|
NewOpcode = X86::CALL64r;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
Inst.setOpcode(NewOpcode);
|
2021-07-30 08:28:51 +08:00
|
|
|
removeAnnotation(Inst, MCPlus::MCAnnotation::kTailCall);
|
2018-03-10 01:45:13 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2022-01-09 09:48:33 +08:00
|
|
|
bool convertCallToIndirectCall(MCInst &Inst, const MCSymbol *TargetLocation,
|
2020-12-02 08:29:39 +08:00
|
|
|
MCContext *Ctx) override {
|
2018-03-10 01:45:13 +08:00
|
|
|
assert((Inst.getOpcode() == X86::CALL64pcrel32 ||
|
2022-05-14 02:56:45 +08:00
|
|
|
(Inst.getOpcode() == X86::JMP_4 && isTailCall(Inst))) &&
|
2018-03-10 01:45:13 +08:00
|
|
|
"64-bit direct (tail) call instruction expected");
|
2022-01-09 09:48:33 +08:00
|
|
|
const auto NewOpcode =
|
|
|
|
(Inst.getOpcode() == X86::CALL64pcrel32) ? X86::CALL64m : X86::JMP32m;
|
2018-03-10 01:45:13 +08:00
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
|
|
|
|
// Replace the first operand and preserve auxiliary operands of
|
|
|
|
// the instruction.
|
|
|
|
Inst.erase(Inst.begin());
|
|
|
|
Inst.insert(Inst.begin(),
|
|
|
|
MCOperand::createReg(X86::NoRegister)); // AddrSegmentReg
|
|
|
|
Inst.insert(Inst.begin(),
|
|
|
|
MCOperand::createExpr( // Displacement
|
2022-01-09 09:48:33 +08:00
|
|
|
MCSymbolRefExpr::create(TargetLocation,
|
|
|
|
MCSymbolRefExpr::VK_None, *Ctx)));
|
2018-03-10 01:45:13 +08:00
|
|
|
Inst.insert(Inst.begin(),
|
|
|
|
MCOperand::createReg(X86::NoRegister)); // IndexReg
|
|
|
|
Inst.insert(Inst.begin(),
|
|
|
|
MCOperand::createImm(1)); // ScaleAmt
|
|
|
|
Inst.insert(Inst.begin(),
|
|
|
|
MCOperand::createReg(X86::RIP)); // BaseReg
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-12-02 08:29:39 +08:00
|
|
|
void convertIndirectCallToLoad(MCInst &Inst, MCPhysReg Reg) override {
|
|
|
|
bool IsTailCall = isTailCall(Inst);
|
|
|
|
if (IsTailCall)
|
2021-07-30 08:28:51 +08:00
|
|
|
removeAnnotation(Inst, MCPlus::MCAnnotation::kTailCall);
|
2019-12-14 09:27:03 +08:00
|
|
|
if (Inst.getOpcode() == X86::CALL64m ||
|
2020-12-02 08:29:39 +08:00
|
|
|
(Inst.getOpcode() == X86::JMP32m && IsTailCall)) {
|
2019-12-14 09:27:03 +08:00
|
|
|
Inst.setOpcode(X86::MOV64rm);
|
|
|
|
Inst.insert(Inst.begin(), MCOperand::createReg(Reg));
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
if (Inst.getOpcode() == X86::CALL64r ||
|
2020-12-02 08:29:39 +08:00
|
|
|
(Inst.getOpcode() == X86::JMP32r && IsTailCall)) {
|
2019-12-14 09:27:03 +08:00
|
|
|
Inst.setOpcode(X86::MOV64rr);
|
|
|
|
Inst.insert(Inst.begin(), MCOperand::createReg(Reg));
|
|
|
|
return;
|
|
|
|
}
|
2020-12-02 08:29:39 +08:00
|
|
|
LLVM_DEBUG(Inst.dump());
|
2019-12-14 09:27:03 +08:00
|
|
|
llvm_unreachable("not implemented");
|
|
|
|
}
|
|
|
|
|
2022-03-17 00:37:54 +08:00
|
|
|
bool shortenInstruction(MCInst &Inst,
|
|
|
|
const MCSubtargetInfo &STI) const override {
|
2018-03-10 01:45:13 +08:00
|
|
|
unsigned OldOpcode = Inst.getOpcode();
|
|
|
|
unsigned NewOpcode = OldOpcode;
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
int MemOpNo = getMemoryOperandNo(Inst);
|
2022-03-17 00:37:54 +08:00
|
|
|
|
|
|
|
// Check and remove redundant Address-Size override prefix.
|
|
|
|
if (opts::X86StripRedundantAddressSize) {
|
|
|
|
uint64_t TSFlags = Info->get(OldOpcode).TSFlags;
|
|
|
|
unsigned Flags = Inst.getFlags();
|
|
|
|
|
|
|
|
if (!X86_MC::needsAddressSizeOverride(Inst, STI, MemOpNo, TSFlags) &&
|
|
|
|
Flags & X86::IP_HAS_AD_SIZE)
|
|
|
|
Inst.setFlags(Flags ^ X86::IP_HAS_AD_SIZE);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check and remove EIZ/RIZ. These cases represent ambiguous cases where
|
|
|
|
// SIB byte is present, but no index is used and modrm alone should have
|
|
|
|
// been enough. Converting to NoRegister effectively removes the SIB byte.
|
2020-12-02 08:29:39 +08:00
|
|
|
if (MemOpNo >= 0) {
|
2021-04-08 15:19:26 +08:00
|
|
|
MCOperand &IndexOp =
|
2020-12-02 08:29:39 +08:00
|
|
|
Inst.getOperand(static_cast<unsigned>(MemOpNo) + X86::AddrIndexReg);
|
2022-01-09 09:48:33 +08:00
|
|
|
if (IndexOp.getReg() == X86::EIZ || IndexOp.getReg() == X86::RIZ)
|
2020-12-02 08:29:39 +08:00
|
|
|
IndexOp = MCOperand::createReg(X86::NoRegister);
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
if (isBranch(Inst)) {
|
|
|
|
NewOpcode = getShortBranchOpcode(OldOpcode);
|
|
|
|
} else if (OldOpcode == X86::MOV64ri) {
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
if (Inst.getOperand(MCPlus::getNumPrimeOperands(Inst) - 1).isImm()) {
|
2021-04-08 15:19:26 +08:00
|
|
|
const int64_t Imm =
|
|
|
|
Inst.getOperand(MCPlus::getNumPrimeOperands(Inst) - 1).getImm();
|
2021-12-29 09:52:08 +08:00
|
|
|
if (int64_t(Imm) == int64_t(int32_t(Imm)))
|
2018-03-10 01:45:13 +08:00
|
|
|
NewOpcode = X86::MOV64ri32;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// If it's arithmetic instruction check if signed operand fits in 1 byte.
|
2021-04-08 15:19:26 +08:00
|
|
|
const unsigned ShortOpcode = getShortArithOpcode(OldOpcode);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (ShortOpcode != OldOpcode &&
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
Inst.getOperand(MCPlus::getNumPrimeOperands(Inst) - 1).isImm()) {
|
2021-04-08 15:19:26 +08:00
|
|
|
int64_t Imm =
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
Inst.getOperand(MCPlus::getNumPrimeOperands(Inst) - 1).getImm();
|
2021-12-29 09:52:08 +08:00
|
|
|
if (int64_t(Imm) == int64_t(int8_t(Imm)))
|
2018-03-10 01:45:13 +08:00
|
|
|
NewOpcode = ShortOpcode;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (NewOpcode == OldOpcode)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
[BOLT] CMOVConversion pass
Convert simple hammocks into cmov based on misprediction rate.
Test Plan:
- Assembly test: `cmov-conversion.s`
- Testing on a binary:
# Bootstrap clang with `-x86-cmov-converter-force-all` and `-Wl,--emit-relocs`
(Release build)
# Collect perf.data:
- `clang++ <opts> bolt/lib/Core/BinaryFunction.cpp -E > bf.cpp`
- `perf record -e cycles:u -j any,u -- clang-15 bf.cpp -O2 -std=c++14 -c -o bf.o`
# Optimize clang-15 with and w/o -cmov-conversion:
- `llvm-bolt clang-15 -p perf.data -o clang-15.bolt`
- `llvm-bolt clang-15 -p perf.data -cmov-conversion -o clang-15.bolt.cmovconv`
# Run perf experiment:
- test: `clang-15.bolt.cmovconv`,
- control: `clang-15.bolt`,
- workload (clang options): `bf.cpp -O2 -std=c++14 -c -o bf.o`
Results:
```
task-clock [delta: -360.21 ± 356.75, delta(%): -1.7760 ± 1.7589, p-value: 0.047951, balance: -6]
instructions [delta: 44061118 ± 13246382, delta(%): 0.0690 ± 0.0207, p-value: 0.000001, balance: 50]
icache-misses [delta: -5534468 ± 2779620, delta(%): -0.4331 ± 0.2175, p-value: 0.028014, balance: -28]
branch-misses [delta: -1624270 ± 1113244, delta(%): -0.3456 ± 0.2368, p-value: 0.030300, balance: -22]
```
Reviewed By: rafauler
Differential Revision: https://reviews.llvm.org/D120177
2022-02-08 12:16:13 +08:00
|
|
|
bool
|
|
|
|
convertMoveToConditionalMove(MCInst &Inst, unsigned CC, bool AllowStackMemOp,
|
|
|
|
bool AllowBasePtrStackMemOp) const override {
|
|
|
|
// - Register-register moves are OK
|
|
|
|
// - Stores are filtered out by opcode (no store CMOV)
|
|
|
|
// - Non-stack loads are prohibited (generally unsafe)
|
|
|
|
// - Stack loads are OK if AllowStackMemOp is true
|
|
|
|
// - Stack loads with RBP are OK if AllowBasePtrStackMemOp is true
|
|
|
|
if (isLoad(Inst)) {
|
|
|
|
// If stack memory operands are not allowed, no loads are allowed
|
|
|
|
if (!AllowStackMemOp)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// If stack memory operands are allowed, check if it's a load from stack
|
|
|
|
bool IsLoad, IsStore, IsStoreFromReg, IsSimple, IsIndexed;
|
|
|
|
MCPhysReg Reg;
|
|
|
|
int32_t SrcImm;
|
|
|
|
uint16_t StackPtrReg;
|
|
|
|
int64_t StackOffset;
|
|
|
|
uint8_t Size;
|
|
|
|
bool IsStackAccess =
|
|
|
|
isStackAccess(Inst, IsLoad, IsStore, IsStoreFromReg, Reg, SrcImm,
|
|
|
|
StackPtrReg, StackOffset, Size, IsSimple, IsIndexed);
|
|
|
|
// Prohibit non-stack-based loads
|
|
|
|
if (!IsStackAccess)
|
|
|
|
return false;
|
|
|
|
// If stack memory operands are allowed, check if it's RBP-based
|
|
|
|
if (!AllowBasePtrStackMemOp &&
|
|
|
|
RegInfo->isSubRegisterEq(X86::RBP, StackPtrReg))
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
unsigned NewOpcode = 0;
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
case X86::MOV16rr:
|
|
|
|
NewOpcode = X86::CMOV16rr;
|
|
|
|
break;
|
|
|
|
case X86::MOV16rm:
|
|
|
|
NewOpcode = X86::CMOV16rm;
|
|
|
|
break;
|
|
|
|
case X86::MOV32rr:
|
|
|
|
NewOpcode = X86::CMOV32rr;
|
|
|
|
break;
|
|
|
|
case X86::MOV32rm:
|
|
|
|
NewOpcode = X86::CMOV32rm;
|
|
|
|
break;
|
|
|
|
case X86::MOV64rr:
|
|
|
|
NewOpcode = X86::CMOV64rr;
|
|
|
|
break;
|
|
|
|
case X86::MOV64rm:
|
|
|
|
NewOpcode = X86::CMOV64rm;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
// Insert CC at the end of prime operands, before annotations
|
|
|
|
Inst.insert(Inst.begin() + MCPlus::getNumPrimeOperands(Inst),
|
|
|
|
MCOperand::createImm(CC));
|
|
|
|
// CMOV is a 3-operand MCInst, so duplicate the destination as src1
|
|
|
|
Inst.insert(Inst.begin(), Inst.getOperand(0));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
bool lowerTailCall(MCInst &Inst) override {
|
2020-12-02 08:29:39 +08:00
|
|
|
if (Inst.getOpcode() == X86::JMP_4 && isTailCall(Inst)) {
|
2018-03-10 01:45:13 +08:00
|
|
|
Inst.setOpcode(X86::JMP_1);
|
2021-07-30 08:28:51 +08:00
|
|
|
removeAnnotation(Inst, MCPlus::MCAnnotation::kTailCall);
|
2018-03-10 01:45:13 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
const MCSymbol *getTargetSymbol(const MCInst &Inst,
|
|
|
|
unsigned OpNum = 0) const override {
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
if (OpNum >= MCPlus::getNumPrimeOperands(Inst))
|
2018-03-10 01:45:13 +08:00
|
|
|
return nullptr;
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCOperand &Op = Inst.getOperand(OpNum);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (!Op.isExpr())
|
|
|
|
return nullptr;
|
|
|
|
|
|
|
|
auto *SymExpr = dyn_cast<MCSymbolRefExpr>(Op.getExpr());
|
|
|
|
if (!SymExpr || SymExpr->getKind() != MCSymbolRefExpr::VK_None)
|
|
|
|
return nullptr;
|
|
|
|
|
|
|
|
return &SymExpr->getSymbol();
|
|
|
|
}
|
|
|
|
|
2020-12-02 08:29:39 +08:00
|
|
|
// This is the same as the base class, but since we are overriding one of
|
|
|
|
// getTargetSymbol's signatures above, we need to override all of them.
|
|
|
|
const MCSymbol *getTargetSymbol(const MCExpr *Expr) const override {
|
|
|
|
return &cast<const MCSymbolRefExpr>(Expr)->getSymbol();
|
|
|
|
}
|
|
|
|
|
2022-01-09 09:48:33 +08:00
|
|
|
bool analyzeBranch(InstructionIterator Begin, InstructionIterator End,
|
|
|
|
const MCSymbol *&TBB, const MCSymbol *&FBB,
|
2018-03-10 01:45:13 +08:00
|
|
|
MCInst *&CondBranch,
|
|
|
|
MCInst *&UncondBranch) const override {
|
|
|
|
auto I = End;
|
|
|
|
|
|
|
|
// Bottom-up analysis
|
|
|
|
while (I != Begin) {
|
|
|
|
--I;
|
|
|
|
|
|
|
|
// Ignore nops and CFIs
|
2021-06-17 03:10:20 +08:00
|
|
|
if (isPseudo(*I))
|
2018-03-10 01:45:13 +08:00
|
|
|
continue;
|
|
|
|
|
|
|
|
// Stop when we find the first non-terminator
|
|
|
|
if (!isTerminator(*I))
|
|
|
|
break;
|
|
|
|
|
|
|
|
if (!isBranch(*I))
|
|
|
|
break;
|
|
|
|
|
|
|
|
// Handle unconditional branches.
|
2022-01-09 09:48:33 +08:00
|
|
|
if ((I->getOpcode() == X86::JMP_1 || I->getOpcode() == X86::JMP_2 ||
|
2020-12-02 08:29:39 +08:00
|
|
|
I->getOpcode() == X86::JMP_4) &&
|
|
|
|
!isTailCall(*I)) {
|
2018-03-10 01:45:13 +08:00
|
|
|
// If any code was seen after this unconditional branch, we've seen
|
|
|
|
// unreachable code. Ignore them.
|
|
|
|
CondBranch = nullptr;
|
|
|
|
UncondBranch = &*I;
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCSymbol *Sym = getTargetSymbol(*I);
|
2018-03-10 01:45:13 +08:00
|
|
|
assert(Sym != nullptr &&
|
|
|
|
"Couldn't extract BB symbol from jump operand");
|
|
|
|
TBB = Sym;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Handle conditional branches and ignore indirect branches
|
2020-07-16 14:02:58 +08:00
|
|
|
if (!isUnsupportedBranch(I->getOpcode()) &&
|
2020-12-02 08:29:39 +08:00
|
|
|
getCondCode(*I) == X86::COND_INVALID) {
|
2018-03-10 01:45:13 +08:00
|
|
|
// Indirect branch
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (CondBranch == nullptr) {
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCSymbol *TargetBB = getTargetSymbol(*I);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (TargetBB == nullptr) {
|
|
|
|
// Unrecognized branch target
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
FBB = TBB;
|
|
|
|
TBB = TargetBB;
|
|
|
|
CondBranch = &*I;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
llvm_unreachable("multiple conditional branches in one BB");
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
template <typename Itr>
|
|
|
|
std::pair<IndirectBranchType, MCInst *>
|
2022-01-09 09:48:33 +08:00
|
|
|
analyzePICJumpTable(Itr II, Itr IE, MCPhysReg R1, MCPhysReg R2) const {
|
2018-03-10 01:45:13 +08:00
|
|
|
// Analyze PIC-style jump table code template:
|
|
|
|
//
|
|
|
|
// lea PIC_JUMP_TABLE(%rip), {%r1|%r2} <- MemLocInstr
|
|
|
|
// mov ({%r1|%r2}, %index, 4), {%r2|%r1}
|
|
|
|
// add %r2, %r1
|
|
|
|
// jmp *%r1
|
|
|
|
//
|
|
|
|
// (with any irrelevant instructions in-between)
|
|
|
|
//
|
|
|
|
// When we call this helper we've already determined %r1 and %r2, and
|
|
|
|
// reverse instruction iterator \p II is pointing to the ADD instruction.
|
|
|
|
//
|
|
|
|
// PIC jump table looks like following:
|
|
|
|
//
|
|
|
|
// JT: ----------
|
|
|
|
// E1:| L1 - JT |
|
|
|
|
// |----------|
|
|
|
|
// E2:| L2 - JT |
|
|
|
|
// |----------|
|
|
|
|
// | |
|
|
|
|
// ......
|
|
|
|
// En:| Ln - JT |
|
|
|
|
// ----------
|
|
|
|
//
|
|
|
|
// Where L1, L2, ..., Ln represent labels in the function.
|
|
|
|
//
|
|
|
|
// The actual relocations in the table will be of the form:
|
|
|
|
//
|
|
|
|
// Ln - JT
|
|
|
|
// = (Ln - En) + (En - JT)
|
|
|
|
// = R_X86_64_PC32(Ln) + En - JT
|
|
|
|
// = R_X86_64_PC32(Ln + offsetof(En))
|
|
|
|
//
|
2020-12-02 08:29:39 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "Checking for PIC jump table\n");
|
2018-03-10 01:45:13 +08:00
|
|
|
MCInst *MemLocInstr = nullptr;
|
|
|
|
const MCInst *MovInstr = nullptr;
|
|
|
|
while (++II != IE) {
|
2021-04-08 15:19:26 +08:00
|
|
|
MCInst &Instr = *II;
|
|
|
|
const MCInstrDesc &InstrDesc = Info->get(Instr.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
if (!InstrDesc.hasDefOfPhysReg(Instr, R1, *RegInfo) &&
|
|
|
|
!InstrDesc.hasDefOfPhysReg(Instr, R2, *RegInfo)) {
|
|
|
|
// Ignore instructions that don't affect R1, R2 registers.
|
|
|
|
continue;
|
2021-12-29 09:52:08 +08:00
|
|
|
}
|
|
|
|
if (!MovInstr) {
|
2018-03-10 01:45:13 +08:00
|
|
|
// Expect to see MOV instruction.
|
|
|
|
if (!isMOVSX64rm32(Instr)) {
|
2020-12-02 08:29:39 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "MOV instruction expected.\n");
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check if it's setting %r1 or %r2. In canonical form it sets %r2.
|
|
|
|
// If it sets %r1 - rename the registers so we have to only check
|
|
|
|
// a single form.
|
2021-04-08 15:19:26 +08:00
|
|
|
unsigned MovDestReg = Instr.getOperand(0).getReg();
|
2018-03-10 01:45:13 +08:00
|
|
|
if (MovDestReg != R2)
|
|
|
|
std::swap(R1, R2);
|
|
|
|
if (MovDestReg != R2) {
|
2020-12-02 08:29:39 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "MOV instruction expected to set %r2\n");
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Verify operands for MOV.
|
|
|
|
unsigned BaseRegNum;
|
|
|
|
int64_t ScaleValue;
|
|
|
|
unsigned IndexRegNum;
|
|
|
|
int64_t DispValue;
|
|
|
|
unsigned SegRegNum;
|
2022-01-09 09:48:33 +08:00
|
|
|
if (!evaluateX86MemoryOperand(Instr, &BaseRegNum, &ScaleValue,
|
|
|
|
&IndexRegNum, &DispValue, &SegRegNum))
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
2022-01-09 09:48:33 +08:00
|
|
|
if (BaseRegNum != R1 || ScaleValue != 4 ||
|
|
|
|
IndexRegNum == X86::NoRegister || DispValue != 0 ||
|
2018-03-10 01:45:13 +08:00
|
|
|
SegRegNum != X86::NoRegister)
|
|
|
|
break;
|
|
|
|
MovInstr = &Instr;
|
|
|
|
} else {
|
|
|
|
if (!InstrDesc.hasDefOfPhysReg(Instr, R1, *RegInfo))
|
|
|
|
continue;
|
|
|
|
if (!isLEA64r(Instr)) {
|
2020-12-02 08:29:39 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "LEA instruction expected\n");
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (Instr.getOperand(0).getReg() != R1) {
|
2020-12-02 08:29:39 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "LEA instruction expected to set %r1\n");
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Verify operands for LEA.
|
|
|
|
unsigned BaseRegNum;
|
|
|
|
int64_t ScaleValue;
|
|
|
|
unsigned IndexRegNum;
|
|
|
|
const MCExpr *DispExpr = nullptr;
|
2018-06-22 02:03:57 +08:00
|
|
|
int64_t DispValue;
|
2018-03-10 01:45:13 +08:00
|
|
|
unsigned SegRegNum;
|
2022-01-09 09:48:33 +08:00
|
|
|
if (!evaluateX86MemoryOperand(Instr, &BaseRegNum, &ScaleValue,
|
|
|
|
&IndexRegNum, &DispValue, &SegRegNum,
|
|
|
|
&DispExpr))
|
2018-03-10 01:45:13 +08:00
|
|
|
break;
|
|
|
|
if (BaseRegNum != RegInfo->getProgramCounter() ||
|
2022-01-09 09:48:33 +08:00
|
|
|
IndexRegNum != X86::NoRegister || SegRegNum != X86::NoRegister ||
|
2018-03-10 01:45:13 +08:00
|
|
|
DispExpr == nullptr)
|
|
|
|
break;
|
|
|
|
MemLocInstr = &Instr;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!MemLocInstr)
|
|
|
|
return std::make_pair(IndirectBranchType::UNKNOWN, nullptr);
|
|
|
|
|
2020-12-02 08:29:39 +08:00
|
|
|
LLVM_DEBUG(dbgs() << "checking potential PIC jump table\n");
|
2018-03-10 01:45:13 +08:00
|
|
|
return std::make_pair(IndirectBranchType::POSSIBLE_PIC_JUMP_TABLE,
|
|
|
|
MemLocInstr);
|
|
|
|
}
|
|
|
|
|
|
|
|
IndirectBranchType analyzeIndirectBranch(
|
2022-01-09 09:48:33 +08:00
|
|
|
MCInst &Instruction, InstructionIterator Begin, InstructionIterator End,
|
|
|
|
const unsigned PtrSize, MCInst *&MemLocInstrOut, unsigned &BaseRegNumOut,
|
|
|
|
unsigned &IndexRegNumOut, int64_t &DispValueOut,
|
|
|
|
const MCExpr *&DispExprOut, MCInst *&PCRelBaseOut) const override {
|
2018-03-10 01:45:13 +08:00
|
|
|
// Try to find a (base) memory location from where the address for
|
|
|
|
// the indirect branch is loaded. For X86-64 the memory will be specified
|
|
|
|
// in the following format:
|
|
|
|
//
|
|
|
|
// {%rip}/{%basereg} + Imm + IndexReg * Scale
|
|
|
|
//
|
|
|
|
// We are interested in the cases where Scale == sizeof(uintptr_t) and
|
2019-06-29 00:21:27 +08:00
|
|
|
// the contents of the memory are presumably an array of pointers to code.
|
2018-03-10 01:45:13 +08:00
|
|
|
//
|
|
|
|
// Normal jump table:
|
|
|
|
//
|
2019-06-29 00:21:27 +08:00
|
|
|
// jmp *(JUMP_TABLE, %index, Scale) <- MemLocInstr
|
2018-03-10 01:45:13 +08:00
|
|
|
//
|
|
|
|
// or
|
|
|
|
//
|
2019-06-29 00:21:27 +08:00
|
|
|
// mov (JUMP_TABLE, %index, Scale), %r1 <- MemLocInstr
|
2018-03-10 01:45:13 +08:00
|
|
|
// ...
|
|
|
|
// jmp %r1
|
|
|
|
//
|
|
|
|
// We handle PIC-style jump tables separately.
|
|
|
|
//
|
|
|
|
MemLocInstrOut = nullptr;
|
|
|
|
BaseRegNumOut = X86::NoRegister;
|
|
|
|
IndexRegNumOut = X86::NoRegister;
|
|
|
|
DispValueOut = 0;
|
|
|
|
DispExprOut = nullptr;
|
|
|
|
|
|
|
|
std::reverse_iterator<InstructionIterator> II(End);
|
|
|
|
std::reverse_iterator<InstructionIterator> IE(Begin);
|
|
|
|
|
|
|
|
IndirectBranchType Type = IndirectBranchType::UNKNOWN;
|
|
|
|
|
|
|
|
// An instruction referencing memory used by jump instruction (directly or
|
|
|
|
// via register). This location could be an array of function pointers
|
|
|
|
// in case of indirect tail call, or a jump table.
|
|
|
|
MCInst *MemLocInstr = nullptr;
|
|
|
|
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
if (MCPlus::getNumPrimeOperands(Instruction) == 1) {
|
2018-03-10 01:45:13 +08:00
|
|
|
// If the indirect jump is on register - try to detect if the
|
|
|
|
// register value is loaded from a memory location.
|
|
|
|
assert(Instruction.getOperand(0).isReg() && "register operand expected");
|
2021-04-08 15:19:26 +08:00
|
|
|
const unsigned R1 = Instruction.getOperand(0).getReg();
|
2018-03-10 01:45:13 +08:00
|
|
|
// Check if one of the previous instructions defines the jump-on register.
|
|
|
|
for (auto PrevII = II; PrevII != IE; ++PrevII) {
|
2021-04-08 15:19:26 +08:00
|
|
|
MCInst &PrevInstr = *PrevII;
|
|
|
|
const MCInstrDesc &PrevInstrDesc = Info->get(PrevInstr.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
if (!PrevInstrDesc.hasDefOfPhysReg(PrevInstr, R1, *RegInfo))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (isMoveMem2Reg(PrevInstr)) {
|
|
|
|
MemLocInstr = &PrevInstr;
|
|
|
|
break;
|
2021-12-29 09:52:08 +08:00
|
|
|
}
|
|
|
|
if (isADD64rr(PrevInstr)) {
|
2021-04-08 15:19:26 +08:00
|
|
|
unsigned R2 = PrevInstr.getOperand(2).getReg();
|
2018-03-10 01:45:13 +08:00
|
|
|
if (R1 == R2)
|
|
|
|
return IndirectBranchType::UNKNOWN;
|
|
|
|
std::tie(Type, MemLocInstr) = analyzePICJumpTable(PrevII, IE, R1, R2);
|
|
|
|
break;
|
|
|
|
}
|
2021-12-29 09:52:08 +08:00
|
|
|
return IndirectBranchType::UNKNOWN;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
if (!MemLocInstr) {
|
|
|
|
// No definition seen for the register in this function so far. Could be
|
|
|
|
// an input parameter - which means it is an external code reference.
|
|
|
|
// It also could be that the definition happens to be in the code that
|
|
|
|
// we haven't processed yet. Since we have to be conservative, return
|
|
|
|
// as UNKNOWN case.
|
|
|
|
return IndirectBranchType::UNKNOWN;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
MemLocInstr = &Instruction;
|
|
|
|
}
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCRegister RIPRegister = RegInfo->getProgramCounter();
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// Analyze the memory location.
|
2022-01-09 09:48:33 +08:00
|
|
|
unsigned BaseRegNum, IndexRegNum, SegRegNum;
|
|
|
|
int64_t ScaleValue, DispValue;
|
2018-03-10 01:45:13 +08:00
|
|
|
const MCExpr *DispExpr;
|
|
|
|
|
2022-01-09 09:48:33 +08:00
|
|
|
if (!evaluateX86MemoryOperand(*MemLocInstr, &BaseRegNum, &ScaleValue,
|
|
|
|
&IndexRegNum, &DispValue, &SegRegNum,
|
2018-03-10 01:45:13 +08:00
|
|
|
&DispExpr))
|
|
|
|
return IndirectBranchType::UNKNOWN;
|
|
|
|
|
2018-08-07 02:22:45 +08:00
|
|
|
BaseRegNumOut = BaseRegNum;
|
|
|
|
IndexRegNumOut = IndexRegNum;
|
|
|
|
DispValueOut = DispValue;
|
|
|
|
DispExprOut = DispExpr;
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
if ((BaseRegNum != X86::NoRegister && BaseRegNum != RIPRegister) ||
|
|
|
|
SegRegNum != X86::NoRegister)
|
|
|
|
return IndirectBranchType::UNKNOWN;
|
|
|
|
|
2018-08-07 02:22:45 +08:00
|
|
|
if (MemLocInstr == &Instruction &&
|
|
|
|
(!ScaleValue || IndexRegNum == X86::NoRegister)) {
|
|
|
|
MemLocInstrOut = MemLocInstr;
|
|
|
|
return IndirectBranchType::POSSIBLE_FIXED_BRANCH;
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
if (Type == IndirectBranchType::POSSIBLE_PIC_JUMP_TABLE &&
|
|
|
|
(ScaleValue != 1 || BaseRegNum != RIPRegister))
|
|
|
|
return IndirectBranchType::UNKNOWN;
|
|
|
|
|
|
|
|
if (Type != IndirectBranchType::POSSIBLE_PIC_JUMP_TABLE &&
|
|
|
|
ScaleValue != PtrSize)
|
|
|
|
return IndirectBranchType::UNKNOWN;
|
|
|
|
|
|
|
|
MemLocInstrOut = MemLocInstr;
|
|
|
|
|
|
|
|
return Type;
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Analyze a callsite to see if it could be a virtual method call. This only
|
|
|
|
/// checks to see if the overall pattern is satisfied, it does not guarantee
|
|
|
|
/// that the callsite is a true virtual method call.
|
|
|
|
/// The format of virtual method calls that are recognized is one of the
|
|
|
|
/// following:
|
|
|
|
///
|
|
|
|
/// Form 1: (found in debug code)
|
|
|
|
/// add METHOD_OFFSET, %VtableReg
|
|
|
|
/// mov (%VtableReg), %MethodReg
|
|
|
|
/// ...
|
|
|
|
/// call or jmp *%MethodReg
|
|
|
|
///
|
|
|
|
/// Form 2:
|
|
|
|
/// mov METHOD_OFFSET(%VtableReg), %MethodReg
|
|
|
|
/// ...
|
|
|
|
/// call or jmp *%MethodReg
|
|
|
|
///
|
|
|
|
/// Form 3:
|
|
|
|
/// ...
|
|
|
|
/// call or jmp *METHOD_OFFSET(%VtableReg)
|
|
|
|
///
|
|
|
|
bool analyzeVirtualMethodCall(InstructionIterator ForwardBegin,
|
|
|
|
InstructionIterator ForwardEnd,
|
|
|
|
std::vector<MCInst *> &MethodFetchInsns,
|
2022-01-09 09:48:33 +08:00
|
|
|
unsigned &VtableRegNum, unsigned &MethodRegNum,
|
2018-03-10 01:45:13 +08:00
|
|
|
uint64_t &MethodOffset) const override {
|
|
|
|
VtableRegNum = X86::NoRegister;
|
|
|
|
MethodRegNum = X86::NoRegister;
|
|
|
|
MethodOffset = 0;
|
|
|
|
|
|
|
|
std::reverse_iterator<InstructionIterator> Itr(ForwardEnd);
|
|
|
|
std::reverse_iterator<InstructionIterator> End(ForwardBegin);
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
MCInst &CallInst = *Itr++;
|
2018-03-10 01:45:13 +08:00
|
|
|
assert(isIndirectBranch(CallInst) || isCall(CallInst));
|
|
|
|
|
|
|
|
unsigned BaseReg, IndexReg, SegmentReg;
|
|
|
|
int64_t Scale, Disp;
|
|
|
|
const MCExpr *DispExpr;
|
|
|
|
|
|
|
|
// The call can just be jmp offset(reg)
|
2022-01-09 09:48:33 +08:00
|
|
|
if (evaluateX86MemoryOperand(CallInst, &BaseReg, &Scale, &IndexReg, &Disp,
|
|
|
|
&SegmentReg, &DispExpr)) {
|
|
|
|
if (!DispExpr && BaseReg != X86::RIP && BaseReg != X86::RBP &&
|
2018-03-10 01:45:13 +08:00
|
|
|
BaseReg != X86::NoRegister) {
|
|
|
|
MethodRegNum = BaseReg;
|
2022-01-09 09:48:33 +08:00
|
|
|
if (Scale == 1 && IndexReg == X86::NoRegister &&
|
2018-03-10 01:45:13 +08:00
|
|
|
SegmentReg == X86::NoRegister) {
|
|
|
|
VtableRegNum = MethodRegNum;
|
|
|
|
MethodOffset = Disp;
|
|
|
|
MethodFetchInsns.push_back(&CallInst);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return false;
|
2021-12-29 09:52:08 +08:00
|
|
|
}
|
|
|
|
if (CallInst.getOperand(0).isReg())
|
2018-03-10 01:45:13 +08:00
|
|
|
MethodRegNum = CallInst.getOperand(0).getReg();
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-03-10 01:45:13 +08:00
|
|
|
return false;
|
|
|
|
|
|
|
|
if (MethodRegNum == X86::RIP || MethodRegNum == X86::RBP) {
|
|
|
|
VtableRegNum = X86::NoRegister;
|
|
|
|
MethodRegNum = X86::NoRegister;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
// find load from vtable, this may or may not include the method offset
|
|
|
|
while (Itr != End) {
|
2021-04-08 15:19:26 +08:00
|
|
|
MCInst &CurInst = *Itr++;
|
|
|
|
const MCInstrDesc &Desc = Info->get(CurInst.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
if (Desc.hasDefOfPhysReg(CurInst, MethodRegNum, *RegInfo)) {
|
|
|
|
if (isLoad(CurInst) &&
|
2022-01-09 09:48:33 +08:00
|
|
|
evaluateX86MemoryOperand(CurInst, &BaseReg, &Scale, &IndexReg,
|
|
|
|
&Disp, &SegmentReg, &DispExpr)) {
|
|
|
|
if (!DispExpr && Scale == 1 && BaseReg != X86::RIP &&
|
|
|
|
BaseReg != X86::RBP && BaseReg != X86::NoRegister &&
|
|
|
|
IndexReg == X86::NoRegister && SegmentReg == X86::NoRegister &&
|
2018-03-10 01:45:13 +08:00
|
|
|
BaseReg != X86::RIP) {
|
|
|
|
VtableRegNum = BaseReg;
|
|
|
|
MethodOffset = Disp;
|
|
|
|
MethodFetchInsns.push_back(&CurInst);
|
|
|
|
if (MethodOffset != 0)
|
|
|
|
return true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!VtableRegNum)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
// look for any adds affecting the method register.
|
|
|
|
while (Itr != End) {
|
2021-04-08 15:19:26 +08:00
|
|
|
MCInst &CurInst = *Itr++;
|
|
|
|
const MCInstrDesc &Desc = Info->get(CurInst.getOpcode());
|
2018-03-10 01:45:13 +08:00
|
|
|
if (Desc.hasDefOfPhysReg(CurInst, VtableRegNum, *RegInfo)) {
|
|
|
|
if (isADDri(CurInst)) {
|
|
|
|
assert(!MethodOffset);
|
|
|
|
MethodOffset = CurInst.getOperand(2).getImm();
|
|
|
|
MethodFetchInsns.insert(MethodFetchInsns.begin(), &CurInst);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool createStackPointerIncrement(MCInst &Inst, int Size,
|
|
|
|
bool NoFlagsClobber) const override {
|
|
|
|
if (NoFlagsClobber) {
|
|
|
|
Inst.setOpcode(X86::LEA64r);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RSP));
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RSP)); // BaseReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(1)); // ScaleAmt
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // IndexReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(-Size)); // Displacement
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // AddrSegmentReg
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
Inst.setOpcode(X86::SUB64ri8);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RSP));
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RSP));
|
|
|
|
Inst.addOperand(MCOperand::createImm(Size));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool createStackPointerDecrement(MCInst &Inst, int Size,
|
|
|
|
bool NoFlagsClobber) const override {
|
|
|
|
if (NoFlagsClobber) {
|
|
|
|
Inst.setOpcode(X86::LEA64r);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RSP));
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RSP)); // BaseReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(1)); // ScaleAmt
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // IndexReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(Size)); // Displacement
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // AddrSegmentReg
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
Inst.setOpcode(X86::ADD64ri8);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RSP));
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RSP));
|
|
|
|
Inst.addOperand(MCOperand::createImm(Size));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool createSaveToStack(MCInst &Inst, const MCPhysReg &StackReg, int Offset,
|
|
|
|
const MCPhysReg &SrcReg, int Size) const override {
|
|
|
|
unsigned NewOpcode;
|
|
|
|
switch (Size) {
|
2022-01-09 09:48:33 +08:00
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case 2: NewOpcode = X86::MOV16mr; break;
|
|
|
|
case 4: NewOpcode = X86::MOV32mr; break;
|
|
|
|
case 8: NewOpcode = X86::MOV64mr; break;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
Inst.clear();
|
2022-01-09 09:48:33 +08:00
|
|
|
Inst.addOperand(MCOperand::createReg(StackReg)); // BaseReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(1)); // ScaleAmt
|
2018-03-10 01:45:13 +08:00
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // IndexReg
|
2022-01-09 09:48:33 +08:00
|
|
|
Inst.addOperand(MCOperand::createImm(Offset)); // Displacement
|
2018-03-10 01:45:13 +08:00
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // AddrSegmentReg
|
|
|
|
Inst.addOperand(MCOperand::createReg(SrcReg));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool createRestoreFromStack(MCInst &Inst, const MCPhysReg &StackReg,
|
|
|
|
int Offset, const MCPhysReg &DstReg,
|
|
|
|
int Size) const override {
|
|
|
|
return createLoad(Inst, StackReg, /*Scale=*/1, /*IndexReg=*/X86::NoRegister,
|
2018-08-04 07:36:06 +08:00
|
|
|
Offset, nullptr, /*AddrSegmentReg=*/X86::NoRegister,
|
|
|
|
DstReg, Size);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
2018-08-04 07:36:06 +08:00
|
|
|
bool createLoad(MCInst &Inst, const MCPhysReg &BaseReg, int64_t Scale,
|
|
|
|
const MCPhysReg &IndexReg, int64_t Offset,
|
|
|
|
const MCExpr *OffsetExpr, const MCPhysReg &AddrSegmentReg,
|
2019-02-01 03:23:02 +08:00
|
|
|
const MCPhysReg &DstReg, int Size) const override {
|
2018-03-10 01:45:13 +08:00
|
|
|
unsigned NewOpcode;
|
|
|
|
switch (Size) {
|
2022-01-09 09:48:33 +08:00
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
case 2: NewOpcode = X86::MOV16rm; break;
|
|
|
|
case 4: NewOpcode = X86::MOV32rm; break;
|
|
|
|
case 8: NewOpcode = X86::MOV64rm; break;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.addOperand(MCOperand::createReg(DstReg));
|
|
|
|
Inst.addOperand(MCOperand::createReg(BaseReg));
|
|
|
|
Inst.addOperand(MCOperand::createImm(Scale));
|
|
|
|
Inst.addOperand(MCOperand::createReg(IndexReg));
|
2018-07-26 10:07:41 +08:00
|
|
|
if (OffsetExpr)
|
|
|
|
Inst.addOperand(MCOperand::createExpr(OffsetExpr)); // Displacement
|
|
|
|
else
|
|
|
|
Inst.addOperand(MCOperand::createImm(Offset)); // Displacement
|
2018-08-04 07:36:06 +08:00
|
|
|
Inst.addOperand(MCOperand::createReg(AddrSegmentReg)); // AddrSegmentReg
|
2018-03-10 01:45:13 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-12-14 09:27:03 +08:00
|
|
|
void createLoadImmediate(MCInst &Inst, const MCPhysReg Dest,
|
2020-02-24 14:27:28 +08:00
|
|
|
uint32_t Imm) const override {
|
2019-12-14 09:27:03 +08:00
|
|
|
Inst.setOpcode(X86::MOV64ri32);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.addOperand(MCOperand::createReg(Dest));
|
|
|
|
Inst.addOperand(MCOperand::createImm(Imm));
|
|
|
|
}
|
|
|
|
|
2019-06-20 11:10:49 +08:00
|
|
|
bool createIncMemory(MCInst &Inst, const MCSymbol *Target,
|
|
|
|
MCContext *Ctx) const override {
|
|
|
|
|
2019-07-03 07:56:41 +08:00
|
|
|
Inst.setOpcode(X86::LOCK_INC64m);
|
2019-06-20 11:10:49 +08:00
|
|
|
Inst.clear();
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RIP)); // BaseReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(1)); // ScaleAmt
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // IndexReg
|
|
|
|
|
|
|
|
Inst.addOperand(MCOperand::createExpr(
|
|
|
|
MCSymbolRefExpr::create(Target, MCSymbolRefExpr::VK_None,
|
|
|
|
*Ctx))); // Displacement
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // AddrSegmentReg
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
bool createIJmp32Frag(SmallVectorImpl<MCInst> &Insts,
|
|
|
|
const MCOperand &BaseReg, const MCOperand &Scale,
|
|
|
|
const MCOperand &IndexReg, const MCOperand &Offset,
|
|
|
|
const MCOperand &TmpReg) const override {
|
|
|
|
// The code fragment we emit here is:
|
|
|
|
//
|
|
|
|
// mov32 (%base, %index, scale), %tmpreg
|
|
|
|
// ijmp *(%tmpreg)
|
|
|
|
//
|
|
|
|
MCInst IJmp;
|
|
|
|
IJmp.setOpcode(X86::JMP64r);
|
|
|
|
IJmp.addOperand(TmpReg);
|
|
|
|
|
|
|
|
MCInst Load;
|
|
|
|
Load.setOpcode(X86::MOV32rm);
|
|
|
|
Load.addOperand(TmpReg);
|
|
|
|
Load.addOperand(BaseReg);
|
|
|
|
Load.addOperand(Scale);
|
|
|
|
Load.addOperand(IndexReg);
|
|
|
|
Load.addOperand(Offset);
|
|
|
|
Load.addOperand(MCOperand::createReg(X86::NoRegister));
|
|
|
|
|
|
|
|
Insts.push_back(Load);
|
|
|
|
Insts.push_back(IJmp);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool createNoop(MCInst &Inst) const override {
|
|
|
|
Inst.setOpcode(X86::NOOP);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool createReturn(MCInst &Inst) const override {
|
2021-10-09 02:47:10 +08:00
|
|
|
Inst.setOpcode(X86::RET64);
|
2018-03-10 01:45:13 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType createInlineMemcpy(bool ReturnEnd) const override {
|
|
|
|
InstructionListType Code;
|
2021-12-29 09:52:08 +08:00
|
|
|
if (ReturnEnd)
|
2018-05-27 03:40:51 +08:00
|
|
|
Code.emplace_back(MCInstBuilder(X86::LEA64r)
|
|
|
|
.addReg(X86::RAX)
|
|
|
|
.addReg(X86::RDI)
|
|
|
|
.addImm(1)
|
|
|
|
.addReg(X86::RDX)
|
|
|
|
.addImm(0)
|
|
|
|
.addReg(X86::NoRegister));
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-05-27 03:40:51 +08:00
|
|
|
Code.emplace_back(MCInstBuilder(X86::MOV64rr)
|
|
|
|
.addReg(X86::RAX)
|
|
|
|
.addReg(X86::RDI));
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2018-05-27 03:40:51 +08:00
|
|
|
Code.emplace_back(MCInstBuilder(X86::MOV32rr)
|
|
|
|
.addReg(X86::ECX)
|
|
|
|
.addReg(X86::EDX));
|
|
|
|
Code.emplace_back(MCInstBuilder(X86::REP_MOVSB_64));
|
|
|
|
|
|
|
|
return Code;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType createOneByteMemcpy() const override {
|
|
|
|
InstructionListType Code;
|
2019-05-21 11:11:40 +08:00
|
|
|
Code.emplace_back(MCInstBuilder(X86::MOV8rm)
|
|
|
|
.addReg(X86::CL)
|
|
|
|
.addReg(X86::RSI)
|
|
|
|
.addImm(0)
|
|
|
|
.addReg(X86::NoRegister)
|
|
|
|
.addImm(0)
|
|
|
|
.addReg(X86::NoRegister));
|
|
|
|
Code.emplace_back(MCInstBuilder(X86::MOV8mr)
|
|
|
|
.addReg(X86::RDI)
|
|
|
|
.addImm(0)
|
|
|
|
.addReg(X86::NoRegister)
|
|
|
|
.addImm(0)
|
|
|
|
.addReg(X86::NoRegister)
|
|
|
|
.addReg(X86::CL));
|
|
|
|
Code.emplace_back(MCInstBuilder(X86::MOV64rr)
|
|
|
|
.addReg(X86::RAX)
|
|
|
|
.addReg(X86::RDI));
|
|
|
|
return Code;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType createCmpJE(MCPhysReg RegNo, int64_t Imm,
|
|
|
|
const MCSymbol *Target,
|
|
|
|
MCContext *Ctx) const override {
|
|
|
|
InstructionListType Code;
|
2019-05-21 11:11:40 +08:00
|
|
|
Code.emplace_back(MCInstBuilder(X86::CMP64ri8)
|
|
|
|
.addReg(RegNo)
|
|
|
|
.addImm(Imm));
|
2020-12-02 08:29:39 +08:00
|
|
|
Code.emplace_back(MCInstBuilder(X86::JCC_1)
|
2019-05-21 11:11:40 +08:00
|
|
|
.addExpr(MCSymbolRefExpr::create(
|
2020-12-02 08:29:39 +08:00
|
|
|
Target, MCSymbolRefExpr::VK_None, *Ctx))
|
|
|
|
.addImm(X86::COND_E));
|
2019-05-21 11:11:40 +08:00
|
|
|
return Code;
|
|
|
|
}
|
|
|
|
|
[BOLT] Support for lite mode with relocations
Summary:
Add '-lite' support for relocations for improved processing time,
memory consumption, and more resilient processing of binaries with
embedded assembly code.
In lite relocation mode, BOLT will skip full processing of functions
without a profile. It will run scanExternalRefs() on such functions
to discover external references and to create internal relocations
to update references to optimized functions.
Note that we could have relied on the compiler/linker to provide
relocations for function references. However, there's no assurance
that all such references are reported. E.g., the compiler can resolve
inter-procedural references internally, leaving no relocations
for the linker.
The scan process takes about <10 seconds per 100MB of code on modern
hardware. It's a reasonable overhead to live with considering the
flexibility it provides.
If BOLT fails to scan or disassemble a function, .e.g., due to a data
object embedded in code, or an unsupported instruction, it enables a
patching mode to guarantee that the failed function will call
optimized/moved versions of functions. The patching happens at original
function entry points.
'-skip=<func1,func2,...>' option now can be used to skip processing of
arbitrary functions in the relocation mode.
With '-use-old-text' or '-strict' we require all functions to be
processed. As such, it is incompatible with '-lite' option,
and '-skip' option will only disable optimizations of listed
functions, not their disassembly and emission.
(cherry picked from FBD22040717)
2020-06-15 15:15:47 +08:00
|
|
|
Optional<Relocation>
|
|
|
|
createRelocation(const MCFixup &Fixup,
|
|
|
|
const MCAsmBackend &MAB) const override {
|
|
|
|
const MCFixupKindInfo &FKI = MAB.getFixupKindInfo(Fixup.getKind());
|
|
|
|
|
|
|
|
assert(FKI.TargetOffset == 0 && "0-bit relocation offset expected");
|
|
|
|
const uint64_t RelOffset = Fixup.getOffset();
|
|
|
|
|
|
|
|
uint64_t RelType;
|
|
|
|
if (FKI.Flags & MCFixupKindInfo::FKF_IsPCRel) {
|
|
|
|
switch (FKI.TargetSize) {
|
|
|
|
default:
|
|
|
|
return NoneType();
|
|
|
|
case 8: RelType = ELF::R_X86_64_PC8; break;
|
|
|
|
case 16: RelType = ELF::R_X86_64_PC16; break;
|
|
|
|
case 32: RelType = ELF::R_X86_64_PC32; break;
|
|
|
|
case 64: RelType = ELF::R_X86_64_PC64; break;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
switch (FKI.TargetSize) {
|
|
|
|
default:
|
|
|
|
return NoneType();
|
|
|
|
case 8: RelType = ELF::R_X86_64_8; break;
|
|
|
|
case 16: RelType = ELF::R_X86_64_16; break;
|
|
|
|
case 32: RelType = ELF::R_X86_64_32; break;
|
|
|
|
case 64: RelType = ELF::R_X86_64_64; break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Extract a symbol and an addend out of the fixup value expression.
|
|
|
|
//
|
|
|
|
// Only the following limited expression types are supported:
|
|
|
|
// Symbol + Addend
|
|
|
|
// Symbol
|
|
|
|
uint64_t Addend = 0;
|
|
|
|
MCSymbol *Symbol = nullptr;
|
|
|
|
const MCExpr *ValueExpr = Fixup.getValue();
|
|
|
|
if (ValueExpr->getKind() == MCExpr::Binary) {
|
|
|
|
const auto *BinaryExpr = cast<MCBinaryExpr>(ValueExpr);
|
|
|
|
assert(BinaryExpr->getOpcode() == MCBinaryExpr::Add &&
|
|
|
|
"unexpected binary expression");
|
|
|
|
const MCExpr *LHS = BinaryExpr->getLHS();
|
|
|
|
assert(LHS->getKind() == MCExpr::SymbolRef && "unexpected LHS");
|
2020-12-02 08:29:39 +08:00
|
|
|
Symbol = const_cast<MCSymbol *>(this->getTargetSymbol(LHS));
|
[BOLT] Support for lite mode with relocations
Summary:
Add '-lite' support for relocations for improved processing time,
memory consumption, and more resilient processing of binaries with
embedded assembly code.
In lite relocation mode, BOLT will skip full processing of functions
without a profile. It will run scanExternalRefs() on such functions
to discover external references and to create internal relocations
to update references to optimized functions.
Note that we could have relied on the compiler/linker to provide
relocations for function references. However, there's no assurance
that all such references are reported. E.g., the compiler can resolve
inter-procedural references internally, leaving no relocations
for the linker.
The scan process takes about <10 seconds per 100MB of code on modern
hardware. It's a reasonable overhead to live with considering the
flexibility it provides.
If BOLT fails to scan or disassemble a function, .e.g., due to a data
object embedded in code, or an unsupported instruction, it enables a
patching mode to guarantee that the failed function will call
optimized/moved versions of functions. The patching happens at original
function entry points.
'-skip=<func1,func2,...>' option now can be used to skip processing of
arbitrary functions in the relocation mode.
With '-use-old-text' or '-strict' we require all functions to be
processed. As such, it is incompatible with '-lite' option,
and '-skip' option will only disable optimizations of listed
functions, not their disassembly and emission.
(cherry picked from FBD22040717)
2020-06-15 15:15:47 +08:00
|
|
|
const MCExpr *RHS = BinaryExpr->getRHS();
|
|
|
|
assert(RHS->getKind() == MCExpr::Constant && "unexpected RHS");
|
|
|
|
Addend = cast<MCConstantExpr>(RHS)->getValue();
|
|
|
|
} else {
|
|
|
|
assert(ValueExpr->getKind() == MCExpr::SymbolRef && "unexpected value");
|
2020-12-02 08:29:39 +08:00
|
|
|
Symbol = const_cast<MCSymbol *>(this->getTargetSymbol(ValueExpr));
|
[BOLT] Support for lite mode with relocations
Summary:
Add '-lite' support for relocations for improved processing time,
memory consumption, and more resilient processing of binaries with
embedded assembly code.
In lite relocation mode, BOLT will skip full processing of functions
without a profile. It will run scanExternalRefs() on such functions
to discover external references and to create internal relocations
to update references to optimized functions.
Note that we could have relied on the compiler/linker to provide
relocations for function references. However, there's no assurance
that all such references are reported. E.g., the compiler can resolve
inter-procedural references internally, leaving no relocations
for the linker.
The scan process takes about <10 seconds per 100MB of code on modern
hardware. It's a reasonable overhead to live with considering the
flexibility it provides.
If BOLT fails to scan or disassemble a function, .e.g., due to a data
object embedded in code, or an unsupported instruction, it enables a
patching mode to guarantee that the failed function will call
optimized/moved versions of functions. The patching happens at original
function entry points.
'-skip=<func1,func2,...>' option now can be used to skip processing of
arbitrary functions in the relocation mode.
With '-use-old-text' or '-strict' we require all functions to be
processed. As such, it is incompatible with '-lite' option,
and '-skip' option will only disable optimizations of listed
functions, not their disassembly and emission.
(cherry picked from FBD22040717)
2020-06-15 15:15:47 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return Relocation({RelOffset, Symbol, RelType, Addend, 0});
|
|
|
|
}
|
|
|
|
|
2019-06-29 00:21:27 +08:00
|
|
|
bool replaceImmWithSymbolRef(MCInst &Inst, const MCSymbol *Symbol,
|
|
|
|
int64_t Addend, MCContext *Ctx, int64_t &Value,
|
2019-06-05 06:30:22 +08:00
|
|
|
uint64_t RelType) const override {
|
2018-03-10 01:45:13 +08:00
|
|
|
unsigned ImmOpNo = -1U;
|
|
|
|
|
2022-01-09 09:48:33 +08:00
|
|
|
for (unsigned Index = 0; Index < MCPlus::getNumPrimeOperands(Inst);
|
|
|
|
++Index) {
|
2018-03-10 01:45:13 +08:00
|
|
|
if (Inst.getOperand(Index).isImm()) {
|
|
|
|
ImmOpNo = Index;
|
|
|
|
// TODO: this is a bit hacky. It finds the correct operand by
|
|
|
|
// searching for a specific immediate value. If no value is
|
|
|
|
// provided it defaults to the last immediate operand found.
|
|
|
|
// This could lead to unexpected results if the instruction
|
|
|
|
// has more than one immediate with the same value.
|
|
|
|
if (Inst.getOperand(ImmOpNo).getImm() == Value)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ImmOpNo == -1U)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
Value = Inst.getOperand(ImmOpNo).getImm();
|
|
|
|
|
2018-07-04 02:57:46 +08:00
|
|
|
setOperandToSymbolRef(Inst, ImmOpNo, Symbol, Addend, Ctx, RelType);
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2021-08-11 01:02:32 +08:00
|
|
|
bool replaceRegWithImm(MCInst &Inst, unsigned Register,
|
|
|
|
int64_t Imm) const override {
|
|
|
|
|
|
|
|
enum CheckSignExt : uint8_t {
|
|
|
|
NOCHECK = 0,
|
|
|
|
CHECK8,
|
|
|
|
CHECK32,
|
|
|
|
};
|
|
|
|
|
|
|
|
using CheckList = std::vector<std::pair<CheckSignExt, unsigned>>;
|
|
|
|
struct InstInfo {
|
|
|
|
// Size in bytes that Inst loads from memory.
|
|
|
|
uint8_t DataSize;
|
|
|
|
|
|
|
|
// True when the target operand has to be duplicated because the opcode
|
|
|
|
// expects a LHS operand.
|
|
|
|
bool HasLHS;
|
|
|
|
|
|
|
|
// List of checks and corresponding opcodes to be used. We try to use the
|
|
|
|
// smallest possible immediate value when various sizes are available,
|
|
|
|
// hence we may need to check whether a larger constant fits in a smaller
|
|
|
|
// immediate.
|
|
|
|
CheckList Checks;
|
|
|
|
};
|
|
|
|
|
|
|
|
InstInfo I;
|
|
|
|
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
default: {
|
|
|
|
switch (getPushSize(Inst)) {
|
|
|
|
|
|
|
|
case 2: I = {2, false, {{CHECK8, X86::PUSH16i8}, {NOCHECK, X86::PUSHi16}}}; break;
|
|
|
|
case 4: I = {4, false, {{CHECK8, X86::PUSH32i8}, {NOCHECK, X86::PUSHi32}}}; break;
|
|
|
|
case 8: I = {8, false, {{CHECK8, X86::PUSH64i8},
|
|
|
|
{CHECK32, X86::PUSH64i32},
|
|
|
|
{NOCHECK, Inst.getOpcode()}}}; break;
|
|
|
|
default: return false;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// MOV
|
|
|
|
case X86::MOV8rr: I = {1, false, {{NOCHECK, X86::MOV8ri}}}; break;
|
|
|
|
case X86::MOV16rr: I = {2, false, {{NOCHECK, X86::MOV16ri}}}; break;
|
|
|
|
case X86::MOV32rr: I = {4, false, {{NOCHECK, X86::MOV32ri}}}; break;
|
|
|
|
case X86::MOV64rr: I = {8, false, {{CHECK32, X86::MOV64ri32},
|
|
|
|
{NOCHECK, X86::MOV64ri}}}; break;
|
|
|
|
|
|
|
|
case X86::MOV8mr: I = {1, false, {{NOCHECK, X86::MOV8mi}}}; break;
|
|
|
|
case X86::MOV16mr: I = {2, false, {{NOCHECK, X86::MOV16mi}}}; break;
|
|
|
|
case X86::MOV32mr: I = {4, false, {{NOCHECK, X86::MOV32mi}}}; break;
|
|
|
|
case X86::MOV64mr: I = {8, false, {{CHECK32, X86::MOV64mi32},
|
|
|
|
{NOCHECK, X86::MOV64mr}}}; break;
|
|
|
|
|
|
|
|
// MOVZX
|
|
|
|
case X86::MOVZX16rr8: I = {1, false, {{NOCHECK, X86::MOV16ri}}}; break;
|
|
|
|
case X86::MOVZX32rr8: I = {1, false, {{NOCHECK, X86::MOV32ri}}}; break;
|
|
|
|
case X86::MOVZX32rr16: I = {2, false, {{NOCHECK, X86::MOV32ri}}}; break;
|
|
|
|
|
|
|
|
// CMP
|
|
|
|
case X86::CMP8rr: I = {1, false, {{NOCHECK, X86::CMP8ri}}}; break;
|
|
|
|
case X86::CMP16rr: I = {2, false, {{CHECK8, X86::CMP16ri8},
|
|
|
|
{NOCHECK, X86::CMP16ri}}}; break;
|
|
|
|
case X86::CMP32rr: I = {4, false, {{CHECK8, X86::CMP32ri8},
|
|
|
|
{NOCHECK, X86::CMP32ri}}}; break;
|
|
|
|
case X86::CMP64rr: I = {8, false, {{CHECK8, X86::CMP64ri8},
|
|
|
|
{CHECK32, X86::CMP64ri32},
|
|
|
|
{NOCHECK, X86::CMP64rr}}}; break;
|
|
|
|
|
|
|
|
// TEST
|
|
|
|
case X86::TEST8rr: I = {1, false, {{NOCHECK, X86::TEST8ri}}}; break;
|
|
|
|
case X86::TEST16rr: I = {2, false, {{NOCHECK, X86::TEST16ri}}}; break;
|
|
|
|
case X86::TEST32rr: I = {4, false, {{NOCHECK, X86::TEST32ri}}}; break;
|
|
|
|
case X86::TEST64rr: I = {8, false, {{CHECK32, X86::TEST64ri32},
|
|
|
|
{NOCHECK, X86::TEST64rr}}}; break;
|
|
|
|
|
|
|
|
// ADD
|
|
|
|
case X86::ADD8rr: I = {1, true, {{NOCHECK, X86::ADD8ri}}}; break;
|
|
|
|
case X86::ADD16rr: I = {2, true, {{CHECK8, X86::ADD16ri8},
|
|
|
|
{NOCHECK, X86::ADD16ri}}}; break;
|
|
|
|
case X86::ADD32rr: I = {4, true, {{CHECK8, X86::ADD32ri8},
|
|
|
|
{NOCHECK, X86::ADD32ri}}}; break;
|
|
|
|
case X86::ADD64rr: I = {8, true, {{CHECK8, X86::ADD64ri8},
|
|
|
|
{CHECK32, X86::ADD64ri32},
|
|
|
|
{NOCHECK, X86::ADD64rr}}}; break;
|
|
|
|
|
|
|
|
// SUB
|
|
|
|
case X86::SUB8rr: I = {1, true, {{NOCHECK, X86::SUB8ri}}}; break;
|
|
|
|
case X86::SUB16rr: I = {2, true, {{CHECK8, X86::SUB16ri8},
|
|
|
|
{NOCHECK, X86::SUB16ri}}}; break;
|
|
|
|
case X86::SUB32rr: I = {4, true, {{CHECK8, X86::SUB32ri8},
|
|
|
|
{NOCHECK, X86::SUB32ri}}}; break;
|
|
|
|
case X86::SUB64rr: I = {8, true, {{CHECK8, X86::SUB64ri8},
|
|
|
|
{CHECK32, X86::SUB64ri32},
|
|
|
|
{NOCHECK, X86::SUB64rr}}}; break;
|
|
|
|
|
|
|
|
// AND
|
|
|
|
case X86::AND8rr: I = {1, true, {{NOCHECK, X86::AND8ri}}}; break;
|
|
|
|
case X86::AND16rr: I = {2, true, {{CHECK8, X86::AND16ri8},
|
|
|
|
{NOCHECK, X86::AND16ri}}}; break;
|
|
|
|
case X86::AND32rr: I = {4, true, {{CHECK8, X86::AND32ri8},
|
|
|
|
{NOCHECK, X86::AND32ri}}}; break;
|
|
|
|
case X86::AND64rr: I = {8, true, {{CHECK8, X86::AND64ri8},
|
|
|
|
{CHECK32, X86::AND64ri32},
|
|
|
|
{NOCHECK, X86::AND64rr}}}; break;
|
|
|
|
|
|
|
|
// OR
|
|
|
|
case X86::OR8rr: I = {1, true, {{NOCHECK, X86::OR8ri}}}; break;
|
|
|
|
case X86::OR16rr: I = {2, true, {{CHECK8, X86::OR16ri8},
|
|
|
|
{NOCHECK, X86::OR16ri}}}; break;
|
|
|
|
case X86::OR32rr: I = {4, true, {{CHECK8, X86::OR32ri8},
|
|
|
|
{NOCHECK, X86::OR32ri}}}; break;
|
|
|
|
case X86::OR64rr: I = {8, true, {{CHECK8, X86::OR64ri8},
|
|
|
|
{CHECK32, X86::OR64ri32},
|
|
|
|
{NOCHECK, X86::OR64rr}}}; break;
|
|
|
|
|
|
|
|
// XOR
|
|
|
|
case X86::XOR8rr: I = {1, true, {{NOCHECK, X86::XOR8ri}}}; break;
|
|
|
|
case X86::XOR16rr: I = {2, true, {{CHECK8, X86::XOR16ri8},
|
|
|
|
{NOCHECK, X86::XOR16ri}}}; break;
|
|
|
|
case X86::XOR32rr: I = {4, true, {{CHECK8, X86::XOR32ri8},
|
|
|
|
{NOCHECK, X86::XOR32ri}}}; break;
|
|
|
|
case X86::XOR64rr: I = {8, true, {{CHECK8, X86::XOR64ri8},
|
|
|
|
{CHECK32, X86::XOR64ri32},
|
|
|
|
{NOCHECK, X86::XOR64rr}}}; break;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Compute the new opcode.
|
|
|
|
unsigned NewOpcode = 0;
|
|
|
|
for (const std::pair<CheckSignExt, unsigned> &Check : I.Checks) {
|
|
|
|
NewOpcode = Check.second;
|
|
|
|
if (Check.first == NOCHECK)
|
|
|
|
break;
|
2022-02-03 01:04:10 +08:00
|
|
|
if (Check.first == CHECK8 && isInt<8>(Imm))
|
2021-08-11 01:02:32 +08:00
|
|
|
break;
|
2022-02-03 01:04:10 +08:00
|
|
|
if (Check.first == CHECK32 && isInt<32>(Imm))
|
2021-08-11 01:02:32 +08:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (NewOpcode == Inst.getOpcode())
|
|
|
|
return false;
|
|
|
|
|
|
|
|
const MCInstrDesc &InstDesc = Info->get(Inst.getOpcode());
|
|
|
|
|
|
|
|
unsigned NumFound = 0;
|
|
|
|
for (unsigned Index = InstDesc.getNumDefs() + (I.HasLHS ? 1 : 0),
|
2022-01-09 09:48:33 +08:00
|
|
|
E = InstDesc.getNumOperands();
|
|
|
|
Index != E; ++Index)
|
2021-08-11 01:02:32 +08:00
|
|
|
if (Inst.getOperand(Index).isReg() &&
|
|
|
|
Inst.getOperand(Index).getReg() == Register)
|
|
|
|
NumFound++;
|
|
|
|
|
|
|
|
if (NumFound != 1)
|
|
|
|
return false;
|
|
|
|
|
2022-02-24 14:30:03 +08:00
|
|
|
MCOperand TargetOp = Inst.getOperand(0);
|
|
|
|
Inst.clear();
|
2021-08-11 01:02:32 +08:00
|
|
|
Inst.setOpcode(NewOpcode);
|
2022-02-24 14:30:03 +08:00
|
|
|
Inst.addOperand(TargetOp);
|
|
|
|
if (I.HasLHS)
|
|
|
|
Inst.addOperand(TargetOp);
|
|
|
|
Inst.addOperand(MCOperand::createImm(Imm));
|
2021-08-11 01:02:32 +08:00
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool replaceRegWithReg(MCInst &Inst, unsigned ToReplace,
|
|
|
|
unsigned ReplaceWith) const override {
|
|
|
|
|
|
|
|
// Get the HasLHS value so that iteration can be done
|
|
|
|
bool HasLHS;
|
2022-03-18 16:48:00 +08:00
|
|
|
if (X86::isAND(Inst.getOpcode()) || X86::isADD(Inst.getOpcode()) ||
|
|
|
|
X86::isSUB(Inst.getOpcode())) {
|
2021-08-11 01:02:32 +08:00
|
|
|
HasLHS = true;
|
2022-03-18 16:48:00 +08:00
|
|
|
} else if (isPop(Inst) || isPush(Inst) || X86::isCMP(Inst.getOpcode()) ||
|
|
|
|
X86::isTEST(Inst.getOpcode())) {
|
2021-08-11 01:02:32 +08:00
|
|
|
HasLHS = false;
|
|
|
|
} else {
|
|
|
|
switch (Inst.getOpcode()) {
|
|
|
|
case X86::MOV8rr:
|
|
|
|
case X86::MOV8rm:
|
|
|
|
case X86::MOV8mr:
|
|
|
|
case X86::MOV8ri:
|
|
|
|
case X86::MOV16rr:
|
|
|
|
case X86::MOV16rm:
|
|
|
|
case X86::MOV16mr:
|
|
|
|
case X86::MOV16ri:
|
|
|
|
case X86::MOV32rr:
|
|
|
|
case X86::MOV32rm:
|
|
|
|
case X86::MOV32mr:
|
|
|
|
case X86::MOV32ri:
|
|
|
|
case X86::MOV64rr:
|
|
|
|
case X86::MOV64rm:
|
|
|
|
case X86::MOV64mr:
|
|
|
|
case X86::MOV64ri:
|
|
|
|
case X86::MOVZX16rr8:
|
|
|
|
case X86::MOVZX32rr8:
|
|
|
|
case X86::MOVZX32rr16:
|
|
|
|
case X86::MOVSX32rm8:
|
|
|
|
case X86::MOVSX32rr8:
|
|
|
|
case X86::MOVSX64rm32:
|
|
|
|
case X86::LEA64r:
|
|
|
|
HasLHS = false;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
const MCInstrDesc &InstDesc = Info->get(Inst.getOpcode());
|
|
|
|
|
|
|
|
bool FoundOne = false;
|
|
|
|
|
|
|
|
// Iterate only through src operands that arent also dest operands
|
|
|
|
for (unsigned Index = InstDesc.getNumDefs() + (HasLHS ? 1 : 0),
|
|
|
|
E = InstDesc.getNumOperands();
|
|
|
|
Index != E; ++Index) {
|
|
|
|
BitVector RegAliases = getAliases(ToReplace, true);
|
|
|
|
if (!Inst.getOperand(Index).isReg() ||
|
|
|
|
!RegAliases.test(Inst.getOperand(Index).getReg()))
|
|
|
|
continue;
|
|
|
|
// Resize register if needed
|
|
|
|
unsigned SizedReplaceWith = getAliasSized(
|
|
|
|
ReplaceWith, getRegSize(Inst.getOperand(Index).getReg()));
|
|
|
|
MCOperand NewOperand = MCOperand::createReg(SizedReplaceWith);
|
|
|
|
Inst.getOperand(Index) = NewOperand;
|
|
|
|
FoundOne = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Return true if at least one operand was replaced
|
|
|
|
return FoundOne;
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
bool createUncondBranch(MCInst &Inst, const MCSymbol *TBB,
|
|
|
|
MCContext *Ctx) const override {
|
|
|
|
Inst.setOpcode(X86::JMP_1);
|
|
|
|
Inst.addOperand(MCOperand::createExpr(
|
|
|
|
MCSymbolRefExpr::create(TBB, MCSymbolRefExpr::VK_None, *Ctx)));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2019-02-01 03:23:02 +08:00
|
|
|
bool createCall(MCInst &Inst, const MCSymbol *Target,
|
|
|
|
MCContext *Ctx) override {
|
|
|
|
Inst.setOpcode(X86::CALL64pcrel32);
|
|
|
|
Inst.addOperand(MCOperand::createExpr(
|
|
|
|
MCSymbolRefExpr::create(Target, MCSymbolRefExpr::VK_None, *Ctx)));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
bool createTailCall(MCInst &Inst, const MCSymbol *Target,
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
MCContext *Ctx) override {
|
2021-06-24 02:24:09 +08:00
|
|
|
return createDirectCall(Inst, Target, Ctx, /*IsTailCall*/ true);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
void createLongTailCall(InstructionListType &Seq, const MCSymbol *Target,
|
2021-10-05 00:17:01 +08:00
|
|
|
MCContext *Ctx) override {
|
|
|
|
Seq.clear();
|
|
|
|
Seq.emplace_back();
|
|
|
|
createDirectCall(Seq.back(), Target, Ctx, /*IsTailCall*/ true);
|
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
bool createTrap(MCInst &Inst) const override {
|
|
|
|
Inst.clear();
|
|
|
|
Inst.setOpcode(X86::TRAP);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool reverseBranchCondition(MCInst &Inst, const MCSymbol *TBB,
|
|
|
|
MCContext *Ctx) const override {
|
2020-12-02 08:29:39 +08:00
|
|
|
unsigned InvCC = getInvertedCondCode(getCondCode(Inst));
|
|
|
|
assert(InvCC != X86::COND_INVALID && "invalid branch instruction");
|
|
|
|
Inst.getOperand(Info->get(Inst.getOpcode()).NumOperands - 1).setImm(InvCC);
|
2018-03-10 01:45:13 +08:00
|
|
|
Inst.getOperand(0) = MCOperand::createExpr(
|
2021-08-18 01:15:21 +08:00
|
|
|
MCSymbolRefExpr::create(TBB, MCSymbolRefExpr::VK_None, *Ctx));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool replaceBranchCondition(MCInst &Inst, const MCSymbol *TBB, MCContext *Ctx,
|
|
|
|
unsigned CC) const override {
|
|
|
|
if (CC == X86::COND_INVALID)
|
|
|
|
return false;
|
|
|
|
Inst.getOperand(Info->get(Inst.getOpcode()).NumOperands - 1).setImm(CC);
|
|
|
|
Inst.getOperand(0) = MCOperand::createExpr(
|
2018-03-10 01:45:13 +08:00
|
|
|
MCSymbolRefExpr::create(TBB, MCSymbolRefExpr::VK_None, *Ctx));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2020-12-02 08:29:39 +08:00
|
|
|
unsigned getCanonicalBranchCondCode(unsigned CC) const override {
|
|
|
|
switch (CC) {
|
|
|
|
default: return X86::COND_INVALID;
|
|
|
|
|
|
|
|
case X86::COND_E: return X86::COND_E;
|
|
|
|
case X86::COND_NE: return X86::COND_E;
|
|
|
|
|
|
|
|
case X86::COND_L: return X86::COND_L;
|
|
|
|
case X86::COND_GE: return X86::COND_L;
|
|
|
|
|
|
|
|
case X86::COND_LE: return X86::COND_G;
|
|
|
|
case X86::COND_G: return X86::COND_G;
|
|
|
|
|
|
|
|
case X86::COND_B: return X86::COND_B;
|
|
|
|
case X86::COND_AE: return X86::COND_B;
|
|
|
|
|
|
|
|
case X86::COND_BE: return X86::COND_A;
|
|
|
|
case X86::COND_A: return X86::COND_A;
|
|
|
|
|
|
|
|
case X86::COND_S: return X86::COND_S;
|
|
|
|
case X86::COND_NS: return X86::COND_S;
|
|
|
|
|
|
|
|
case X86::COND_P: return X86::COND_P;
|
|
|
|
case X86::COND_NP: return X86::COND_P;
|
|
|
|
|
|
|
|
case X86::COND_O: return X86::COND_O;
|
|
|
|
case X86::COND_NO: return X86::COND_O;
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
bool replaceBranchTarget(MCInst &Inst, const MCSymbol *TBB,
|
|
|
|
MCContext *Ctx) const override {
|
|
|
|
assert((isCall(Inst) || isBranch(Inst)) && !isIndirectBranch(Inst) &&
|
|
|
|
"Invalid instruction");
|
|
|
|
Inst.getOperand(0) = MCOperand::createExpr(
|
|
|
|
MCSymbolRefExpr::create(TBB, MCSymbolRefExpr::VK_None, *Ctx));
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2022-01-09 09:48:33 +08:00
|
|
|
MCPhysReg getX86R11() const override { return X86::R11; }
|
2018-07-26 10:07:41 +08:00
|
|
|
|
2019-05-21 11:11:40 +08:00
|
|
|
MCPhysReg getIntArgRegister(unsigned ArgNo) const override {
|
|
|
|
// FIXME: this should depend on the calling convention.
|
|
|
|
switch (ArgNo) {
|
|
|
|
case 0: return X86::RDI;
|
|
|
|
case 1: return X86::RSI;
|
|
|
|
case 2: return X86::RDX;
|
|
|
|
case 3: return X86::RCX;
|
|
|
|
case 4: return X86::R8;
|
|
|
|
case 5: return X86::R9;
|
|
|
|
default: return getNoRegister();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-07-26 10:07:41 +08:00
|
|
|
void createPause(MCInst &Inst) const override {
|
|
|
|
Inst.clear();
|
|
|
|
Inst.setOpcode(X86::PAUSE);
|
|
|
|
}
|
|
|
|
|
|
|
|
void createLfence(MCInst &Inst) const override {
|
|
|
|
Inst.clear();
|
|
|
|
Inst.setOpcode(X86::LFENCE);
|
|
|
|
}
|
|
|
|
|
2021-06-24 02:24:09 +08:00
|
|
|
bool createDirectCall(MCInst &Inst, const MCSymbol *Target, MCContext *Ctx,
|
|
|
|
bool IsTailCall) override {
|
2018-07-26 10:07:41 +08:00
|
|
|
Inst.clear();
|
2021-06-24 02:24:09 +08:00
|
|
|
Inst.setOpcode(IsTailCall ? X86::JMP_4 : X86::CALL64pcrel32);
|
2018-07-26 10:07:41 +08:00
|
|
|
Inst.addOperand(MCOperand::createExpr(
|
|
|
|
MCSymbolRefExpr::create(Target, MCSymbolRefExpr::VK_None, *Ctx)));
|
2021-06-24 02:24:09 +08:00
|
|
|
if (IsTailCall)
|
|
|
|
setTailCall(Inst);
|
2018-07-26 10:07:41 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
void createShortJmp(InstructionListType &Seq, const MCSymbol *Target,
|
2021-10-05 00:17:01 +08:00
|
|
|
MCContext *Ctx, bool IsTailCall) override {
|
2018-07-26 10:07:41 +08:00
|
|
|
Seq.clear();
|
|
|
|
MCInst Inst;
|
|
|
|
Inst.setOpcode(X86::JMP_1);
|
|
|
|
Inst.addOperand(MCOperand::createExpr(
|
|
|
|
MCSymbolRefExpr::create(Target, MCSymbolRefExpr::VK_None, *Ctx)));
|
2021-10-05 00:17:01 +08:00
|
|
|
if (IsTailCall)
|
|
|
|
setTailCall(Inst);
|
2018-07-26 10:07:41 +08:00
|
|
|
Seq.emplace_back(Inst);
|
|
|
|
}
|
|
|
|
|
2021-08-11 01:02:32 +08:00
|
|
|
bool isConditionalMove(const MCInst &Inst) const override {
|
|
|
|
unsigned OpCode = Inst.getOpcode();
|
|
|
|
return (OpCode == X86::CMOV16rr || OpCode == X86::CMOV32rr ||
|
|
|
|
OpCode == X86::CMOV64rr);
|
|
|
|
}
|
|
|
|
|
2018-07-26 10:07:41 +08:00
|
|
|
bool isBranchOnMem(const MCInst &Inst) const override {
|
2021-04-08 15:19:26 +08:00
|
|
|
unsigned OpCode = Inst.getOpcode();
|
2020-12-02 08:29:39 +08:00
|
|
|
if (OpCode == X86::CALL64m || (OpCode == X86::JMP32m && isTailCall(Inst)) ||
|
2018-07-26 10:07:41 +08:00
|
|
|
OpCode == X86::JMP64m)
|
|
|
|
return true;
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool isBranchOnReg(const MCInst &Inst) const override {
|
2021-04-08 15:19:26 +08:00
|
|
|
unsigned OpCode = Inst.getOpcode();
|
2020-12-02 08:29:39 +08:00
|
|
|
if (OpCode == X86::CALL64r || (OpCode == X86::JMP32r && isTailCall(Inst)) ||
|
2018-07-26 10:07:41 +08:00
|
|
|
OpCode == X86::JMP64r)
|
|
|
|
return true;
|
|
|
|
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
void createPushRegister(MCInst &Inst, MCPhysReg Reg,
|
|
|
|
unsigned Size) const override {
|
|
|
|
Inst.clear();
|
|
|
|
unsigned NewOpcode = 0;
|
2019-06-20 11:10:49 +08:00
|
|
|
if (Reg == X86::EFLAGS) {
|
|
|
|
switch (Size) {
|
|
|
|
case 2: NewOpcode = X86::PUSHF16; break;
|
|
|
|
case 4: NewOpcode = X86::PUSHF32; break;
|
|
|
|
case 8: NewOpcode = X86::PUSHF64; break;
|
|
|
|
default:
|
2022-01-09 09:48:33 +08:00
|
|
|
llvm_unreachable("Unexpected size");
|
2019-06-20 11:10:49 +08:00
|
|
|
}
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
return;
|
|
|
|
}
|
2018-07-26 10:07:41 +08:00
|
|
|
switch (Size) {
|
|
|
|
case 2: NewOpcode = X86::PUSH16r; break;
|
|
|
|
case 4: NewOpcode = X86::PUSH32r; break;
|
|
|
|
case 8: NewOpcode = X86::PUSH64r; break;
|
|
|
|
default:
|
2022-01-09 09:48:33 +08:00
|
|
|
llvm_unreachable("Unexpected size");
|
2018-07-26 10:07:41 +08:00
|
|
|
}
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
Inst.addOperand(MCOperand::createReg(Reg));
|
|
|
|
}
|
|
|
|
|
|
|
|
void createPopRegister(MCInst &Inst, MCPhysReg Reg,
|
|
|
|
unsigned Size) const override {
|
|
|
|
Inst.clear();
|
|
|
|
unsigned NewOpcode = 0;
|
2019-06-20 11:10:49 +08:00
|
|
|
if (Reg == X86::EFLAGS) {
|
|
|
|
switch (Size) {
|
|
|
|
case 2: NewOpcode = X86::POPF16; break;
|
|
|
|
case 4: NewOpcode = X86::POPF32; break;
|
|
|
|
case 8: NewOpcode = X86::POPF64; break;
|
|
|
|
default:
|
2022-01-09 09:48:33 +08:00
|
|
|
llvm_unreachable("Unexpected size");
|
2019-06-20 11:10:49 +08:00
|
|
|
}
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
return;
|
|
|
|
}
|
2018-07-26 10:07:41 +08:00
|
|
|
switch (Size) {
|
|
|
|
case 2: NewOpcode = X86::POP16r; break;
|
|
|
|
case 4: NewOpcode = X86::POP32r; break;
|
|
|
|
case 8: NewOpcode = X86::POP64r; break;
|
|
|
|
default:
|
2022-01-09 09:48:33 +08:00
|
|
|
llvm_unreachable("Unexpected size");
|
2018-07-26 10:07:41 +08:00
|
|
|
}
|
|
|
|
Inst.setOpcode(NewOpcode);
|
|
|
|
Inst.addOperand(MCOperand::createReg(Reg));
|
|
|
|
}
|
|
|
|
|
2019-06-20 11:10:49 +08:00
|
|
|
void createPushFlags(MCInst &Inst, unsigned Size) const override {
|
|
|
|
return createPushRegister(Inst, X86::EFLAGS, Size);
|
|
|
|
}
|
|
|
|
|
|
|
|
void createPopFlags(MCInst &Inst, unsigned Size) const override {
|
|
|
|
return createPopRegister(Inst, X86::EFLAGS, Size);
|
|
|
|
}
|
|
|
|
|
2021-10-12 00:10:06 +08:00
|
|
|
void createAddRegImm(MCInst &Inst, MCPhysReg Reg, int64_t Value,
|
|
|
|
unsigned Size) const {
|
|
|
|
unsigned int Opcode;
|
|
|
|
switch (Size) {
|
|
|
|
case 1: Opcode = X86::ADD8ri; break;
|
|
|
|
case 2: Opcode = X86::ADD16ri; break;
|
|
|
|
case 4: Opcode = X86::ADD32ri; break;
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
Inst.setOpcode(Opcode);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.addOperand(MCOperand::createReg(Reg));
|
|
|
|
Inst.addOperand(MCOperand::createReg(Reg));
|
|
|
|
Inst.addOperand(MCOperand::createImm(Value));
|
|
|
|
}
|
|
|
|
|
|
|
|
void createClearRegWithNoEFlagsUpdate(MCInst &Inst, MCPhysReg Reg,
|
|
|
|
unsigned Size) const {
|
|
|
|
unsigned int Opcode;
|
|
|
|
switch (Size) {
|
|
|
|
case 1: Opcode = X86::MOV8ri; break;
|
|
|
|
case 2: Opcode = X86::MOV16ri; break;
|
|
|
|
case 4: Opcode = X86::MOV32ri; break;
|
|
|
|
case 8: Opcode = X86::MOV64ri; break;
|
|
|
|
default:
|
|
|
|
llvm_unreachable("Unexpected size");
|
|
|
|
}
|
|
|
|
Inst.setOpcode(Opcode);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.addOperand(MCOperand::createReg(Reg));
|
|
|
|
Inst.addOperand(MCOperand::createImm(0));
|
|
|
|
}
|
|
|
|
|
|
|
|
void createX86SaveOVFlagToRegister(MCInst &Inst, MCPhysReg Reg) const {
|
|
|
|
Inst.setOpcode(X86::SETCCr);
|
|
|
|
Inst.clear();
|
|
|
|
Inst.addOperand(MCOperand::createReg(Reg));
|
|
|
|
Inst.addOperand(MCOperand::createImm(X86::COND_O));
|
|
|
|
}
|
|
|
|
|
|
|
|
void createX86Lahf(MCInst &Inst) const {
|
|
|
|
Inst.setOpcode(X86::LAHF);
|
|
|
|
Inst.clear();
|
|
|
|
}
|
|
|
|
|
|
|
|
void createX86Sahf(MCInst &Inst) const {
|
|
|
|
Inst.setOpcode(X86::SAHF);
|
|
|
|
Inst.clear();
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
void createInstrIncMemory(InstructionListType &Instrs, const MCSymbol *Target,
|
2021-10-12 00:10:06 +08:00
|
|
|
MCContext *Ctx, bool IsLeaf) const override {
|
|
|
|
unsigned int I = 0;
|
|
|
|
|
|
|
|
Instrs.resize(IsLeaf ? 13 : 11);
|
|
|
|
// Don't clobber application red zone (ABI dependent)
|
|
|
|
if (IsLeaf)
|
|
|
|
createStackPointerIncrement(Instrs[I++], 128,
|
|
|
|
/*NoFlagsClobber=*/true);
|
|
|
|
|
|
|
|
// Performance improvements based on the optimization discussed at
|
|
|
|
// https://reviews.llvm.org/D6629
|
|
|
|
// LAHF/SAHF are used instead of PUSHF/POPF
|
|
|
|
// PUSHF
|
|
|
|
createPushRegister(Instrs[I++], X86::RAX, 8);
|
|
|
|
createClearRegWithNoEFlagsUpdate(Instrs[I++], X86::RAX, 8);
|
|
|
|
createX86Lahf(Instrs[I++]);
|
|
|
|
createPushRegister(Instrs[I++], X86::RAX, 8);
|
|
|
|
createClearRegWithNoEFlagsUpdate(Instrs[I++], X86::RAX, 8);
|
|
|
|
createX86SaveOVFlagToRegister(Instrs[I++], X86::AL);
|
|
|
|
// LOCK INC
|
|
|
|
createIncMemory(Instrs[I++], Target, Ctx);
|
|
|
|
// POPF
|
|
|
|
createAddRegImm(Instrs[I++], X86::AL, 127, 1);
|
|
|
|
createPopRegister(Instrs[I++], X86::RAX, 8);
|
|
|
|
createX86Sahf(Instrs[I++]);
|
|
|
|
createPopRegister(Instrs[I++], X86::RAX, 8);
|
|
|
|
|
|
|
|
if (IsLeaf)
|
|
|
|
createStackPointerDecrement(Instrs[I], 128,
|
|
|
|
/*NoFlagsClobber=*/true);
|
|
|
|
}
|
|
|
|
|
2019-12-14 09:27:03 +08:00
|
|
|
void createSwap(MCInst &Inst, MCPhysReg Source, MCPhysReg MemBaseReg,
|
|
|
|
int64_t Disp) const {
|
|
|
|
Inst.setOpcode(X86::XCHG64rm);
|
|
|
|
Inst.addOperand(MCOperand::createReg(Source));
|
|
|
|
Inst.addOperand(MCOperand::createReg(Source));
|
|
|
|
Inst.addOperand(MCOperand::createReg(MemBaseReg)); // BaseReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(1)); // ScaleAmt
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // IndexReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(Disp)); // Displacement
|
2022-01-09 09:48:33 +08:00
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // AddrSegmentReg
|
2019-12-14 09:27:03 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void createIndirectBranch(MCInst &Inst, MCPhysReg MemBaseReg,
|
|
|
|
int64_t Disp) const {
|
|
|
|
Inst.setOpcode(X86::JMP64m);
|
|
|
|
Inst.addOperand(MCOperand::createReg(MemBaseReg)); // BaseReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(1)); // ScaleAmt
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // IndexReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(Disp)); // Displacement
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // AddrSegmentReg
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType createInstrumentedIndirectCall(const MCInst &CallInst,
|
|
|
|
bool TailCall,
|
|
|
|
MCSymbol *HandlerFuncAddr,
|
|
|
|
int CallSiteID,
|
|
|
|
MCContext *Ctx) override {
|
2020-04-07 08:38:11 +08:00
|
|
|
// Check if the target address expression used in the original indirect call
|
|
|
|
// uses the stack pointer, which we are going to clobber.
|
|
|
|
static BitVector SPAliases(getAliases(X86::RSP));
|
2021-05-14 01:50:47 +08:00
|
|
|
bool UsesSP = false;
|
2020-04-07 08:38:11 +08:00
|
|
|
// Skip defs.
|
|
|
|
for (unsigned I = Info->get(CallInst.getOpcode()).getNumDefs(),
|
2022-01-09 09:48:33 +08:00
|
|
|
E = MCPlus::getNumPrimeOperands(CallInst);
|
|
|
|
I != E; ++I) {
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCOperand &Operand = CallInst.getOperand(I);
|
2020-04-07 08:38:11 +08:00
|
|
|
if (Operand.isReg() && SPAliases[Operand.getReg()]) {
|
|
|
|
UsesSP = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType Insts;
|
2019-12-14 09:27:03 +08:00
|
|
|
MCPhysReg TempReg = getIntArgRegister(0);
|
|
|
|
// Code sequence used to enter indirect call instrumentation helper:
|
|
|
|
// push %rdi
|
2020-04-07 08:38:11 +08:00
|
|
|
// add $8, %rsp ;; $rsp may be used in target, so fix it to prev val
|
2019-12-14 09:27:03 +08:00
|
|
|
// movq target, %rdi ;; via convertIndirectCallTargetToLoad
|
2020-04-07 08:38:11 +08:00
|
|
|
// sub $8, %rsp ;; restore correct stack value
|
2019-12-14 09:27:03 +08:00
|
|
|
// push %rdi
|
|
|
|
// movq $CallSiteID, %rdi
|
|
|
|
// push %rdi
|
2021-06-24 02:24:09 +08:00
|
|
|
// callq/jmp HandlerFuncAddr
|
2020-04-07 08:38:11 +08:00
|
|
|
Insts.emplace_back();
|
|
|
|
createPushRegister(Insts.back(), TempReg, 8);
|
|
|
|
if (UsesSP) { // Only adjust SP if we really need to
|
|
|
|
Insts.emplace_back();
|
|
|
|
createStackPointerDecrement(Insts.back(), 8, /*NoFlagsClobber=*/false);
|
|
|
|
}
|
|
|
|
Insts.emplace_back(CallInst);
|
2020-12-02 08:29:39 +08:00
|
|
|
// Insts.back() and CallInst now share the same annotation instruction.
|
|
|
|
// Strip it from Insts.back(), only preserving tail call annotation.
|
|
|
|
stripAnnotations(Insts.back(), /*KeepTC=*/true);
|
2020-04-07 08:38:11 +08:00
|
|
|
convertIndirectCallToLoad(Insts.back(), TempReg);
|
|
|
|
if (UsesSP) {
|
|
|
|
Insts.emplace_back();
|
|
|
|
createStackPointerIncrement(Insts.back(), 8, /*NoFlagsClobber=*/false);
|
|
|
|
}
|
|
|
|
Insts.emplace_back();
|
|
|
|
createPushRegister(Insts.back(), TempReg, 8);
|
|
|
|
Insts.emplace_back();
|
|
|
|
createLoadImmediate(Insts.back(), TempReg, CallSiteID);
|
|
|
|
Insts.emplace_back();
|
|
|
|
createPushRegister(Insts.back(), TempReg, 8);
|
|
|
|
Insts.emplace_back();
|
2021-06-24 02:24:09 +08:00
|
|
|
createDirectCall(Insts.back(), HandlerFuncAddr, Ctx,
|
|
|
|
/*TailCall=*/TailCall);
|
2019-12-14 09:27:03 +08:00
|
|
|
// Carry over metadata
|
|
|
|
for (int I = MCPlus::getNumPrimeOperands(CallInst),
|
|
|
|
E = CallInst.getNumOperands();
|
2021-12-29 09:52:08 +08:00
|
|
|
I != E; ++I)
|
2020-04-07 08:38:11 +08:00
|
|
|
Insts.back().addOperand(CallInst.getOperand(I));
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2019-12-14 09:27:03 +08:00
|
|
|
return Insts;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType createInstrumentedIndCallHandlerExitBB() const override {
|
2019-12-14 09:27:03 +08:00
|
|
|
const MCPhysReg TempReg = getIntArgRegister(0);
|
2021-06-24 02:24:09 +08:00
|
|
|
// We just need to undo the sequence created for every ind call in
|
2019-12-14 09:27:03 +08:00
|
|
|
// instrumentIndirectTarget(), which can be accomplished minimally with:
|
2021-06-24 02:24:09 +08:00
|
|
|
// popfq
|
2019-12-14 09:27:03 +08:00
|
|
|
// pop %rdi
|
|
|
|
// add $16, %rsp
|
|
|
|
// xchg (%rsp), %rdi
|
|
|
|
// jmp *-8(%rsp)
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType Insts(5);
|
2021-06-24 02:24:09 +08:00
|
|
|
createPopFlags(Insts[0], 8);
|
|
|
|
createPopRegister(Insts[1], TempReg, 8);
|
|
|
|
createStackPointerDecrement(Insts[2], 16, /*NoFlagsClobber=*/false);
|
|
|
|
createSwap(Insts[3], TempReg, X86::RSP, 0);
|
|
|
|
createIndirectBranch(Insts[4], X86::RSP, -8);
|
2019-12-14 09:27:03 +08:00
|
|
|
return Insts;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType
|
2021-06-24 02:24:09 +08:00
|
|
|
createInstrumentedIndTailCallHandlerExitBB() const override {
|
2019-12-14 09:27:03 +08:00
|
|
|
const MCPhysReg TempReg = getIntArgRegister(0);
|
|
|
|
// Same thing as above, but for tail calls
|
2021-06-24 02:24:09 +08:00
|
|
|
// popfq
|
2019-12-14 09:27:03 +08:00
|
|
|
// add $16, %rsp
|
|
|
|
// pop %rdi
|
|
|
|
// jmp *-16(%rsp)
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType Insts(4);
|
2021-06-24 02:24:09 +08:00
|
|
|
createPopFlags(Insts[0], 8);
|
|
|
|
createStackPointerDecrement(Insts[1], 16, /*NoFlagsClobber=*/false);
|
|
|
|
createPopRegister(Insts[2], TempReg, 8);
|
|
|
|
createIndirectBranch(Insts[3], X86::RSP, -16);
|
|
|
|
return Insts;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType
|
2021-06-24 02:24:09 +08:00
|
|
|
createInstrumentedIndCallHandlerEntryBB(const MCSymbol *InstrTrampoline,
|
|
|
|
const MCSymbol *IndCallHandler,
|
|
|
|
MCContext *Ctx) override {
|
|
|
|
const MCPhysReg TempReg = getIntArgRegister(0);
|
|
|
|
// Code sequence used to check whether InstrTampoline was initialized
|
|
|
|
// and call it if so, returns via IndCallHandler.
|
|
|
|
// pushfq
|
|
|
|
// mov InstrTrampoline,%rdi
|
|
|
|
// cmp $0x0,%rdi
|
|
|
|
// je IndCallHandler
|
|
|
|
// callq *%rdi
|
|
|
|
// jmpq IndCallHandler
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType Insts;
|
2021-06-24 02:24:09 +08:00
|
|
|
Insts.emplace_back();
|
|
|
|
createPushFlags(Insts.back(), 8);
|
|
|
|
Insts.emplace_back();
|
|
|
|
createMove(Insts.back(), InstrTrampoline, TempReg, Ctx);
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType cmpJmp = createCmpJE(TempReg, 0, IndCallHandler, Ctx);
|
2021-06-24 02:24:09 +08:00
|
|
|
Insts.insert(Insts.end(), cmpJmp.begin(), cmpJmp.end());
|
|
|
|
Insts.emplace_back();
|
|
|
|
Insts.back().setOpcode(X86::CALL64r);
|
|
|
|
Insts.back().addOperand(MCOperand::createReg(TempReg));
|
|
|
|
Insts.emplace_back();
|
|
|
|
createDirectCall(Insts.back(), IndCallHandler, Ctx, /*IsTailCall*/ true);
|
2019-12-14 09:27:03 +08:00
|
|
|
return Insts;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType createNumCountersGetter(MCContext *Ctx) const override {
|
|
|
|
InstructionListType Insts(2);
|
2021-01-29 04:32:03 +08:00
|
|
|
MCSymbol *NumLocs = Ctx->getOrCreateSymbol("__bolt_num_counters");
|
|
|
|
createMove(Insts[0], NumLocs, X86::EAX, Ctx);
|
|
|
|
createReturn(Insts[1]);
|
|
|
|
return Insts;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType
|
2021-06-24 02:24:09 +08:00
|
|
|
createInstrLocationsGetter(MCContext *Ctx) const override {
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType Insts(2);
|
2021-01-29 04:32:03 +08:00
|
|
|
MCSymbol *Locs = Ctx->getOrCreateSymbol("__bolt_instr_locations");
|
|
|
|
createLea(Insts[0], Locs, X86::EAX, Ctx);
|
|
|
|
createReturn(Insts[1]);
|
|
|
|
return Insts;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType createInstrTablesGetter(MCContext *Ctx) const override {
|
|
|
|
InstructionListType Insts(2);
|
2021-01-29 04:44:14 +08:00
|
|
|
MCSymbol *Locs = Ctx->getOrCreateSymbol("__bolt_instr_tables");
|
|
|
|
createLea(Insts[0], Locs, X86::EAX, Ctx);
|
|
|
|
createReturn(Insts[1]);
|
|
|
|
return Insts;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType createInstrNumFuncsGetter(MCContext *Ctx) const override {
|
|
|
|
InstructionListType Insts(2);
|
2021-01-29 04:44:14 +08:00
|
|
|
MCSymbol *NumFuncs = Ctx->getOrCreateSymbol("__bolt_instr_num_funcs");
|
|
|
|
createMove(Insts[0], NumFuncs, X86::EAX, Ctx);
|
|
|
|
createReturn(Insts[1]);
|
|
|
|
return Insts;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType createSymbolTrampoline(const MCSymbol *TgtSym,
|
2021-06-19 04:08:35 +08:00
|
|
|
MCContext *Ctx) const override {
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType Insts(1);
|
2021-06-19 04:08:35 +08:00
|
|
|
createUncondBranch(Insts[0], TgtSym, Ctx);
|
|
|
|
return Insts;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType createDummyReturnFunction(MCContext *Ctx) const override {
|
|
|
|
InstructionListType Insts(1);
|
2021-06-21 01:59:38 +08:00
|
|
|
createReturn(Insts[0]);
|
|
|
|
return Insts;
|
|
|
|
}
|
|
|
|
|
2019-06-20 11:10:49 +08:00
|
|
|
BlocksVectorTy indirectCallPromotion(
|
2022-01-09 09:48:33 +08:00
|
|
|
const MCInst &CallInst,
|
|
|
|
const std::vector<std::pair<MCSymbol *, uint64_t>> &Targets,
|
|
|
|
const std::vector<std::pair<MCSymbol *, uint64_t>> &VtableSyms,
|
|
|
|
const std::vector<MCInst *> &MethodFetchInsns,
|
|
|
|
const bool MinimizeCodeSize, MCContext *Ctx) override {
|
2018-03-10 01:45:13 +08:00
|
|
|
const bool IsTailCall = isTailCall(CallInst);
|
|
|
|
const bool IsJumpTable = getJumpTable(CallInst) != 0;
|
2019-06-20 11:10:49 +08:00
|
|
|
BlocksVectorTy Results;
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// Label for the current code block.
|
2021-12-10 03:53:12 +08:00
|
|
|
MCSymbol *NextTarget = nullptr;
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// The join block which contains all the instructions following CallInst.
|
|
|
|
// MergeBlock remains null if CallInst is a tail call.
|
2021-12-10 03:53:12 +08:00
|
|
|
MCSymbol *MergeBlock = nullptr;
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
unsigned FuncAddrReg = X86::R10;
|
|
|
|
|
2018-04-21 11:03:31 +08:00
|
|
|
const bool LoadElim = !VtableSyms.empty();
|
|
|
|
assert((!LoadElim || VtableSyms.size() == Targets.size()) &&
|
2018-03-10 01:45:13 +08:00
|
|
|
"There must be a vtable entry for every method "
|
|
|
|
"in the targets vector.");
|
|
|
|
|
|
|
|
if (MinimizeCodeSize && !LoadElim) {
|
|
|
|
std::set<unsigned> UsedRegs;
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
for (unsigned int I = 0; I < MCPlus::getNumPrimeOperands(CallInst); ++I) {
|
|
|
|
const MCOperand &Op = CallInst.getOperand(I);
|
2021-12-29 09:52:08 +08:00
|
|
|
if (Op.isReg())
|
2018-03-10 01:45:13 +08:00
|
|
|
UsedRegs.insert(Op.getReg());
|
|
|
|
}
|
|
|
|
|
|
|
|
if (UsedRegs.count(X86::R10) == 0)
|
|
|
|
FuncAddrReg = X86::R10;
|
|
|
|
else if (UsedRegs.count(X86::R11) == 0)
|
|
|
|
FuncAddrReg = X86::R11;
|
|
|
|
else
|
|
|
|
return Results;
|
|
|
|
}
|
|
|
|
|
2021-12-10 03:53:12 +08:00
|
|
|
const auto jumpToMergeBlock = [&](InstructionListType &NewCall) {
|
2018-03-10 01:45:13 +08:00
|
|
|
assert(MergeBlock);
|
2018-03-30 09:42:06 +08:00
|
|
|
NewCall.push_back(CallInst);
|
|
|
|
MCInst &Merge = NewCall.back();
|
2018-03-10 01:45:13 +08:00
|
|
|
Merge.clear();
|
|
|
|
createUncondBranch(Merge, MergeBlock, Ctx);
|
|
|
|
};
|
|
|
|
|
|
|
|
for (unsigned int i = 0; i < Targets.size(); ++i) {
|
2021-12-10 03:53:12 +08:00
|
|
|
Results.emplace_back(NextTarget, InstructionListType());
|
|
|
|
InstructionListType *NewCall = &Results.back().second;
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
if (MinimizeCodeSize && !LoadElim) {
|
|
|
|
// Load the call target into FuncAddrReg.
|
2022-01-09 09:48:33 +08:00
|
|
|
NewCall->push_back(CallInst); // Copy CallInst in order to get SMLoc
|
2018-03-30 09:42:06 +08:00
|
|
|
MCInst &Target = NewCall->back();
|
2018-03-10 01:45:13 +08:00
|
|
|
Target.clear();
|
|
|
|
Target.setOpcode(X86::MOV64ri32);
|
|
|
|
Target.addOperand(MCOperand::createReg(FuncAddrReg));
|
|
|
|
if (Targets[i].first) {
|
|
|
|
// Is this OK?
|
2022-01-09 09:48:33 +08:00
|
|
|
Target.addOperand(MCOperand::createExpr(MCSymbolRefExpr::create(
|
|
|
|
Targets[i].first, MCSymbolRefExpr::VK_None, *Ctx)));
|
2018-03-10 01:45:13 +08:00
|
|
|
} else {
|
2021-04-08 15:19:26 +08:00
|
|
|
const uint64_t Addr = Targets[i].second;
|
2018-03-10 01:45:13 +08:00
|
|
|
// Immediate address is out of sign extended 32 bit range.
|
2021-12-29 09:52:08 +08:00
|
|
|
if (int64_t(Addr) != int64_t(int32_t(Addr)))
|
2019-06-20 11:10:49 +08:00
|
|
|
return BlocksVectorTy();
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
Target.addOperand(MCOperand::createImm(Addr));
|
|
|
|
}
|
|
|
|
|
|
|
|
// Compare current call target to a specific address.
|
|
|
|
NewCall->push_back(CallInst);
|
2018-03-30 09:42:06 +08:00
|
|
|
MCInst &Compare = NewCall->back();
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.clear();
|
2021-12-29 09:52:08 +08:00
|
|
|
if (isBranchOnReg(CallInst))
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.setOpcode(X86::CMP64rr);
|
2021-12-29 09:52:08 +08:00
|
|
|
else if (CallInst.getOpcode() == X86::CALL64pcrel32)
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.setOpcode(X86::CMP64ri32);
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.setOpcode(X86::CMP64rm);
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.addOperand(MCOperand::createReg(FuncAddrReg));
|
|
|
|
|
|
|
|
// TODO: Would be preferable to only load this value once.
|
|
|
|
for (unsigned i = 0;
|
2022-01-09 09:48:33 +08:00
|
|
|
i < Info->get(CallInst.getOpcode()).getNumOperands(); ++i)
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
if (!CallInst.getOperand(i).isInst())
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.addOperand(CallInst.getOperand(i));
|
|
|
|
} else {
|
|
|
|
// Compare current call target to a specific address.
|
|
|
|
NewCall->push_back(CallInst);
|
2018-03-30 09:42:06 +08:00
|
|
|
MCInst &Compare = NewCall->back();
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.clear();
|
2021-12-29 09:52:08 +08:00
|
|
|
if (isBranchOnReg(CallInst))
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.setOpcode(X86::CMP64ri32);
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.setOpcode(X86::CMP64mi32);
|
|
|
|
|
|
|
|
// Original call address.
|
|
|
|
for (unsigned i = 0;
|
2022-01-09 09:48:33 +08:00
|
|
|
i < Info->get(CallInst.getOpcode()).getNumOperands(); ++i)
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
if (!CallInst.getOperand(i).isInst())
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.addOperand(CallInst.getOperand(i));
|
|
|
|
|
|
|
|
// Target address.
|
2018-04-21 11:03:31 +08:00
|
|
|
if (Targets[i].first || LoadElim) {
|
2021-04-08 15:19:26 +08:00
|
|
|
const MCSymbol *Sym =
|
|
|
|
LoadElim ? VtableSyms[i].first : Targets[i].first;
|
|
|
|
const uint64_t Addend = LoadElim ? VtableSyms[i].second : 0;
|
2018-04-21 11:03:31 +08:00
|
|
|
const MCExpr *Expr = MCSymbolRefExpr::create(Sym, *Ctx);
|
2021-12-29 09:52:08 +08:00
|
|
|
if (Addend)
|
2022-01-09 09:48:33 +08:00
|
|
|
Expr = MCBinaryExpr::createAdd(
|
|
|
|
Expr, MCConstantExpr::create(Addend, *Ctx), *Ctx);
|
2018-04-21 11:03:31 +08:00
|
|
|
Compare.addOperand(MCOperand::createExpr(Expr));
|
2018-03-10 01:45:13 +08:00
|
|
|
} else {
|
2021-04-08 15:19:26 +08:00
|
|
|
const uint64_t Addr = Targets[i].second;
|
2018-03-10 01:45:13 +08:00
|
|
|
// Immediate address is out of sign extended 32 bit range.
|
2021-12-29 09:52:08 +08:00
|
|
|
if (int64_t(Addr) != int64_t(int32_t(Addr)))
|
2019-06-20 11:10:49 +08:00
|
|
|
return BlocksVectorTy();
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
Compare.addOperand(MCOperand::createImm(Addr));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// jump to next target compare.
|
2020-12-02 08:29:39 +08:00
|
|
|
NextTarget =
|
|
|
|
Ctx->createNamedTempSymbol(); // generate label for the next block
|
2018-03-10 01:45:13 +08:00
|
|
|
NewCall->push_back(CallInst);
|
|
|
|
|
|
|
|
if (IsJumpTable) {
|
2018-03-30 09:42:06 +08:00
|
|
|
MCInst &Je = NewCall->back();
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// Jump to next compare if target addresses don't match.
|
|
|
|
Je.clear();
|
2020-12-02 08:29:39 +08:00
|
|
|
Je.setOpcode(X86::JCC_1);
|
2021-12-29 09:52:08 +08:00
|
|
|
if (Targets[i].first)
|
2022-01-09 09:48:33 +08:00
|
|
|
Je.addOperand(MCOperand::createExpr(MCSymbolRefExpr::create(
|
|
|
|
Targets[i].first, MCSymbolRefExpr::VK_None, *Ctx)));
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-03-10 01:45:13 +08:00
|
|
|
Je.addOperand(MCOperand::createImm(Targets[i].second));
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2020-12-02 08:29:39 +08:00
|
|
|
Je.addOperand(MCOperand::createImm(X86::COND_E));
|
2018-03-10 01:45:13 +08:00
|
|
|
assert(!isInvoke(CallInst));
|
|
|
|
} else {
|
2018-03-30 09:42:06 +08:00
|
|
|
MCInst &Jne = NewCall->back();
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// Jump to next compare if target addresses don't match.
|
|
|
|
Jne.clear();
|
2020-12-02 08:29:39 +08:00
|
|
|
Jne.setOpcode(X86::JCC_1);
|
2018-03-10 01:45:13 +08:00
|
|
|
Jne.addOperand(MCOperand::createExpr(MCSymbolRefExpr::create(
|
|
|
|
NextTarget, MCSymbolRefExpr::VK_None, *Ctx)));
|
2020-12-02 08:29:39 +08:00
|
|
|
Jne.addOperand(MCOperand::createImm(X86::COND_NE));
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// Call specific target directly.
|
2021-05-08 09:43:25 +08:00
|
|
|
Results.emplace_back(Ctx->createNamedTempSymbol(),
|
2021-12-10 03:53:12 +08:00
|
|
|
InstructionListType());
|
2018-03-10 01:45:13 +08:00
|
|
|
NewCall = &Results.back().second;
|
|
|
|
NewCall->push_back(CallInst);
|
|
|
|
MCInst &CallOrJmp = NewCall->back();
|
|
|
|
|
|
|
|
CallOrJmp.clear();
|
|
|
|
|
|
|
|
if (MinimizeCodeSize && !LoadElim) {
|
2020-12-02 08:29:39 +08:00
|
|
|
CallOrJmp.setOpcode(IsTailCall ? X86::JMP32r : X86::CALL64r);
|
2018-03-10 01:45:13 +08:00
|
|
|
CallOrJmp.addOperand(MCOperand::createReg(FuncAddrReg));
|
|
|
|
} else {
|
2020-12-02 08:29:39 +08:00
|
|
|
CallOrJmp.setOpcode(IsTailCall ? X86::JMP_4 : X86::CALL64pcrel32);
|
2018-03-10 01:45:13 +08:00
|
|
|
|
2021-12-29 09:52:08 +08:00
|
|
|
if (Targets[i].first)
|
2018-03-10 01:45:13 +08:00
|
|
|
CallOrJmp.addOperand(MCOperand::createExpr(MCSymbolRefExpr::create(
|
|
|
|
Targets[i].first, MCSymbolRefExpr::VK_None, *Ctx)));
|
2021-12-29 09:52:08 +08:00
|
|
|
else
|
2018-03-10 01:45:13 +08:00
|
|
|
CallOrJmp.addOperand(MCOperand::createImm(Targets[i].second));
|
|
|
|
}
|
2020-12-02 08:29:39 +08:00
|
|
|
if (IsTailCall)
|
2021-07-30 08:28:51 +08:00
|
|
|
setTailCall(CallOrJmp);
|
2018-03-10 01:45:13 +08:00
|
|
|
|
2021-07-17 07:05:18 +08:00
|
|
|
if (CallOrJmp.getOpcode() == X86::CALL64r ||
|
|
|
|
CallOrJmp.getOpcode() == X86::CALL64pcrel32) {
|
2021-08-04 08:53:32 +08:00
|
|
|
if (Optional<uint32_t> Offset = getOffset(CallInst))
|
2021-07-17 07:05:18 +08:00
|
|
|
// Annotated as duplicated call
|
2021-08-04 08:53:32 +08:00
|
|
|
setOffset(CallOrJmp, *Offset);
|
2021-07-17 07:05:18 +08:00
|
|
|
}
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
if (isInvoke(CallInst) && !isInvoke(CallOrJmp)) {
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
// Copy over any EH or GNU args size information from the original
|
|
|
|
// call.
|
2021-04-08 15:19:26 +08:00
|
|
|
Optional<MCPlus::MCLandingPad> EHInfo = getEHInfo(CallInst);
|
[BOLT][Refactoring] Isolate changes to MC layer
Summary:
Changes that we made to MCInst, MCOperand, MCExpr, etc. are now all
moved into tools/llvm-bolt. That required a change to the way we handle
annotations and any extra operands for MCInst.
Any MCPlus information is now attached via an extra operand of type
MCInst with an opcode ANNOTATION_LABEL. Since this operand is MCInst, we
attach extra info as operands to this instruction. For first-level
annotations use functions to access the information, such as
getConditionalTailCall() or getEHInfo(), etc. For the rest, optional or
second-class annotations, use a general named-annotation interface such
as getAnnotationAs<uint64_t>(Inst, "Count").
I did a test on HHVM binary, and a memory consumption went down a little
bit while the runtime remained the same.
(cherry picked from FBD7405412)
2018-03-20 09:32:12 +08:00
|
|
|
if (EHInfo)
|
|
|
|
addEHInfo(CallOrJmp, *EHInfo);
|
2021-04-08 15:19:26 +08:00
|
|
|
int64_t GnuArgsSize = getGnuArgsSize(CallInst);
|
2018-03-10 01:45:13 +08:00
|
|
|
if (GnuArgsSize >= 0)
|
|
|
|
addGnuArgsSize(CallOrJmp, GnuArgsSize);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!IsTailCall) {
|
|
|
|
// The fallthrough block for the most common target should be
|
|
|
|
// the merge block.
|
|
|
|
if (i == 0) {
|
|
|
|
// Fallthrough to merge block.
|
2020-12-02 08:29:39 +08:00
|
|
|
MergeBlock = Ctx->createNamedTempSymbol();
|
2018-03-10 01:45:13 +08:00
|
|
|
} else {
|
|
|
|
// Insert jump to the merge block if we are not doing a fallthrough.
|
2018-03-30 09:42:06 +08:00
|
|
|
jumpToMergeBlock(*NewCall);
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Cold call block.
|
2021-12-10 03:53:12 +08:00
|
|
|
Results.emplace_back(NextTarget, InstructionListType());
|
|
|
|
InstructionListType &NewCall = Results.back().second;
|
2021-12-29 09:52:08 +08:00
|
|
|
for (const MCInst *Inst : MethodFetchInsns)
|
2018-03-10 01:45:13 +08:00
|
|
|
if (Inst != &CallInst)
|
|
|
|
NewCall.push_back(*Inst);
|
|
|
|
NewCall.push_back(CallInst);
|
|
|
|
|
|
|
|
// Jump to merge block from cold call block
|
|
|
|
if (!IsTailCall && !IsJumpTable) {
|
2018-03-30 09:42:06 +08:00
|
|
|
jumpToMergeBlock(NewCall);
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// Record merge block
|
2021-12-10 03:53:12 +08:00
|
|
|
Results.emplace_back(MergeBlock, InstructionListType());
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return Results;
|
|
|
|
}
|
|
|
|
|
2019-06-20 11:10:49 +08:00
|
|
|
BlocksVectorTy jumpTablePromotion(
|
2022-01-09 09:48:33 +08:00
|
|
|
const MCInst &IJmpInst,
|
|
|
|
const std::vector<std::pair<MCSymbol *, uint64_t>> &Targets,
|
|
|
|
const std::vector<MCInst *> &TargetFetchInsns,
|
|
|
|
MCContext *Ctx) const override {
|
2018-03-10 01:45:13 +08:00
|
|
|
assert(getJumpTable(IJmpInst) != 0);
|
|
|
|
uint16_t IndexReg = getAnnotationAs<uint16_t>(IJmpInst, "JTIndexReg");
|
|
|
|
if (IndexReg == 0)
|
2019-06-20 11:10:49 +08:00
|
|
|
return BlocksVectorTy();
|
2018-03-10 01:45:13 +08:00
|
|
|
|
2019-06-20 11:10:49 +08:00
|
|
|
BlocksVectorTy Results;
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// Label for the current code block.
|
2021-12-10 03:53:12 +08:00
|
|
|
MCSymbol *NextTarget = nullptr;
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
for (unsigned int i = 0; i < Targets.size(); ++i) {
|
2021-12-10 03:53:12 +08:00
|
|
|
Results.emplace_back(NextTarget, InstructionListType());
|
|
|
|
InstructionListType *CurBB = &Results.back().second;
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// Compare current index to a specific index.
|
|
|
|
CurBB->emplace_back(MCInst());
|
|
|
|
MCInst &CompareInst = CurBB->back();
|
|
|
|
CompareInst.setLoc(IJmpInst.getLoc());
|
|
|
|
CompareInst.setOpcode(X86::CMP64ri32);
|
|
|
|
CompareInst.addOperand(MCOperand::createReg(IndexReg));
|
|
|
|
|
2021-04-08 15:19:26 +08:00
|
|
|
const uint64_t CaseIdx = Targets[i].second;
|
2018-03-10 01:45:13 +08:00
|
|
|
// Immediate address is out of sign extended 32 bit range.
|
2021-12-29 09:52:08 +08:00
|
|
|
if (int64_t(CaseIdx) != int64_t(int32_t(CaseIdx)))
|
2019-06-20 11:10:49 +08:00
|
|
|
return BlocksVectorTy();
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
CompareInst.addOperand(MCOperand::createImm(CaseIdx));
|
2022-03-17 00:37:54 +08:00
|
|
|
shortenInstruction(CompareInst, *Ctx->getSubtargetInfo());
|
2018-03-10 01:45:13 +08:00
|
|
|
|
|
|
|
// jump to next target compare.
|
2020-12-02 08:29:39 +08:00
|
|
|
NextTarget =
|
|
|
|
Ctx->createNamedTempSymbol(); // generate label for the next block
|
2018-03-10 01:45:13 +08:00
|
|
|
CurBB->push_back(MCInst());
|
|
|
|
|
|
|
|
MCInst &JEInst = CurBB->back();
|
|
|
|
JEInst.setLoc(IJmpInst.getLoc());
|
|
|
|
|
|
|
|
// Jump to target if indices match
|
2020-12-02 08:29:39 +08:00
|
|
|
JEInst.setOpcode(X86::JCC_1);
|
2018-03-10 01:45:13 +08:00
|
|
|
JEInst.addOperand(MCOperand::createExpr(MCSymbolRefExpr::create(
|
|
|
|
Targets[i].first, MCSymbolRefExpr::VK_None, *Ctx)));
|
2020-12-02 08:29:39 +08:00
|
|
|
JEInst.addOperand(MCOperand::createImm(X86::COND_E));
|
2018-03-10 01:45:13 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Cold call block.
|
2021-12-10 03:53:12 +08:00
|
|
|
Results.emplace_back(NextTarget, InstructionListType());
|
|
|
|
InstructionListType &CurBB = Results.back().second;
|
2021-12-29 09:52:08 +08:00
|
|
|
for (const MCInst *Inst : TargetFetchInsns)
|
2018-03-10 01:45:13 +08:00
|
|
|
if (Inst != &IJmpInst)
|
|
|
|
CurBB.push_back(*Inst);
|
2021-12-29 09:52:08 +08:00
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
CurBB.push_back(IJmpInst);
|
|
|
|
|
|
|
|
return Results;
|
|
|
|
}
|
|
|
|
|
2021-01-29 04:32:03 +08:00
|
|
|
private:
|
|
|
|
bool createMove(MCInst &Inst, const MCSymbol *Src, unsigned Reg,
|
|
|
|
MCContext *Ctx) const {
|
|
|
|
Inst.setOpcode(X86::MOV64rm);
|
|
|
|
Inst.addOperand(MCOperand::createReg(Reg));
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RIP)); // BaseReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(1)); // ScaleAmt
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // IndexReg
|
|
|
|
Inst.addOperand(MCOperand::createExpr(
|
|
|
|
MCSymbolRefExpr::create(Src, MCSymbolRefExpr::VK_None,
|
|
|
|
*Ctx))); // Displacement
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // AddrSegmentReg
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool createLea(MCInst &Inst, const MCSymbol *Src, unsigned Reg,
|
|
|
|
MCContext *Ctx) const {
|
|
|
|
Inst.setOpcode(X86::LEA64r);
|
|
|
|
Inst.addOperand(MCOperand::createReg(Reg));
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::RIP)); // BaseReg
|
|
|
|
Inst.addOperand(MCOperand::createImm(1)); // ScaleAmt
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // IndexReg
|
|
|
|
Inst.addOperand(MCOperand::createExpr(
|
|
|
|
MCSymbolRefExpr::create(Src, MCSymbolRefExpr::VK_None,
|
|
|
|
*Ctx))); // Displacement
|
|
|
|
Inst.addOperand(MCOperand::createReg(X86::NoRegister)); // AddrSegmentReg
|
|
|
|
return true;
|
|
|
|
}
|
2018-03-10 01:45:13 +08:00
|
|
|
};
|
|
|
|
|
2022-01-09 09:48:33 +08:00
|
|
|
} // namespace
|
2018-03-10 01:45:13 +08:00
|
|
|
|
[BOLT] Support for lite mode with relocations
Summary:
Add '-lite' support for relocations for improved processing time,
memory consumption, and more resilient processing of binaries with
embedded assembly code.
In lite relocation mode, BOLT will skip full processing of functions
without a profile. It will run scanExternalRefs() on such functions
to discover external references and to create internal relocations
to update references to optimized functions.
Note that we could have relied on the compiler/linker to provide
relocations for function references. However, there's no assurance
that all such references are reported. E.g., the compiler can resolve
inter-procedural references internally, leaving no relocations
for the linker.
The scan process takes about <10 seconds per 100MB of code on modern
hardware. It's a reasonable overhead to live with considering the
flexibility it provides.
If BOLT fails to scan or disassemble a function, .e.g., due to a data
object embedded in code, or an unsupported instruction, it enables a
patching mode to guarantee that the failed function will call
optimized/moved versions of functions. The patching happens at original
function entry points.
'-skip=<func1,func2,...>' option now can be used to skip processing of
arbitrary functions in the relocation mode.
With '-use-old-text' or '-strict' we require all functions to be
processed. As such, it is incompatible with '-lite' option,
and '-skip' option will only disable optimizations of listed
functions, not their disassembly and emission.
(cherry picked from FBD22040717)
2020-06-15 15:15:47 +08:00
|
|
|
namespace llvm {
|
|
|
|
namespace bolt {
|
|
|
|
|
2018-03-10 01:45:13 +08:00
|
|
|
MCPlusBuilder *createX86MCPlusBuilder(const MCInstrAnalysis *Analysis,
|
|
|
|
const MCInstrInfo *Info,
|
|
|
|
const MCRegisterInfo *RegInfo) {
|
|
|
|
return new X86MCPlusBuilder(Analysis, Info, RegInfo);
|
|
|
|
}
|
[BOLT] Support for lite mode with relocations
Summary:
Add '-lite' support for relocations for improved processing time,
memory consumption, and more resilient processing of binaries with
embedded assembly code.
In lite relocation mode, BOLT will skip full processing of functions
without a profile. It will run scanExternalRefs() on such functions
to discover external references and to create internal relocations
to update references to optimized functions.
Note that we could have relied on the compiler/linker to provide
relocations for function references. However, there's no assurance
that all such references are reported. E.g., the compiler can resolve
inter-procedural references internally, leaving no relocations
for the linker.
The scan process takes about <10 seconds per 100MB of code on modern
hardware. It's a reasonable overhead to live with considering the
flexibility it provides.
If BOLT fails to scan or disassemble a function, .e.g., due to a data
object embedded in code, or an unsupported instruction, it enables a
patching mode to guarantee that the failed function will call
optimized/moved versions of functions. The patching happens at original
function entry points.
'-skip=<func1,func2,...>' option now can be used to skip processing of
arbitrary functions in the relocation mode.
With '-use-old-text' or '-strict' we require all functions to be
processed. As such, it is incompatible with '-lite' option,
and '-skip' option will only disable optimizations of listed
functions, not their disassembly and emission.
(cherry picked from FBD22040717)
2020-06-15 15:15:47 +08:00
|
|
|
|
2022-01-09 09:48:33 +08:00
|
|
|
} // namespace bolt
|
|
|
|
} // namespace llvm
|